HGCTransformer: Hybrid Gated CNN-Transformer for Breast Cancer Image Classification

Authors

  • Shengpei Ye

DOI:

https://doi.org/10.62051/ijcsit.v8n3.07

Keywords:

Breast cancer classification, Histopathological image, Convolutional neural network, Transformer

Abstract

Breast cancer is the most common malignant tumor among women, and histopathological images play a crucial role in the differential diagnosis of benign and malignant lesions. Although Convolutional Neural Networks (CNNs) and Transformers have been widely used in medical image classification, CNNs often struggle to capture global structural patterns, while Transformers may underperform in modeling fine-grained local features. To address this, we propose a Hybrid Gated CNN-Transformer (HGCTransformer) model for breast tumor histopathological image classification. The model introduces a Dual-Branch Convolution and Attention Residual Module (DBCARM) into the Transformer block to integrate local texture and global contextual information. Additionally, a Gated Multi-Scale Feed-forward Network (GMSFN) is designed to enhance the discrimination of multi-scale malignant features, such as nuclear atypia and architectural disarray. Experimental results on public breast histopathology dataset indicate that the proposed method achieves a classification accuracy of 99.76%, underscoring its potential for computer-aided diagnosis of breast cancer.

Downloads

Download data is not yet available.

References

[1] BRAY F, LAVERSANNE M, SUNG H, et al. Global cancer statistics 2022: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries [J]. CA: a cancer journal for clinicians, 2024, 74(3): 229-63.

[2] OBEAGU E I, OBEAGU G U. Breast cancer: A review of risk factors and diagnosis [J]. Medicine, 2024, 103(3): e36905.

[3] DE MARCO P, RICCIARDI V, MONTESANO M, et al. Transfer learning classification of suspicious lesions on breast ultrasound: is there room to avoid biopsies of benign lesions? [J]. European Radiology Experimental, 2024, 8(1): 121.

[4] TSENG L-J, MATSUYAMA A, MACDONALD-DICKINSON V. Histology: The gold standard for diagnosis? [J]. The Canadian Veterinary Journal, 2023, 64(4): 389.

[5] SILVERWOOD S M, BACKER G, GALLOWAY A, et al. Assessing the rates of false-positive ovarian cancer screenings and surgical interventions associated with screening tools: a systematic review [J]. BMJ oncology, 2024, 3(1): e000404.

[6] BOMEISL P, GILMORE H. Spectrum of atypical ductal hyperplasia (ADH) and ductal carcinoma in-situ (DCIS): Diagnostic challenges; proceedings of the Seminars in Diagnostic Pathology, F, 2024 [C]. Elsevier.

[7] JIMENEZ-DEL-TORO O, OTáLORA S, ANDERSSON M, et al. Analysis of histopathology images: From traditional machine learning to deep learning [M]. Biomedical texture analysis. Elsevier. 2017: 281-314.

[8] MANKKI J-J, BOCHENINA K. Vision Transformers in Brain Image Segmentation [J]. 2025.

[9] IQBAL N, MUMTAZ R, SHAFI U, et al. Gray level co-occurrence matrix (GLCM) texture based crop classification using low altitude remote sensing platforms [J]. PeerJ Computer Science, 2021, 7: e536.

[10] SERTE S, DEMIREL H. Gabor wavelet-based deep learning for skin lesion classification [J]. Computers in biology and medicine, 2019, 113: 103423.

[11] SHEN D, JI Y, LI P, et al. Ranet: Region attention network for semantic segmentation [J]. Advances in Neural Information Processing Systems, 2020, 33: 13927-38.

[12] ALQAHTANI Y, MANDAWKAR U, SHARMA A, et al. Breast cancer pathological image classification based on the multiscale CNN squeeze model [J]. Computational Intelligence and Neuroscience, 2022, 2022(1): 7075408.

[13] HAN K, WANG Y, CHEN H, et al. A survey on vision transformer [J]. IEEE transactions on pattern analysis and machine intelligence, 2022, 45(1): 87-110.

[14] WU H, XIAO B, CODELLA N, et al. Cvt: Introducing convolutions to vision transformers; proceedings of the Proceedings of the IEEE/CVF international conference on computer vision, F, 2021 [C].

[15] TU Z, TALEBI H, ZHANG H, et al. Maxvit: Multi-axis vision transformer; proceedings of the European conference on computer vision, F, 2022 [C]. Springer.

[16] LI W, WANG X, XIA X, et al. Sepvit: Separable vision transformer [J]. arXiv preprint arXiv:220315380, 2022.

[17] MANZARI O N, AHMADABADI H, KASHIANI H, et al. MedViT: a robust vision transformer for generalized medical image classification [J]. Computers in biology and medicine, 2023, 157: 106791.

[18] HU S, GAO F, ZHOU X, et al. Hybrid convolutional and attention network for hyperspectral image denoising [J]. IEEE Geoscience and Remote Sensing Letters, 2024, 21: 1-5.

[19] NEYESTANAK M S, JAHANI H, KHODARAHMI M, et al. A quantitative comparison between focal loss and binary cross-entropy loss in brain tumor auto-segmentation using U-Net [J]. Journal of Biostatistics and Epidemiology, 2025, 11(1): 15-35.

[20] HE J. Gradient reweighting: Towards imbalanced class-incremental learning; proceedings of the Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, F, 2024 [C].

[21] PEREIRA G A, HUSSAIN M. A review of transformer-based models for computer vision tasks: Capturing global context and spatial relationships [J]. arXiv preprint arXiv:240815178, 2024.

[22] MIAO J, ZHU W. Precision–recall curve (PRC) classification trees [J]. Evolutionary intelligence, 2022, 15(3): 1545-69.

[23] RAINIO O, TAMMINEN J, VENäLäINEN M S, et al. Comparison of thresholds for a convolutional neural network classifying medical images [J]. International Journal of Data Science and Analytics, 2024: 1-7.

[24] BIGGERI A, FORNI S, BRAGA M. The risk of over-diagnosis in serological testing. Implications for communications strategies [J]. Epidemiologia e Prevenzione, 2020, 44(56): 184-92.

[25] TOMA T A, BISWAS S, MIAH M S, et al. Breast cancer detection based on simplified deep learning technique with histopathological image using BreaKHis database [J]. Radio Science, 2023, 58(11): 1-18.

[26] GUPTA I, NAYAK S R, GUPTA S, et al. A deep learning based approach to detect IDC in histopathology images [J]. Multimedia Tools and Applications, 2022, 81(25): 36309-30.

[27] YANG J, LONG Q. A modification of adaptive moment estimation (adam) for machine learning [J]. Journal of Industrial and Management Optimization, 2024, 20(7): 2516-40.

[28] THAKUR A, GUPTA M, SINHA D K, et al. Transformative breast Cancer diagnosis using CNNs with optimized ReduceLROnPlateau and Early stopping Enhancements [J]. International Journal of Computational Intelligence Systems, 2024, 17(1): 1-18.

[29] HE K, ZHANG X, REN S, et al. Deep residual learning for image recognition; proceedings of the Proceedings of the IEEE conference on computer vision and pattern recognition, F, 2016 [C].

[30] DENG J, DONG W, SOCHER R, et al. Imagenet: A large-scale hierarchical image database; proceedings of the 2009 IEEE conference on computer vision and pattern recognition, F, 2009 [C]. Ieee.

[31] ZHAO Y, FU C, XU S, et al. HCT-Net: A hybrid CNN-Transformer network for multi-class cervical cell classification [J]. Biomedical Signal Processing and Control, 2026, 112: 108383.

[32] YAN C, FAN X, FAN J, et al. Hyformer: Hybrid transformer and cnn for pixel-level multispectral image land cover classification [J]. International Journal of Environmental Research and Public Health, 2023, 20(4): 3059.

[33] LI P, YIN L, LIU S. Mix-ln: Unleashing the power of deeper layers by combining pre-ln and post-ln [J]. arXiv preprint arXiv:241213795, 2024.

Downloads

Published

20-03-2026

Issue

Section

Articles

How to Cite

Ye, S. (2026). HGCTransformer: Hybrid Gated CNN-Transformer for Breast Cancer Image Classification. International Journal of Computer Science and Information Technology, 8(3), 56-72. https://doi.org/10.62051/ijcsit.v8n3.07