Most recent publication updates can be found on my [Google Scholar] profile.
2025
📌 HyFormer-Net: A Synergistic CNN-Transformer with Interpretable Multi-Scale Fusion for Breast Lesion Segmentation and Classification in Ultrasound Images [under review]
Authors: Mohammad Amanour Rahman
Journal: arXiv preprint (arXiv 2025)
[Abstract] [PDF] [Code & Dataset] [Citation]
Breast cancer early detection heavily relies on ultrasound imaging, but accurate diagnosis is often hindered by speckle noise, operator dependency, and indistinct lesion boundaries. Existing deep learning methods lack hierarchical multi-scale integration, quantitative interpretability validation, and cross-dataset generalization analysis, hindering clinical adoption. In this work, we propose HyFormer-Net, a hybrid CNN-Transformer framework integrating EfficientNet-B3 and Swin Transformer via multi-scale hierarchical fusion blocks at four encoder stages. The attention-gated decoder enables dual-pipeline interpretability: intrinsic attention validation (quantitative IoU verification) and Grad-CAM classification reasoning. We evaluated performance on BUSI dataset and conducted systematic cross-dataset validation on BUS-UCLM with progressive fine-tuning. HyFormer-Net achieved 76.1% Dice score and 93.2% classification accuracy on BUSI, with clinically critical 92.1% Malignant Recall. Cross-dataset evaluation showed zero-shot failure, but 10% target-domain fine-tuning recovered 92.5% performance.
Cite as: arXiv:2511.01013