Mammogram Analysis with YOLO Models on an Affordable Embedded System
Simple Summary
Abstract
1. Introduction
- We create a mammography dataset (a web-based annotation tool) and train lightweight versions of YOLO and a Hybrid Convolutional–Transformer Architecture (YOLOv5, YOLOv8, YOLOv10, YOLOv11, and RT-DETR) for breast lesion detection, providing a comprehensive performance comparison.
- We deploy the optimized model on a NVIDIA Jetson Nano, demonstrating real-time inference capability on the device.
- We discuss the feasibility of this low-cost system for use in resource-limited clinical settings.
2. Related Work
2.1. CAD Systems for Mammogram Detection
2.2. Challenges of the YOLO Model Until the Present
3. Materials and Methods
3.1. Graphical User Interface of the Custom Web-Based Labeling Tool
- Patient ID listing,
- Listing of breast compositions, which enumerate four breast densities: almost entirely fatty, scattered areas of fibroglandular density, heterogeneously dense, and extremely dense,
- Scrolling feature for individual breasts,
- Scrolling feature for dual breasts,
- Creating a bounding box,
- Resetting the image to its original view,
- Displaying the lesion type in detail,
- Displaying the mammograms,
- Quadruple view selection (CC: Craniocaudal, MLO: Mediolateral Oblique, LL: Left–Left, and RR: Right–Right) of the mammograms, and
- Saving annotated data.

3.2. Mammogram Dataset
3.2.1. Private Dataset
3.2.2. VinDr-Mammo Dataset
3.3. Preprocessing Tasks
3.3.1. Image Resizing and Fidelity Preservation
3.3.2. Image Augmentation
3.4. Jetson Nano Configuration
3.5. Performance Metrics
3.6. Implementation of the Model
3.6.1. Initial Models Training
3.6.2. Binary Evaluation Models
3.6.3. Deploying a Mammogram on the NVIDIA Jetson Nano Board
4. Results and Discussion
4.1. Experiment 1: Initial Model Training
4.2. Experiment 2: Model Evaluation
4.2.1. Impact of Model Complexity on Generalization (Private Dataset)
4.2.2. Architectural Trade-Offs: Sensitivity vs. Specificity
4.2.3. Role of Attention Mechanisms in the RT-DETR Model (Private Dataset)
4.2.4. Impact of Model Complexity on Generalization (VinDr-Mammo)
4.2.5. Trade-Off Between Detection and Localization Accuracy
4.2.6. Role of Attention Mechanisms in the RT-DETR Model (VinDr-Mammo)
4.3. Experiment 3: Deploying a Mammogram on the NVIDIA Jetson Nano Board
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Giaquinto, A.N.; Sung, H.; Newman, L.A.; Freedman, R.A.; Smith, R.A.; Star, J.; Jemal, A.; Siegel, R.L. Breast cancer statistics 2024. CA A Cancer J. Clin. 2024, 74, 477–495. [Google Scholar] [CrossRef]
- Karaca Aydemir, B.K.; Telatar, Z.; Güney, S.; Dengiz, B. Detecting and classifying breast masses via YOLO-based deep learning. Neural Comput. Appl. 2025, 37, 11555–11582. [Google Scholar] [CrossRef]
- Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
- Li, C.; Li, L.; Jiang, H.; Weng, K.; Geng, Y.; Li, L.; Ke, Z.; Li, Q.; Cheng, M.; Nie, W. YOLOv6: A single-stage object detection framework for industrial applications. arXiv 2022, arXiv:2209.02976. [Google Scholar] [CrossRef]
- Wang, C.-Y.; Bochkovskiy, A.; Liao, H.-Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023. [Google Scholar]
- Wang, C.-Y.; Yeh, I.-H.; Mark Liao, H.-Y. Yolov9: Learning what you want to learn using programmable gradient information. In Proceedings of the European Conference on Computer Vision, Milan, Italy, 29 September–4 October 2024; pp. 1–21. [Google Scholar]
- Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J. Yolov10: Real-time end-to-end object detection. Adv. Neural Inf. Process. Syst. 2024, 37, 107984–108011. [Google Scholar]
- Khanam, R.; Hussain, M. Yolov11: An overview of the key architectural enhancements. arXiv 2024, arXiv:2410.17725. [Google Scholar] [CrossRef]
- Alif, M.A.R.; Hussain, M. Yolov12: A breakdown of the key architectural features. arXiv 2025, arXiv:2502.14740. [Google Scholar] [CrossRef]
- Wang, C.-Y.; Liao, H.-Y.M. YOLOv1 to YOLOv10: The fastest and most accurate real-time object detection systems. APSIPA Trans. Signal Inf. Process. 2024, 13, e29. [Google Scholar] [CrossRef]
- Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A. Language models are few-shot learners. Adv. Neural Inf. Process. Syst. 2020, 33, 1877–1901. [Google Scholar]
- Achiam, J.; Adler, S.; Agarwal, S.; Ahmad, L.; Akkaya, I.; Aleman, F.L.; Almeida, D.; Altenschmidt, J.; Altman, S.; Anadkat, S. Gpt-4 technical report. arXiv 2023, arXiv:2303.08774. [Google Scholar] [CrossRef]
- Team, G.; Anil, R.; Borgeaud, S.; Alayrac, J.-B.; Yu, J.; Soricut, R.; Schalkwyk, J.; Dai, A.M.; Hauth, A.; Millican, K. Gemini: A family of highly capable multimodal models. arXiv 2023, arXiv:2312.11805. [Google Scholar] [CrossRef]
- Team, G.; Georgiev, P.; Lei, V.I.; Burnell, R.; Bai, L.; Gulati, A.; Tanzer, G.; Vincent, D.; Pan, Z.; Wang, S. Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context. arXiv 2024, arXiv:2403.05530. [Google Scholar] [CrossRef]
- Bai, Y.; Kadavath, S.; Kundu, S.; Askell, A.; Kernion, J.; Jones, A.; Chen, A.; Goldie, A.; Mirhoseini, A.; McKinnon, C. Constitutional ai: Harmlessness from ai feedback. arXiv 2022, arXiv:2212.08073. [Google Scholar] [CrossRef]
- Touvron, H.; Lavril, T.; Izacard, G.; Martinet, X.; Lachaux, M.-A.; Lacroix, T.; Rozière, B.; Goyal, N.; Hambro, E.; Azhar, F. Llama: Open and efficient foundation language models. arXiv 2023, arXiv:2302.13971. [Google Scholar] [CrossRef]
- Touvron, H.; Martin, L.; Stone, K.; Albert, P.; Almahairi, A.; Babaei, Y.; Bashlykov, N.; Batra, S.; Bhargava, P.; Bhosale, S. Llama 2: Open foundation and fine-tuned chat models. arXiv 2023, arXiv:2307.09288. [Google Scholar] [CrossRef]
- Chowdhery, A.; Narang, S.; Devlin, J.; Bosma, M.; Mishra, G.; Roberts, A.; Barham, P.; Chung, H.W.; Sutton, C.; Gehrmann, S. Palm: Scaling language modeling with pathways. J. Mach. Learn. Res. 2023, 24, 240. [Google Scholar]
- Anil, R.; Dai, A.M.; Firat, O.; Johnson, M.; Lepikhin, D.; Passos, A.; Shakeri, S.; Taropa, E.; Bailey, P.; Chen, Z. Palm 2 technical report. arXiv 2023, arXiv:2305.10403. [Google Scholar] [CrossRef]
- Raffel, C.; Shazeer, N.; Roberts, A.; Lee, K.; Narang, S.; Matena, M.; Zhou, Y.; Li, W.; Liu, P.J. Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 2020, 21, 140. [Google Scholar]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; Curran Associates Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
- Dosovitskiy, A. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
- Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
- Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 213–229. [Google Scholar]
- Li, J.; Li, D.; Xiong, C.; Hoi, S. Blip: Bootstrapping language-image pre-training for unified vision-language understanding and generation. In Proceedings of the International Conference on Machine Learning, Baltimore, MD, USA, 17–23 July 2022; pp. 12888–12900. [Google Scholar]
- Alayrac, J.-B.; Donahue, J.; Luc, P.; Miech, A.; Barr, I.; Hasson, Y.; Lenc, K.; Mensch, A.; Millican, K.; Reynolds, M. Flamingo: A visual language model for few-shot learning. Adv. Neural Inf. Process. Syst. 2022, 35, 23716–23736. [Google Scholar]
- Radford, A.; Kim, J.W.; Xu, T.; Brockman, G.; McLeavey, C.; Sutskever, I. Robust speech recognition via large-scale weak supervision. In Proceedings of the International Conference on Machine Learning, Honolulu HI, USA, 23–29 July 2023; pp. 28492–28518. [Google Scholar]
- Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.-Y. Segment anything. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 4015–4026. [Google Scholar]
- Sarker, S.; Sarker, P.; Bebis, G.; Tavakkoli, A. MV-Swin-T: Mammogram classification with multi-view swin transformer. In Proceedings of the 2024 IEEE International Symposium on Biomedical Imaging (ISBI), Athens, Greece, 27–30 May 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1–5. [Google Scholar]
- Al Mansour, A.G.; Alshomrani, F.; Alfahaid, A.; Almutairi, A.T. MammoViT: A custom vision transformer architecture for accurate BIRADS classification in mammogram analysis. Diagnostics 2025, 15, 285. [Google Scholar] [CrossRef]
- Chen, X.; Zhang, K.; Abdoli, N.; Gilley, P.W.; Wang, X.; Liu, H.; Zheng, B.; Qiu, Y. Transformers improve breast cancer diagnosis from unregistered multi-view mammograms. Diagnostics 2022, 12, 1549. [Google Scholar] [CrossRef]
- Ayana, G.; Dese, K.; Dereje, Y.; Kebede, Y.; Barki, H.; Amdissa, D.; Husen, N.; Mulugeta, F.; Habtamu, B.; Choe, S.-W. Vision-transformer-based transfer learning for mammogram classification. Diagnostics 2023, 13, 178. [Google Scholar] [CrossRef]
- Hassan, N.M.; Hamad, S.; Mahar, K. YOLO-based CAD framework with ViT transformer for breast mass detection and classification in CESM and FFDM images. Neural Comput. Appl. 2024, 36, 6467–6496. [Google Scholar] [CrossRef]
- Manigrasso, F.; Milazzo, R.; Russo, A.S.; Lamberti, F.; Strand, F.; Pagnani, A.; Morra, L. Mammography classification with multi-view deep learning techniques: Investigating graph and transformer-based architectures. Med. Image Anal. 2025, 99, 103320. [Google Scholar] [CrossRef]
- Available online: https://www.topbots.com/important-cnn-architectures/ (accessed on 2 June 2025).
- Chen, H.; Martel, A.L. Enhancing breast cancer detection on screening mammogram using self-supervised learning and a hybrid deep model of Swin Transformer and convolutional neural networks. J. Med. Imaging 2025, 12, S22007. [Google Scholar] [CrossRef]
- Berment, H.; Becette, V.; Mohallem, M.; Ferreira, F.; Chérel, P. Masses in mammography: What are the underlying anatomopathological lesions? Diagn. Interv. Imaging 2014, 95, 124–133. [Google Scholar] [CrossRef]
- Hernández, P.A.; Estrada, T.T.; Pizarro, A.L.; Cisternas, M.D.; Tapia, C.S. Breast calcifications: Description and classification according to bi-rads 5th edition. Rev. Chil. Radiol. 2016, 22, 80–91. [Google Scholar]
- Kim, S.; Tran, T.X.M.; Song, H.; Park, B. Microcalcifications, mammographic breast density, and risk of breast cancer: A cohort study. Breast Cancer Res. 2022, 24, 96. [Google Scholar] [CrossRef]
- Al-Masni, M.A.; Al-Antari, M.A.; Park, J.M.; Gi, G.; Kim, T.Y.; Rivera, P.; Valarezo, E.; Choi, M.T.; Han, S.M.; Kim, T.S. Simultaneous detection and classification of breast masses in digital mammograms via a deep learning YOLO-based CAD system. Comput. Methods Programs Biomed. 2018, 157, 85–94. (In English) [Google Scholar] [CrossRef]
- Aly, G.H.; Marey, M.; El-Sayed, S.A.; Tolba, M.F. YOLO Based Breast Masses Detection and Classification in Full-Field Digital Mammograms. Comput. Methods Programs Biomed. 2021, 200, 105823. [Google Scholar] [CrossRef]
- Prinzi, F.; Insalaco, M.; Orlando, A.; Gaglio, S.; Vitabile, S. A yolo-based model for breast cancer detection in mammograms. Cogn. Comput. 2024, 16, 107–120. [Google Scholar] [CrossRef]
- Shia, W.-C.; Ku, T.-H. Enhancing Microcalcification Detection in Mammography with YOLO-v8 Performance and Clinical Implications. Diagnostics 2024, 14, 2875. [Google Scholar] [CrossRef]
- Aishwarya, N.; Kannaa, G.Y.; Seemakurthy, K. YOLOSkin: A fusion framework for improved skin cancer diagnosis using YOLO detectors on Nvidia Jetson Nano. Biomed. Signal Process. Control 2025, 100, 107093. [Google Scholar] [CrossRef]
- Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
- Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar] [CrossRef]
- Bochkovskiy, A.; Wang, C.-Y.; Liao, H.-Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar] [CrossRef]
- Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path aggregation network for instance segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8759–8768. [Google Scholar]
- Ding, X.; Zhang, X.; Ma, N.; Han, J.; Ding, G.; Sun, J. Repvgg: Making vgg-style convnets great again. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13733–13742. [Google Scholar]
- Nguyen, H.T.; Nguyen, H.Q.; Pham, H.H.; Lam, K.; Le, L.T.; Dao, M.; Vu, V. VinDr-Mammo: A large-scale benchmark dataset for computer-aided diagnosis in full-field digital mammography. Sci. Data 2023, 10, 277. [Google Scholar] [CrossRef]
- Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
- Zhang, R.; Isola, P.; Efros, A.A.; Shechtman, E.; Wang, O. The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 586–595. [Google Scholar]
- Ultralytics. Configuration. Available online: https://docs.ultralytics.com/usage/cfg/ (accessed on 15 December 2025).
- Khanam, R.; Hussain, M. What is YOLOv5: A deep look into the internal features of the popular object detector. arXiv 2024, arXiv:2407.20892. [Google Scholar] [CrossRef]
- Wang, Y.; Sun, H.; Luo, K.; Zhu, Q.; Li, H.; Sun, Y.; Wu, Z.; Wang, G. A lightweight YOLOv11-based framework for small steel defect detection with a newly enhanced feature fusion module. Sci. Rep. 2025, 15, 34322. [Google Scholar] [CrossRef] [PubMed]
- Tian, Z.; Shen, C.; Chen, H.; He, T. Fcos: Fully convolutional one-stage object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9627–9636. [Google Scholar]
- Ghosh, A. YOLO11: Redefining Real-Time Object Detection. LearnOpenCV 2024. Available online: https://learnopencv.com/yolo11/ (accessed on 2 June 2025).
- Intasam, A.; Promworn, Y.; Juhong, A.; Thanasitthichai, S.; Khwayotha, S.; Jiranantanakorn, T.; Thawornwanchai, S.; Piyawattanametha, W. Optimizing the hyperparameter tuning of yolov5 for breast cancer detection. In Proceedings of the 2023 9th International Conference on Engineering, Applied Sciences, and Technology (ICEAST), Vientiane, Laos, 1–4 June 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 184–187. [Google Scholar]
- Su, Y.; Liu, Q.; Xie, W.; Hu, P. YOLO-LOGO: A transformer-based YOLO segmentation model for breast mass detection and segmentation in digital mammograms. Comput. Methods Programs Biomed. 2022, 221, 106903. [Google Scholar] [CrossRef]
- Marchi, R.; Hau, S.; Suryaningrum, K.M.; Yunanda, R. Comparing yolov8 and yolov9 algorithm on breast cancer detection case. Procedia Comput. Sci. 2024, 245, 239–246. [Google Scholar] [CrossRef]
- Muhammad, M.B.; Yeasin, M. Eigen-cam: Class activation map using principal components. In Proceedings of the 2020 International Joint Conference on Neural Networks (IJCNN), Glasgow, UK, 19–24 July 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–7. [Google Scholar]










| Model | Backbone |
|---|---|
| YOLOv5 | YOLOv5 CSPDarknet (CBS, C3, and SPPF) |
| YOLOv7 | YOLOv7 (CBS, ELAN, and MP1) |
| YOLOv8 | YOLOv8 CSPDarknet (CBS, C3, and SPPF) |
| YOLOv10 | YOLOv10 CSPnet (C2f-Compact Inverted Block, SPPF, and PSA) |
| YOLOv11 | YOLOv11 (CBS, C3k2, SPPF, and C2PSA) |
| RT-DETR | CNN-Transformer hybrid |
| Type of Lesion | Number of Lesions |
|---|---|
| MB | 565 |
| MM | 413 |
| CB | 412 |
| CM | 143 |
| AFB | 310 |
| AFM | 115 |
| Parameter | Value |
|---|---|
| Epochs | 150 |
| Learning rate | 0.01 |
| Batch size | 16 |
| Momentum | 0.937 |
| Weight decay | 0.0005 |
| Augmentation Feature | v5 (Value) | v8 (Value) | v10 (Value) | v11 (Value) |
|---|---|---|---|---|
| Image translation | 0.1 | 0.1 | 0.1 | 0.1 |
| Image scale | 0.9 | 0.5 | 0.5 | 0.5 |
| Image flip, left and right | 0.5 | 0.5 | 0.5 | 0.5 |
| Image mosaic | 1.0 | 1.0 | 1.0 | 1.0 |
| Image mixup | 0.1 | - | - | - |
| Image segment copy-paste | 0.1 | - | - | - |
| Model | Number of Layers | Number of Parameters | Billions (109) of Floating-Point Operations per Forward Pass (GFLOPs) |
|---|---|---|---|
| Yolov5x | 498 | 61,992,179 | 154.4 |
| Yolov8n | 168 | 3,006,818 | 8.1 |
| Yolov10s | 293 | 8,039,604 | 24.5 |
| Yolov11n | 238 | 2,583,322 | 6.3 |
| RT-DETR Large | 302 | 31,996,070 | 103.5 |
| Model | mAP@0.5 | mAP@0.5:0.95 | Precision | Recall | F1-Score |
|---|---|---|---|---|---|
| YOLOv5n | 0.78 | 0.43 | 0.80 | 0.72 | 0.76 |
| YOLOv5s | 0.91 | 0.62 | 0.92 | 0.85 | 0.88 |
| YOLOv5m | 0.96 | 0.73 | 0.97 | 0.90 | 0.93 |
| YOLOv5l | 0.96 | 0.76 | 0.98 | 0.89 | 0.94 |
| YOLOv5x | 0.96 | 0.79 | 0.98 | 0.89 | 0.81 |
| YOLOv7 | 0.85 | 0.88 | 0.75 | 0.59 | 0.77 |
| YOLOv7x | 0.69 | 0.42 | 0.73 | 0.63 | 0.67 |
| YOLOv8n | 0.86 | 0.60 | 0.87 | 0.80 | 0.83 |
| YOLOv8s | 0.94 | 0.74 | 0.96 | 0.88 | 0.91 |
| YOLOv8m | 0.92 | 0.76 | 0.95 | 0.85 | 0.90 |
| YOLOv8l | 0.90 | 0.74 | 0.91 | 0.80 | 0.86 |
| YOLOv8x | 0.91 | 0.77 | 0.95 | 0.83 | 0.89 |
| YOLOv10n | 0.83 | 0.54 | 0.83 | 0.75 | 0.79 |
| YOLOv10s | 0.75 | 0.47 | 0.78 | 0.67 | 0.72 |
| YOLOv10m | 0.68 | 0.44 | 0.72 | 0.63 | 0.67 |
| YOLOv10l | 0.65 | 0.43 | 0.76 | 0.59 | 0.66 |
| YOLOv10x | 0.77 | 0.57 | 0.91 | 0.69 | 0.78 |
| YOLOv11n | 0.80 | 0.52 | 0.82 | 0.73 | 0.77 |
| YOLOv11s | 0.89 | 0.62 | 0.90 | 0.81 | 0.85 |
| YOLOv11m | 0.87 | 0.63 | 0.90 | 0.78 | 0.84 |
| YOLOv11l | 0.90 | 0.68 | 0.94 | 0.79 | 0.86 |
| YOLOv11x | 0.91 | 0.71 | 0.93 | 0.83 | 0.89 |
| RT-DETR Large | 0.85 | 0.55 | 0.89 | 0.79 | 0.84 |
| Model | mAP@0.5 | mAP@0.5:0.95 | Precision | Recall | F1-Score |
|---|---|---|---|---|---|
| YOLOv5n | 0.94 | 0.67 | 0.92 | 0.88 | 0.90 |
| YOLOv5s | 0.90 | 0.66 | 0.88 | 0.86 | 0.87 |
| YOLOv5m | 0.91 | 0.74 | 0.92 | 0.85 | 0.88 |
| YOLOv5l | 0.91 | 0.76 | 0.91 | 0.86 | 0.88 |
| YOLOv5x | 0.90 | 0.77 | 0.91 | 0.85 | 0.88 |
| YOLOv7 | 0.86 | 0.65 | 0.87 | 0.80 | 0.83 |
| YOLOv7x | 0.74 | 0.48 | 0.73 | 0.70 | 0.71 |
| YOLOv8n | 0.91 | 0.72 | 0.90 | 0.85 | 0.87 |
| YOLOv8s | 0.85 | 0.60 | 0.84 | 0.78 | 0.81 |
| YOLOv8m | 0.85 | 0.60 | 0.89 | 0.75 | 0.81 |
| YOLOv8l | 0.85 | 0.60 | 0.86 | 0.77 | 0.81 |
| YOLOv8x | 0.84 | 0.60 | 0.85 | 0.77 | 0.81 |
| YOLOv10n | 0.83 | 0.58 | 0.86 | 0.73 | 0.79 |
| YOLOv10s | 0.85 | 0.67 | 0.87 | 0.78 | 0.82 |
| YOLOv10m | 0.84 | 0.67 | 0.90 | 0.79 | 0.84 |
| YOLOv10l | 0.84 | 0.71 | 0.88 | 0.78 | 0.83 |
| YOLOv10x | 0.85 | 0.72 | 0.89 | 0.79 | 0.84 |
| YOLOv11n | 0.90 | 0.68 | 0.89 | 0.83 | 0.86 |
| YOLOv11s | 0.88 | 0.70 | 0.90 | 0.80 | 0.85 |
| YOLOv11m | 0.86 | 0.66 | 0.85 | 0.82 | 0.83 |
| YOLOv11l | 0.83 | 0.65 | 0.86 | 0.77 | 0.81 |
| YOLOv11x | 0.85 | 0.69 | 0.88 | 0.77 | 0.82 |
| RT-DETR Large | 0.88 | 0.69 | 0.92 | 0.83 | 0.87 |
| Model | TP | TN | FP | FN | Sensitivity | Specificity | Accuracy |
|---|---|---|---|---|---|---|---|
| YOLOv5n | 84 | 79 | 21 | 16 | 0.84 | 0.79 | 0.82 |
| YOLOv5s | 78 | 79 | 21 | 22 | 0.78 | 0.79 | 0.79 |
| YOLOv5m | 68 | 88 | 12 | 32 | 0.68 | 0.88 | 0.78 |
| YOLOv5l | 75 | 90 | 10 | 25 | 0.75 | 0.90 | 0.83 |
| YOLOv5x | 87 | 89 | 11 | 13 | 0.87 | 0.89 | 0.88 |
| YOLOv8n | 72 | 78 | 22 | 28 | 0.72 | 0.78 | 0.75 |
| YOLOv8s | 69 | 96 | 4 | 31 | 0.69 | 0.96 | 0.83 |
| YOLOv8m | 55 | 94 | 6 | 45 | 0.55 | 0.94 | 0.75 |
| YOLOv8l | 59 | 96 | 4 | 41 | 0.59 | 0.96 | 0.78 |
| YOLOv8x | 58 | 93 | 7 | 42 | 0.58 | 0.93 | 0.76 |
| YOLOv10n | 66 | 90 | 10 | 34 | 0.66 | 0.90 | 0.82 |
| YOLOv10s | 76 | 79 | 21 | 24 | 0.76 | 0.79 | 0.78 |
| YOLOv10m | 69 | 87 | 13 | 31 | 0.69 | 0.87 | 0.78 |
| YOLOv10l | 68 | 95 | 5 | 32 | 0.68 | 0.95 | 0.82 |
| YOLOv10x | 53 | 88 | 12 | 47 | 0.53 | 0.88 | 0.71 |
| YOLOv11n | 84 | 91 | 9 | 16 | 0.84 | 0.91 | 0.86 |
| YOLOv11s | 69 | 89 | 11 | 31 | 0.69 | 0.89 | 0.79 |
| YOLOv11m | 62 | 95 | 5 | 38 | 0.62 | 0.95 | 0.79 |
| YOLOv11l | 58 | 92 | 8 | 42 | 0.58 | 0.92 | 0.75 |
| YOLOv11x | 55 | 95 | 5 | 45 | 0.55 | 0.95 | 0.75 |
| RT-DETR Large | 67 | 99 | 1 | 33 | 0.67 | 0.95 | 0.83 |
| Model | mAP@0.5 | mAP@0.5:0.95 | Precision | Recall | F1-Score |
|---|---|---|---|---|---|
| YOLOv5n | 0.92 | 0.65 | 0.88 | 0.86 | 0.87 |
| YOLOv5s | 0.88 | 0.65 | 0.82 | 0.85 | 0.83 |
| YOLOv5m | 0.90 | 0.73 | 0.84 | 0.86 | 0.85 |
| YOLOv5l | 0.89 | 0.75 | 0.85 | 0.85 | 0.85 |
| YOLOv5x | 0.87 | 0.76 | 0.85 | 0.85 | 0.85 |
| YOLOv8n | 0.89 | 0.71 | 0.89 | 0.8 | 0.84 |
| YOLOv8s | 0.82 | 0.58 | 0.84 | 0.77 | 0.80 |
| YOLOv8m | 0.83 | 0.58 | 0.85 | 0.74 | 0.79 |
| YOLOv8l | 0.82 | 0.59 | 0.81 | 0.77 | 0.79 |
| YOLOv8x | 0.83 | 0.58 | 0.83 | 0.78 | 0.80 |
| YOLOv10n | 0.84 | 0.64 | 0.87 | 0.76 | 0.81 |
| YOLOv10s | 0.83 | 0.66 | 0.80 | 0.82 | 0.81 |
| YOLOv10m | 0.83 | 0.67 | 0.80 | 0.81 | 0.80 |
| YOLOv10l | 0.84 | 0.71 | 0.84 | 0.80 | 0.82 |
| YOLOv10x | 0.85 | 0.72 | 0.86 | 0.79 | 0.82 |
| YOLOv11n | 0.88 | 0.67 | 0.83 | 0.85 | 0.84 |
| YOLOv11s | 0.87 | 0.69 | 0.85 | 0.79 | 0.82 |
| YOLOv11m | 0.84 | 0.66 | 0.88 | 0.76 | 0.82 |
| YOLOv11l | 0.83 | 0.65 | 0.82 | 0.78 | 0.80 |
| YOLOv11x | 0.86 | 0.69 | 0.85 | 0.80 | 0.82 |
| RT-DETR Large | 0.83 | 0.67 | 0.86 | 0.80 | 0.83 |
| Platform | Model | Inference Time (ms/Image) (Average ± Std) | Frames per Second (FPS) (Average ± Std) |
|---|---|---|---|
| RTX 3090 GPU | YOLOv5x | 66.79 ± 0.57 | 14.97 ± 0.13 |
| YOLOv11n | 19.21 ± 0.11 | 52.06 ± 0.30 | |
| NVIDIA Jetson Nano Board | YOLOv5x | 1367.7 ± 40.95 | 0.72 ± 1.55 |
| YOLOv11n | 162.82 ± 8.90 | 6.16 ± 0.31 |
| Metric | GPU Computer (ms) | NVIDIA Jetson Nano (ms) | ||
|---|---|---|---|---|
| Benign | Malignant | Benign | Malignant | |
| Sample Size (n) | 100 patients | 100 patients | 100 patients | 100 patients |
| Mean (µ) | 16.44 | 15.96 | 152.35 | 142.39 |
| Median (M) | 15.75 | 15.75 | 130.63 | 130.44 |
| Standard deviation (σ) | 7.75 | 2.86 | 41.62 | 26.57 |
| Minimum | 7.50 | 7.50 | 129.38 | 129.35 |
| Maximum | 30.45 | 46.90 | 392.85 | 256.83 |
| First Quartile (Q1) | 9.13 | 15.50 | 130.13 | 129.93 |
| Third Quartile (Q3) | 20.03 | 16.50 | 160.38 | 139.71 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Intasam, A.; Piyawattanametha, N.; Promworn, Y.; Jiranantanakorn, T.; Thawornwanchai, S.; Pichayakul, P.; Sriwanichwiphat, S.; Thanasitthichai, S.; Khwayotha, S.; Lertkowit, M.; et al. Mammogram Analysis with YOLO Models on an Affordable Embedded System. Cancers 2026, 18, 70. https://doi.org/10.3390/cancers18010070
Intasam A, Piyawattanametha N, Promworn Y, Jiranantanakorn T, Thawornwanchai S, Pichayakul P, Sriwanichwiphat S, Thanasitthichai S, Khwayotha S, Lertkowit M, et al. Mammogram Analysis with YOLO Models on an Affordable Embedded System. Cancers. 2026; 18(1):70. https://doi.org/10.3390/cancers18010070
Chicago/Turabian StyleIntasam, Anongnat, Nicholas Piyawattanametha, Yuttachon Promworn, Titipon Jiranantanakorn, Soonthorn Thawornwanchai, Pakpawee Pichayakul, Sarawan Sriwanichwiphat, Somchai Thanasitthichai, Sirihattaya Khwayotha, Methininat Lertkowit, and et al. 2026. "Mammogram Analysis with YOLO Models on an Affordable Embedded System" Cancers 18, no. 1: 70. https://doi.org/10.3390/cancers18010070
APA StyleIntasam, A., Piyawattanametha, N., Promworn, Y., Jiranantanakorn, T., Thawornwanchai, S., Pichayakul, P., Sriwanichwiphat, S., Thanasitthichai, S., Khwayotha, S., Lertkowit, M., Phakwapee, N., Juhong, A., & Piyawattanametha, W. (2026). Mammogram Analysis with YOLO Models on an Affordable Embedded System. Cancers, 18(1), 70. https://doi.org/10.3390/cancers18010070

