Spatial Attention Mechanism and Cascade Feature Extraction in a U-Net Model for Enhancing Breast Tumor Segmentation
Abstract
:1. Introduction
2. Methodology
2.1. Attention Module
2.2. Suggested U-Net Model
- Improved accuracy: The use of an attention mechanism can help to highlight the most important features in the feature maps of the U-Net model, allowing the model to focus on the relevant areas and improve its accuracy. Additionally, the cascade feature extraction block can help to extract more meaningful features from the input data, further improving the accuracy of the segmentation;
- Reduced false positives and false negatives: By using the cascade feature extraction block and attention mechanism, the model can better distinguish between healthy tissue and tumor tissue, reducing the likelihood of false positives and false negatives in the segmentation results;
- Robustness to noise and variability: The use of a U-Net model provides a robust framework for breast tumor segmentation, as it is able to handle noisy and variable input data. By combining this with the attention mechanism and cascade feature extraction, the model can better adapt to different types of input data and improve its segmentation accuracy;
- Faster convergence and training: The attention mechanism and cascade feature extraction block can help to reduce the number of training iterations required for the model to converge, which can save time and computational resources during the training process;
- Generalizability: The use of a U-Net model, combined with the attention mechanism and cascade feature extraction, can provide a generalizable framework for breast tumor segmentation that can be applied to different types of data and imaging modalities. This can be particularly useful for clinical applications, where different types of imaging data may be available for different patients.
= 1 − (2 × sum(y_true × y_pred) + alpha)/(sum(y_true × y_pred) + beta ×
sum(y_true × (1 − y_pred)) + (1 − alpha) × sum((1 − y_true) × y_pred) + smooth)
- Learning Rate: The learning rate determines the step size at which the model adjusts its parameters during training. A higher learning rate may result in faster convergence but could also lead to overshooting and instability. On the other hand, a lower learning rate might result in slower convergence or becoming stuck in local optima. It is essential to find an optimal learning rate that balances convergence speed and stability for the specific model and dataset. We used 0.001 as the value for the Learning Rate;
- Batch Size: The batch size determines the number of samples processed in each training iteration. A larger batch size may lead to faster convergence but requires more memory and computational resources. Conversely, a smaller batch size can result in more noise during training but may help the model generalize better. Finding the right balance between batch size and convergence speed is crucial. Batch Size was 10 in our study;
- Network Architecture: The architecture of the U-Net model itself is an important hyperparameter. The number of layers, layer sizes, skip connections, and other architectural choices can significantly impact the model’s capacity to learn and its ability to handle the complexity of breast tumor segmentation. Experimenting with different network architectures, such as varying the number of layers or adjusting the number of filters in each layer, can help optimize performance;
- Regularization Techniques: Regularization techniques such as dropout, batch normalization, or weight decay can help prevent overfitting and improve generalization. The choice of regularization parameters, such as dropout rates or weight decay coefficients, can influence the model’s ability to generalize to unseen data. Careful tuning and experimentation with these regularization techniques are essential to achieve optimal performance;
- Data Augmentation: Data augmentation techniques, such as rotation, scaling, or flipping, can help increase the robustness of the model by providing additional training examples. The choice and extent of data augmentation, including the range of rotations or scales applied, can impact the model’s ability to generalize and handle variations in tumor shapes, sizes, and orientations;
- Loss Function: The choice of loss function plays a critical role in training an accurate segmentation model. Common choices include Dice loss, cross-entropy loss, or a combination of both. The selection of appropriate loss function and its associated parameters can impact the model’s ability to handle fuzzy boundaries, vague shapes, and class imbalance in the dataset.
3. Experiments
3.1. Dataset
3.2. Data Augmentation
3.3. Evaluation Metrics
3.4. Experimental Results
- Limited evaluation: The paper only evaluates the HCOW method on a single dataset, the MIAS database, and does not compare its performance with other state-of-the-art methods. This limits the generalizability of the proposed method to other datasets and makes it difficult to compare its performance with other methods;
- Complexity: The HCOW method consists of multiple steps, including wavelet decomposition, histogram-based clustering, and adaptive multi-thresholding. This complexity may make it difficult for the method to be implemented in real-world clinical settings, where speed and ease of use are crucial;
- Lack of interpretability: The HCOW method does not provide any visual or quantitative explanation of how it makes its segmentation decisions. This lack of interpretability may limit its clinical utility, as clinicians may be hesitant to rely on a method that cannot explain its reasoning;
- Inadequate validation: The paper only reports sensitivity, specificity, and accuracy as performance metrics, which are inadequate for evaluating the performance of a segmentation method. Metrics such as Dice similarity coefficient and intersection over union are commonly used in the literature and provide a more comprehensive evaluation of the method’s performance;
- Over-segmentation: The HCOW method has a tendency to over-segment regions, resulting in the inclusion of healthy tissue in the segmented regions. This may reduce the specificity of the method and increase the number of false positives.
- Data representation: Ensure that the training dataset includes an adequate representation of different breast cancer types, including in situ and invasive breast cancers. A balanced dataset that covers the various types of breast cancer helps mitigate potential bias;
- Evaluation metrics: Use evaluation metrics that consider the performance across different cancer types. For example, calculating precision, recall, or accuracy separately for each cancer type can provide insights into any disparities in performance;
- Fairness assessment: Conduct fairness assessments to identify and address any potential biases that may arise in the model. This includes examining whether the model’s predictions exhibit disparities in accuracy, sensitivity, specificity, or other performance metrics across different breast cancer types.
- Improved Segmentation Accuracy: The combination of a U-Net model with a spatial attention mechanism in the proposed approach is designed to enhance the accuracy of breast tumor segmentation. By incorporating cascade feature extraction and spatial attention, the model can capture subtle tumor features and focus on important regions while suppressing irrelevant areas. This improved accuracy can aid in precise tumor localization and boundary delineation, which are crucial for accurate diagnosis and treatment planning;
- Addressing Challenges in Breast Tumor Segmentation: Breast tumor segmentation in medical imaging poses challenges such as fuzzy boundaries, vague tumor shapes, variation in tumor size, and illumination variations. The proposed method specifically addresses these challenges by leveraging the U-Net model and attention mechanism. By enhancing subtle features and focusing on important regions, the model aims to overcome these difficulties and provide more accurate and reliable tumor segmentation results;
- State-of-the-Art Performance: The abstract states that the proposed methodology demonstrates state-of-the-art performance on the Mini-MIAS dataset, surpassing existing methods in terms of accuracy, sensitivity, and specificity. Achieving high accuracy and sensitivity is crucial in clinical applications to ensure that breast tumors are correctly identified and localized. The improved performance of the proposed model suggests its potential to provide reliable and accurate results for clinical decision-making;
- Efficiency in Segmentation Process: In addition to accuracy, the proposed methodology aims to improve the efficiency of the segmentation process. By utilizing the combination of the U-Net model and attention mechanism, the model can streamline and optimize the segmentation process, potentially reducing the time required for manual analysis. This increased efficiency can have practical implications in clinical settings, where timely diagnosis and treatment planning are critical.
4. Conclusions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Houssein, E.H.; Emam, M.M.; Ali, A.A.; Suganthan, P.N. Deep and machine learning techniques for medical imaging-based breast cancer: A comprehensive review. Expert Syst. Appl. 2020, 167, 114161. [Google Scholar] [CrossRef]
- Valdora, F.; Houssami, N.; Rossi, F.; Calabrese, M.; Tagliafico, A.S. Rapid review: Radiomics and breast cancer. Breast Cancer Res. Treat. 2018, 169, 217–229. [Google Scholar] [CrossRef] [PubMed]
- Ghoushchi, S.J.; Ranjbarzadeh, R.; Najafabadi, S.A.; Osgooei, E.; Tirkolaee, E.B. An extended approach to the diagnosis of tumour location in breast cancer using deep learning. J. Ambient. Intell. Humaniz. Comput. 2021, 14, 8487–8497. [Google Scholar] [CrossRef]
- Ranjbarzadeh, R.; Sarshar, N.T.; Ghoushchi, S.J.; Esfahani, M.S.; Parhizkar, M.; Pourasad, Y.; Anari, S.; Bendechache, M. MRFE-CNN: Multi-route feature extraction model for breast tumor segmentation in Mammograms using a convolutional neural network. Ann. Oper. Res. 2022, 1–22. [Google Scholar] [CrossRef]
- Zhang, J.; Saha, A.; Zhu, Z.; Mazurowski, M.A. Hierarchical Convolutional Neural Networks for Segmentation of Breast Tumors in MRI with Application to Radiogenomics. IEEE Trans. Med. Imaging 2019, 38, 435–447. [Google Scholar] [CrossRef]
- Raherinirina, A.; Randriamandroso, A.; Hajalalaina, A.R.; Rakotoarivelo, R.A.; Rafamatantantsoa, F. A Gaussian Multivariate Hidden Markov Model for Breast Tumor Diagnosis. Appl. Math. 2021, 12, 679–693. [Google Scholar] [CrossRef]
- Ranjbarzadeh, R.; Dorosti, S.; Ghoushchi, S.J.; Caputo, A.; Tirkolaee, E.B.; Ali, S.S.; Arshadi, Z.; Bendechache, M. Breast tumor localization and segmentation using machine learning techniques: Overview of datasets, findings, and methods. Comput. Biol. Med. 2023, 152, 106443. [Google Scholar] [CrossRef]
- Wang, S.; Sun, K.; Wang, L.; Qu, L.; Yan, F.; Wang, Q.; Shen, D. Breast Tumor Segmentation in DCE-MRI With Tumor Sensitive Synthesis. IEEE Trans. Neural Netw. Learn. Syst. 2021, 1–12. [Google Scholar] [CrossRef]
- Le, P.T.; Pham, B.-T.; Chang, C.-C.; Hsu, Y.-C.; Tai, T.-C.; Li, Y.-H.; Wang, J.-C. Anti-Aliasing Attention U-net Model for Skin Lesion Segmentation. Diagnostics 2023, 13, 1460. [Google Scholar] [CrossRef]
- Le, P.T.; Pham, T.; Hsu, Y.-C.; Wang, J.-C. Convolutional Blur Attention Network for Cell Nuclei Segmentation. Sensors 2022, 22, 1586. [Google Scholar] [CrossRef]
- Ranjbarzadeh, R.; Zarbakhsh, P.; Caputo, A.; Tirkolaee, E.B.; Bendechache, M. Brain Tumor Segmentation based on an Optimized Convolutional Neural Network and an Improved Chimp Optimization Algorithm. 2022; preprint. [Google Scholar] [CrossRef]
- Vakanski, A.; Xian, M.; Freer, P.E. Attention-Enriched Deep Learning Model for Breast Tumor Segmentation in Ultrasound Images. Ultrasound Med. Biol. 2020, 46, 2819–2833. [Google Scholar] [CrossRef]
- Yan, Y.; Liu, Y.; Wu, Y.; Zhang, H.; Zhang, Y.; Meng, L. Accurate segmentation of breast tumors using AE U-net with HDC model in ultrasound images. Biomed. Signal Process. Control 2021, 72, 103299. [Google Scholar] [CrossRef]
- Rezaei, Z. A review on image-based approaches for breast cancer detection, segmentation, and classification. Expert Syst. Appl. 2021, 182, 115204. [Google Scholar] [CrossRef]
- Pezeshki, H. Breast tumor segmentation in digital mammograms using spiculated regions. Biomed. Signal Process. Control 2022, 76, 103652. [Google Scholar] [CrossRef]
- Hussain, S.; Xi, X.; Ullah, I.; Inam, S.A.; Naz, F.; Shaheed, K.; Ali, S.A.; Tian, C. A Discriminative Level Set Method with Deep Supervision for Breast Tumor Segmentation. Comput. Biol. Med. 2022, 149, 105995. [Google Scholar] [CrossRef] [PubMed]
- Kavitha, T.; Mathai, P.P.; Karthikeyan, C.; Ashok, M.; Kohar, R.; Avanija, J.; Neelakandan, S. Deep Learning Based Capsule Neural Network Model for Breast Cancer Diagnosis Using Mammogram Images. Interdiscip. Sci. Comput. Life Sci. 2022, 14, 113–129. [Google Scholar] [CrossRef] [PubMed]
- Zebari, D.A.; Ibrahim, D.A.; Al-Zebari, A. Suspicious Region Segmentation Using Deep Features in Breast Cancer Mammogram Images. In Proceedings of the 2nd 2022 International Conference on Computer Science and Software Engineering, CSASE, Duhok, Iraq, 15–17 March 2022; pp. 253–258. [Google Scholar] [CrossRef]
- Toz, G.; Erdogmus, P. A Novel Hybrid Image Segmentation Method for Detection of Suspicious Regions in Mammograms Based on Adaptive Multi-Thresholding (HCOW). IEEE Access 2021, 9, 85377–85391. [Google Scholar] [CrossRef]
- Ranjbarzadeh, R.; Caputo, A.; Tirkolaee, E.B.; Ghoushchi, S.J.; Bendechache, M. Brain tumor segmentation of MRI images: A comprehensive review on the application of artificial intelligence tools. Comput. Biol. Med. 2023, 152, 106405. [Google Scholar] [CrossRef]
- Jain, D.K.; Jain, R.; Upadhyay, Y.; Kathuria, A.; Lan, X. Deep Refinement: Capsule network with attention mechanism-based system for text classification. Neural Comput. Appl. 2020, 32, 1839–1856. [Google Scholar] [CrossRef]
- Guo, M.-H.; Xu, T.-X.; Liu, J.-J.; Liu, Z.-N.; Jiang, P.-T.; Mu, T.-J.; Zhang, S.-H.; Martin, R.R.; Cheng, M.-M.; Hu, S.-M. Attention mechanisms in computer vision: A survey. Comput. Vis. Media 2022, 8, 331–368. [Google Scholar] [CrossRef]
- Qu, Y.; Baghbaderani, R.K.; Qi, H.; Kwan, C. Unsupervised Pansharpening Based on Self-Attention Mechanism. IEEE Trans. Geosci. Remote. Sens. 2020, 59, 3192–3208. [Google Scholar] [CrossRef]
- Ranjbarzadeh, R.; Kasgari, A.B.; Ghoushchi, S.J.; Anari, S.; Naseri, M.; Bendechache, M. Brain tumor segmentation based on deep learning and an attention mechanism using MRI multi-modalities brain images. Sci. Rep. 2021, 11, 10930. [Google Scholar] [CrossRef] [PubMed]
- Shuang, K.; Ren, X.; Guo, H.; Loo, J.; Xu, P. Effective Strategies for Combining Attention Mechanism with LSTM for Aspect-Level Sentiment Classification. In Advances in Intelligent Systems and Computing; Springer: Berlin/Heidelberg, Germany, 2018; pp. 841–850. [Google Scholar] [CrossRef]
- Yao, H.; Zhang, X.; Zhou, X.; Liu, S. Parallel Structure Deep Neural Network Using CNN and RNN with an Attention Mechanism for Breast Cancer Histology Image Classification. Cancers 2019, 11, 1901. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Zhou, T.; Ruan, S.; Guo, Y.; Canu, S. A Multi-Modality Fusion Network Based on Attention Mechanism for Brain Tumor Segmentation. In Proceedings of the International Symposium on Biomedical Imaging, IEEE Computer Society, Iowa City, IA, USA, 3–7 April 2020; pp. 377–380. [Google Scholar] [CrossRef]
- Liu, H.; Huo, G.; Li, Q.; Guan, X.; Tseng, M.-L. Multiscale lightweight 3D segmentation algorithm with attention mechanism: Brain tumor image segmentation. Expert Syst. Appl. 2023, 214, 119166. [Google Scholar] [CrossRef]
- Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation; Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar] [CrossRef] [Green Version]
- Zhao, J.; Dang, M.; Chen, Z.; Wan, L. DSU-Net: Distraction-Sensitive U-Net for 3D lung tumor segmentation. Eng. Appl. Artif. Intell. 2022, 109, 104649. [Google Scholar] [CrossRef]
- Liu, J.; Wang, J.; Ruan, W.; Lin, C.; Chen, D. Diagnostic and Gradation Model of Osteoporosis Based on Improved Deep U-Net Network. J. Med. Syst. 2019, 44, 15. [Google Scholar] [CrossRef]
- Raza, R.; Bajwa, U.I.; Mehmood, Y.; Anwar, M.W.; Jamal, M.H. dResU-Net: 3D deep residual U-Net based brain tumor segmentation from multimodal MRI. Biomed. Signal Process. Control 2023, 79, 103861. [Google Scholar] [CrossRef]
- Liu, Z.; Song, Y.-Q.; Sheng, V.S.; Wang, L.; Jiang, R.; Zhang, X.; Yuan, D. Liver CT sequence segmentation based with improved U-Net and graph cut. Expert Syst. Appl. 2019, 126, 54–63. [Google Scholar] [CrossRef]
- Khaled, R.; Vidal, J.; Vilanova, J.C.; Martí, R. A U-Net Ensemble for breast lesion segmentation in DCE MRI. Comput. Biol. Med. 2022, 140, 105093. [Google Scholar] [CrossRef]
- Lee, B.; Yamanakkanavar, N.; Choi, J.Y. Automatic segmentation of brain MRI using a novel patch-wise U-net deep architecture. PLoS ONE 2020, 15, e0236493. [Google Scholar] [CrossRef]
- Li, Q.; Fan, S.; Chen, C. An Intelligent Segmentation and Diagnosis Method for Diabetic Retinopathy Based on Improved U-NET Network. J. Med. Syst. 2019, 43, 304. [Google Scholar] [CrossRef] [PubMed]
- Mo, S.; Cho, M.; Shin, J. Freeze the Discriminator: A Simple Baseline for Fine-Tuning GANs. arXiv 2020, arXiv:2002.10964. [Google Scholar]
- Zhang, Z.; Wen, G.; Chen, S. Weld image deep learning-based on-line defects detection using convolutional neural networks for Al alloy in robotic arc welding. J. Manuf. Process. 2019, 45, 208–216. [Google Scholar] [CrossRef]
- Rajapriya, R.; Rajeswari, K.; Thiruvengadam, S.J. Deep learning and machine learning techniques to improve hand movement classification in myoelectric control system. Biocybern. Biomed. Eng. 2021, 41, 554–571. [Google Scholar] [CrossRef]
- Naiemi, F.; Ghods, V.; Khalesi, H. A novel pipeline framework for multi oriented scene text image detection and recognition. Expert Syst. Appl. 2020, 170, 114549. [Google Scholar] [CrossRef]
- Ranjbarzadeh, R.; Jafarzadeh Ghoushchi, S.; Anari, S.; Safavi, S.; Tataei Sarshar, N.; Babaee Tirkolaee, E.; Bendechache, M. A Deep Learning Approach for Robust, Multi-oriented, and Curved Text Detection. Cognit. Comput. 2022, 1, 1–13. [Google Scholar] [CrossRef]
- Parhizkar, M.; Amirfakhrian, M. Recognizing the Damaged Surface Parts of Cars in the Real Scene Using a Deep Learning Framework. Math. Probl. Eng. 2022, 2022, 5004129. [Google Scholar] [CrossRef]
- Sau, P.C. Retinal Blood Vessel Segmentation Using Attention Module and Tversky Loss Function; Springer: Singapore, 2022; Volume 435, pp. 503–513. [Google Scholar] [CrossRef]
- Ke, Z.; Xu, X.; Zhou, K.; Guo, J. A scale-aware UNet++ model combined with attentional context supervision and adaptive Tversky loss for accurate airway segmentation. Appl. Intell. 2023, 53, 18138–18154. [Google Scholar] [CrossRef]
- Nour, M.; Öcal, H.; Alhudhaif, A.; Polat, K. Skin Lesion Segmentation Based on Edge Attention Vnet with Balanced Focal Tversky Loss. Math. Probl. Eng. 2022, 2022, 4677044. [Google Scholar] [CrossRef]
- Kasgari, A.B.; Safavi, S.; Nouri, M.; Hou, J.; Sarshar, N.T.; Ranjbarzadeh, R. Point-of-Interest Preference Model Using an Attention Mechanism in a Convolutional Neural Network. Bioengineering 2023, 10, 495. [Google Scholar] [CrossRef]
- Ranjbarzadeh, R.; Ghoushchi, S.J.; Sarshar, N.T.; Tirkolaee, E.B.; Ali, S.S.; Kumar, T.; Bendechache, M. ME-CCNN: Multi-encoded images and a cascade convolutional neural network for breast tumor segmentation and recognition. Artif. Intell. Rev. 2023, 1–38. [Google Scholar] [CrossRef]
- Valizadeh, A.; Ghoushchi, S.J.; Ranjbarzadeh, R.; Pourasad, Y. Presentation of a Segmentation Method for a Diabetic Retinopathy Patient’s Fundus Region Detection Using a Convolutional Neural Network. Comput. Intell. Neurosci. 2021, 2021, 7714351. [Google Scholar] [CrossRef] [PubMed]
- Ghoushchi, S.J.; Ranjbarzadeh, R.; Dadkhah, A.H.; Pourasad, Y.; Bendechache, M. An Extended Approach to Predict Retinopathy in Diabetic Patients Using the Genetic Algorithm and Fuzzy C-Means. BioMed Res. Int. 2021, 2021, 5597222. [Google Scholar] [CrossRef]
- Ranjbarzadeh, R.; Dorosti, S.; Ghoushchi, S.J.; Safavi, S.; Razmjooy, N.; Sarshar, N.T.; Anari, S.; Bendechache, M. Nerve optic segmentation in CT images using a deep learning model and a texture descriptor. Complex Intell. Syst. 2022, 8, 3543–3557. [Google Scholar] [CrossRef]
- The Mini-MIAS Database of Mammograms. Available online: http://peipa.essex.ac.uk/info/mias.html (accessed on 8 October 2022).
- Tsochatzidis, L.; Costaridou, L.; Pratikakis, I. Deep Learning for Breast Cancer Diagnosis from Mammograms—A Comparative Study. J. Imaging 2019, 5, 37. [Google Scholar] [CrossRef] [Green Version]
- Lei, Y.; He, X.; Yao, J.; Wang, T.; Wang, L.; Li, W.; Curran, W.J.; Liu, T.; Xu, D.; Yang, X. Breast tumor segmentation in 3D automatic breast ultrasound using Mask scoring R-CNN. Med. Phys. 2021, 48, 204–214. [Google Scholar] [CrossRef] [PubMed]
- Oyelade, O.N.; Ezugwu, A.E. A novel wavelet decomposition and transformation convolutional neural network with data augmentation for breast cancer detection using digital mammogram. Sci. Rep. 2022, 12, 5913. [Google Scholar] [CrossRef] [PubMed]
- Nalepa, J.; Marcinkiewicz, M.; Kawulok, M. Data Augmentation for Brain-Tumor Segmentation: A Review. Front. Comput. Neurosci. 2019, 13, 83. [Google Scholar] [CrossRef] [Green Version]
- Wang, F.; Zhong, S.-H.; Peng, J.; Jiang, J.; Liu, Y. Data Augmentation for EEG-Based Emotion Recognition with Deep Convolutional Neural Networks; Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Berlin/Heidelberg, Germany, 2018; pp. 82–93. [Google Scholar] [CrossRef]
- Shorten, C.; Khoshgoftaar, T.M. A survey on Image Data Augmentation for Deep Learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef] [Green Version]
- Zeiser, F.A.; da Costa, C.A.; Zonta, T.; Marques, N.M.C.; Roehe, A.V.; Moreno, M.; da Righi, R.R. Segmentation of Masses on Mammograms Using Data Augmentation and Deep Learning. J. Digit. Imaging 2020, 33, 858–868. [Google Scholar] [CrossRef]
- Dvornik, N.; Mairal, J.; Schmid, C. On the Importance of Visual Context for Data Augmentation in Scene Understanding. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 2014–2028. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Singh, A.; Ranjbarzadeh, R.; Raj, K.; Kumar, T.; Roy, A.M. Understanding EEG signals for subject-wise Definition of Armoni Activities. arXiv 2023, arXiv:2301.00948. [Google Scholar]
- Mousavi, S.M.; Asgharzadeh-Bonab, A.; Ranjbarzadeh, R. Time-Frequency Analysis of EEG Signals and GLCM Features for Depth of Anesthesia Monitoring. Comput. Intell. Neurosci. 2021, 2021, 8430565. [Google Scholar] [CrossRef] [PubMed]
- Haseli, G.; Ranjbarzadeh, R.; Hajiaghaei-Keshteli, M.; Ghoushchi, S.J.; Hasani, A.; Deveci, M.; Ding, W. HECON: Weight assessment of the product loyalty criteria considering the customer decision's halo effect using the convolutional neural networks. Inf. Sci. 2023, 623, 184–205. [Google Scholar] [CrossRef]
- Peng, C.; Zhang, Y.; Meng, Y.; Yang, Y.; Qiu, B.; Cao, Y.; Zheng, J. LMA-Net: A lesion morphology aware network for medical image segmentation towards breast tumors. Comput. Biol. Med. 2022, 147, 105685. [Google Scholar] [CrossRef]
- Kaitouni, S.E.I.E.; Abbad, A.; Tairi, H. A breast tumors segmentation and elimination of pectoral muscle based on hidden markov and region growing. Multimedia Tools Appl. 2018, 77, 31347–31362. [Google Scholar] [CrossRef]
- Safavi, S.; Jalali, M. RecPOID: POI Recommendation with Friendship Aware and Deep CNN. Futur. Internet 2021, 13, 79. [Google Scholar] [CrossRef]
- Safavi, S.; Jalali, M. DeePOF: A hybrid approach of deep convolutional neural network and friendship to Point-of-Interest (POI) recommendation system in location-based social networks. Concurr. Comput. Pract. Exp. 2022, 34, e6981. [Google Scholar] [CrossRef]
- Ranjbarzadeh, R.; Saadi, S.B. Corrigendum to “Automated liver and tumor segmentation based on concave and convex points using fuzzy c-means and mean shift clustering” [Measurement 150 (2020) 107086]. Measurement 2020, 151, 107230. [Google Scholar] [CrossRef]
- Saadi, S.B.; Sarshar, N.T.; Sadeghi, S.; Ranjbarzadeh, R.; Forooshani, M.K.; Bendechache, M. Investigation of Effectiveness of Shuffled Frog-Leaping Optimizer in Training a Convolution Neural Network. J. Health Eng. 2022, 2022, 4703682. [Google Scholar] [CrossRef]
- Anari, S.; Sarshar, N.T.; Mahjoori, N.; Dorosti, S.; Rezaie, A. Review of Deep Learning Approaches for Thyroid Cancer Diagnosis. Math. Probl. Eng. 2022, 2022, 5052435. [Google Scholar] [CrossRef]
Structures | Sensitivity | Accuracy | Specificity | ||||||
---|---|---|---|---|---|---|---|---|---|
Normal | Malignant | Benign | Normal | Malignant | Benign | Normal | Malignant | Benign | |
Deep Supervision [16] | 0.90 | 0.85 | 0.87 | 0.89 | 0.83 | 0.88 | 0.91 | 0.87 | 0.90 |
Capsule Neural Network [17] | 0.91 | 0.87 | 0.89 | 0.91 | 0.88 | 0.89 | 0.90 | 0.86 | 0.88 |
Deep Features [18] | 0.87 | 0.83 | 0.85 | 0.89 | 0.84 | 0.86 | 0.88 | 0.85 | 0.86 |
HCOW [19] | 0.89 | 0.84 | 0.87 | 0.88 | 0.86 | 0.88 | 0.89 | 0.86 | 0.87 |
Our model | 0.93 | 0.90 | 0.91 | 0.94 | 0.89 | 0.90 | 0.94 | 0.91 | 0.94 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zarbakhsh, P. Spatial Attention Mechanism and Cascade Feature Extraction in a U-Net Model for Enhancing Breast Tumor Segmentation. Appl. Sci. 2023, 13, 8758. https://doi.org/10.3390/app13158758
Zarbakhsh P. Spatial Attention Mechanism and Cascade Feature Extraction in a U-Net Model for Enhancing Breast Tumor Segmentation. Applied Sciences. 2023; 13(15):8758. https://doi.org/10.3390/app13158758
Chicago/Turabian StyleZarbakhsh, Payam. 2023. "Spatial Attention Mechanism and Cascade Feature Extraction in a U-Net Model for Enhancing Breast Tumor Segmentation" Applied Sciences 13, no. 15: 8758. https://doi.org/10.3390/app13158758
APA StyleZarbakhsh, P. (2023). Spatial Attention Mechanism and Cascade Feature Extraction in a U-Net Model for Enhancing Breast Tumor Segmentation. Applied Sciences, 13(15), 8758. https://doi.org/10.3390/app13158758