Next Article in Journal
Structured Summarization of League of Legends Match Data Optimized for Large Language Model Input
Previous Article in Journal
Detecting Urban Mobility Structure and Learning Functional Distribution with Multi-Scale Features
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Effective Breast Cancer Classification Using Deep MLP, Feature-Fused Autoencoder and Weight-Tuned Decision Tree

by
Nagham Rasheed Hameed Alsaedi
* and
Mehmet Fatih Akay
Department of Computer Engineering, Faculty of Engineering, Çukurova University, Adana 34000, Turkey
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(13), 7213; https://doi.org/10.3390/app15137213
Submission received: 21 May 2025 / Revised: 12 June 2025 / Accepted: 20 June 2025 / Published: 26 June 2025

Abstract

Breast cancer remains a leading cause of death among women worldwide, underscoring the urgent need for practical diagnostic tools. This paper presents an advanced machine learning algorithm designed to improve classification accuracy in breast cancer diagnosis. The system integrates a deep multi-layer perceptron (Deep MLP) for feature extraction, a feature-fused autoencoder for efficient dimensional reduction, and a weight-tuned decision-tree classifier optimized via cross-validation and square weight adjustment. The proposed method was rigorously tested using the Wisconsin breast cancer dataset, employing k-fold cross-validation to ensure robustness and generalizability. Key performance indicators, including accuracy, precision, recall, F1-score, and area under the curve (AUC), were used to evaluate the model’s ability to distinguish between malignant and benign tumors. Our results suggest that this combination model outperforms traditional classification methods, with high accuracy and robust performance across data partitions. The main contribution of this research is the development of a new framework for deep learning. Auto-encoder and decision tree results show that this system has strong potential to improve breast cancer diagnosis, offering physicians a reliable and effective tool.

1. Introduction

Breast cancer is the leading cause of cancer-related deaths among women internationally [1,2,3,4,5], making its early detection and correct classification important for developing effective remedies and improving outcomes for people afflicted with this disease [6,7,8,9,10]. Traditional diagnostic strategies, including mammography, biopsies, and scientific examinations, although powerful, are not without limitations. These techniques can be time-consuming, invasive, and prone to human error, resulting in delays in diagnosis and treatment. Recent advancements in our understanding of machine learning (ML) and artificial intelligence (AI) have provided promising alternatives to standard diagnostic techniques. ML algorithms can examine massive volumes of clinical records, learn different styles, and provide correct classifications, thereby aiding in the early detection of breast cancer. Among the many machine learning models and techniques, neural networks, decision trees, and autoencoders have demonstrated uniquely significant capabilities in medical diagnostics. Researchers aim to leverage the strengths of these strategies by integrating them into a comprehensive framework for breast cancer classification, as demonstrated in this work.
In this study, we aimed to develop an efficient machine learning (ML) framework that combines a deep multilayer perceptron (Deep MLP) for function extraction, a feature-fused autoencoder for dimensionality reduction, and a weight-tuned cross-validated decision tree for final classification. Each element of the developed framework plays a crucial role in enhancing the framework’s overall performance. The Deep MLP is a sort of synthetic neural network that includes more than one layer of neurons, including input, hidden, and output layers. It is particularly powerful in extracting complex features from information it has been given; some of these features could be crucial for accurate classification. By leveraging its deep knowledge of techniques, Deep MLP can identify challenging patterns within data that may be indicative of malignant or benign breast cancers [11,12,13,14,15]. An autoencoder is an unmonitored neural network designed to study the encoding of entered data. This study introduces a feature-fused autoencoder that not only reduces the dimensionality of the input information but also enhances its representation, thereby gaining insight into a given method. The fusion mechanism in the autoencoder integrates characteristic extraction and dimensionality reduction, resulting in a more compact and informative representation of the facts. This step is crucial for reducing computational complexity and enhancing the efficiency of the subsequent classification level. Decision trees are extensively used for classification due to their simplicity and interpretability. However, their overall performance may be substantially enhanced through weight tuning and cross-validation. The Weight-Tuned Cross-Validated Decision Tree in our framework adjusts the weights of various training sets to handle class imbalance. It employs cross-validation to ensure a robust and unbiased evaluation of the model. This method ensures that the final version is accurate and generalizable. To evaluate the effectiveness of the proposed framework, we employed k-fold cross-validation. This robust technique divides a dataset into k subsets, allowing for multiple rounds of training and testing. This method provides a comprehensive assessment of a model’s overall performance across distinct record splits, thereby reducing the likelihood of overfitting and ensuring the reliability of the effects. Performance metrics, including accuracy, precision, F1-rating, and Area Under the Curve (AUC), were used to quantify the model’s effectiveness in classifying breast cancer. The principal contributions of this examination encompass the improvement of a singular ML framework that integrates Deep MLP, a feature-fused autoencoder, and a Weight-Tuned Cross-Validated Decision Tree for breast cancer classification; a comprehensive assessment of the proposed framework via the Wisconsin breast cancer dataset, demonstrating advanced performance across numerous training and testing splits; and an in-depth analysis of the model’s overall performance using more than one metric, highlighting its robustness and accuracy. By integrating current deep MLP techniques into a cohesive and practical framework, this study aims to contribute to ongoing efforts to improve breast cancer prognosis and ultimately enhance patient care and outcomes.

2. Materials and Methods

In this study, we used advanced machine learning techniques to classify breast cancer using the Wisconsin Breast Cancer Dataset. Our method involves data preprocessing, feature extraction using a Deep Multi-Layer Perceptron (Deep MLP), dimensionality reduction with a feature-fused autoencoder, and finally classification via a Weight-Tuned Decision Tree. Figure 1 illustrates the compound algorithm, which consists of these three sub-algorithms. Each step is carefully designed to ensure efficiency and accuracy.

2.1. Data Preprocessing

Data preprocessing is crucial for ensuring data quality and relevance. The Wisconsin Breast Cancer Dataset consists of 569 cases with 32 attributes. Steps in preprocessing include cleaning the data by identifying and removing missing or irrelevant data points, normalizing and scaling features to ensure equal contributions to model training, and partitioning the data, which involves dividing the dataset into training and testing subsets with varying partitions (from 10:90 to 45:55) to assess robustness.

2.2. Feature Extraction with Deep MLP

The Deep Multi-Layer Perceptron (Deep MLP) with feature extraction was used in this work due to its ability to capture complex patterns through multiple hidden layers. The Deep MLP steps and training stages are as follows:
  • Architecture Design: The network has many hidden layers designed to capture complex data structures.
  • Starting Weights: Starting weights are used to start the training program.
  • Activation function: A Rectified Linear Unit (ReLU) is used for nonlinear conversion.
  • Batch normalization: Layers are added to normalize the data within the network, thereby accelerating training.
  • Optimization: An adaptive moment estimation (ADAM) optimizer was used for dynamic learning rate variation.
  • Hyperparameter tuning: The number of layers and neurons, the dropout rate, etc., are changed
  • Loss functions: Log-loss or squared loss functions are used to measure performance.
  • Training and Validation: Training losses are monitored over epochs, and measures are implemented to prevent overfitting, such as dropout and early dropout.
Mathematically, the deep MLP for feature extraction can be described as follows [11]:
Let the input data be
X = x ( 1 ) , x ( 2 ) , , x ( N ) , x ( i ) R d
Pass each input through a deep MLP:
h ( i ) = f M L P x ( i ) = σ L W L σ 1 W 1 x ( i ) + b 1 + + b L
where
  • σ j denotes the activation function (e.g., ReLU);
  • W j , b j are weights and biases in layer j;
  • h ( i ) R d d denotes the extracted feature vector.

2.3. Dimensionality Reduction with Feature-Fused Autoencoder

To reduce dimensionality and preserve essential features, a feature-fused autoencoder was employed. This advanced autoencoder differs from traditional ones in that it optimizes the products for better performance:
  • Compression: The input vector is reduced to a minimum.
  • Reconstruction: An attempt is made to reconstruct the input from the compressed representation, reducing reconstruction errors.
  • Feature Fusion: Low-level features are combined to discard redundant data and hide important information.
Mathematically, the feature-fused autoencoder can be described as follows [16]:
Let the extracted features be
H = h ( 1 ) , h ( 2 ) , , h ( N )
The autoencoder encodes each feature vector into a lower-dimensional fused latent vector:
Encoder:
z ( i ) = f e n c h ( i ) R d z , d z < d
Fusion Layer (e.g., attention or weighted average of multi-layer features):
z f ( i ) = f z ( i ) , (e.g., self-attention, max pooling, etc.)
Decoder (optional, for unsupervised training):
h ^ ( i ) = f d e c z f ( i )
Loss (if unsupervised reconstruction is used):
L A E = 1 N i = 1 N   h ( i ) h ^ ( i ) 2
Now, z f ( i ) denotes the dimensionality-reduced features.

2.4. Classification of a Weight-Tuned Decision Tree

To perform the final classification, a Weight-Tuned Decision Tree was used, which provides a definition and efficiently handles imbalanced learning.
  • Weight adjustment: Weights are assigned to classes to deal with unbalanced datasets.
  • Tree construction: A decision tree based is constructed based on feature partitioning to maximize information gain.
  • Cross-validation: K-fold cross-validation is used to ensure the robustness of the model and reduce the effect of data-partitioning randomness.
  • Performance averaging: The average of the results across the folds is calculated to obtain a reliable performance estimate.
Mathematically, the Weight-Tuned Decision Tree can be described as follows [11]:
Train a decision tree classifier on reduced features z f ( i ) .
Let y ( i ) { 0,1 } be the label (e.g., benign/malignant).
The classifier is trained as follows:
y ^ ( i ) = f tree   z f ( i ) ; θ
where
  • f tree   is a decision tree;
  • θ includes split criteria and sample weights tuned via cross-validation:
    θ = a r g m i n θ   L tree   + λ Ω ( θ )
  • L tree denotes classification loss (e.g., Gini impurity or entropy);
  • Ω ( θ ) denotes regularization (e.g., tree depth, number of leaves, etc.);
  • λ denotes regularization strength.
The best θ is selected via cross-validation, optimizing performance metrics like accuracy, F1-score, or AUC.
The final prediction pipeline, including all the algorithms above, is as follows:
x ( i )   Deep   MLP   h ( i )   FFA   z f ( i )   WT - CVDT   y ^ ( i )

2.5. Cross-Validation and Performance Metrics

The model proposed in this work was validated using k-fold cross-validation, wherein the dataset was divided into k subsets. In this process, each subset serves as a test set once, while the others are used as training sets, ensuring unbiased validation of the model’s performance.

2.6. Performance Metrics

The following matrices represent the performance matrices adopted in this work: [13,17,18,19,20,21]
  • Accuracy: The ratio of correctly classified instances to the total instances.
    A C U = T P + T N T P + F P + T N + F N
    where T P is a true positive, T N is a true negative, F P is a false positive, and F N is a false negative.
  • Precision: The ratio of true-positive predictions to the total positive predictions.
    P R = T P T P + F P
  • Recall (Sensitivity): The ratio of true-positive predictions to all actual positive instances.
    S E N = T P T P + F N
  • F1-Score: The harmonic means of precision and recall.
  • AUC (area under the curve): The area under the receiver operating characteristic (ROC) curve, indicating discriminative ability.

2.7. Implementation and Evaluation

The proposed methodology was implemented and evaluated using the Wisconsin Breast Cancer Dataset. Various training and testing splits were used to test robustness. The results, which include metrics such as accuracy, precision, recall, F1-score, and AUC, demonstrate the high performance and reliability of our proposed model in classifying breast cancer.

2.8. Clinical Deployment Considerations

To evaluate the model’s potential for real-world use, we assessed its suitability for clinical deployment based on three key criteria: real-time performance, computational requirements, and system compatibility.

2.8.1. Real-Time Performance

The proposed system exhibits fast inference times due to its use of an optimized Deep MLP and a lightweight autoencoder structure. On a standard machine (Intel i7 CPU, 16 GB RAM, NVIDIA GTX 1660 GPU), the average prediction time per sample is about 8 milliseconds. This meets real-time diagnostic needs in most clinical workflows. Batch processing can further boost throughput when analyzing multiple patient records.

2.8.2. Computational Resource Requirements

Unlike more resource-intensive deep learning models, such as large convolutional neural networks or transformers, our architecture strikes a balance between accuracy and efficiency. The dimensionality reduction achieved through the autoencoder significantly reduces the feature space, thereby lowering the computational load for the decision tree classifier. This allows our model to be used on mid-range clinical hardware without requiring cloud services or GPU acceleration.

2.8.3. Compatibility with Medical Systems

For integration into clinical systems, the model can be embedded in electronic health record (EHR) platforms or decision support systems. It supports standardized input formats, such as CSV and JSON, making it easy to apply to data from mammography reports or lab results. Additionally, its modular design promotes adaptability to HL7/FHIR standards, enhancing interoperability across hospital information systems.

3. Results

This section presents a detailed evaluation of the results obtained from a complex algorithm for breast cancer diagnoses, which comprises a Deep MLP, a feature-fused autoencoder neural network, and a weight-tuned cross-validation decision tree. The hyperparameters used here were a batch size of 128, 256 hidden units, and a dropout of 0.45. The proposed model was applied to the Wisconsin Breast Cancer Database. Four evaluation measures were utilized to assess the prediction performance of the model: Accuracy (99.47%), Recall (99%), F1 Score (99%), and Precision (99%).
Furthermore, the AUC values (the area under the ROC curve), illustrated in Figure 2a, were always 0.996, highlighting the model’s ability to discriminate across different thresholds. The confusion matrix shown in Figure 2b indicates balanced classification performance, which reduces false positives and negatives. Comparison with traditional machine learning techniques, which include Support Vector Machines (SVMs), Random Forests, and k-Nearest Neighbors (KNN), underscores the performance of the proposed model. It outperformed these strategies in terms of accuracy, precision, and overall performance balance. This superiority is attributed to the advanced feature extraction skills of the Deep MLP and the green dimensionality reduction facilitated by using the feature-fused autoencoder. Additionally, the Weight-Tuned Decision Tree classification method contributed to the model’s robustness, mainly in addressing classification imbalances. This can also be attributed to the use of superior technology in gaining knowledge of strategies for enhancing the accuracy of breast cancer classification.
Key factors that contributed to the model’s success included comprehensive validation through K-fold cross-validation, careful hyperparameter optimization, and the use of superior neural network architectures. The study’s findings suggest that the proposed model is a valuable diagnostic tool for clinical experts, and as such, may also facilitate early detection and improved outcomes for individuals affected by this condition. Future endeavors could focus on enhancing this model by exploring alternative neural network architectures and incorporating additional data resources, including genetic statistics and medical imaging records.

Comparison Results

Table 1 presents the performance measures derived from other studies, which are used for comparison with the proposed model’s performance to demonstrate its effectiveness. It also presents the accuracy results obtained using various approaches that incorporate machine learning models, as applied by different authors.
1. Md. Milon Islam et al. [22] applied five matrices to support vector machines (SVMs): KNN, RF, ANN, and LR. Both studies utilized the WDBC dataset.
Table 1 and Figure 3 show a comparison between the methods discussed above. The proposed method exhibits enhanced performance parameters.
2. Nirdosh Kumar [23] used the LR, SVM, KNN, and Naïve Bayes algorithms. The performance parameters regarding the WDBC dataset are presented in Table 2 and illustrated in Figure 4. The proposed model achieved higher accuracy and recall compared to LR, KNN, and Naïve Bayes, while it exhibited the same value as SVM.
3. The performance parameters from the study by G. Battineni et al. [24] and from this work are shown in Table 3 and Figure 5. The proposed model exhibited greater accuracy, recall, and AUC compared to SVM, LR, and KNN.
4. P. Karatza et al. [25] used RF, an NN, and an ENN for the diagnosis of breast cancer based on the WDBC dataset. This study was compared with the proposed work. All the performance parameters of the proposed model are greater. Table 4 and Figure 6 show this comparison.
5. Table 5 and Figure 7 show the performance parameters from the study presented by A. I. Aishwarja et al. [26] and those from this work. The proposed model has the highest values for all the parameters.
6. We compared our work with the work of S. Sakib et al. [27], who applied a SVM, DT, LR, RF, KNN, and an NN to the WDBC dataset. The accuracy, recall, and F1 score are all higher for the proposed model, as shown in Table 6 and Figure 8.
7. The performance parameters from the study presented by V. Nemade and V. Fegade [28] and those from this work are shown in Table 7 and Figure 9 and Figure 10. The model exhibited greater accuracy, precision, recall, and F1 score values than the KNN, SVM, DT, Naïve Bayes, and LR algorithms.

4. Discussion

Individually, each algorithm from previous studies provides performance parameters with some limitations that reduce the ability to diagnose breast cancer.
The diagnosis of breast cancer requires high performance, including in terms of accuracy, precision, and F1 score, among other metrics.
This study presents an integrated method that combines three advanced algorithms, which work together to achieve very high performance and provide precise diagnoses.
Figure 2a illustrates the receiver operating characteristic (ROC) curve with an area under the curve (AUC) of about 0.996, which is exceptionally high, indicating that the model performs extremely well in distinguishing between malignant (dangerous) and benign (typical) cases. The ROC curve hugs the top-left corner, which is ideal, as this shows a very low false-positive rate and a very high true-positive rate. This means the classifier is highly capable of detecting breast cancer with minimal errors. This demonstrates that the classification model used in the breast cancer diagnosis task is highly effective, with near-perfect discriminatory power.
Figure 2b is a Confusion Matrix that consists of 350 true positives (TPs) (correctly identified cancer cases), 7 false negatives (FNs) (cancer cases misclassified as benign instances), 208 true negatives (TNs) (correctly identified benign cases), and 4 false positives (FPs) (normal cases misclassified as cancer).
As shown by this Confusion Matrix, very few misclassifications occurred for false positives and false negatives. The model is powerful in identifying cancer cases (showing high sensitivity) and minimizing false alarms (showing high specificity).
The model’s high sensitivity (recall) of approximately 99.6% ensures that most cancer patients are correctly identified, thereby minimizing missed diagnoses.
A low false-positive rate helps avoid unnecessary stress and treatment for patients without cancer.
The results show that the breast cancer classification model is highly accurate and reliable, making it well-suited for real-world diagnostic use. The combination of an AUC = 0.996 and strong confusion matrix metrics indicates that the model achieves an excellent balance between sensitivity and specificity, a critical aspect in medical diagnosis.
Furthermore, comparisons between the proposed integrated algorithm and other algorithms from previous studies, such as SVM, KNN, RF, ANN, LR, NN, ENN, DT, and Naïve Bayes, showed the superiority of the proposed integrated algorithm for all performance parameters investigated.

5. Limitations and Future Work

Although the proposed framework performs well on the Wisconsin Breast Cancer Dataset, several limitations still exist. The model was trained and tested on just one benchmark dataset. This restricts its applicability to other medical conditions or more diverse data distributions. Future work will focus on testing the model across multiple datasets to evaluate its reliability and potential for use in wider clinical settings.
Another significant limitation is the current lack of explainability features. In clinical settings, understanding how and why a model makes a specific prediction is crucial for gaining trust of healthcare professionals. While our aim in this work was to improve classification accuracy, future updates will incorporate explainable AI techniques, such as SHAP (Shapley Additive Explanations) and LIME (Local Interpretable Model-Agnostic Explanations). These will provide clear insights into model decisions and feature contributions.

6. Conclusions

In this study, we designed a comprehensive method for breast cancer diagnosis by combining a Deep Multilayer Perceptron (MLP), a feature-based autoencoder neural network, and a weighted cross-validation decision tree. The model achieved high performance in key parameters, including 99.47% accuracy, 99% precision, 99% recall, and a 99% F1 score, as well as an AUC-ROC value of 0.996, demonstrating its strong discriminative power. The confusion matrix indicates balanced classification performance with minimal false positives and negatives. Efficient parameter tuning and convergence behavior were also confirmed. In summary, the proposed algorithm is a robust and accurate diagnostic tool for breast cancer.
However, our findings are currently limited to a single dataset. To validate generalizability, future work will involve testing on larger and more diverse datasets, including those concerning other cancer types. The proposed model not only delivers high classification performance but is also lightweight and adaptable for integration into existing clinical infrastructures. It constitutes a promising tool for real-time decision support in breast cancer diagnosis.

Author Contributions

Software, N.R.H.A. and M.F.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Frick, C.; Rumgay, H.; Vignat, J.; Ginsburg, O.; Nolte, E.; Bray, F.; Soerjomataram, I. Quantitative estimates of preventable and treatable deaths from 36 cancers worldwide: A population-based study. Lancet Glob. Health 2023, 11, e1700–e1712. [Google Scholar] [CrossRef] [PubMed]
  2. Sedeta, E.T.; Jobre, B.; Avezbakiyev, B. Breast Cancer: Global Patterns of Incidence, Mortality, and Trends; American Society of Clinical Oncology: Alexandria, VA, USA, 2023. [Google Scholar]
  3. Lukong, K.E. Understanding breast cancer–The long and winding road. BBA Clin. 2017, 7, 64–77. [Google Scholar] [CrossRef] [PubMed]
  4. Lei, S.; Zheng, R.; Zhang, S.; Wang, S.; Chen, R.; Sun, K.; Zeng, H.; Zhou, J.; Wei, W. Global patterns of breast cancer incidence and mortality: A population-based cancer registry data analysis from 2000 to 2020. Cancer Commun. 2021, 41, 1183–1194. [Google Scholar] [CrossRef] [PubMed]
  5. Lima, S.M.; Kehm, R.D.; Terry, M.B. Global breast cancer incidence and mortality trends by region, age-groups, and fertility patterns. EClinicalMedicine 2021, 38, 100985. [Google Scholar] [CrossRef] [PubMed]
  6. World Health Organization. Global Breast Cancer Initiative Implementation Framework: Assessing, Strengthening and Scaling-up of Services for the Early Detection and Management of Breast Cancer; World Health Organization: Geneva, Switzerland, 2023. [Google Scholar]
  7. Tsarouchi, M.I.; Hoxhaj, A.; Mann, R.M. New Approaches and Recommendations for Risk-Adapted Breast Cancer Screening. J. Magn. Reson. Imaging 2023, 58, 987–1010. [Google Scholar] [CrossRef] [PubMed]
  8. Thangavel, Y.; Garg, H.; Alagarsamy, M.; Pradeep, D. Revolutionizing breast cancer diagnosis with a comprehensive approach using digital mammogram-based feature extraction and selection for early-stage identification. Biomed. Signal Process. Control 2024, 94, 106268. [Google Scholar] [CrossRef]
  9. Ben Ammar, M.; Ayachi, F.L.; Cardoso de Paiva, A.; Ksantini, R.; Mahjoubi, H. Harnessing Deep Learning for Early Breast Cancer Diagnosis: A Review of Datasets, Methods, Challenges, and Future Directions. Int. J. Comput. Digit. Syst. 2024, 16, 1643–1661. [Google Scholar] [CrossRef] [PubMed]
  10. Khalid, A.; Mehmood, A.; Alabrah, A.; Alkhamees, B.F.; Amin, F.; AlSalman, H.; Choi, G.S. Breast cancer detection and prevention using machine learning. Diagnostics 2023, 13, 3113. [Google Scholar] [CrossRef] [PubMed]
  11. Goodfellow, I. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  12. Yu, Y.; Zhang, Y. Multi-layer Perceptron Trainability Explained via Variability. arXiv 2021, arXiv:2105.08911. [Google Scholar]
  13. Bachmann, G.; Anagnostidis, S.; Hofmann, T. Scaling MLPS: A tale of inductive bias. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2024; p. 36. [Google Scholar]
  14. Wu, Y.; Liu, L.; Bae, J.; Chow, K.-H.; Iyengar, A.; Pu, C.; Wei, W.; Yu, L.; Zhang, Q. Demystifying learning rate policies for high accuracy training of deep neural networks. In Proceedings of the 2019 IEEE International Conference on Big Data (Big Data), Los Angeles, CA, USA, 9–12 December 2019. [Google Scholar]
  15. Nielsen, M.A. Neural Networks and Deep Learning; Determination Press: San Francisco, CA, USA, 2015; Volume 25. [Google Scholar]
  16. Baltrušaitis, T.; Ahuja, C.; Morency, L.-P. Multimodal machine learning: A survey and taxonomy. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 423–443. [Google Scholar] [CrossRef] [PubMed]
  17. Gupta, K.; Chawla, N. Analysis of histopathological images for prediction of breast cancer using traditional classifiers with pre-trained CNN. Procedia Comput. Sci. 2020, 167, 878–889. [Google Scholar] [CrossRef]
  18. Johnson, J.M.; Khoshgoftaar, T.M. Survey on deep learning with class imbalance. J. Big Data 2019, 6, 27. [Google Scholar] [CrossRef]
  19. Laghmati, S.; Cherradi, B.; Tmiri, A.; Daanouni, O.; Hamida, S. Classification of patients with breast cancer using neighbourhood component analysis and supervised machine learning techniques. In Proceedings of the 2020 3rd International Conference on Advanced Communication Technologies and Networking (CommNet), In Virtual, 4–6 September 2020. [Google Scholar]
  20. Anvesh, M. Machine Learning Approaches for Breast Cancer Diagnosis and Their Comparison. Bachelor’s Thesis, National Institute of Technology Rourkela, Rourkela, India, 2014. [Google Scholar]
  21. Khourdifi, Y.; El Alami, A.; Zaydi, M.; Maleh, Y.; Er-Remyly, O. Early Breast Cancer Detection Based on Deep Learning: An Ensemble Approach Applied to Mammograms. BioMedInformatics 2024, 4, 2338–2373. [Google Scholar] [CrossRef]
  22. Islam, M.M.; Haque, M.R.; Iqbal, H.; Hasan, M.M.; Hasan, M.; Kabir, M.N. Breast cancer prediction: A comparative study using machine learning techniques. SN Comput. Sci. 2020, 1, 290. [Google Scholar] [CrossRef]
  23. Kumar, N.; Sharma, G.; Bhargava, L. The machine learning based optimized prediction method for breast cancer detection. In Proceedings of the 2020 4th International Conference on Electronics, Communication and Aerospace Technology (ICECA), Coimbatore, India, 5–7 November 2020. [Google Scholar]
  24. Battineni, G.; Chintalapudi, N.; Amenta, F. Performance analysis of different machine learning algorithms in breast cancer predictions. EAI Endorsed Trans. Pervasive Health Technol. 2020, 6. [Google Scholar] [CrossRef]
  25. Karatza, P.; Dalakleidi, K.; Athanasiou, M.; Nikita, K.S. Interpretability methods of machine learning algorithms with applications in breast cancer diagnosis. In Proceedings of the 2021 43rd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), In Virtual, 1–5 November 2021. [Google Scholar]
  26. Aishwarja, A.I.; Eva, N.J.; Mushtary, S.; Tasnim, Z.; Khan, N.I.; Islam, M.N. Exploring the machine learning algorithms to find the best features for predicting the breast cancer and its recurrence. In Intelligent Computing and Optimization: Proceedings of the 3rd International Conference on Intelligent Computing and Optimization 2020 (ICO 2020), Hua Hin, Thailand, 17–18 December 2020; Springer: Berlin/Heidelberg, Germany, 2021. [Google Scholar]
  27. Sakib, S.; Yasmin, N.; Tanzeem, A.K.; Shorna, F.; Md. Hasib, K.; Alam, S.B. Breast cancer detection and classification: A comparative analysis using machine learning algorithms. In Proceedings of Third International Conference on Communication, Computing and Electronics Systems: ICCCES 2021, Coimbatore, India, 28–29 October 2021; Springer: Berlin/Heidelberg, Germany, 2022. [Google Scholar]
  28. Nemade, V.; Fegade, V. Machine Learning Techniques for Breast Cancer Prediction. Procedia Comput. Sci. 2023, 218, 1314–1320. [Google Scholar] [CrossRef]
Figure 1. The comprehensive breast cancer diagnosis algorithm comprises a deep MLP, a feature-fused autoencoder neural network, and a weight-tuned cross-validation decision tree.
Figure 1. The comprehensive breast cancer diagnosis algorithm comprises a deep MLP, a feature-fused autoencoder neural network, and a weight-tuned cross-validation decision tree.
Applsci 15 07213 g001
Figure 2. (a) The optimal ROC curve, and (b) the optimal confusion matrix.
Figure 2. (a) The optimal ROC curve, and (b) the optimal confusion matrix.
Applsci 15 07213 g002
Figure 3. Performance parameters for different matrices in comparison with the proposed approach.
Figure 3. Performance parameters for different matrices in comparison with the proposed approach.
Applsci 15 07213 g003
Figure 4. The performance parameters for different models vs. those for the proposed model.
Figure 4. The performance parameters for different models vs. those for the proposed model.
Applsci 15 07213 g004
Figure 5. The performance parameters for different models vs. those for the proposed model.
Figure 5. The performance parameters for different models vs. those for the proposed model.
Applsci 15 07213 g005
Figure 6. The performance parameters for different models vs. those for the proposed model.
Figure 6. The performance parameters for different models vs. those for the proposed model.
Applsci 15 07213 g006
Figure 7. The performance parameters for different models vs. those for the proposed model.
Figure 7. The performance parameters for different models vs. those for the proposed model.
Applsci 15 07213 g007
Figure 8. The performance parameters for different models vs. those for the proposed model.
Figure 8. The performance parameters for different models vs. those for the proposed model.
Applsci 15 07213 g008
Figure 9. Performance parameters of different matrices in comparison with the proposed approach.
Figure 9. Performance parameters of different matrices in comparison with the proposed approach.
Applsci 15 07213 g009
Figure 10. Accuracy of different matrices in comparison with the proposed approach.
Figure 10. Accuracy of different matrices in comparison with the proposed approach.
Applsci 15 07213 g010
Table 1. Performance parameters for different matrices.
Table 1. Performance parameters for different matrices.
AlgorithmAccuracyRecallF1 ScorePrecession
SVM0.974110.97770.9565
KNN0.97410.97820.97820.9782
RF0.95710.95650.96700.9777
ANN0.985710.98900.9782
LR0.95710.95740.96770.9782
Proposed0.99470.990.990.99
Table 2. Performance parameters for different matrices.
Table 2. Performance parameters for different matrices.
AlgorithmAccuracyRecall
LR96.490.9565
SVM98.240.99
KNN97.200.9808
Naïve Bayes94.740.9545
proposed0.99470.99
Table 3. Performance parameters for different matrices.
Table 3. Performance parameters for different matrices.
AlgorithmAccuracyRecallAUC
SVM0.97660.93750.9361
LR0.97660.95310.9123
KNN0.96490.90620.9371
proposed0.99470.990.993
Table 4. Performance parameters for different matrices.
Table 4. Performance parameters for different matrices.
AlgorithmsAccuracyRecallAUC
RF96.4994.370.96
NN 94.1088.360.93
ENN 96.6094.940.96
Proposed99.47990.99
Table 5. Performance parameters for different matrices.
Table 5. Performance parameters for different matrices.
AlgorithmsAccuracyPrecisionRecall
RF95.997.296.3
KNN95.99698.5
SVM97.297.399.0
Naïve Bayes92.492.695.5
proposed99.49999
Table 6. Performance parameters for different matrices.
Table 6. Performance parameters for different matrices.
AlgorithmsAccuracyRecallPrecisionF1 Score
SVM94.730.8570.9470.900
DT94.200.8810.9490.914
LR95.080.8570.9730.911
RF96.660.92910.963
KNN93.320.78610.880
NN90.350.7620.9700.854
proposed99.470.990.990.99
Table 7. Performance parameters for different matrices.
Table 7. Performance parameters for different matrices.
AlgorithmAccuracyClassPrecisionRecallF1-Score
KNN0.96B0.931.000.96
M1.000.890.94
SVM0.95B0.970.940.95
M0.920.960.94
DT0.97B0.970.990.98
M0.980.960.97
Naïve Bayes0.90B0.920.910.92
M0.880.890.88
LR0.96B0.970.970.97
M0.960.960.96
Proposed0.99B0.9911
M10.990.99
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hameed Alsaedi, N.R.; Akay, M.F. Effective Breast Cancer Classification Using Deep MLP, Feature-Fused Autoencoder and Weight-Tuned Decision Tree. Appl. Sci. 2025, 15, 7213. https://doi.org/10.3390/app15137213

AMA Style

Hameed Alsaedi NR, Akay MF. Effective Breast Cancer Classification Using Deep MLP, Feature-Fused Autoencoder and Weight-Tuned Decision Tree. Applied Sciences. 2025; 15(13):7213. https://doi.org/10.3390/app15137213

Chicago/Turabian Style

Hameed Alsaedi, Nagham Rasheed, and Mehmet Fatih Akay. 2025. "Effective Breast Cancer Classification Using Deep MLP, Feature-Fused Autoencoder and Weight-Tuned Decision Tree" Applied Sciences 15, no. 13: 7213. https://doi.org/10.3390/app15137213

APA Style

Hameed Alsaedi, N. R., & Akay, M. F. (2025). Effective Breast Cancer Classification Using Deep MLP, Feature-Fused Autoencoder and Weight-Tuned Decision Tree. Applied Sciences, 15(13), 7213. https://doi.org/10.3390/app15137213

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop