Next Article in Journal
Monolithic Integration of a Dual-Mode On-Chip Antenna with a Ferroelectric Hafnium Zirconium Oxide Varactor for Reprogrammable Radio-Frequency Front Ends
Next Article in Special Issue
Advances, Trends and Challenges for Determining the Condition of Railway Rolling Stock Using Automatic Classifiers: A Systematic Review
Previous Article in Journal
Optimizing Power Control in Generation Units: LSTM-Based Machine Learning for Enhanced Stability in Virtual Synchronous Generators
Previous Article in Special Issue
RoadMark-cGAN: Generative Conditional Learning to Directly Map Road Marking Lines from Aerial Orthophotos via Image-to-Image Translation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Advances in Medical Image Processing for Early Breast Cancer Detection: Classical Techniques and Deep Learning Perspectives

by
Wenxian Jin
and
Barmak Honarvar Shakibaei Asli
*
Faculty of Engineering and Applied Sciences, Cranfield University, Cranfield, Bedfordshire MK43 0AL, UK
*
Author to whom correspondence should be addressed.
Electronics 2026, 15(4), 790; https://doi.org/10.3390/electronics15040790
Submission received: 18 January 2026 / Revised: 4 February 2026 / Accepted: 10 February 2026 / Published: 12 February 2026

Abstract

Breast cancer is the most common malignancy among women and a leading cause of cancer-related mortality, making early and accurate detection essential. This review summarises advances in breast imaging and computational diagnostics across mammography, ultrasound, and magnetic resonance imaging (MRI), highlighting challenges in differentiating benign from malignant lesions and identifying rarer tumour types. Key preprocessing steps—denoising, deblurring, and contrast enhancement—are reviewed as they improve image quality prior to analysis. Classical methods (e.g., thresholding, edge detection, and region growing) are compared with deep learning approaches for segmentation and classification. CNNs, RNNs, and emerging transformer-based models consistently outperform handcrafted pipelines, with representative studies reporting 5–15% gains in AUC/accuracy and deep models achieving AUC > 0.85–0.95 on several benchmarks. The review also discusses dataset constraints, common evaluation metrics (AUC, Dice, sensitivity, specificity), and clinical translation barriers such as interpretability and domain shift. Overall, AI-driven methods show strong potential to enhance early detection and support improved breast cancer outcomes.

1. Introduction

Breast cancer continues to be a major global health challenge for women, accounting for approximately 2.3 million new cases and 666,103 deaths annually worldwide [1]. Despite significant advances in screening and treatment, early and accurate detection remains critical to improving survival rates and reducing mortality. Medical imaging plays a central role in this process, with modalities such as magnetic resonance imaging (MRI), ultrasound, and mammography serving as indispensable tools for tumour detection and characterisation [2]. Each modality, however, presents distinct advantages and limitations that affect diagnostic accuracy and clinical applicability.
Mammography remains the most widely used screening tool, offering high sensitivity for micro-calcifications and architectural distortions, but suffers from reduced effectiveness in women with dense breast tissue [3]. Ultrasound provides real-time lesion assessment and improved performance in dense breasts, yet its operator dependency and higher false-positive rates limit reliability [4]. Dynamic contrast-enhanced MRI (DCE-MRI) achieves sensitivity levels approaching 90–95%, but its high cost, long acquisition time, and susceptibility to false positives often lead to unnecessary biopsies [5]. These challenges highlight the need for advanced computational approaches to augment human interpretation, enhance diagnostic precision, and reduce inter-observer variability.
Traditional computer-aided detection (CAD) systems relied on classical image processing methods such as thresholding, region growing, and active contour models for tumour segmentation [6]. While these approaches demonstrated value, their dependence on handcrafted features and vulnerability to noise and artefacts limited their scalability in complex clinical settings. Recent breakthroughs in artificial intelligence (AI), particularly deep learning (DL), have revolutionised breast imaging by enabling automated feature extraction and significantly improving diagnostic accuracy [7]. Convolutional neural networks (CNNs) have achieved notable success in segmentation and classification tasks, with architectures such as U-Net [8] and ResNet [9] achieving area under the curve (AUC) values exceeding 0.95 in distinguishing benign from malignant lesions [10]. Recurrent neural networks (RNNs) and hybrid CNN–RNN models have further improved diagnostic performance by capturing temporal patterns in DCE-MRI and sequential ultrasound imaging [11].
Despite these advances, several challenges remain. Many studies are constrained by limited dataset size and diversity, which restricts generalisability across patient populations and imaging platforms [12]. Moreover, the “black-box” nature of deep learning models raises concerns regarding transparency, interpretability, and clinical adoption [13]. Addressing these challenges requires the development of explainable and robust AI frameworks, integration of multimodal imaging data, and careful consideration of ethical and regulatory factors.
This review is structured as follows: Section 2 presents breast tumour characteristics across imaging modalities. Section 3 discusses image preprocessing techniques for quality enhancement. Section 4 reviews classical image processing methods, while Section 5 examines deep learning approaches. Section 6 discussed challenges, opportunities, and future directions in breast cancer detection. By critically evaluating existing literature, this paper highlights the transformative potential of AI in breast imaging while identifying key areas for future research. To better clarify the structure, Figure 1 presents the flow of this review.

2. Breast Tumour Characteristics in Medical Images

Classical pathology divides breast tumours into several categories based on their overall morphology and architectural organisation. Invasive ductal carcinoma, not otherwise specified, (IDC NOS) is the most common type observed and reported, which consists of about 75% of cases; then invasive lobular carcinoma (ILC) represents about 10% of cases, the next most frequent histologic type of breast cancer. Together, the two categories make up the majority of breast cancer [14]. Whilst most tumours in breast cancer are oestrogen receptor (ER) positive (86%), well or moderately differentiated (73%), small 20 mm (74%) and stage I/II (91%) [15]. However, rare breast cancer is associated with a higher risk of mortality due to its tumour being characterised by poor clinical and pathologic characteristics, especially regarding inflammatory carcinomas [16]. Thus, despite the fact that the typical image features of IDC-NOS and ILC are well known, the appearance of unusual tumours is not well-established. Unusual breast tumours are the subtypes of IDC, including mucinous, tubular, medullary, and papillary carcinomas, which account for 10% of invasive breast tumours. Other very rare malignant breast lesions include lymphoid and hematopoietic malignancies, mesenchymal stromal tumours (such as malignant phyllodes tumours and sarcomas), and metastatic tumours.

2.1. MRI

Breast magnetic resonance imaging (MRI) is a highly sensitive imaging modality that provides excellent soft-tissue contrast and enables detailed evaluation of breast morphology and enhancement patterns. Unlike mammography and ultrasound, MRI can capture dynamic vascular information through contrast-enhanced sequences, making it particularly valuable for detecting small lesions, assessing tumour extent, and identifying multifocal or multicentric disease. It is widely used in high-risk screening, pre-operative staging, and treatment response monitoring, especially in women with dense breast tissue. However, most of these tumours lack morphologic and/or dynamic signs that typically predict malignancy; therefore, it causes the difficulty of differential diagnosis from benign breast masses [17]. In addition, characteristically benign kinetics can be present in malignant lesions. As invasive breast cancers approach 1 cm, the features on MRI become increasingly typical of malignancy, but those 5 mm cancers are difficult to distinguish from more common benign lesions, which have similar features. As a result, this makes the diagnosis more difficult [18]. The typical breast MRI appearance of invasive ductal carcinoma can be a focal mass with an irregular shape, plateau curve (type III) or washout curve (type IV) with strong enhancement in the early phase. A tumour with well-defined margins and a rounded shape with smooth borders is less often seen in breast mammography, but they may make up to one-third of IDC cases, supported by the research data. In the MRI image case of ductal carcinoma in situ (DCIS), a non-invasive early stage breast cancer type, demonstrates a non-mass enhancement, which usually appears as areas of an increased signal without a clear border or mass-like shape [19]. Different to the IDC NOS, medullary invasive cancers usually demonstrated a roundish configuration with pushing margins and no evidence of infiltrating disease in both imaging studies and gross pathology [20,21]. Lobular invasive breast cancer, due to its growth pattern, on MRI, the majority of cancers are present as very irregular masses, and in 20% of cases, instead of an actual mass, diffuse infiltration of the parenchyma manifests itself as diffuse asymmetric enhancement [22]. Figure 2 is an MRI image for the preoperative work-up for breast cancer in the upper right quadrant. As shown in the pre- and post-treatment MRI (Figure 3), the patient’s invasive lobular carcinoma lesion significantly shrank following neoadjuvant chemotherapy, though areas of irregular enhancement persisted. For those other unusual tumours, Table 1 summarises their characteristics in the MRI images.

2.2. Ultrasound

For ultrasound images, research finds that invasive cancers are more common in breast cancers with irregular shapes, non-parallel orientations and hypoechoic or complex echo patterns, which are the typical malignant features of solid breast masses, than in cases of ductal carcinoma in situ (DCIS) cases [25]. A previous study described that high-grade invasive cancers were more likely to demonstrate posterior acoustic enhancement and a well-defined margin on ultrasound [26]. The majority of ultrasound-detected breast cancers occur in women with dense breast tissue (BI-RADS grade 3 or 4). Compared to mammography-detected cancers, ultrasound-detected cancers were more frequently categorised as BI-RADS category 4 and less likely to be classified as category 5. The tumours are generally smaller, with a mean size of 1.32 cm, and 41.5% of them are ≤1 cm in size; they are more likely to be invasive, node-negative, and well or moderately differentiated [27]. As shown in Figure 4, an IDC tumour typically presents as a hypoechoic mass with an indistinct margin and irregular border. The nodule appears markedly hypoechoic, has a taller-than-wide shape, and casts a strong acoustic shadow. Furthermore, studies indicate that margin characteristics correlate with tumour grade. Tumours with spiculated or angular margins are commonly low-grade (Grade I or II), while those with well-defined or lobulated margins are typically high-grade (Grade III). The data shows that posterior shadowing is highly indicative of ER-positive tumours (97% likelihood) and is more frequently observed in low-grade tumours (92%), suggesting a less aggressive nature with a better prognosis. In contrast, posterior enhancement is strongly linked to ER-negative, PR-negative (triple-negative) tumours, commonly found in high-grade (Grade III) cases (76.4%), indicating a more aggressive tumour with a poorer prognosis [28]. Figure 5 is an estrogen receptor-positive/progesterone receptor-positive breast cancer case found in a 41-year-old woman. Ultrasound is a valuable tool for distinguishing malignant lesions, which are typically hypoechoic with poorly defined margins. Table 2 summarises the key ultrasound characteristics of malignant breast tissues [29]. Rare cancer types like triple-negatives have different tumour characteristics in ultrasound images. Figure 6 illustrates the triple negative breast tumours with regular shape and circumscribed margins: (A) invasive ductal carcinoma in a 35-year-old female patient (BI-RADS: 4A, grade III, Ki67 30%); (B) invasive ductal carcinoma in a 55-year-old female patient (BI-RADS: 4A, grade III, Ki67 30%); (C) invasive ductal carcinoma in a 33-year-old female patient (BI-RADS: 4A, grade III, Ki67 80%); (D) invasive ductal carcinoma in a 48-year-old female patient (BI-RADS: 5, grade III, Ki67 80%).

2.3. Mammography

Sickles provides insights into both conventional and subtle indirect signs that indicate malignancy when discussing mammographic findings in breast tumour detection. A common early sign of breast cancer is clustered calcifications, which are often seen as small (≤0.5 mm) calcifications. Malignant calcifications tend to have thin linear, curvilinear, or branching shapes, whereas round or oval shapes are more likely benign. Another conventional mammographic sign of breast cancer is a poorly defined mass; those irregular, ill-defined masses are indicative of malignancy, especially spiculated or multinodular (knobby) margins, which strongly suggest cancer. However, many early cancers appear as non-spiculated masses with poorly defined borders, making them hard to distinguish from benign lesions. Moreover, there are several indirect (subtle) mammographic signs of breast cancer, including single duct, architectural distortion, asymmetry, and developing density [32]. Most masses, those with calcification and architectural distortion, are caused by invasive cancer, and the most common malignancy appearance in mammography of DCIS is calcification [33]. Different from common breast cancer, a study focused on the mammographic characteristics of tubular carcinoma of the breast, a rare and less aggressive subtype of breast cancer, stated that most tubular carcinomas are small and spiculated [34]. Figure 7 provides three different types of breast tumour appearances in mammogram images: palpable, spiculated, and calcified in various situations of breast density. And Figure 8 gives another clear visualisation of the mammography image of an IDC case, a large mass with irregular and spiculated margins can be easily seen in the image.
In Table 3, we summarised the different characteristics of three medical image modalities that are used most frequently in breast cancer diagnostics. Also, we compared three modalities and came up with the advantages and disadvantages of each. Nevertheless, there are other medical image modalities that we have not discussed in this section, which are also useful and popular in breast cancer diagnostics. For example, PET-CT combines a positron emission tomography (PET) scan and a computed tomography (CT) scan to provide detailed, 3D images of the body’s structure and function; histopathology involves the microscopic examination of breast tissue to diagnose and characterise breast diseases, most commonly breast cancer.

3. Image Preprocessing for Medical Images

The very first steps of image processing include several different operations that are known as image preprocessing, such as non-linear characteristics corrections, adjusting brightness and/or contrast, radiometric and geometric corrections, etc. Image preprocessing is always a crucial procedure to improve the performance and accuracy of further analysis. It can include image enhancement (deblurring, denoising, etc.), cropping, and resizing. Medical images have features of low contrast and high noise. One should pay attention to which image system is used to apply image enhancement techniques to medical images. Figure 9 introduces the pipeline for the image preprocessing procedure and lists both mandatory and optional actions that can be taken in the process.

3.1. Denoising and Deblurring

Denoising and/or deblurring processing is often applied to MRI and ultrasound images. A unified Bayesian framework was proposed to reduce noise in medical images, particularly for both additive noise like Gaussian and multiplicative noise like Poisson or Rayleigh. By expressing the Maximum a Posteriori (MAP) estimate as a solution to the Sylvester–Lyapunov equation A X + X B = C , enabling robust and computationally efficient noise reduction while preserving image features such as edges, the method outperformed traditional filters in better maintaining edge details in both synthetic and real datasets. The detailed methods are shown below to estimate the clean image X from a noisy observation Y using a Maximum a Posteriori (MAP) estimation framework. The MAP estimator minimises the negative log-posterior, given by
X * = arg min X log P ( Y X ) log P ( X ) .
This leads to the energy functional
E ( X ) = E d ( X , Y ) + λ E p ( X ) ,
where E d ( X , Y ) is the data fidelity term, E p ( X ) is the regularisation term (prior), and λ is the regularisation parameter.
In the case of additive white Gaussian noise, the data fidelity term takes the form
E d ( X , Y ) = 1 2 σ 2 || X Y || 2 .
When dealing with other noise models, such as Poisson or Rayleigh noise, surrogate approximations are typically used to adapt the MAP framework.
Several priors are considered to enforce image regularity. A common choice is the Gaussian (L2) prior:
E p ( X ) = i , j ( 𝜕 x X i , j ) 2 + ( 𝜕 y X i , j ) 2 .
To better preserve edges, the total variation (TV) prior is used:
E p ( X ) = i , j ( 𝜕 x X i , j ) 2 + ( 𝜕 y X i , j ) 2 .
An alternative is the Benford prior, which applies a logarithmic penalty on gradients:
E p ( X ) = i , j log 1 + X i , j .
Under the assumption of Gaussian noise and a Gaussian prior, the MAP estimation leads to an Euler–Lagrange equation:
E ( X ) = 1 σ 2 ( X Y ) + λ L X = 0 .
which results in a Sylvester–Lyapunov equation of the form
A X + X B = C ,
where A = I + λ D x , B = λ D y , and C = Y . Here, D x and D y represent discrete derivative operators in horizontal and vertical directions, respectively.
For nonlinear priors like TV and Benford, the optimisation becomes non-quadratic. A Majorise–Minimise (MM) approach is employed, using a quadratic surrogate energy at each iteration. The update rule at iteration k is given by
A ( k ) X ( k + 1 ) + X ( k + 1 ) B ( k ) = C ( k ) ,
which is again solved using Sylvester solvers [37].
For the ultrasound images denosing process, some traditional homomorphic techniques usually convert the multiplicative speckle noise into additive noise using a logarithmic transformation and then filters (e.g., wavelet denoising) are applied based on the assumption that the resulting noise is white Gaussian noise (WGN). However, this assumption has been proven invalid; the log-transformed noise is non-Gaussian, often following a Fisher–Tippett distribution with a spiky, heavy-tailed nature. Thus, a two-stage preprocessing pipeline is introduced to overcome the problem. Spectral Equalisation: A decorrelation filter flattens the power spectral density of the image to reduce spatial correlation. Assuming a convolutional image formation model:
g ( n , m ) = f ( n , m ) h ( n , m ) + u ( n , m ) ,
where g is the RF image, f the tissue reflectivity, h the point spread function (PSF), and u additive noise, the power spectral density is
P g ( ω 1 , ω 2 ) = σ f 2 | H ( ω 1 , ω 2 ) | 2 + σ u 2 ,
with σ f 2 and σ u 2 denoting variances of the reflectivity and noise, respectively.
The spectral equalisation filter is defined by
L ( ω 1 , ω 2 ) = | H ( ω 1 , ω 2 ) | 2 + ε 1 / 2 ,
where ε = σ u 2 / σ f 2 is a tunable parameter controlling high-frequency amplification. Applying this filter flattens the power spectrum and reduces spatial correlation. Then, Outlier Shrinkage is the nonlinear suppression of spiky noise components using robust residuals from median-filtered images. After log-transforming the decorrelated envelope image G ( n , m ) , the spiky noise component is suppressed using robust residuals:
R ( n , m ) = sign ( Δ G ( n , m ) ) · | Δ G ( n , m ) | λ + ,
where Δ G ( n , m ) = G ( n , m ) MedianFilter ( G ) ( n , m ) , λ is a threshold (typically covering 93–95% of values), and
( x ) + = x , x > 0 , 0 , otherwise .
The preprocessed image is then:
G ( n , m ) = G ( n , m ) R ( n , m ) ,
which approximates white Gaussian noise (WGN) behaviour, allowing optimal performance of standard denoising filters [38].
A wavelet-domain denoising method using a Bivariate Laplacian Mixture Model (BLMM) in the dual-tree complex wavelet domain effectively captures interscale and intrascale wavelet coefficient dependencies. The clean wavelet coefficient vector is modelled as a local mixture of bivariate Laplacian distributions:
p ( w k ) = i = 1 I a i , k · p i ( w k ) ,
where each component p i ( w k ) is defined as
p i ( w k ) = 1 2 σ i , w σ i , p exp 2 | w k | σ i , w + | w p , k | σ i , p ,
with w k = ( w k , w p , k ) representing a wavelet coefficient and its parent.
With EM-based estimation of weights and variances. Closed-form MAP and MMSE estimators are derived for Gaussian and Rayleigh noise, enabling spatially adaptive and statistically grounded shrinkage. The method shows strong performance on MR, CT and ultrasound imaging, preserving structural features while reducing noise [39].

3.2. Image Quality Enhancement

Medical images usually show limited contrast and quality, which can affect the accuracy of diagnostic results. Many studies have worked on methods to improve image quality using filters or histograms. To enhance medical images, an adaptive fractional-order differential (AFDA) operator is developed that intelligently balances between sharpening edges, preserving textures, and maintaining smooth areas. The core mathematical theory behind the AFDA relies primarily on fractional calculus. Fractional derivatives generalise integer-order derivatives to non-integer (fractional) orders, defined by the Grünwald–Letnikov fractional derivative [40]:
D b v a f ( t ) = lim h 0 h v j = 0 ( b a ) / h ( 1 ) j v j f ( t j h ) ,
In the frequency domain, fractional derivatives behave as power-law filters, emphasising high-frequency features based on fractional order v:
F { D v s } ( ω ) = ( i ω ) v S ( ω ) = | ω | v e i sgn ( ω ) π v 2 S ( ω ) .
To make the operator adaptive, the fractional order v ( i , j ) at each pixel position ( i , j ) is computed by using local gradient magnitude M ( i , j ) and a threshold t:
v ( i , j ) = M ( i , j ) t M ( i , j ) , M ( i , j ) t and M ( i , j ) t M ( i , j ) v 1 , v 1 , M ( i , j ) t and M ( i , j ) t M ( i , j ) < v 1 , v 2 , 2 < M ( i , j ) < t and M ( i , j ) t v 2 , M ( i , j ) t , 2 < M ( i , j ) < t and M ( i , j ) t < v 2 , 0 , M ( i , j ) 2 ,
where thresholds v 1 and v 2 are defined using global mean gradient Q and local means of edge and texture gradients M edge ¯ and M tex ¯ :
v 1 = M edge ¯ Q M edge ¯ , v 2 = Q M tex ¯ Q .
An improved Otsu threshold based on gradient histograms is used to achieve effective segmentation to determine these thresholds. For a given gradient histogram p ( k ) , the Otsu thresholds optimise the variance between classes:
σ B 2 ( t ) = w 1 ( μ 1 μ ) 2 + w 2 ( μ 2 μ ) 2 , where μ = w 1 μ 1 + w 2 μ 2 ,
with class weights w 1 , w 2 and class means μ 1 , μ 2 calculated from the histogram:
w 1 ( t ) = i = 0 t p ( i ) , w 2 ( t ) = i = t + 1 N max p ( i ) , μ 1 = 1 w 1 i = 0 t i p ( i ) , μ 2 = 1 w 2 i = t + 1 N max i p ( i ) .
A histogram equalisation is a classical method in digital image processing, which can be done globally in medical images to improve contrast. Shams explores various histogram-based algorithms aimed at improving the contrast and visual quality of medical images, crucial for accurate medical diagnoses. The study specifically evaluates and compares four main histogram-based enhancement techniques. Histogram-based enhancement methods are based primarily on the distribution of pixel intensity levels. For a digital image, the histogram and its normalised form are defined as
h ( r k ) = n k , p ( r k ) = n k M N , k = 0 , 1 , , L 1
where n k is the count of pixels at intensity level r k , and M N is the total pixel count.
The Histogram Equalisation (HE) algorithm transforms the intensity using the cumulative distribution function (CDF):
cdf ( k ) = i = 0 k p ( r i ) , s k = T ( r k ) = floor ( L 1 ) · cdf ( k )
Cumulative Histogram Equalisation (CHE) uses the general formula:
e h ( k ) = round ( L 1 ) cdf ( k ) cdf min M N cdf min
Quadrant Dynamic Histogram Equalisation (QDHE) partitions histograms dynamically using medians as follows:
m 1 = I width × height × 0.25 , m 2 = I width × height × 0.50 , m 3 = I width × height × 0.75
It clips histogram intensities exceeding a threshold T c , computed as:
T c = average intensity of the image
The dynamic intensity redistribution is performed by
range i = ( L 1 ) span i k = 1 4 span k
Contrast Limited Adaptive Histogram Equalisation (CLAHE) applies local enhancements controlled by a clip limit β :
β = M N 1 + α 100 ( S max 1 )
Performance evaluation metrics for these enhancement algorithms include Mean Squared Error (MSE), Peak Signal-to-Noise Ratio (PSNR), and Standard Deviation (SD):
MSE = 1 M N i = 1 M j = 1 N [ I ( i , j ) I ( i , j ) ] 2 , PSNR = 10 log 10 ( L 1 ) 2 MSE
SD = 1 n 1 k = 1 n ( f ( i , j ) X ) 2
Compared experimental results via different metrics, including PSNR, MSE, and SD, QDHE generally outperformed other methods, providing the best overall contrast enhancement and preserving image brightness effectively; however, CLAHE was particularly effective on images with non-uniform intensity distributions [41].
Apart from the four methods compared in the above study, there are other histogram-based methods, including bi-histogram equalisation (BHE), brightness-preserving bi-histogram fuzzy equalisation (BPFHE), recursive mean-separate histogram equalisation (RMSHE), adaptive histogram equalisation (AHE), dualistic sub-image histogram equalisation (DSIHE), and recursive sub-image histogram equalisation (RSIHE). BHE uses the greyscale average to divide the histogram into two sub-histograms and equalises the separated histograms to enhance contrast. Meanwhile, RMSHE is a modified form of BHE where the histogram is recursively divided into multiple sub-histograms based on the average brightness [42]. BPFHE has two stages. First, the fuzzy histogram is calculated based on fuzzy set theory, which can handle the inaccuracy of greyscale values better than the classical sharp histogram. Two well-known parameters, such as entropy or average information content (AIC) and feature similarity index matrix (FSIM), are used to evaluate the proposed BPBFHE algorithm for different greyscale images [43]. AHE is a technique that operates on smaller areas or tiles of an image rather than the entire area, and each tile is processed to enhance contrast [44]. DSIHE and RSIHE both decompose the image, but using different variable median values and the cumulative density function [45]. Figure 10 presents the results of images that are enhanced by different histogram techniques.

3.3. Frequency Domain Enhancement

Beyond histogram-based methods for medical image enhancement, frequency domain techniques offer powerful alternatives by transforming images into frequency representations, where specific components can be selectively enhanced or suppressed. These methods include widely adopted transforms such as the Discrete Fourier Transform (DFT), Fast Fourier Transform (FFT), Discrete Cosine Transform (DCT), Hartley Transform, Haar Transform, Slant Transform, Wavelet Transform, Curvelet Transform, and Contourlet Transform. Each transform is tailored to extract different types of features: FFT efficiently converts spatial data into frequency components, DCT emphasises energy compaction for compression, and the Hartley Transform retains real-valued computations for speed and efficiency. The Haar and Slant transforms emphasise abrupt and smooth transitions, respectively, while Wavelet-based approaches excel at multi-resolution analysis and local adaptation. In the reviewed literature, Tang et al. [46] proposed a DCT-based method that enhances contrast by grouping and adjusting coefficients across frequency bands, enabling block-wise contrast adaptation. Jin et al. [47] advanced this concept by applying Local Adaptive Histogram Equalisation (LAHE) within wavelet sub-bands, effectively improving local contrast in medical images. Toet [48] introduced an adaptive multi-scale enhancement framework based on pyramid recombination, facilitating scale-specific detail amplification, while Mukhopadhyay and Chanda [49,50] used morphological top-hat transforms to isolate and enhance local features. More recently, Iqbal et al. [51] developed a Dual-Tree Complex Wavelet Transform with Nonlocal Means (DT-CWT-NLM) enhancement method that combines spatial and frequency information to boost edge detail while minimising noise. Supporting these findings, Singh and Mittal [52] emphasised the capability of frequency-based transforms such as DCT, DFT, and wavelets to enhance edges, suppress periodic noise, and preserve structure through transform coefficient manipulation. Collectively, these frequency domain techniques provide a rich and flexible framework for medical image enhancement, offering substantial improvements in contrast, detail preservation, and noise resilience compared to traditional spatial-domain approaches [53]. Figure 11 demonstrates the pipeline of how the frequency domain enhancement works in image processing. And Figure 12 is a set of result images that used frequency-domain methods, including chest X-ray, fast MRI and a dataset from NYU.

3.4. Image Quality Assessment

Image quality assessment (IQA) in medical imaging can be divided into full-reference (FR) methods, which require a pristine ‘ground-truth’ image, and no-reference (NR) or blind methods, which predict perceptual quality without any reference. FR metrics quantify the fidelity of a test image by comparing it directly to a high-quality reference. Classical pixel error measures, like the peak signal-to-noise ratio (PSNR) and mean squared error, correlate poorly with human perception in medical scans, motivating more sophisticated approaches. The Structural Similarity Index (SSIM) models image degradation via three components: luminance, contrast, and structure computed over local windows and then averaged to yield a global similarity score. Extensions include Multiscale SSIM (MS-SSIM), which evaluates SSIM at multiple resolutions, and Information Weighted SSIM (IW-SSIM), which weights local SSIM scores by the information content of each region, both of which improve agreement with subjective ratings in MR images [55]. Feature similarity measures such as FSIM and its Riesz-transform variant (RFSIM) exploit phase congruency and gradient magnitude to emphasise edge-related distortions, achieving top performance across general and medical image datasets. Information-theoretic criteria such as the Information Fidelity Criterion (IFC) and its visual extension (VIF) treat the reference–test pair as input and output of a communication channel, estimating the mutual information preserved and producing perceptually meaningful scores, albeit at a higher computational cost [56].
In routine clinical MRI, there is rarely a true reference, which requires NR IQA. Natural-scene statistics (NSS)-based indices model undistorted images by fitting statistical distributions to transform or spatial-domain coefficients, then quantify how distortions perturb these statistics. BRISQUE, a popular spatial-domain NSS model, applies mean-subtracted contrast normalisation (MSCN) and fits the resulting coefficients to a generalised Gaussian distribution, with deviations from the learned model indicating quality loss [57]. Wavelet domain NSS methods (e.g., DIIVINE) first identify the distortion type, then apply distortion-specific wavelet-based features, while DCT-based schemes like BLIINDS-II employ Bayesian modelling of transform coefficient statistics to predict quality from a compact feature set [58]. Transform-free spatial domain models exploit local image features such as gradients: gradient similarity measures compute pixel-wise comparison of gradient magnitude or orientation between distorted images and an implicit edge model to capture contrast and structural changes, providing fast, real-time blind IQA [59]. Learning-based NR methods further integrate handcrafted features (e.g., high-boost-filtered local descriptors) into regression models, such as support vector regression, to predict quality scores closely matching radiologist opinions. Table 4 summarises the IQA methods of both FR and NR reviewed in this section.

3.5. Evaluation Metrics

To comprehensively assess processing performance, a suite of evaluation metrics has been established to quantify results across multiple dimensions. Table 5 summarises the most commonly used metrics for medical image segmentation quality assessment. These metrics, along with recommendations for their appropriate application and notes on possible interpretation biases, are drawn from the standardised framework proposed by Müller et al. [60].

4. Medical Image Processing Using Classical Methods

Petrou and Kamata stated in their book that image processing attempts to solve three main problems related to texture: Classification, which involves analysing the texture in a sample image of the surface to identify what a surface represents; Segmentation, which involves dividing an image into regions of different textures; and Defect detection, which consists in determining whether the texture is as expected or contains defects. They also identified that texture has two important properties: texture is scale-dependent, and image texture depends on the imaging geometry [61]. Furthermore, image processing is not a one-step process; it requires distinguishing several steps that must be performed sequentially to extract the data of interest from the observed scene. Since the 21st century, many methods have been developed for the early detection of breast cancer. These methods include threshold segmentation, active contour model, region growth, watershed, template matching, level set, and marker-controlled watershed methods.

4.1. MRI Analysis

MRI is a common diagnostic tool for breast cancer, providing clear cross-sectional images with high resolution [62]. CAD (computer-aided diagnosis) applications with image processing tools that include classification and segmentation help improve the efficiency and precision of breast cancer detection. A study conducted a texture analysis using co-occurrence matrix (COM) features, which quantify the heterogeneity of pixel intensity distributions in 200 cases. Classification was performed using a cross-validated k-nearest neighbour ( k = 3 ) model on retrospective training data and prospectively applied to a test set. Significant differences in entropy characteristics were observed between invasive lobular and ductal cancers, suggesting distinct internal heterogeneity patterns and demonstrating that the textural features can reflect the aggressiveness of the tumours, achieving up to 75% accuracy [63]. Yao et al. emphasise texture analysis and wavelet transform, demonstrating high accuracy in classification by applying advanced texture features combined with wavelet transforms, using an SVM committee for robust classification [64]. While Levman et al. systematically explore various feature extraction strategies and kernel functions in SVMs, demonstrating the flexibility and robustness of SVM-based methods, using principal component visualisation for better interpretability [65]. In the context of rapid T2*-weighted imaging, Torheim et al. evaluate several classifiers. Their primary finding is the effectiveness of the minimum enhancement threshold method. Then, noise reduction methods were incorporated, which slightly improved their results [66]. The straightforward yet effective 2D image processing techniques, including Otsu’s thresholding, median filtering, and morphological operations, can achieve high accuracy (97.33%) with minimal human intervention. This method prioritises simplicity, speed, and practical usability, though it may be less robust for complex cases [67]. A sophisticated and fully automated 3D segmentation framework that utilises a probabilistic atlas and statistical (Expectation Maximisation) modelling achieved a high segmentation accuracy (Dice Similarity Coefficient of up to 0.94), which is robust and suitable for clinical assessments and potential use in large-scale studies [68]. A study combined structural segmentation techniques and K-means clustering for precise tumour identification, alongside effective 3D reconstruction (Marching Cubes and Maximum Intensity Projection) for enhanced visualisation. This approach is particularly valuable for detailed clinical visualisation and diagnosis, but lacks a rigorous quantitative evaluation for further application [69].

4.2. Mammogram Analysis

Mammogram segmentation based on classical methods can be divided into three categories: region-based segmentation, edge-based segmentation, and threshold-based segmentation. Classical segmentation divides the image into non-overlapping regions with specific properties. Discussing the region-based methods (RM), Senthilkumar et al. proposed a region growing segmentation algorithm for breast cancer detection, which applied a median filter to remove noise from the target mammogram image and then enhanced the image through contrast-limited adaptive histogram equalisation (CLAHE) to improve the segmentation accuracy. In this method, the researchers used the Harris region point detector to obtain the growing seeds and used CG-1 to extract the qualitative concept of the selected window for calculating Ex, En, and He of the window. The methods achieved a segmentation accuracy of 93% [70]. Berber et al. introduced a breast mass contour segmentation method (BMCS)for digital mammograms based on classical seeded region growth, which enhances traditional seeded region-growing methods by incorporating an adaptive thresholding strategy. BMCS is specifically designed to improve segmentation accuracy in digital mammograms by dynamically adjusting threshold values to prevent common issues like under- and over-segmentation [71]. Petrick presented an effective approach to breast mass segmentation that integrates density-weighted contrast enhancement (DWCE) and an adaptive region-growing algorithm for digitised mammograms. The DWCE firstly applies an adaptive filtering to enhance local contrast, accentuating mammographic structures while suppressing background noise and subtle intensity variations. Then, Laplacian–Gaussian edge detection is applied, yielding initial regions that highlight potential breast masses. This preprocessing efficiently identifies mass candidates, but typically results in undersegmented structures. To refine segmentation, region-growing is applied based on greyscale and gradient information, utilising gradient-based thresholds that are adaptively updated during iterative expansions. The technique begins by identifying seed regions from local intensity maxima within areas highlighted by DWCE. As a result, it prevents over- and under-segmentation and significantly improves the precision of the mass contour [72]. An automated approach to segment and classify breast masses from mammograms is proposed by combining optimised region growth and Dragon Fly Optimisation (DFO). The initial seed selection and optimal thresholds are adaptively determined using the DFO, a swarm-based optimisation technique. The method maximises Rényi’s entropy to select the optimal segmentation parameters, significantly enhancing segmentation accuracy and robustness against intensity variations in mammograms. Using Grey Level Co-occurrence Matrix (GLCM) and Grey Level Run Length Matrix (GLRLM) to extract texture features from the segmented region, and feed the features into a Feed Forward Neural Network (FFNN) classifier trained via back-propagation (Levenberg–Marquardt algorithm), so the classifier is able to differentiate between benign and malignant breast masses based on extracted texture features [73]. Figure 13 demonstrates the architecture of the optimised region growing system.
Seed selection is very crucial for seed-based region growing segmentation. An automated approach for selecting the initial seed point is presented to detect microcalcifications in mammogram images. The method involves regional maxima identification to extract candidate regions where microcalcifications typically appear; eliminating plateau pixels to ensure accurate seed points by refining local maxima through morphological techniques; and determination of optimal local maxima, where the average position of these maxima is selected as the initial seed. By utilising mathematical morphology, specifically morphological operations (dilation, erosion, opening, closing, and regional maxima), optimal initial seed points can be automatically determined without human intervention [74]. Isa and Siong proposed an automated method for segmenting breast microcalcifications in mammograms using a seeded region-growing algorithm with morphologically based automatic seed selection. The technique involves preprocessing to enhance contrast, automatic seed identification through morphological operations, and adaptive region-growing based on intensity similarity. The method achieves robust, accurate segmentation with high consistency compared to expert radiologist annotations, highlighting its effectiveness for early detection of breast cancer [75]. Pereira et al. introduced a computational framework combining wavelet multiresolution analysis and a genetic algorithm (GA) for automated segmentation and detection of breast cancer in mammograms. Initially, artefacts are removed from mammographic images using morphological operations. Subsequently, image enhancement is performed through wavelet decomposition and adaptive Wiener filtering to improve contrast and reduce noise. For segmentation, a multilevel thresholding approach is implemented, where optimal thresholds are determined using a genetic algorithm (GA) applied to wavelet-reduced histograms. The genetic algorithm efficiently selects the optimal number and values of thresholds, thereby accurately delineating suspicious regions. Post-processing further reduces false positives by comparing suspicious areas across cranio-caudal (CC) and mediolateral oblique (MLO) views [76]. Table 6 summarises the three main categories of segmentation using classical methods for mammogram images, listing both advantages and disadvantages for each category. Furthermore, Table 7, Table 8 and Table 9 summarise various works on classical segmentation methods applied to breast mammogram images in recent decades, presenting a timeline of the development of these techniques.

4.3. Ultrasound Analysis

Traditional segmentation methods for breast ultrasound (BUS) images—namely thresholding, region growing, and watershed algorithms—have formed the foundation of early computer-aided diagnostic approaches, offering interpretable and computationally efficient solutions. Maolood et al. [106] introduced a two-stage fuzzy entropy thresholding integrated with a level-set algorithm, which enhanced boundary precision under noisy conditions but required sensitive parameter tuning. Similarly, Gómez-Flores et al. [107] proposed a fully automated segmentation framework using iterative thresholding combined with texture-based preprocessing and radial derivative enhancement, achieving robust results across a large dataset, though its effectiveness declined with low-contrast lesions. Region-growing techniques, as applied by Massich et al. [108], offered a simple strategy to define lesion boundaries by combining Gaussian-based preprocessing with seeded region expansion; however, performance heavily depended on accurate initialisation. For boundary-focused approaches, Gomez et al. [109] employed marker-controlled watershed segmentation, leveraging anisotropic diffusion and contrast-limited adaptive histogram equalisation (CLAHE) to guide contour extraction, which improved localisation accuracy but at the cost of computational complexity. Extending this line of work, Gu et al. [110] incorporated Sobel gradient magnitude and morphological reconstruction within a 3D BUS segmentation pipeline, enabling multi-slice tissue classification with moderate success, although its sensitivity to small lesion detection remained a limitation. Together, these methods laid the foundation for BUS segmentation, emphasising the balance between algorithmic simplicity, boundary accuracy, and robustness to noise and image variability.
Several early studies on breast ultrasound (BUS) image classification employed traditional, non-learning-based approaches that relied on rule-based heuristics, expert-defined thresholds, and statistical decision logic. Bartolotta et al. [111] exemplify classification schemes grounded in radiological descriptors and BI-RADS-based feature mapping, the latter integrating the S-Detect system for structured lesion characterisation without involving any learning algorithm. Similarly, Pesce et al. [112] used fixed threshold values derived from shear wave elastography (SWE) metrics to differentiate malignant from benign lesions, offering quantitative but non-adaptive decision support. Moon et al. [113] introduced fuzzy clustering for elastography analysis, where classification relied on membership functions rather than trained models, providing flexibility in handling ambiguous cases but limited by its sensitivity to cluster initialisation. Ding et al. [114] applied a citation-kNN method. This memory-based logic system used local distance metrics and neighbourhood influence for classification, albeit lacking generalisation capability due to its reliance on stored instances. Thus, these classical methods offer interpretability and simplicity, making them clinically appealing in early CAD systems, but rigid decision rules and limited adaptability to complex imaging variations constrain their performance. Table 10 summarises some more related work in classical segmentation for ultrasound images.

5. Medical Image Processing Using Deep Learning Methods

With recent advances in artificial intelligence, deep learning has emerged as a powerful tool for breast cancer detection, enabling earlier diagnoses and improving patient survival rates. Unlike traditional machine learning methods, deep networks can automatically learn rich hierarchical features from imaging data, reducing the need for manual intervention in feature extraction. The most popular methods applied in the diagnosis and detection of breast cancer are the convolutional neural network (CNN) and the recurrent neural network (RNN). Figure 14 shows the architecture of a CNN for the classification of medical images, and Figure 15 gives a basic architecture of an RNN.

5.1. Convolutional Neural Network (CNN)

Convolutional neural networks (CNNs) have become a cornerstone of medical image analysis, particularly for segmentation and classification, because their hierarchical representations capture complex patterns in high-dimensional imaging data. Among CNN-based segmentation models, U-Net remains one of the most influential architectures. Originally developed for biomedical image segmentation, U-Net employs an encoder–decoder design with skip connections that preserve fine-grained spatial information during reconstruction. Numerous extensions, including 3D U-Net for volumetric imaging and attention U-Net for selectively emphasising salient regions, have further improved performance in clinically realistic settings [126]. Suzuki [127] provides a detailed comparison between classical machine learning pipelines and deep CNNs, highlighting CNNs’ key advantage in end-to-end learning directly from raw pixel intensities without dependence on handcrafted features or prior segmentation. Relative to Massive-Training Artificial Neural Networks (MTANNs), CNNs offer greater architectural flexibility and scalability, though they typically require larger annotated datasets to realise their full potential. Nevertheless, CNNs have become widely adopted due to their ability to reduce false positives in lesion detection while improving sensitivity in classification tasks such as malignancy prediction. Lundervold [128] further summarise CNN integration across the MRI workflow, spanning reconstruction, denoising, and downstream disease detection, and report the strong performance of architectures such as U-Net and V-Net for tissue segmentation and pathology localisation in applications including brain and prostate imaging. More broadly, CNNs outperform traditional image processing approaches by learning task-specific features from data, enabling state-of-the-art performance in tumour segmentation, anomaly detection, and super-resolution MRI reconstruction. Importantly, CNNs have also been extended to quantitative imaging problems such as MR fingerprinting and quantitative susceptibility mapping (QSM), demonstrating their capacity to approximate complex inverse mappings in imaging physics. For example, QSMnet, a 3D U-Net–style CNN, enables robust susceptibility estimation from single-orientation measurements, supporting more efficient clinical workflows. Ravi et al. also emphasise CNN effectiveness across modalities, including CT, X-ray, and MRI, noting that GPU-accelerated optimisation and transfer learning have reduced reliance on domain-specific feature engineering and made high-performance classification more accessible across institutions. Raghu et al. [129] critically examine transfer learning in medical imaging and show that although pre-trained backbones (e.g., ResNet and VGG) can accelerate convergence and improve performance when labels are scarce, careful fine-tuning remains essential to mitigate domain shift between natural images and medical data. In breast imaging, Wang et al. [130] demonstrate that CNNs outperform conventional models for mammography and ultrasound classification, with fine-tuned ResNet-50 and InceptionV3 achieving strong sensitivity and specificity in distinguishing malignant lesions. In breast MRI, CNN-based pipelines are also widely employed for lesion segmentation, exploiting both spatial texture and temporal kinetics (e.g., from DCE-MRI). Although earlier studies combined wavelet-transformed features with classifiers such as SVMs, modern CNN models increasingly provide automated, robust segmentation and improved generalisation under limited annotation regimes [131]. Figure 16, as shown, is a comparative study of different state-of-the-art methods that use CNN models.
Across dynamic contrast-enhanced MRI (DCE-MRI), ultrasound, mammography, thermography, and histopathology, a consistent pattern emerges: explicitly incorporating lesion localisation and richer feature representations into learning pipelines improves diagnostic performance across modalities. Recent CNN-based frameworks for breast cancer diagnosis increasingly integrate segmentation, attention, and lesion-guided learning to enhance accuracy and interpretability. In ultrasound, Hossain et al. propose RKO-UNet, which introduces spatial and channel self-attention for precise tumour segmentation and subsequently feeds segmented regions into a fine-tuned CNN, achieving 98.41% accuracy [132]. Similarly, Chiao et al. demonstrate that Mask R-CNN can perform joint detection, pixel-level segmentation, and benign/malignant classification on sonograms, reporting m A P 0.75 and 85% accuracy [133]. In mammography, Salama and Aly combine a modified U-Net with InceptionV3 in an end-to-end framework, reaching 98.87% accuracy with A U C 0.988 on DDSM [134]. Complementing mask-driven pipelines, Singh et al. employ a conditional GAN to generate high-fidelity mass masks that inform a shape-aware CNN, achieving Dice = 0.94, I o U = 0.87 , and 80% accuracy for shape classification [135]; Figure 17 illustrates the workflow of this approach. For DCE-MRI segmentation, El Adoui et al. report that U-Net outperforms SegNet (mean IoU 76.14% vs. 68.88%) [136]. Earlier hybrid pipelines, such as Rouhi et al. [137], coupled region-growing and cellular neural network segmentation with genetic algorithm feature selection, achieving 96.47% accuracy on mammograms and foreshadowing today’s lesion-guided learning strategies. Beyond X-ray imaging, thermal CAD has also benefited from CNNs, with a custom architecture reaching 92% accuracy through aggressive augmentation [138]. Meanwhile, U-Net has shown strong performance for automated mammographic mass segmentation (Dice = 0.951, I o U = 0.909 ) [139]. In histopathology, subtype classification has likewise been strengthened by transfer learning, where a fine-tuned DenseNet achieved 95.4% accuracy across multiple tumour categories [140].
More recent developments further strengthen the evidence that segmentation-informed supervision improves classification. Tsochatzidis et al. enhance ResNet-50 by concatenating segmentation masks at each layer and optimising a spatially aware loss, increasing mass-classification AUC from approximately 0.84 to 0.90 on DDSM and CBIS-DDSM [141]. Heenaye-Mamode Khan et al. introduce an adaptive learning-rate strategy within ResNet-50 to improve four-way abnormality classification (masses, calcifications, carcinomas, and asymmetry), raising accuracy from 81.5% to 88% [142]. Aumente-Maestro et al. propose a multi-task framework that jointly learns segmentation and classification, demonstrating that shared representations can improve both interpretability and diagnostic prediction while cautioning that dataset integrity issues (e.g., duplication in BUSI) may inflate reported performance [143]. In a modular workflow closer to clinical deployment, Islam et al. combine U-Net segmentation with an ensemble CNN classifier (MobileNet and Xception), reinforcing the value of lesion-focused refinement while enabling visual explanation via Grad-CAM [144]. UCapsNet extends this two-stage paradigm by integrating an enhanced U-Net for lesion delineation with a capsule-based classifier, aiming to preserve spatial hierarchies and reduce information loss associated with pooling-based CNNs [145]. Beyond architectural refinements, CNN-centric pipelines are increasingly strengthened through training strategies that address limited labels and distribution shift. For example, self-supervised learning for breast tumour segmentation in DCE-MRI incorporates pretext tasks such as global content perception and peritumoral restoration to improve representation quality under scarce supervision. Likewise, contrastive learning has been explored as an effective domain adaptation mechanism for 2D mammography classification, where supervised contrastive objectives improve cross-domain feature alignment. Relatedly, SelfAdaptNet formalises a CNN-centred self-supervised adversarial adaptation framework that combines contrastive representation learning (e.g., SimCLR/BYOL-style pretraining) with adversarial domain alignment to improve generalisation in breast cancer detection. Table 11 has summarised the key takeaways for the studies reviewed in this section. Collectively, these studies indicate that contemporary gains in CNN-based breast imaging arise not solely from deeper backbones, but from lesion-aware supervision, multi-task optimisation, ensemble learning, contrastive representation learning, and domain adaptation—reinforcing review findings that position CNNs as dominant while also reflecting a clear shift toward hybrid and self-supervised systems in modern breast cancer diagnosis.
Figure 16. Comparison study of different state-of-the-art methods. (a) Image index, (b) the original FFDM MG images from the INbreast database (the red rectangles show the location of the BBs of the ground truth lesions), (c) the GTMs given by radiologists, (d) the FCN model, (e) the Dilated-Net model, (f) the SegNet model, (g) the RG method, (h) U-Net model trained with the augmented dataset, (i) the proposed Vanilla U-Net model without augmentation, (j) the proposed Vanilla U-Net model trained with the augmented dataset, and finally (k) the Faster R-CNN model trained with the augmented dataset [139].
Figure 16. Comparison study of different state-of-the-art methods. (a) Image index, (b) the original FFDM MG images from the INbreast database (the red rectangles show the location of the BBs of the ground truth lesions), (c) the GTMs given by radiologists, (d) the FCN model, (e) the Dilated-Net model, (f) the SegNet model, (g) the RG method, (h) U-Net model trained with the augmented dataset, (i) the proposed Vanilla U-Net model without augmentation, (j) the proposed Vanilla U-Net model trained with the augmented dataset, and finally (k) the Faster R-CNN model trained with the augmented dataset [139].
Electronics 15 00790 g016
Figure 17. Automatic Automatic workflow for breast tumour segmentation and shape classification using a conditional generative adversarial network (cGAN) and a convolutional neural network (CNN). Stage I performs mass segmentation, including preprocessing, cGAN-based encoder–decoder segmentation, and postprocessing to obtain a binary tumour mask. Stage II performs shape classification based on the segmented tumour. Different colours denote functional modules in the pipeline, while the green box lists the final tumour shape categories (irregular, lobular, oval, and round) predicted by the CNN [135].
Figure 17. Automatic Automatic workflow for breast tumour segmentation and shape classification using a conditional generative adversarial network (cGAN) and a convolutional neural network (CNN). Stage I performs mass segmentation, including preprocessing, cGAN-based encoder–decoder segmentation, and postprocessing to obtain a binary tumour mask. Stage II performs shape classification based on the segmented tumour. Different colours denote functional modules in the pipeline, while the green box lists the final tumour shape categories (irregular, lobular, oval, and round) predicted by the CNN [135].
Electronics 15 00790 g017

5.2. Recurrent Neural Network (RNN)

Several recent studies have demonstrated the efficacy of recurrent neural network (RNN) architectures—particularly LSTM and BiLSTM variants—when combined with convolutional and attention mechanisms for breast cancer imaging. Patil and Biradar [146] introduced a hybrid CNN–RNN model in which an RNN branch processes sequential GLCM/GLRM texture features alongside a CNN branch on segmented mammograms, with both components’ architectures jointly optimised via a firefly–chicken swarm algorithm, yielding marked accuracy gains on the MIAS dataset. Pan et al. [147] embedded a spatial–channel attention bidirectional LSTM module (SC-FCN-BLSTM) within a fully convolutional encoder–decoder for automated whole-breast ultrasound, capturing inter-slice context to achieve a Dice score of 0.8178. Figure 18 shows the whole architecture of this novel model. Tripathi et al. [148] proposed a context-based patch modelling approach in histopathology, ordering variable-size patches into sequences fed to a BiLSTM, an end-to-end scheme that reached 90% accuracy on microscopy images and 84% on whole-slide regions. Building on this, Srikantamurthy et al. [149] leveraged ImageNet-pretrained CNN features as inputs to an LSTM for multi-magnification subtype classification on BreakHis, achieving 99% binary and 92.5% four-class accuracy. Paul and Preethi [150] combined an LS-CED–based segmentation pipeline with a wildebeest herd–optimised RNN classifier (WHO-RNN), reporting 97.9% overall accuracy in their histopathology cohort. Zheng et al. [151] proposed a DLA-EABA pipeline that combined stacked sparse autoencoders, an LSTM layer, and an AdaBoost ensemble, achieving 97.2% precision in the Cancer Imaging Archive dataset. Aslan et al. [152] demonstrated that augmenting a CNN with a bidirectional LSTM increased the mammogram classification accuracy from 97.6% (CNN alone) to 98.56% in MIAS and INbreast. Saha and Chakraborty’s Her2Net incorporated trapezoidal LSTM blocks into a semantic-segmentation encoder-decoder for HER2-stained histopathology, yielding 98.3% accuracy and a 96.8% F1-score [153]. Yao et al. designed a parallel CNN–LSTM architecture with attention and switchable normalisation, surpassing IRRCNN benchmarks (e.g., >98.6% accuracy) on BACH2018 and Bioimaging2015 [154]. Finally, Malebary and Hashmi fused k-means segmentation, CNN feature extraction, and an LSTM-based ensemble classifier to achieve 96–98% accuracies and AUCs above 0.94 on DDSM and MIAS [155]. Duraisamy and Emperumal first applied Chan–Vese level-set segmentation to delineate fuzzy lesion boundaries and then refined these regions using a CNN augmented with LSTM modules for benign-vs-malignant discrimination [156]. Patil et al. introduced a holistically nested edge detector to generate attention maps, followed by region growing with adaptive fuzzy C-means—optimally tuned via a grasshopper-inspired metaheuristic—and an RNN classifier, yielding several points of improvement over traditional thresholding and neural baselines [157]. Ranjbarzadeh et al. leveraged U-Net++’s dense skip-connections for fine-grained ultrasound segmentation (and ported to mammograms), fusing multiple encoded feature maps through a cascade of lightweight CNNs for enhanced lesion delineation [158]. On the feature engineering front, Begum and Lakshmi extracted dominant wavelet-based textures and then selected them via an oppositional GSA before RNN-based binary classification, demonstrating the effectiveness of hand-crafted features in lightweight models [159]. Finally, Kaddes et al. achieved near-state-of-the-art results (around 99.9% accuracy) by combining CNN-derived spatial representations with an LSTM to model sequential patch dependencies across the mammogram, underscoring the power of hybrid CNN–LSTM architectures when large annotated datasets are available [160]. Across public benchmarks such as BUSI, MIAS, and DDSM, these pipelines report accuracy and AUC gains—often exceeding 95%—highlighting how attention-guided segmentation, meta-heuristic optimisation, and sequence modelling collectively push mammogram CAD toward clinical viability. Figure 19 lists the segmented results of breast mammogram images that use a novel Grey-Level Co-occurrence Matrix (GLCM) model. Different to the CNN methods, RNN methods are often applied in combination with CNN or other deep learning models and specifically in breast diagnosis, RNN is more commonly applied for pathology image analysis.

5.3. Transformers

Transformers, originally developed for natural language processing, have rapidly gained traction in medical imaging because of their ability to capture long-range dependencies and global contextual information that are often under-modelled by convolutional neural networks (CNNs). Figure 20 shows an example of input patches generation for the vision transformer. Using the standard Vision Transformer (ViT) architecture as an example, ViT first splits the input image into a sequence of fixed-size patches (e.g., n × 16 × 16 ). Each patch is then flattened and projected into a fixed-length embedding vector. A learnable class token is appended to the patch embeddings, after which positional embeddings are added to preserve spatial information across the sequence. And Figure 21 illustrates the workflow of the transformer; multiple encoder blocks can be stacked sequentially, and the total number of stacked encoders is a key hyperparameter that largely determines the transformer’s depth. Within the multi-head attention module, the number of attention heads is another crucial parameter, and together with the encoder depth, it defines the overall model capacity and scale of the transformer [162]. In breast cancer diagnostics, this capability has proven particularly valuable across imaging modalities, including mammography, thermography, digital breast tomosynthesis (DBT), histopathology, and MRI, where malignancy-related cues can be subtle, spatially distributed, and dependent on broader tissue structure rather than purely local texture patterns. Early work in mammography highlighted the limitations of CNN-centric approaches in learning global inter-view relationships, motivating transformer-based multi-view reasoning; for instance, Chen et al. [163] proposed a multi-view vision transformer (MVT) to process unregistered four-view mammograms jointly by combining local transformer blocks for intra-view dependencies with global blocks for inter-view dependency modelling, achieving an AUC of 0.818 and outperforming multi-view CNN baselines (AUC 0.784), thereby demonstrating the clinical relevance of attention-based contextual fusion for routine screening. In DBT, where volumetric imaging increases lesion visibility but complicates analysis due to the multi-slice structure, Kassis et al. [164] employed Swin Transformers for tumour detection using slice-level processing and cross-slice contextual reasoning, reporting an AUC of 0.934 at high resolution and outperforming both ResNet101 and vanilla ViT, while also showing that resolution scaling significantly enhances transformer performance and supports scalability for high-dimensional medical data. In parallel, hybrid Transformer–CNN paradigms have become increasingly prominent to balance CNNs’ strong local inductive bias with transformers’ global reasoning ability; Mahoro and Akhloufi [165] integrated TransUNet-based segmentation in breast thermography to remove irrelevant background before downstream classification, achieving accuracies above 97% for distinguishing healthy, sick, and uncertain cases, while Khan et al. [166] combined ViT-L16 attention representations with ResNet50 and EfficientNetB1 using a ProDense block and stack-ensemble strategy to reach 98.08% accuracy on the INbreast dataset. In mammography, PatchCascade-ViT proposes a self-supervised Vision Transformer pipeline combined with cascade learning for BI-RADS mammographic classification, reflecting an emerging trend where patch-level pretraining and staged decision refinement are used to improve diagnostic robustness and label efficiency [167]. Similarly, HybMNet presents a hybrid architecture that integrates Swin Transformer components with CNN feature extraction under a self-supervised pretraining strategy, illustrating how hierarchical Transformers (Swin-style window attention) can complement convolutional locality to support mammography classification [168]. This wider trend is consistent with transformer-centric evidence from the papers uploaded today, where transformer variants are empirically benchmarked for breast histopathology classification: a comparative evaluation of ViT and multiple transformer architectures (PiT, CvT, CrossFormer, CrossViT, NesT, MaxViT, and SepViT) demonstrated that hierarchical or structurally aware designs such as MaxViT, NesT, and CvT offer stronger and more reliable classification performance, while standard ViT-like models may exhibit weaker generalisation under certain training conditions [169].
Furthermore, narrative synthesis within recent state-of-the-art imaging reviews also emphasises the emergence of vision transformers as a major trend in breast imaging research, while noting persistent barriers to real-world deployment such as interpretability, limited prospective clinical validation, and variability across scanners and cohorts [170]. Beyond mammography and histopathology, transformer models are also beginning to extend into MRI-based breast cancer analysis, where emerging transformer architectures may provide a more effective mechanism for integrating spatiotemporal relationships in DCE-MRI compared to earlier CNN- and SVM-dominated feature learning. Table 12 summarises representative transformer-based breast cancer imaging studies reviewed in this section, including datasets, tasks, and reported performance. Collectively, these studies establish transformers as a powerful alternative or complement to CNNs in breast cancer diagnostics, primarily through: (1) modelling long-range and multi-view dependencies without explicit image registration, (2) adapting effectively to 2D, 3D, and sequential imaging settings, and (3) enabling hybrid or ensemble combinations that integrate global attention with local feature extraction; however, challenges remain in data scarcity, interpretability, and clinical generalisation, reinforcing the need for large-scale pretraining, domain-robust evaluation, and multi-centre validation.

5.4. State-of-the-Art and Foundation Models

Recent evaluations of state-of-the-art (SOTA) deep learning and foundation models reveal a shift from narrowly optimised, task-specific performance toward robustness, label efficiency, and cross-domain generalisation. Conventional supervised architectures, including CNN- and U-Net-based models, continue to report strong Dice and AUC values on curated benchmarks, yet their evaluation typically remains confined to single-task, in-distribution testing. Frameworks such as nnU-Net demonstrate that automated configuration can yield consistently high segmentation accuracy across diverse datasets; however, these evaluations still rely on dataset-specific retraining, limiting insight into generalisability under domain shift [171]. Foundation models introduce broader evaluation criteria that emphasise cross-task robustness and reduced annotation dependence. Promptable models derived from SAM have been extensively evaluated on medical datasets, where out-of-the-box performance is inconsistent—particularly for low-contrast lesions and weak boundaries—highlighting that zero-shot capability alone is insufficient for clinical-grade deployment without domain adaptation [172]. In contrast, MedSAM, fine-tuned on more than 1.57 million medical image–mask pairs, demonstrates improved stability and accuracy across 86 internal and 60 external validation tasks, frequently matching or surpassing modality-specific specialist models. The scale and diversity of these evaluations provide stronger evidence of robustness than traditional single-dataset benchmarks [173].
More recent extensions, such as MedSAM-2 further broaden evaluation paradigms by incorporating minimal-prompt and cross-slice generalisation criteria. By reformulating 2D and 3D medical segmentation as a video-tracking problem, MedSAM-2 achieves new SOTA performance across multiple benchmarks while substantially reducing user interaction. Importantly, its evaluations include breast cancer datasets in both 2D and 3D settings, demonstrating reliable tumour delineation from a single prompt across entire volumes, an operational advantage rarely captured in classical segmentation studies [174]. Self-supervised foundation models shift evaluation emphasis toward label efficiency and out-of-distribution performance. The 3DINO-ViT framework, pretrained on nearly 100,000 unlabeled CT and MRI volumes, consistently outperforms supervised and transfer-learning baselines across segmentation and classification tasks, particularly when trained with limited annotated data. Reported Dice improvements of up to 13–55% over random initialisation indicate that large-scale self-supervised pretraining provides measurable gains beyond architectural choice alone. Notably, 3DINO-ViT generalises effectively to 3D breast ultrasound tumour segmentation, supporting its suitability for breast imaging scenarios where annotations are scarce or costly [175].
Across breast cancer modalities—including mammography, ultrasound, and MRI—evaluation evidence suggests that foundation models are less distinguished by marginal gains in peak accuracy than by stable cross-modality performance, improved label efficiency, and reduced interaction cost. These properties directly address long-standing challenges in breast cancer imaging, such as heterogeneous acquisition protocols, rare tumour subtypes, and inter-institutional variability. Nevertheless, current evaluations remain largely retrospective, with limited prospective or multi-centre clinical validation. Consequently, recent reviews emphasise that foundation models should be regarded as complementary to task-specific models, with careful fine-tuning, validation, and governance required for safety-critical clinical deployment [176].

5.5. Datasets and Evaluation Metrics

Although deep learning studies in breast cancer imaging frequently report strong quantitative performance across diverse datasets and modalities, the clinical interpretation of these results requires careful consideration. In particular, breast cancer screening and diagnostic datasets are often highly imbalanced, meaning that accuracy can be misleading, as high values may simply reflect correct predictions on the dominant negative/benign class while masking poor malignant-case detection. For this reason, most studies additionally report sensitivity, specificity, and F1-score, alongside segmentation measures such as Dice and IoU and general discriminative indicators such as AUC. Among these, AUC remains one of the most commonly reported metrics because it provides a threshold-independent summary of performance and enables comparison across studies. However, AUC alone is not sufficient for clinical translation, as it does not reflect performance at a clinically relevant operating point and may remain high despite inadequate sensitivity in the low false-positive regions required for screening. Moreover, improvements in offline metrics do not necessarily imply real-world benefit, since deployment depends on whether an AI system enhances radiologist decision-making under realistic constraints. Consequently, evaluation frameworks increasingly emphasise reader studies and task-based observer performance models, which assess diagnostic performance with and without AI assistance, capturing inter-reader variability and the practical impact on workflow outcomes such as recall rates, reading time, and missed-cancer reduction. Additionally, the clinical usefulness of AI outputs depends not only on discrimination but also on probability calibration and threshold selection, as poorly calibrated models may produce overconfident malignancy risk estimates that distort decision-making. Therefore, clinically meaningful validation should include threshold-based reporting (e.g., at high-sensitivity operating points), calibration assessments, and external testing across multi-centre cohorts. These considerations highlight why strong performance on public benchmarks may not generalise to real-world populations, reinforcing the need for large-scale, diverse, and consistently annotated datasets, as summarised in Table 13.

6. Discussion

6.1. Classical and Deep Learning Methods: Performance and Trade-Offs

The comparative analysis of classical image processing and modern deep learning methods in breast cancer imaging highlights a series of important trade-offs between interpretability, accuracy, and clinical usability. Classical methods—including thresholding, texture analysis, and wavelet-based transforms—remain attractive due to their lower computational cost, robustness in smaller datasets, and clinician-friendly interpretability. These characteristics make them suitable as strong baselines in data-scarce or resource-constrained environments. However, their performance often saturates when faced with the heterogeneity of tumour appearance and complex background tissues, limiting their scalability in real-world clinical settings. Moreover, many classical pipelines rely on manually engineered descriptors and empirically selected hyperparameters, which may be sensitive to differences in image acquisition, scanner settings, and population characteristics, thereby reducing reproducibility across institutions.
Recent work in mammography further confirms the shift toward deep learning-based methods. Tsochatzidis et al. showed that convolutional neural networks (CNNs) consistently outperform handcrafted descriptors for lesion classification, particularly when fine-tuning pre-trained architectures on medical imaging datasets [141,177]. Their analysis indicates that fine-tuned backbones such as ResNet and VGG generally achieve higher AUCs than models trained from scratch, which is particularly important given the limited availability of large, fully annotated mammography cohorts. This dependence on transfer learning reflects both the flexibility and the data demands of deep learning, as strong performance typically requires either large-scale pre-training or carefully designed regularisation strategies to prevent overfitting. Nevertheless, transfer learning alone does not eliminate generalisation limitations, as pre-trained representations may still struggle with subtle malignant patterns such as microcalcifications, architectural distortions, or lesions embedded in dense breast tissue.
A similar trend is observed in histopathological imaging. Sharma and Mehra compared handcrafted features (Hu moments, Haralick textures, and colour histograms) with transfer learning pipelines based on VGG16, VGG19, and ResNet50 [178]. Their results suggest that classical features combined with standard classifiers (e.g., SVM and Random Forest) can deliver competitive baseline accuracies (approximately 85–90%), while transfer learning approaches achieve consistently higher performance, exceeding 93% in patch-level classification. Although augmentation and balanced sampling help reduce overfitting and mitigate class imbalance, patch-level performance does not fully capture clinical diagnostic reasoning, which is inherently multi-scale and contextual, requiring integration of tissue architecture, cellular morphology, and tumour microenvironment cues.
Boumaraf et al. extended these comparisons by combining high-performing CNN classifiers with explainability mechanisms such as Grad-CAM visualisation [179]. While CNN-based approaches substantially outperform conventional machine learning models on datasets such as BreaKHis and KIMIA Path960 (up to 98% binary accuracy compared to 87–89% with classical classifiers), explainability methods introduce additional trade-offs. Visual saliency maps may support transparency and clinician engagement, yet they remain sensitive to perturbations and do not necessarily correspond to causal evidence, meaning that interpretability should be treated as supportive rather than definitive validation of clinical reliability.

6.2. Clinical Reality: Evaluation Gaps and Failure Modes

Despite steadily improving diagnostic and segmentation performance reported in the literature, clinical evidence indicates a substantial gap between algorithmic performance in retrospective studies and robust behaviour in real-world practice. Many deep learning models are evaluated on carefully curated datasets with balanced classes and homogeneous acquisition protocols, conditions that rarely reflect routine screening or diagnostic workflows. As a result, reported AUC or Dice scores may overestimate robustness when models are deployed across institutions.
Clinical reviews emphasise that overfitting to institutional data, class imbalance favouring malignant findings, and implicit biases in dataset composition frequently lead to reduced generalisation when models are applied across centres or populations [180,181]. In breast imaging, these limitations are amplified by variability in breast density, scanner vendors, compression techniques, and noise characteristics, which can result in unstable false-positive rates and reduced clinician trust despite high reported performance metrics [182].
Beyond algorithmic accuracy, robust clinical validation and regulatory readiness are critical for adoption. Many investigations remain retrospective and benchmark-focused, with limited external validation across scanner vendors, acquisition protocols, and patient demographics. Clinically deployable systems require multi-centre evaluation and prospective evidence demonstrating consistent benefit under realistic workflow conditions. These requirements are closely related to regulatory pathways for medical AI, including FDA clearance and CE marking, which increasingly emphasise transparency, robustness, calibration, and post-deployment monitoring to manage performance drift. In parallel, fairness considerations are increasingly important, as dataset composition can introduce hidden biases related to age, ethnicity, breast density distribution, and imaging vendor variation. Without systematic auditing and stratified evaluation, models risk reduced reliability in underrepresented subgroups, thereby amplifying disparities in diagnostic accuracy.

6.3. From Deep Learning to Foundation Models: Limitations and Future Directions

Taken together, the studies reviewed in this work indicate that classical methods, task-specific deep learning models, and emerging foundation models should be viewed as complementary rather than mutually exclusive. Classical approaches offer interpretability and stability but fail to scale to heterogeneous clinical data. Task-specific deep learning architectures represent the current state of the art in performance but remain sensitive to annotation availability, domain shift, and dataset bias. Foundation models promise improved generalisation and label efficiency, yet introduce new trade-offs, including increased computational cost, reduced transparency, and more complex deployment and governance requirements.
Several structural barriers continue to limit real-world translation. Data sharing restrictions and privacy concerns remain major obstacles to building sufficiently diverse training cohorts, and while federated learning offers a promising solution, its adoption in breast imaging remains relatively limited. Breast cancer diagnosis is inherently multimodal, yet many studies remain focused on single-modality inputs, underscoring the need for multimodal fusion strategies that integrate imaging with MRI multiparametric sequences, ultrasound correlation, pathology confirmation, radiology reports, and genetic risk factors. Furthermore, heavy reliance on fully supervised learning persists despite the scarcity of high-quality pixel-level or slide-level annotations, highlighting the importance of weakly supervised, semi-supervised, and self-supervised learning to improve scalability and reduce annotation burden. Uncertainty quantification also remains underexplored, even though uncertainty-aware modelling could support safer clinical decision-making by flagging ambiguous cases and reducing overconfident predictions.
Overall, the key challenge is not only to close the gap between performance and interpretability, but also to bridge the divide between retrospective benchmark success and clinically safe deployment. Future research should prioritise multi-centre data strategies, privacy-preserving learning, multimodal integration, annotation-efficient training, uncertainty-aware decision support, and fairness-aware validation. These advances will be essential for transforming computational breast cancer imaging methods from research prototypes into trustworthy clinical tools. Table 14 summarises the main differences between classical and deep learning image processing methods and the trade-offs between them.

7. Conclusions

This review examined medical image processing approaches for early breast cancer detection across mammography, ultrasound, and MRI, covering both classical image processing techniques and modern deep learning models. Classical methods remain clinically attractive due to their interpretability, low computational demand, and robustness in small or constrained datasets. However, their dependence on handcrafted features and manually tuned parameters limits scalability and performance in heterogeneous clinical settings.
Deep learning approaches, particularly convolutional neural networks and emerging transformer-based models, consistently demonstrate superior performance on benchmark datasets by enabling automated feature learning and improved contextual modelling. Despite these advances, high reported metrics such as AUC and Dice do not necessarily translate into reliable clinical deployment. Many studies remain retrospective and single-centre, with limited dataset diversity, making models vulnerable to overfitting, replication challenges, class imbalance bias, and performance degradation under domain shift. The lack of large-scale prospective validation and limited interpretability further constrain clinical adoption.
Bridging the gap between research performance and clinical utility requires a shift toward standardised evaluation, multi-centre validation, and greater emphasis on robustness and transparency. Future research is expected to focus on multimodal learning that integrates imaging with clinical data, federated learning frameworks that enable privacy-preserving collaboration, weakly supervised and self-supervised methods to reduce annotation burden, and fairness-aware model development to ensure equitable performance across populations.
In summary, AI-driven image processing has strong potential to enhance early breast cancer detection, but its clinical impact will depend on moving beyond benchmark optimisation toward robust, interpretable, and clinically validated systems.

Author Contributions

Conceptualisation, W.J. and B.H.S.A.; methodology, W.J. and B.H.S.A.; resources, B.H.S.A. and W.J.; writing—original draft preparation, B.H.S.A. and W.J.; writing—review and editing, B.H.S.A. and W.J.; supervision, B.H.S.A.; visualization, B.H.S.A. and W.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AHDAverage Hausdorff Distance
AHEAdaptive Histogram Equalisation
AFDAAdaptive Fractional-Order Differential Approach
AUCArea Under the Curve
BHEBi-Histogram Equalisation
BI-RADSBreast Imaging Reporting and Data System
BLIINDS-IIBlind Image Integrity Notator using DCT Statistics-II
BLMMBivariate Laplacian Mixture Model
BRISQUEBlind/Referenceless Image Spatial Quality Evaluator
BUSBreast Ultrasound
CADComputer-Aided Diagnosis
CCCraniocaudal View
CHECumulative Histogram Equalisation
CLAHEContrast-Limited Adaptive Histogram Equalisation
CNNConvolutional Neural Network
COMCo-occurrence Matrix
CTComputed Tomography
DBTDigital Breast Tomosynthesis
DCISDuctal Carcinoma In Situ
DCE-MRIDynamic Contrast-Enhanced Magnetic Resonance Imaging
DCTDiscrete Cosine Transform
DFTDiscrete Fourier Transform
DIIVINEDistortion Identification-based Image Verity and Integrity Evaluation
DSCDice Similarity Coefficient
DT-CWT-NLMDual-Tree Complex Wavelet Transform with Nonlocal Means
DWCEDensity-Weighted Contrast Enhancement
EMExpectation-Maximisation
EREstrogen Receptor
FFNNFeed Forward Neural Network
FNFalse Negative
FPFalse Positive
FRFull-Reference
FSIMFeature SIMilarity Index
FFTFast Fourier Transform
GAGenetic Algorithm
GLCMGrey-Level Co-occurrence Matrix
GLRLMGrey-Level Run Length Matrix
HEHistogram Equalisation
IDCInvasive Ductal Carcinoma
IDC NOSInvasive Ductal Carcinoma, Not Otherwise Specified
IFCInformation Fidelity Criterion
ILCInvasive Lobular Carcinoma
IoUIntersection-over-Union
IQAImage Quality Assessment
IW-SSIMInformation-Weighted Structural Similarity Index
LAHELocal Adaptive Histogram Equalisation
MAPMaximum a Posteriori
MMMajorise–Minimise
MLOMediolateral Oblique View
mAPmean Average Precision
MRIMagnetic Resonance Imaging
MS-SSIMMultiscale Structural Similarity Index
MSCNMean-Subtracted Contrast Normalised
MSEMean Squared Error
MVTMulti-view Vision Transformer
NRNo-Reference
NSSNatural Scene Statistics
PETPositron Emission Tomography
PET-CTPositron Emission Tomography–Computed Tomography
PRProgesterone Receptor
PSFPoint Spread Function
PSNRPeak Signal-to-Noise Ratio
QDHEQuadrant Dynamic Histogram Equalisation
RFRadio Frequency
RFSIMRiesz Feature Similarity Index
RMSHERecursive Mean-Separate Histogram Equalisation
RNNRecurrent Neural Network
ROCReceiver Operating Characteristic
ROIRegion of Interest
RSIHERecursive Sub-Image Histogram Equalisation
SDStandard Deviation
SSIMStructural Similarity Index
SVMSupport Vector Machine
TNTrue Negative
TPTrue Positive
TVTotal Variation
ViTVision Transformer
VIFVisual Information Fidelity
WGNWhite Gaussian Noise

References

  1. Ferlay, J.; Ervik, M.; Lam, F.; Laversanne, M.; Colombet, M.; Mery, L.; Piñeros, M.; Znaor, A.; Soerjomataram, I.; Bray, F. Global Cancer Observatory: Cancer Today; International Agency for Research on Cancer: Lyon, France, 2024. [Google Scholar]
  2. Giaquinto, A.N.; Sung, H.; Miller, K.D.; Kramer, J.L.; Newman, L.A.; Minihan, A.; Jemal, A.; Siegel, R.L. Breast cancer statistics, 2022. CA A Cancer J. Clin. 2022, 72, 524–541. [Google Scholar] [CrossRef]
  3. Boyd, N.F.; Guo, H.; Martin, L.J.; Sun, L.; Stone, J.; Fishell, E.; Jong, R.A.; Hislop, G.; Chiarelli, A.; Minkin, S.; et al. Mammographic density and the risk and detection of breast cancer. N. Engl. J. Med. 2007, 356, 227–236. [Google Scholar] [CrossRef] [PubMed]
  4. Berg, W.A.; Blume, J.D.; Cormack, J.B.; Mendelson, E.B.; Lehrer, D.; Böhm-Vélez, M.; Pisano, E.D.; Jong, R.A.; Evans, W.P.; Morton, M.J.; et al. Combined screening with ultrasound and mammography vs mammography alone in women at elevated risk of breast cancer. JAMA 2008, 299, 2151–2163. [Google Scholar] [CrossRef] [PubMed]
  5. Kuhl, C.K.; Schrading, S.; Leutner, C.C.; Morakkabati-Spitz, N.; Wardelmann, E.; Fimmers, R.; Kuhn, W.; Schild, H.H. Mammography, breast ultrasound, and magnetic resonance imaging for surveillance of women at high familial risk for breast cancer. J. Clin. Oncol. 2005, 23, 8469–8476. [Google Scholar] [CrossRef] [PubMed]
  6. Cheng, H.D.; Shan, J.; Ju, W.; Guo, Y.; Zhang, L. Automated breast cancer detection and classification using ultrasound images: A survey. Pattern Recognit. 2010, 43, 299–317. [Google Scholar] [CrossRef]
  7. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  8. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015: Proceedings of the 18th International Conference, Munich, Germany, 5–9 October 2015; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  9. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  10. Prevedello, L.M.; Halabi, S.S.; Shih, G.; Wu, C.C.; Kohli, M.D.; Chokshi, F.H.; Erickson, B.J.; Kalpathy-Cramer, J.; Andriole, K.P.; Flanders, A.E. Challenges related to artificial intelligence research in medical imaging and the importance of image analysis competitions. Radiol. Artif. Intell. 2019, 1, e180031. [Google Scholar] [CrossRef]
  11. Antropova, N.; Huynh, B.; Giger, M. Recurrent neural networks for breast lesion classification based on DCE-MRIs. In Proceedings of the Medical Imaging 2018: Computer-Aided Diagnosis, Houston, TX, USA, 10–15 February 2018; SPIE: Bellingham, WA, USA, 2018; Volume 10575, pp. 593–598. [Google Scholar]
  12. May, M. Eight ways machine learning is assisting medicine. Nat. Med. 2021, 27, 2–3. [Google Scholar] [CrossRef]
  13. Holzinger, A.; Saranti, A.; Molnar, C.; Biecek, P.; Samek, W. Explainable AI Methods—A Brief Overview. In xxAI-Beyond Explainable AI: International Workshop, Held in Conjunction with ICML 2020, July 18, 2020, Vienna, Austria, Revised and Extended Papers; Springer International Publishing: Cham, Switzerland, 2022; pp. 13–38. [Google Scholar] [CrossRef]
  14. Bertos, N.R.; Park, M. Breast cancer—one term, many entities? J. Clin. Investig. 2011, 121, 3789–3796. [Google Scholar] [CrossRef]
  15. Figueroa, J.D.; Gierach, G.L.; Duggan, M.A.; Fan, S.; Pfeiffer, R.M.; Wang, Y.; Falk, R.T.; Loudig, O.; Abubakar, M.; Ginsberg, M.; et al. Risk factors for breast cancer development by tumor characteristics among women with benign breast disease. Breast Cancer Res. 2021, 23, 34. [Google Scholar] [CrossRef]
  16. Li, C.; Uribe, D.; Daling1, J. Clinical characteristics of different histologic types of breast cancer. Br. J. Cancer 2005, 93, 1046–1052. [Google Scholar] [CrossRef] [PubMed]
  17. Linda, A.; Zuiani, C.; Girometti, R.; Londero, V.; Machin, P.; Brondani, G.; Bazzocchi, M. Unusual malignant tumors of the breast: MRI features and pathologic correlation. Eur. J. Radiol. 2010, 75, 178–184. [Google Scholar] [CrossRef]
  18. Meissnitzer, M.; Dershaw, D.D.; Feigin, K.; Bernard-Davila, B.; Barra, F.; Morris, E.A. MRI appearance of invasive subcentimetre breast carcinoma: Benign characteristics are common. Br. J. Radiol. 2017, 90, 20170102. [Google Scholar] [CrossRef] [PubMed]
  19. Nunes, L.W.; Schnall, M.D.; Orel, S.G.; Hochman, M.G.; Langlotz, C.P.; Reynolds, C.A.; Torosian, M.H. Correlation of Lesion Appearance and Histologic Findings for the Nodes of a Breast MR Imaging Interpretation Model. RadioGraphics 1999, 19, 79–92. [Google Scholar] [CrossRef]
  20. Kuhl, C.K. MRI of breast tumors. Eur. Radiol. 2000, 10, 46–58. [Google Scholar] [CrossRef]
  21. Kusama, R.; Takayama, F.; Tsuchiya, S.i. MRI of the breast: Comparison of MRI signals and histological characteristics of the same slices. Med Mol. Morphol. 2005, 38, 204–215. [Google Scholar] [CrossRef] [PubMed]
  22. Dixon, J.M.; Anderson, T.J.; Page, D.L.; Lee, D.; Duffy, S.W.; Stewart, H.J. Infiltrating lobular carcinoma of the breast: An evaluation of the incidence and consequence of bilateral disease. Br. J. Surg. 1983, 70, 513–516. [Google Scholar] [CrossRef]
  23. Schubert, R. Breast Lipoma and Cancer. Case Study. Radiopaedia.org, 24 September 2011. Available online: https://radiopaedia.org/cases/15118 (accessed on 9 February 2026).
  24. Li, Y.L.; Zhang, X.P.; Li, J.; Cao, K.; Cui, Y.; Li, X.T.; Sun, Y.S. MRI in diagnosis of pathological complete response in breast cancer patients after neoadjuvant chemotherapy. Eur. J. Radiol. 2015, 84, 242–249. [Google Scholar] [CrossRef]
  25. Kim, S.H.; Bo Kyoung Seo, J.L.; Kim, S.J.; Cho, K.R.; Lee, K.Y.; Je, B.K.; Kim, H.Y.; Kim, Y.S.; Lee, J.H. Correlation of ultrasound findings with histology, tumor grade, and biological markers in breast cancer. Acta Oncol. 2008, 47, 1531–1538. [Google Scholar] [CrossRef]
  26. Lamb, P.M.; Perry, N.M.; Vinnicombe, S.J.; Wells, C.A. Correlation Between Ultrasound Characteristics, Mammographic Findings and Histological Grade in Patients with Invasive Ductal Carcinoma of the Breast. Clin. Radiol. 2000, 55, 40–44. [Google Scholar] [CrossRef]
  27. Bae, M.S.; Han, W.; Koo, H.R.; Cho, N.; Chang, J.M.; Yi, A.; Park, I.A.; Noh, D.Y.; Choi, W.S.; Moon, W.K. Characteristics of breast cancers detected by ultrasound screening in women with negative mammograms. Cancer Sci. 2011, 102, 1862–1867. [Google Scholar] [CrossRef]
  28. Irshad, A.; Leddy, R.; Pisano, E.; Baker, N.; Lewis, M.; Ackerman, S.; Campbell, A. Assessing the Role of Ultrasound in Predicting the Biological Behavior of Breast Cancer. Am. J. Roentgenol. 2013, 200, 284–290. [Google Scholar] [CrossRef]
  29. Gokhale, S. Ultrasound characterization of breast masses. Indian J. Radiol. Imaging 2009, 19, 242–247. [Google Scholar] [CrossRef]
  30. Sehgal, C.; Weinstein, S.; Arger, P.; Conant, E.F. A Review of Breast Ultrasound. J. Mammary Gland Biol. Neoplasia 2006, 11, 113–123. [Google Scholar] [CrossRef] [PubMed]
  31. Li, J.w.; Zhang, K.; Shi, Z.t.; Zhang, X.; Xie, J.; Liu, J.y.; Chang, C. Triple-negative invasive breast carcinoma: The association between the sonographic appearances with clinicopathological feature. Sci. Rep. 2018, 8, 9040. [Google Scholar] [CrossRef] [PubMed]
  32. Sickles, E. Mammographic features of “early” breast cancer. Am. J. Roentgenol. 1984, 143, 461–464. [Google Scholar] [CrossRef]
  33. Gajdos, C.; Tartter, P.I.; Bleiweiss, I.J.; Hermann, G.; De Csepel, J.; Estabrook, A.; Rademaker, A.W. Mammographic appearance of nonpalpable breast cancer reflects pathologic characteristics. Ann. Surg. 2002, 235, 246–251. [Google Scholar] [CrossRef] [PubMed]
  34. Leibman, A.J.; Lewis, M.; Kruse, B. Tubular carcinoma of the breast: Mammographic appearance. Am. J. Roentgenol. 1993, 160, 263–265. [Google Scholar] [CrossRef][Green Version]
  35. Sturesdotter, L.; Sandsveden, M.; Johnson, K.; Larsson, A.M.; Zackrisson, S.; Sartor, H. Mammographic tumour appearance is related to clinicopathological factors and surrogate molecular breast cancer subtype. Sci. Rep. 2020, 10, 20814. [Google Scholar] [CrossRef]
  36. Niknejad, M.T. Breast Mass (BI-RADS 5). Case Study. Radiopaedia.org, 20 November 2021. Available online: https://radiopaedia.org/cases/94040 (accessed on 9 February 2026).
  37. Sanches, J.M.; Nascimento, J.C.; Marques, J.S. Medical Image Noise Reduction Using the Sylvester–Lyapunov Equation. IEEE Trans. Image Process. 2008, 17, 1522–1539. [Google Scholar] [CrossRef]
  38. Michailovich, O.V.; Tannenbaum, A. Despeckling of medical ultrasound images. IEEE Trans. Ultrason. Ferroelectr. Freq. Control 2006, 53, 64–78. [Google Scholar] [CrossRef]
  39. Rabbani, H.; Nezafat, R.; Gazor, S. Wavelet-Domain Medical Image Denoising Using Bivariate Laplacian Mixture Model. IEEE Trans. Biomed. Eng. 2009, 56, 2826–2837. [Google Scholar] [CrossRef]
  40. Li, B.; Xie, W. Adaptive fractional differential approach and its application to medical image enhancement. Comput. Electr. Eng. 2015, 45, 324–335. [Google Scholar] [CrossRef]
  41. Salem, N.; Malik, H.; Shams, A. Medical image enhancement based on histogram algorithms. Procedia Comput. Sci. 2019, 163, 300–311. [Google Scholar] [CrossRef]
  42. Rana, S.B.; Rana, S. A review of medical image enhancement techniques for image processing. Int. J. Curr. Eng. Technol. 2015, 5, 1282–1286. [Google Scholar] [CrossRef]
  43. Magudeeswaran, V.; Ravichandran, C.G.; Thirumurugan, P. Brightness preserving bi-level fuzzy histogram equalization for MRI brain image contrast enhancement. Int. J. Imaging Syst. Technol. 2017, 27, 153–161. [Google Scholar] [CrossRef]
  44. Bhairannawar, S.S. Efficient medical image enhancement technique using transform HSV space and adaptive histogram equalization. In Soft Computing Based Medical Image Analysis; Academic Press: Cambridge, MA, USA, 2018; pp. 51–60. [Google Scholar]
  45. Patel, S.; Bharath, K.; Balaji, S.; Muthu, R.K. Comparative study on histogram equalization techniques for medical image enhancement. In Soft Computing for Problem Solving: SocProS 2018, Volume 1; Springer: Singapore, 2020; pp. 657–669. [Google Scholar]
  46. Tang, J. A contrast based image fusion technique in the DCT domain. Digit. Signal Process. 2004, 14, 218–226. [Google Scholar] [CrossRef]
  47. Jin, Y.; Fayad, L.M.; Laine, A.F. Contrast enhancement by multiscale adaptive histogram equalization. Wavelets Appl. Signal Image Process. IX 2001, 4478, 206–213. [Google Scholar]
  48. Toet, A. Adaptive multi-scale contrast enhancement through non-linear pyramid recombination. Pattern Recognit. Lett. 1990, 11, 735–742. [Google Scholar] [CrossRef]
  49. Mukhopadhyay, S.; Chanda, B. An edge preserving noise smoothing technique using multiscale morphology. Signal Process. 2002, 82, 527–544. [Google Scholar] [CrossRef]
  50. Mukhopadhyay, S.; Chanda, B. Multiscale morphological segmentation of gray-scale images. IEEE Trans. Image Process. 2003, 12, 533–549. [Google Scholar] [CrossRef]
  51. Kumar, T.S.; Swapna, C. Segmentation of image with DT-CWT and NLM filtering using fuzzy c-means clustering. In Proceedings of the 2014 IEEE International Conference on Computational Intelligence and Computing Research, Coimbatore, India, 18–20 December 2014; pp. 1–4. [Google Scholar]
  52. Singh, G.; Mittal, A. Various image enhancement techniques—A critical review. Int. J. Innov. Sci. Res. 2014, 10, 267–274. [Google Scholar]
  53. Radhika, R.; Mahajan, R. Medical Image Enhancement: A Review. In Proceedings of International Conference on Data Science and Applications; Springer: Singapore, 2022; pp. 105–118. [Google Scholar]
  54. Voronin, V.; Zelensky, A.; Agaian, S. 3-D Block-Rooting Scheme with Application to Medical Image Enhancement. IEEE Access 2020, 9, 3880–3893. [Google Scholar] [CrossRef]
  55. Chow, L.S.; Paramesran, R. Review of medical image quality assessment. Biomed. Signal Process. Control 2016, 27, 145–154. [Google Scholar] [CrossRef]
  56. Zhang, L.; Zhang, L.; Mou, X.; Zhang, D. A Comprehensive Evaluation of Full Reference Image Quality Assessment Algorithms. In Proceedings of the 2012 IEEE International Conference on Image Processing (ICIP), Orlando, FL, USA, 30 September–3 October 2012; pp. 1477–1480. [Google Scholar] [CrossRef]
  57. Mittal, A.; Moorthy, A.K.; Bovik, A.C. No-Reference Image Quality Assessment in the Spatial Domain. IEEE Trans. Image Process. 2012, 21, 4695–4708. [Google Scholar] [CrossRef]
  58. Oszust, M.; Piórkowski, A.; Obuchowicz, R. No-reference image quality assessment of magnetic resonance images with high-boost filtering and local features. Magn. Reson. Med. 2020, 84, 1648–1660. [Google Scholar] [CrossRef] [PubMed]
  59. Liu, A.; Lin, W.; Narwaria, M. Image Quality Assessment Based on Gradient Similarity. IEEE Trans. Image Process. 2012, 21, 1500–1512. [Google Scholar] [CrossRef]
  60. Müller, D.; Soto-Rey, I.; Kramer, F. Towards a guideline for evaluation metrics in medical image segmentation. BMC Res. Notes 2022, 15, 210. [Google Scholar] [CrossRef]
  61. Petrou, M.; Bosdogianni, P. Image Processing: The Fundamental; John Wiley & Sons, Ltd.: Hoboken, NJ, USA, 2001; Chapter 1; pp. 1–20. [Google Scholar] [CrossRef]
  62. Kandlikar, S.G.; Perez-Raya, I.; Raghupathi, P.A.; Gonzalez-Hernandez, J.L.; Dabydeen, D.; Medeiros, L.; Phatak, P. Infrared imaging technology for breast cancer detection—Current status, protocols and new directions. Int. J. Heat Mass Transf. 2017, 108, 2303–2320. [Google Scholar] [CrossRef]
  63. Waugh, S.; Purdie, C.; Jordan, L.; Vinnicombe, S.; Lerski, R.; Martin, P.; Thompson, A. Magnetic resonance imaging texture analysis classification of primary breast cancer. Eur. Radiol. 2016, 26, 322–330. [Google Scholar] [CrossRef]
  64. Yao, J.; Chen, J.; Chow, C. Breast tumor analysis in dynamic contrast enhanced MRI using texture features and wavelet transform. IEEE J. Sel. Top. Signal Process. 2009, 3, 94–100. [Google Scholar] [CrossRef]
  65. Levman, J.; Leung, T.; Causer, P.; Plewes, D.; Martel, A.L. Classification of Dynamic Contrast-Enhanced Magnetic Resonance Breast Lesions by Support Vector Machines. IEEE Trans. Med Imaging 2008, 27, 688–696. [Google Scholar] [CrossRef]
  66. Torheim, G.; Godtliebsen, F.; Axelson, D.; Kvistad, K.; Haraldseth, A.; Rinck, P. Feature extraction and classification of dynamic contrast-enhanced T2*-weighted breast image data. IEEE Trans. Med Imaging 2001, 20, 1293–1301. [Google Scholar] [CrossRef] [PubMed]
  67. Rahman, M.; Hussain, M.G.; Hasan, M.R.; Sultana, B.; Akter, S. Detection and Segmentation of Breast Tumor from MRI Images Using Image Processing Techniques. In Proceedings of the 2020 Fourth International Conference on Computing Methodologies and Communication (ICCMC), Erode, India, 11–13 March 2020; pp. 720–724. [Google Scholar] [CrossRef]
  68. Gubern-Mérida, A.; Kallenberg, M.; Mann, R.M.; Martí, R.; Karssemeijer, N. Breast Segmentation and Density Estimation in Breast MRI: A Fully Automatic Framework. IEEE J. Biomed. Health Inform. 2015, 19, 349–357. [Google Scholar] [CrossRef]
  69. Gnonnou, C.; Smaoui, N. Segmentation and 3D reconstruction of MRI images for breast cancer detection. In Proceedings of the International Image Processing, Applications and Systems Conference, Sfax, Tunisia, 5–7 November 2014; pp. 1–6. [Google Scholar]
  70. Senthilkumar, B.; Umamaheswari, G.; Karthik, J. A novel region growing segmentation algorithm for the detection of breast cancer. In Proceedings of the 2010 IEEE International Conference on Computational Intelligence and Computing Research, Coimbatore, India, 28–29 December 2010; pp. 1–4. [Google Scholar]
  71. Berber, T.; Alpkocak, A.; Balci, P.; Dicle, O. Breast mass contour segmentation algorithm in digital mammograms. Comput. Methods Programs Biomed. 2013, 110, 150–159. [Google Scholar] [CrossRef]
  72. Petrick, N.; Chan, H.P.; Sahiner, B.; Helvie, M.A. Combined adaptive enhancement and region-growing segmentation of breast masses on digitized mammograms. Med Phys. 1999, 26, 1642–1654. [Google Scholar] [CrossRef]
  73. Punitha, S.; Amuthan, A.; Joseph, K.S. Benign and malignant breast cancer segmentation using optimized region growing technique. Future Comput. Inform. J. 2018, 3, 348–358. [Google Scholar] [CrossRef]
  74. Malek, A.A.; Rahman, W.E.Z.W.A.; Yasiran, S.S.; Jumaat, A.K.; Jalil, U.M.A. Seed point selection for seed-based region growing in segmenting microcalcifications. In Proceedings of the 2012 International Conference on Statistics in Science, Business and Engineering (ICSSBE), Langkawi, Malaysia, 10–12 September 2012; pp. 1–5. [Google Scholar]
  75. Isa, N.A.M.; Siong, T.S. Automatic segmentation and detection of mass in digital mammograms. In Recent Researches in Communications, Signals and Information Technology; WSEAS: Athens, Greece, 2012; pp. 143–146. [Google Scholar]
  76. Pereira, D.C.; Ramos, R.P.; Do Nascimento, M.Z. Segmentation and detection of breast cancer in mammograms combining wavelet analysis and genetic algorithm. Comput. Methods Programs Biomed. 2014, 114, 88–101. [Google Scholar] [CrossRef]
  77. Michael, E.; Ma, H.; Li, H.; Kulwa, F.; Li, J. Breast cancer segmentation methods: Current status and future potentials. BioMed Res. Int. 2021, 2021, 9962109. [Google Scholar] [CrossRef]
  78. Qian, W.; Sun, X.; Song, D.; Clark, R.A. Digital mammography: Wavelet transform and Kalman-filtering neural network in mass segmentation and detection. Acad. Radiol. 2001, 8, 1074–1082. [Google Scholar] [CrossRef] [PubMed]
  79. Hatanaka, Y.; Hara, T.; Fujita, H.; Kasai, S.; Endo, T.; Iwase, T. Development of an automated method for detecting mammographic masses with a partial loss of region. IEEE Trans. Med Imaging 2001, 20, 1209–1214. [Google Scholar] [CrossRef] [PubMed]
  80. Zhang, H.; Foo, S.W.; Krishnan, S.M.; Thng, C.H. Automated breast masses segmentation in digitized mammograms. In Proceedings of the IEEE International Workshop on Biomedical Circuits and Systems (BioCAS 2004), Singapore, 1–3 December 2004; IEEE: Piscataway, NJ, USA, 2004; pp. S2/2-1–S2/2-4. [Google Scholar] [CrossRef]
  81. Cascio, D.; Fauci, F.; Magro, R.; Raso, G.; Bellotti, R.; De Carlo, F.; Tangaro, S.; De Nunzio, G.; Quarta, M.; Forni, G.; et al. Mammogram segmentation by contour searching and mass lesions classification with neural network. IEEE Trans. Nucl. Sci. 2006, 53, 2827–2833. [Google Scholar] [CrossRef]
  82. Wei, K.; Guangzhi, W.; Hui, D. Segmentation of the breast region in mammograms using watershed transformation. In Proceedings of the 2005 IEEE Engineering in Medicine and Biology 27th Annual Conference, Shanghai, China, 1–4 September 2005; IEEE: Piscataway, NJ, USA, 2006; pp. 6500–6503. [Google Scholar]
  83. Pei, C.; Wang, C.; Xu, S. Segmentation of the breast region in mammograms using marker-controlled watershed transform. In Proceedings of the 2nd International Conference on Information Science and Engineering, Hangzhou, China, 4–6 December 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 2371–2374. [Google Scholar]
  84. Sahakyan, A.; Sarukhanyan, H. Segmentation of the breast region in digital mammograms and detection of masses. Int. J. Adv. Comput. Sci. Appl. 2012, 3, 102–105. [Google Scholar] [CrossRef]
  85. Liu, F.; Zhang, F.; Gong, Z.; Chen, Y.; Chai, W. A fully automated scheme for mass detection and segmentation in mammograms. In Proceedings of the 2012 5th International Conference on BioMedical Engineering and Informatics, Chongqing, China, 16–18 October 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 140–144. [Google Scholar]
  86. Desai, S.D.; Megha, G.; Avinash, B.; Sudhanva, K.; Rasiya, S.; Linganagouda, K. Detection of microcalcification in digital mammograms by improved-MMGW segmentation algorithm. In Proceedings of the 2013 International Conference on Cloud & Ubiquitous Computing & Emerging Technologies, Pune, India, 15–16 November 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 213–218. [Google Scholar]
  87. Shareef, S.R. Breast cancer detection based on watershed transformation. IJCSI Int. J. Comput. Sci. Issues 2014, 11, 237–245. [Google Scholar]
  88. Al-Najdawi, N.; Biltawi, M.; Tedmori, S. Mammogram image visual enhancement, mass segmentation and classification. Appl. Soft Comput. 2015, 35, 175–185. [Google Scholar] [CrossRef]
  89. Ojala, T.; Näppi, J.; Nevalainen, O. Accurate segmentation of the breast region from digitized mammograms. Comput. Med Imaging Graph. 2001, 25, 47–59. [Google Scholar] [CrossRef]
  90. Sivaramakrishna, R.; Obuchowski, N.A.; Chilcote, W.A.; Powell, K.A. Automatic segmentation of mammographic density. Acad. Radiol. 2001, 8, 250–256. [Google Scholar] [CrossRef] [PubMed]
  91. Khuzi, A.M.; Besar, R.; Zaki, W.W.; Ahmad, N. Identification of masses in digital mammogram using gray level co-occurrence matrices. Biomed. Imaging Interv. J. 2009, 5, e17. [Google Scholar] [CrossRef]
  92. Hu, K.; Gao, X.; Li, F. Detection of suspicious lesions by adaptive thresholding based on multiresolution analysis in mammograms. IEEE Trans. Instrum. Meas. 2010, 60, 462–472. [Google Scholar] [CrossRef]
  93. Liu, C.C.; Tsai, C.Y.; Liu, J.; Yu, C.Y.; Yu, S.S. A pectoral muscle segmentation algorithm for digital mammograms using Otsu thresholding and multiple regression analysis. Comput. Math. Appl. 2012, 64, 1100–1107. [Google Scholar] [CrossRef]
  94. Chen, Z.; Zwiggelaar, R. A combined method for automatic identification of the breast boundary in mammograms. In Proceedings of the 2012 5th International Conference on BioMedical Engineering and Informatics, Chongqing, China, 16–18 October 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 121–125. [Google Scholar]
  95. Al-Bayati, M.; El-Zaart, A. Mammogram images thresholding for breast cancer detection using different thresholding methods. Adv. Breast Cancer Res. 2013, 2, 72–77. [Google Scholar] [CrossRef]
  96. Shanmugavadivu, P.; Sivakumar, V. A novel technique for mammogram mass segmentation using fractal adaptive thresholding. In Proceedings of the 8th International Conference on Robotic, Vision, Signal Processing & Power Applications: Innovation Excellence Towards Humanistic Technology; Springer: Singapore, 2014; pp. 213–220. [Google Scholar]
  97. Neto, O.P.S.; Carvalho, O.; Sampaio, W.; Corrêa, A.; Paiva, A. Automatic segmentation of masses in digital mammograms using particle swarm optimization and graph clustering. In Proceedings of the 2015 International Conference on Systems, Signals and Image Processing (IWSSIP), London, UK, 10–12 September 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 109–112. [Google Scholar]
  98. Qayyum, A.; Basit, A. Automatic breast segmentation and cancer detection via SVM in mammograms. In Proceedings of the 2016 International Conference on Emerging Technologies (ICET), Islamabad, Pakistan, 18–19 October 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 1–6. [Google Scholar]
  99. Makandar, A.; Halalli, B. Threshold based segmentation technique for mass detection in mammography. J. Comput. 2016, 11, 472–478. [Google Scholar] [CrossRef]
  100. Kwok, S.M.; Chandrasekhar, R.; Attikiouzel, Y.; Rickard, M.T. Automatic pectoral muscle segmentation on mediolateral oblique view mammograms. IEEE Trans. Med Imaging 2004, 23, 1129–1140. [Google Scholar] [CrossRef] [PubMed]
  101. Maitra, I.K.; Nag, S.; Bandyopadhyay, S.K. Accurate breast contour detection algorithms in digital mammogram. Int. J. Comput. Appl. 2011, 25, 1–13. [Google Scholar] [CrossRef]
  102. Yasiran, S.S.; Jumaat, A.K.; Malek, A.A.; Hashim, F.H.; Nasrir, N.D.; Hassan, S.N.A.S.; Ahmad, N.; Mahmud, R. Microcalcifications segmentation using three edge detection techniques. In Proceedings of the 2012 IEEE International Conference on Electronics Design, Systems and Applications (ICEDSA), Kuala Lumpur, Malaysia, 5–6 November 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 207–211. [Google Scholar]
  103. El Fahssi, K.; Abenaou, A.; Jai-Andaloussi, S.; Sekkaki, A. Method and algorithm for mass segmentation in mammograms based on the minimisation of energy and active contour model. In Proceedings of the 2014 8th International Symposium on Medical Information and Communication Technology (ISMICT), Firenze, Italy, 2–4 April 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 1–5. [Google Scholar]
  104. Angayarkanni, S.P.; Kamal, N.B.; Thangaiya, R.J. Dynamic graph cut based segmentation of mammogram. SpringerPlus 2015, 4, 591. [Google Scholar] [CrossRef][Green Version]
  105. Chakraborty, S.; Bhowmik, M.K.; Ghosh, A.K.; Pal, T. Automated edge detection of breast masses on mammograms. In Proceedings of the 2016 IEEE Region 10 Conference (TENCON), Singapore, 22–25 November 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 1241–1245. [Google Scholar]
  106. Maolood, I.Y.; Al-Salhi, Y.E.A.; Lu, S. Thresholding for medical image segmentation for cancer using fuzzy entropy with level set algorithm. Open Med. 2018, 13, 374–383. [Google Scholar] [CrossRef]
  107. Gomez-Flores, W.; Ruiz-Ortega, B.A. New fully automated method for segmentation of breast lesions on ultrasound based on texture analysis. Ultrasound Med. Biol. 2016, 42, 1637–1650. [Google Scholar] [CrossRef]
  108. Massich, J.; Meriaudeau, F.; Pérez, E.; Martí, R.; Oliver, A.; Martí, J. Lesion segmentation in breast sonography. In Digital Mammography: 10th International Workshop, IWDM 2010, Girona, Catalonia, Spain, 16–18 June 2010; Proceedings 10; Springer: Berlin/Heidelberg, Germany, 2010; pp. 39–45. [Google Scholar]
  109. Gómez, W.; Leija, L.; Alvarenga, A.V.; Infantosi, A.F.C.; Pereira, W.C.A. Computerized lesion segmentation of breast ultrasound based on marker-controlled watershed transformation. Med Phys. 2010, 37, 82–95. [Google Scholar] [CrossRef]
  110. Gu, P.; Lee, W.M.; Roubidoux, M.A.; Yuan, J.; Wang, X.; Carson, P.L. Automated 3D ultrasound image segmentation to aid breast cancer image interpretation. Ultrasonics 2016, 65, 51–58. [Google Scholar] [CrossRef]
  111. Bartolotta, T.V.; Orlando, A.A.; Spatafora, L.; Taibbi, A.; Midiri, M.; Lagalla, R. S-Detect characterization of focal breast lesions according to the US BI-RADS lexicon: A pictorial essay. J. Ultrasound 2020, 23, 207–215. [Google Scholar] [CrossRef] [PubMed]
  112. Pesce, K.; Binder, F.; Chico, M.J.; Crimì, F.; Orlandi, D.; Derchi, L.E.; Sconfienza, L.M. Diagnostic performance of shear wave elastography in discriminating malignant and benign breast lesions. J. Ultrasound 2020, 23, 575–583. [Google Scholar] [CrossRef]
  113. Moon, W.K.; Chang, S.C.; Huang, C.S.; Chang, R.F. Breast tumor classification using fuzzy clustering for breast elastography. Ultrasound Med. Biol. 2011, 37, 700–708. [Google Scholar] [CrossRef]
  114. Ding, J.; Cheng, H.; Xian, M.; Zhang, Y.; Xu, F. Local-weighted Citation-kNN algorithm for breast ultrasound image classification. Optik 2015, 126, 5188–5193. [Google Scholar] [CrossRef]
  115. Wang, W.; Zhu, L.; Qin, J.; Chui, Y.P.; Li, B.N.; Heng, P.A. Multiscale geodesic active contours for ultrasound image segmentation using speckle reducing anisotropic diffusion. Opt. Lasers Eng. 2014, 54, 105–116. [Google Scholar] [CrossRef]
  116. Xian, M.; Zhang, Y.; Cheng, H.D. Fully automatic segmentation of breast ultrasound images based on breast characteristics in space and frequency domains. Pattern Recognit. 2015, 48, 485–497. [Google Scholar] [CrossRef]
  117. Huang, Q.; Yang, F.; Liu, L.; Li, X. Automatic segmentation of breast lesions for interaction in ultrasonic computer-aided diagnosis. Inf. Sci. 2015, 314, 293–310. [Google Scholar] [CrossRef]
  118. Rodrigues, R.; Braz, R.; Pereira, M.; Moutinho, J.; Pinheiro, A.M. A two-step segmentation method for breast ultrasound masses based on multi-resolution analysis. Ultrasound Med. Biol. 2015, 41, 1737–1748. [Google Scholar] [CrossRef] [PubMed]
  119. Rodtook, A.; Kirimasthong, K.; Lohitvisate, W.; Makhanov, S.S. Automatic initialization of active contours and level set method in ultrasound images of breast abnormalities. Pattern Recognit. 2018, 79, 172–182. [Google Scholar] [CrossRef]
  120. Keatmanee, C.; Chaumrattanakul, U.; Kotani, K.; Makhanov, S.S. Initialization of active contours for segmentation of breast cancer via fusion of ultrasound, Doppler, and elasticity images. Ultrasonics 2019, 94, 438–453. [Google Scholar] [CrossRef]
  121. Ramadan, H.; Lachqar, C.; Tairi, H. Saliency-guided automatic detection and segmentation of tumor in breast ultrasound images. Biomed. Signal Process. Control 2020, 60, 101945. [Google Scholar] [CrossRef]
  122. Karunanayake, N.; Aimmanee, P.; Lohitvisate, W.; Makhanov, S.S. Particle method for segmentation of breast tumors in ultrasound images. Math. Comput. Simul. 2020, 170, 257–284. [Google Scholar] [CrossRef]
  123. Huang, Q.; Huang, Y.; Luo, Y.; Yuan, F.; Li, X. Segmentation of breast ultrasound image with semantic classification of superpixels. Med. Image Anal. 2020, 61, 101657. [Google Scholar] [CrossRef]
  124. Anwar, S.; Majid, M.; Qayyum, A.; Awais, M.; Alnowami, M.; Khan, K. Medical Image Analysis using Convolutional Neural Networks: A Review. J. Med. Syst. 2018, 42, 226. [Google Scholar] [CrossRef]
  125. Mienye, I.D.; Swart, T.G.; Obaido, G. Recurrent Neural Networks: A Comprehensive Review of Architectures, Variants, and Applications. Information 2024, 15, 517. [Google Scholar] [CrossRef]
  126. Castiglioni, I.; Rundo, L.; Codari, M.; Di Leo, G.; Salvatore, C.; Interlenghi, M.; Gallivanone, F.; Cozzi, A.; D’Amico, N.C.; Sardanelli, F. AI applications to medical images: From machine learning to deep learning. Phys. Medica 2021, 83, 9–24. [Google Scholar] [CrossRef] [PubMed]
  127. Suzuki, K. Overview of deep learning in medical imaging. Radiol. Phys. Technol. 2017, 10, 257–273. [Google Scholar] [CrossRef]
  128. Lundervold, A.S.; Lundervold, A. An overview of deep learning in medical imaging focusing on MRI. Z. Für Med. Phys. 2019, 29, 102–127. [Google Scholar] [CrossRef]
  129. Raghu, M.; Zhang, C.; Kleinberg, J.; Bengio, S. Transfusion: Understanding transfer learning for medical imaging. arXiv 2019, arXiv:1902.07208. [Google Scholar] [CrossRef]
  130. Wang, J.; Zhu, H.; Wang, S.H.; Zhang, Y.D. A Review of Deep Learning on Medical Image Analysis. Mob. Netw. Appl. 2021, 26, 351–380. [Google Scholar] [CrossRef]
  131. Ben Ammar, M.; Ayachi, F.L.; Cardoso de Paiva, A.; Ksantini, R.; Mahjoubi, H. Harnessing Deep Learning for Early Breast Cancer Diagnosis: A Review of Datasets, Methods, Challenges, and Future Directions. Int. J. Comput. Digit. Syst. 2024, 16, 1643–1661. [Google Scholar] [CrossRef]
  132. Hossain, S.; Azam, S.; Montaha, S.; Karim, A.; Chowa, S.S.; Mondol, C.; Hasan, M.Z.; Jonkman, M. Automated breast tumor ultrasound image segmentation with hybrid UNet and classification using fine-tuned CNN model. Heliyon 2023, 9, e21369. [Google Scholar] [CrossRef] [PubMed]
  133. Chiao, J.Y.; Chen, K.Y.; Liao, K.Y.K.; Hsieh, P.H.; Zhang, G.; Huang, T.C. Detection and classification the breast tumors using mask R-CNN on sonograms. Medicine 2019, 98, e15200. [Google Scholar] [CrossRef]
  134. Salama, W.M.; Aly, M.H. Deep learning in mammography images segmentation and classification: Automated CNN approach. Alex. Eng. J. 2021, 60, 4701–4709. [Google Scholar] [CrossRef]
  135. Singh, V.K.; Rashwan, H.A.; Romani, S.; Akram, F.; Pandey, N.; Sarker, M.M.K.; Saleh, A.; Arenas, M.; Arquez, M.; Puig, D.; et al. Breast tumor segmentation and shape classification in mammograms using generative adversarial and convolutional neural network. Expert Syst. Appl. 2020, 139, 112855. [Google Scholar] [CrossRef]
  136. El Adoui, M.; Mahmoudi, S.A.; Larhmam, M.A.; Benjelloun, M. MRI Breast Tumor Segmentation Using Different Encoder and Decoder CNN Architectures. Computers 2019, 8, 52. [Google Scholar] [CrossRef]
  137. Rouhi, R.; Jafari, M.; Kasaei, S.; Keshavarzian, P. Benign and malignant breast tumors classification based on region growing and CNN segmentation. Expert Syst. Appl. 2015, 42, 990–1002. [Google Scholar] [CrossRef]
  138. Zuluaga-Gomez, J.; Masry, Z.A.; Benaggoune, K.; Meraghni, S.; Zerhouni, N. A CNN-based methodology for breast cancer diagnosis using thermal images. Comput. Methods Biomech. Biomed. Eng. Imaging Vis. 2021, 9, 131–145. [Google Scholar] [CrossRef]
  139. Abdelhafiz, D.; Bi, J.; Ammar, R.; Yang, C.; Nabavi, S. Convolutional neural network for automated mass segmentation in mammography. BMC Bioinform. 2020, 21, 192. [Google Scholar] [CrossRef] [PubMed]
  140. Nawaz, M.; Sewissy, A.A.; Soliman, T.H.A. Multi-Class Breast Cancer Classification using Deep Learning Convolutional Neural Network. Int. J. Adv. Comput. Sci. Appl. 2018, 9, 316–324. [Google Scholar] [CrossRef]
  141. Tsochatzidis, L.; Koutla, P.; Costaridou, L.; Pratikakis, I. Integrating segmentation information into CNN for breast cancer diagnosis of mammographic masses. Comput. Methods Programs Biomed. 2021, 200, 105913. [Google Scholar] [CrossRef]
  142. Khan, M.H.M.; Boodoo-Jahangeer, N.; Dullull, W.; Nathire, S.; Gao, X.; Sinha, G.R.; Nagwanshi, K.K. Multi-class classification of breast cancer abnormalities using Deep Convolutional Neural Network (CNN). PLoS ONE 2021, 16, e0256500. [Google Scholar] [CrossRef]
  143. Aumente-Maestro, C.; Díez, J.; Remeseiro, B. A multi-task framework for breast cancer segmentation and classification in ultrasound imaging. Comput. Methods Programs Biomed. 2025, 260, 108540. [Google Scholar] [CrossRef] [PubMed]
  144. Islam, M.R.; Rahman, M.M.; Ali, M.S.; Nafi, A.A.N.; Alam, M.S.; Godder, T.K.; Miah, M.S.; Islam, M.K. Enhancing breast cancer segmentation and classification: An Ensemble Deep Convolutional Neural Network and U-net approach on ultrasound images. Mach. Learn. Appl. 2024, 16, 100555. [Google Scholar] [CrossRef]
  145. Madhu, G.; Meher Bonasi, A.; Kautish, S.; Almazyad, A.S.; Mohamed, A.W.; Werner, F.; Hosseinzadeh, M.; Shokouhifar, M. UCapsNet: A Two-Stage Deep Learning Model Using U-Net and Capsule Network for Breast Cancer Segmentation and Classification in Ultrasound Imaging. Cancers 2024, 16, 3777. [Google Scholar] [CrossRef]
  146. Patil, R.S.; Biradar, N. Automated mammogram breast cancer detection using the optimized combination of convolutional and recurrent neural network. Evol. Intell. 2021, 14, 1459–1474. [Google Scholar] [CrossRef]
  147. Pan, P.; Chen, H.; Li, Y.; Cai, N.; Cheng, L.; Wang, S. Tumor segmentation in automated whole breast ultrasound using bidirectional LSTM neural network and attention mechanism. Ultrasonics 2021, 110, 106271. [Google Scholar] [CrossRef]
  148. Tripathi, S.; Singh, S.K.; Lee, H.K. An end-to-end breast tumour classification model using context-based patch modelling–A BiLSTM approach for image classification. Comput. Med Imaging Graph. 2021, 87, 101838. [Google Scholar] [CrossRef]
  149. Srikantamurthy, M.M.; Rallabandi, V.P.S.; Dudekula, D.B.; Natarajan, S.; Park, J. Classification of benign and malignant subtypes of breast cancer histopathology imaging using hybrid CNN-LSTM based transfer learning. BMC Med Imaging 2023, 23, 19. [Google Scholar] [CrossRef]
  150. Paul, G.R.; Preethi, J. A novel breast cancer detection system using SDM-WHO-RNN classifier with LS-CED segmentation. Expert Syst. Appl. 2024, 238, 121781. [Google Scholar] [CrossRef]
  151. Zheng, J.; Lin, D.; Gao, Z.; Wang, S.; He, M.; Fan, J. Deep Learning Assisted Efficient AdaBoost Algorithm for Breast Cancer Detection and Early Diagnosis. IEEE Access 2020, 8, 96946–96954. [Google Scholar] [CrossRef]
  152. Aslan, M.F. A Hybrid End-to-End Learning Approach for Breast Cancer Diagnosis: Convolutional BiLSTM Classification. Comput. Electr. Eng. 2023, 105, 108562. [Google Scholar] [CrossRef]
  153. Saha, M.; Chakraborty, C. Her2Net: A Deep Framework for Semantic Segmentation and Classification of Cell Membranes and Nuclei in Breast Cancer Evaluation. IEEE Trans. Image Process. 2018, 27, 2189–2199. [Google Scholar] [CrossRef]
  154. Yao, H.; Zhang, X.; Zhou, X.; Liu, S. Parallel Structure Deep Neural Network Using CNN and RNN with an Attention Mechanism for Breast Cancer Histology Image Classification. Cancers 2019, 11, 1901. [Google Scholar] [CrossRef]
  155. Malebary, S.J.; Hashmi, A. Automated Breast Mass Classification System Using Deep Learning and Ensemble Learning in Digital Mammogram. IEEE Access 2021, 9, 55316–55326. [Google Scholar] [CrossRef]
  156. Duraisamy, S.; Emperumal, S. Computer-aided mammogram diagnosis system using deep learning convolutional fully complex-valued relaxation neural network classifier. IET Comput. Vis. 2017, 11, 656–662. [Google Scholar] [CrossRef]
  157. Patil, A.R.B.; Tripathi, G.; Rothe, J.; Gupta, D.; Chaudhari, N.; Rakesh, K.K. Deep Learning-Based Intelligent Diagnostics Framework For The Labeling Of Mammogram Image. South East. Eur. J. Public Health 2024, 909–921. [Google Scholar] [CrossRef]
  158. Ranjbarzadeh, R.; Saadi, E.; Hamzenejad, M. ME-CCNN: Multi-encoded images and a cascade convolutional neural network for fine-grained lesion segmentation in breast mammograms. Appl. Intell. 2023, 53, 10112–10123. [Google Scholar] [CrossRef]
  159. Begum, S.S.; Lakshmi, D.R. Combining optimal wavelet statistical texture and recurrent neural network for tumour detection and classification over MRI. Multimed. Tools Appl. 2020, 79, 14009–14030. [Google Scholar] [CrossRef]
  160. Kaddes, M.; Ayid, Y.M.; Elshewey, A.M.; Fouad, Y. Breast cancer classification based on hybrid CNN with LSTM model. Sci. Rep. 2025, 15, 4409. [Google Scholar] [CrossRef]
  161. Ramesh, S.; Sasikala, S.; Gomathi, S.; Geetha, V.; Anbumani, V. Segmentation and classification of breast cancer using novel deep learning architecture. Neural Comput. Appl. 2022, 34, 16533–16545. [Google Scholar] [CrossRef]
  162. Zeng, X.; Abdullah, N.; Sumari, P. Self-supervised learning framework application for medical image analysis: A review and summary. BioMed. Eng. OnLine 2024, 23, 107. [Google Scholar] [CrossRef] [PubMed]
  163. Chen, X.; Zhang, K.; Abdoli, N.; Gilley, P.W.; Wang, X.; Liu, H.; Zheng, B.; Qiu, Y. Transformers Improve Breast Cancer Diagnosis from Unregistered Multi-View Mammograms. Diagnostics 2022, 12, 1549. [Google Scholar] [CrossRef]
  164. Kassis, I.; Lederman, D.; Ben-Arie, G.; Giladi Rosenthal, M.; Shelef, I.; Zigel, Y. Detection of breast cancer in digital breast tomosynthesis with vision transformers. Sci. Rep. 2024, 14, 22149. [Google Scholar] [CrossRef]
  165. Mahoro, E.; Akhloufi, M.A. Breast cancer classification on thermograms using deep CNN and transformers. Quant. InfraRed Thermogr. J. 2024, 21, 30–49. [Google Scholar] [CrossRef]
  166. Khan, S.U.R.; Asif, S.; Bilal, O. Ensemble Architecture of Vision Transformer and CNNs for Breast Cancer Tumor Detection from Mammograms. Int. J. Imaging Syst. Technol. 2025, 35, e70090. [Google Scholar] [CrossRef]
  167. Chen, H.; Martel, A.L. Enhancing breast cancer detection on screening mammogram using self-supervised learning and a hybrid deep model of Swin Transformer and convolutional neural networks. J. Med. Imaging 2025, 12, S22007. [Google Scholar] [CrossRef]
  168. Abdallah, A.; Kasem, M.S.; Abdelhalim, I.; Alghamdi, N.S.; El-Baz, A. Improving BI-RADS mammographic classification with self-supervised vision transformers and cascade learning. IEEE Access 2025, 13, 135500–135514. [Google Scholar] [CrossRef]
  169. Sriwastawa, A.; Arul Jothi, J.A. Vision transformer and its variants for image classification in digital breast cancer histopathology: A comparative study. Multimed. Tools Appl. 2024, 83, 39731–39753. [Google Scholar] [CrossRef]
  170. Carriero, A.; Groenhoff, L.; Vologina, E.; Basile, P.; Albera, M. Deep Learning in Breast Cancer Imaging: State of the Art and Recent Advancements in Early 2024. Diagnostics 2024, 14, 848. [Google Scholar] [CrossRef]
  171. Isensee, F.; Jaeger, P.F.; Kohl, S.A.; Petersen, J.; Maier-Hein, K.H. nnU-Net: A self-configuring method for deep learning-based biomedical image segmentation. Nat. Methods 2021, 18, 203–211. [Google Scholar] [CrossRef]
  172. Ren, Z.; Zhang, Y.; Wang, S. Large foundation model for cancer segmentation. Technol. Cancer Res. Treat. 2024, 23, 1–3. [Google Scholar] [CrossRef]
  173. Ma, J.; He, Y.; Li, F.; Han, L.; You, C.; Wang, B. Segment anything in medical images. Nat. Commun. 2024, 15, 654. [Google Scholar] [CrossRef]
  174. Zhu, J.; Hamdi, A.; Qi, Y.; Jin, Y.; Wu, J. Medical sam 2: Segment medical images as video via segment anything model 2. arXiv 2024, arXiv:2408.00874. [Google Scholar] [CrossRef]
  175. Xu, T.; Hosseini, S.; Anderson, C.; Rinaldi, A.; Krishnan, R.G.; Martel, A.L.; Goubran, M. A generalizable 3D framework and model for self-supervised learning in medical imaging. npj Digit. Med. 2025, 8, 639. [Google Scholar] [CrossRef]
  176. Bian, Y.; Li, J.; Ye, C.; Jia, X.; Yang, Q. Artificial intelligence in medical imaging: From task-specific models to large-scale foundation models. Chin. Med J. 2025, 138, 651–663. [Google Scholar] [CrossRef]
  177. Tsochatzidis, L.; Costaridou, L.; Pratikakis, I. Deep Learning for Breast Cancer Diagnosis from Mammograms—A Comparative Study. J. Imaging 2019, 5, 37. [Google Scholar] [CrossRef] [PubMed]
  178. Sharma, S.; Mehra, R. Conventional Machine Learning and Deep Learning Approach for Multi-Classification of Breast Cancer Histopathology Images—A Comparative Insight. J. Digit. Imaging 2020, 33, 632–654. [Google Scholar] [CrossRef] [PubMed]
  179. Boumaraf, S.; Liu, X.; Wan, Y.; Zheng, Z.; Ferkous, C.; Ma, X.; Li, Z.; Bardou, D. Conventional Machine Learning versus Deep Learning for Magnification Dependent Histopathological Breast Cancer Image Classification: A Comparative Study with Visual Explanation. Diagnostics 2021, 11, 528. [Google Scholar] [CrossRef]
  180. Yu, Y.; Li, M.; Liu, L.; Li, Y.; Wang, J. Clinical big data and deep learning: Applications, challenges, and future outlooks. Big Data Min. Anal. 2019, 2, 288–305. [Google Scholar] [CrossRef]
  181. Roy, S.; Meena, T.; Lim, S.J. Demystifying supervised learning in healthcare 4.0: A new reality of transforming diagnostic medicine. Diagnostics 2022, 12, 2549. [Google Scholar] [CrossRef]
  182. Kumar, R.R.; Priyadarshi, R. Denoising and segmentation in medical image analysis: A comprehensive review on machine learning and deep learning approaches. Multimed. Tools Appl. 2025, 84, 10817–10875. [Google Scholar] [CrossRef]
Figure 1. Paper structure.
Figure 1. Paper structure.
Electronics 15 00790 g001
Figure 2. Invasive ductal carcinoma, 2 cm in diameter, with a small satellite nodule at the inferior tumor margin, in a 45-year-old woman [23].
Figure 2. Invasive ductal carcinoma, 2 cm in diameter, with a small satellite nodule at the inferior tumor margin, in a 45-year-old woman [23].
Electronics 15 00790 g002
Figure 3. A 53-year-old female patient with invasive lobular carcinoma, MRI before (left) and after treatment (right). The MRI showed that the lesion had significantly shrunk after treatment, and there were still irregular enhancement areas after neoadjuvant chemotherapy [24].
Figure 3. A 53-year-old female patient with invasive lobular carcinoma, MRI before (left) and after treatment (right). The MRI showed that the lesion had significantly shrunk after treatment, and there were still irregular enhancement areas after neoadjuvant chemotherapy [24].
Electronics 15 00790 g003
Figure 4. The infiltrating ductal carcinoma on the left has an indistinct margin with an irregular border. The nodule is markedly hypoechoic, taller than wide, and casts a strong shadow [30].
Figure 4. The infiltrating ductal carcinoma on the left has an indistinct margin with an irregular border. The nodule is markedly hypoechoic, taller than wide, and casts a strong shadow [30].
Electronics 15 00790 g004
Figure 5. In the left breast of a 41-year-old woman with estrogen receptor-positive/progesterone receptor-positive breast cancer, the ultrasound image shows a slightly hypoechoic mass (callipers) with relatively circumscribed margins. The predominant feature posterior to the mass is no acoustic change [28].
Figure 5. In the left breast of a 41-year-old woman with estrogen receptor-positive/progesterone receptor-positive breast cancer, the ultrasound image shows a slightly hypoechoic mass (callipers) with relatively circumscribed margins. The predominant feature posterior to the mass is no acoustic change [28].
Electronics 15 00790 g005
Figure 6. Ultrasound images for triple-negative breast cancer with a regular shape, a circumscribed margin, and posterior acoustic enhancement. (A) Invasive ductal carcinoma in a 35-year-old female patient; (B) Invasive ductal carcinoma in a 55-year-old female patient; (C) Invasive ductal carcinoma in a 33-year-old female patient; (D) Invasive ductal carcinoma in a 48-year-old female patient [31].
Figure 6. Ultrasound images for triple-negative breast cancer with a regular shape, a circumscribed margin, and posterior acoustic enhancement. (A) Invasive ductal carcinoma in a 35-year-old female patient; (B) Invasive ductal carcinoma in a 55-year-old female patient; (C) Invasive ductal carcinoma in a 33-year-old female patient; (D) Invasive ductal carcinoma in a 48-year-old female patient [31].
Electronics 15 00790 g006
Figure 7. Mammographic breast tumour appearances. From left to right: A palpable mass in the left breast of an 80-year-old woman with moderate breast density. A spiculated mass in the right breast of a 79-year-old woman with fatty degeneration of the breast. Calcifications in the left breast of a 77-year-old woman with moderate breast density. All images were acquired in the cephalopod projection [35].
Figure 7. Mammographic breast tumour appearances. From left to right: A palpable mass in the left breast of an 80-year-old woman with moderate breast density. A spiculated mass in the right breast of a 79-year-old woman with fatty degeneration of the breast. Calcifications in the left breast of a 77-year-old woman with moderate breast density. All images were acquired in the cephalopod projection [35].
Electronics 15 00790 g007
Figure 8. A large mass with irregular and spiculated margin, internal microcalcifications, surrounding parenchymal distortion, and adjacent skin thickening and retraction is noted at the central portion of the left breast in both bilateral craniocaudal (CC) and mediolateral oblique (MLO) views, was found in a 55 years woman who has been confirmed to have invasive ductal carcinoma [36].
Figure 8. A large mass with irregular and spiculated margin, internal microcalcifications, surrounding parenchymal distortion, and adjacent skin thickening and retraction is noted at the central portion of the left breast in both bilateral craniocaudal (CC) and mediolateral oblique (MLO) views, was found in a 55 years woman who has been confirmed to have invasive ductal carcinoma [36].
Electronics 15 00790 g008
Figure 9. Preprocessing pipeline. Legend: Green = Mandatory processing steps; Orange dashed = Optional steps; Blue = Merge/decision point; Solid arrows = Primary workflow; Dashed arrows = Optional/alternative paths. Abbreviations: HE (Histogram Equalisation), CLAHE (Contrast-Limited Adaptive Histogram Equalisation), DCT (Discrete Cosine Transform), IQA (Image Quality Assessment), PSNR (Peak Signal-to-Noise Ratio), SSIM (Structural Similarity Index), BRISQUE (Blind/Referenceless Image Spatial Quality Evaluator), ROI (Region of Interest), REG (Registration).
Figure 9. Preprocessing pipeline. Legend: Green = Mandatory processing steps; Orange dashed = Optional steps; Blue = Merge/decision point; Solid arrows = Primary workflow; Dashed arrows = Optional/alternative paths. Abbreviations: HE (Histogram Equalisation), CLAHE (Contrast-Limited Adaptive Histogram Equalisation), DCT (Discrete Cosine Transform), IQA (Image Quality Assessment), PSNR (Peak Signal-to-Noise Ratio), SSIM (Structural Similarity Index), BRISQUE (Blind/Referenceless Image Spatial Quality Evaluator), ROI (Region of Interest), REG (Registration).
Electronics 15 00790 g009
Figure 10. The 1st (top) row displays the five original medical images, the following rows display the enhanced images by the techniques HE, CHE, QDHE and CLAHE, in sequence [41].
Figure 10. The 1st (top) row displays the five original medical images, the following rows display the enhanced images by the techniques HE, CHE, QDHE and CLAHE, in sequence [41].
Electronics 15 00790 g010
Figure 11. Filtering technique in the frequency domain.
Figure 11. Filtering technique in the frequency domain.
Electronics 15 00790 g011
Figure 12. Result images using frequency domain enhancement [54].
Figure 12. Result images using frequency domain enhancement [54].
Electronics 15 00790 g012
Figure 13. Architecture of the system using optimised region growing.
Figure 13. Architecture of the system using optimised region growing.
Electronics 15 00790 g013
Figure 14. A typical A typical convolutional neural network (CNN) architecture for medical image classification. The diagram shows successive convolutional, pooling, and fully connected layers. Yellow boxes and connecting lines indicate the progressive receptive field mapping from the input image to deeper feature representations [124].
Figure 14. A typical A typical convolutional neural network (CNN) architecture for medical image classification. The diagram shows successive convolutional, pooling, and fully connected layers. Yellow boxes and connecting lines indicate the progressive receptive field mapping from the input image to deeper feature representations [124].
Electronics 15 00790 g014
Figure 15. A basic basic recurrent neural network (RNN) architecture. At each time step t, the input x t is processed together with the previous hidden state h t 1 to produce the current hidden state h t and the corresponding output y t . Different node colours distinguish inputs, hidden states, and outputs, while arrows indicate temporal information flow across time steps [125].
Figure 15. A basic basic recurrent neural network (RNN) architecture. At each time step t, the input x t is processed together with the previous hidden state h t 1 to produce the current hidden state h t and the corresponding output y t . Different node colours distinguish inputs, hidden states, and outputs, while arrows indicate temporal information flow across time steps [125].
Electronics 15 00790 g015
Figure 18. Architecture of SC-FCN-BLSTM [147].
Figure 18. Architecture of SC-FCN-BLSTM [147].
Electronics 15 00790 g018
Figure 19. Segmented results of breast mammogram by using RNN [161].
Figure 19. Segmented results of breast mammogram by using RNN [161].
Electronics 15 00790 g019
Figure 20. Example of input patches generation for Vision Transformer [162].
Figure 20. Example of input patches generation for Vision Transformer [162].
Electronics 15 00790 g020
Figure 21. The Workflow of a Vision Transformer. Arrows indicate the forward flow of patch embeddings through layer normalisation, multi-head self-attention, and MLP blocks. The circled “+” symbols denote residual (skip) connections that add the block input to its output, facilitating stable optimisation.
Figure 21. The Workflow of a Vision Transformer. Arrows indicate the forward flow of patch embeddings through layer normalisation, multi-head self-attention, and MLP blocks. The circled “+” symbols denote residual (skip) connections that add the block input to its output, facilitating stable optimisation.
Electronics 15 00790 g021
Table 1. MRI features of unusual breast tumours.
Table 1. MRI features of unusual breast tumours.
MucinousMedullaryTubularPapillary Intraductal
ShapeLobulated, oval, roundLobulated, oval, roundSpiculatedNon-mass-like enhancement with a segmental distribution
MarginsSmoothSmoothIII-defined
Kinetic curveType IType II or IIIType IType I
InflammatoryPapillary IntracysticPapillary Invasive
ShapeIrregularRound, ovalRound, oval
MarginsIrregular, spiculatedSmoothSmooth
Kinetic curveType III or IIType IIIType III
Table 2. Suspicious malignant lesions on ultrasound.
Table 2. Suspicious malignant lesions on ultrasound.
Solid Nodule CharacteristicsPositive Predictive Value
Speculation91.8
Taller than wide81.2
Angular margins67.5
Shadowing64.9
Branching pattern64.0
Hypo echogenicity60.1
Calcifications59.6
Duct extension50.8
Micro lobulations48.2
Table 3. Comparison of imaging modalities for breast cancer diagnosis.
Table 3. Comparison of imaging modalities for breast cancer diagnosis.
AspectMRIMammogramUltrasound
MethodsUses radio waves and a strong magnetic field to create detailed imagesUses specialised X-ray imaging to capture breast tissue for cancer detectionUses high-frequency sound waves to create images of breast tissue
ResolutionHighHighLow
CostHighLowLow
Clinical useHigh-risk screening, investigating abnormalitiesRoutine screening for breast cancer in asymptomatic individualsUsed to investigate symptoms and characterise palpable lumps
AdvantagesHigh sensitivity, no radiation exposureLow cost, effective for early detectionEffective in dense breast tissue, no radiation exposure
DisadvantagesHigh cost, noisy environment, longer scan timeDiscomfort during compression, radiation exposure, increased false positives in dense tissueOperator-dependent, limited coverage of the whole breast
Table 4. Comparison of full-reference and no-reference image quality assessment methods.
Table 4. Comparison of full-reference and no-reference image quality assessment methods.
Full-Reference (FR) MethodsNo-Reference (NR) Methods
PSNRBRISQUE
Peak Signal-to-Noise Ratio; measures pixel-wise fidelity via log-scaled MSE.Blind/Referenceless Image Spatial Quality Evaluator; models deviations from natural scene statistics in the spatial domain.
SSIMDIIVINE
Structural Similarity Index; compares luminance, contrast and structure between test and reference.Distortion Identification-based Image Verity and Integrity Evaluation; wavelet-domain NSS with distortion classification.
MS-SSIMBLIINDS-II
Multiscale SSIM evaluates SSIM over multiple resolutions for scale-aware quality assessment.Blind Image Integrity Notator using DCT Statistics; Bayesian modelling of DCT coefficient distributions.
IW-SSIMGradient Similarity
Information-Weighted SSIM: weights local SSIM scores by regional information content.Compares gradient magnitude/orientation to an implicit edge model to detect blur and structural distortions.
FSIM/RFSIMFeature-Based NR
Feature SIMilarity index; uses phase congruency and gradient features (Riesz-transform variant in RFSIM) to emphasise edge fidelity.Extracts artefact-sensitive features (e.g., blockiness, noise, ringing) and uses regression to predict perceptual quality.
IFC/VIFTask-Based Observer Models
Information Fidelity Criterion/Visual Information Fidelity; treats IQA as mutual-information preservation in a communication channel.Models diagnostic detectability (e.g., Channelised Hotelling Observer) to assess quality based on task performance.
Table 5. Evaluation metrics in medical image processing.
Table 5. Evaluation metrics in medical image processing.
MetricFormulaDescription
Dice Similarity Coefficient (DSC) DSC ( A , B ) = 2 | A B | | A | + | B | Measures overlap between the prediction and the ground truth.
Intersection-over-Union (IoU) IoU = T P T P + F P + F N Ratio of intersection to union of predicted and true regions.
Sensitivity (Recall) Sens = T P T P + F N Proportion of actual positives correctly identified.
Specificity Spec = T N T N + F P Proportion of actual negatives correctly identified.
Accuracy Acc = T P + T N T P + T N + F P + F N Overall proportion of correctly classified pixels.
Area Under ROC Curve (AUC) AUC = 1 1 2 F P F P + T N + F N F N + T P Measure of separability between positive and negative classes.
Cohen’s Kappa ( κ ) κ = ( T P + T N ) f c ( T P + T N + F P + F N ) f c ,Agreement between prediction and ground truth beyond chance.
f c = ( T N + F N ) ( T N + F P ) + ( F P + T P ) ( F N + T P ) T P + T N + F P + F N
Average Hausdorff Distance (AHD) AHD = max d ( A , B ) , d ( B , A ) ,Measures the boundary discrepancy between two contours.
d ( A , B ) = 1 | A | a A min b B a b
Notes: TP = true positives; FP = false positives; FN = false negatives; TN = true negatives; A = set of predicted boundary points; B = set of ground truth boundary points.
Table 6. Comparison of classical segmentation methods: merits and demerits [77].
Table 6. Comparison of classical segmentation methods: merits and demerits [77].
CategoryMeritsDemerits
Edge-based segmentation methodsWorks well when the edge is prominentSensitive to noise
Easy to find the local edge orientationReduce overall contrast in mammograms
Produce unsatisfactory results when it detects the fake and weak edges in mammograms
Not suitable for mammogram images having smooth edges
Threshold-based segmentation methodsSimple and easy to implementIt is not applicable if the tumour area ratio is unknown
FasterSensitive to noise in mammograms
InexpensiveGives poor results when mammograms have low contrast
Difficulties in fixing the threshold value if the number of regions increases
Not easy to process the mammogram whose histograms are nearly unimodal
Region-based segmentation methodsConnected regions are guaranteedCauses over-segmentation if mammograms are noisy
Multiple criteria and give good results with less noiseCannot distinguish the shading of the real mammograms
Time-consuming due to the high resolution of mammograms
Not suitable for noisy mammograms
Seed points must be selected
Table 7. Summary of reviewed works related to classical segmentation in mammogram images: region-growth based methods.
Table 7. Summary of reviewed works related to classical segmentation in mammogram images: region-growth based methods.
Ref.YearTechniqueFilterDatabaseEvaluation Metric
[72]1999Adaptive and region growingGaussianUMH98.0% accuracy
[78]2001Region growingKalmanDDSMROC: 93.0%, without adaptive: 86.0%
[79]2001Partial loss of regionSobelJapanese97.0% true positive
[80]2004Region growingMIAS90.0% TPR, 1.3 FTR/image
[81]2006Contour searchingMAGIC-5ROC: 85.6 ± 0.8%
[82]2006Morphological algorithmMedianMIAS95.0% detection rate
[83]2010WatershedMorphologicalDDSMMean ± SD = 0.93 ± 0.03
[75]2012Region growingContrastMIAS94.59% sensitivity, 3.90% false positive
[84]2012MorphologicalMedianMIAS95.0% detection rate
[85]2012Region growingAdaptiveDDSM97.2% sensitivity, 1.83% false positive
[74]2012Seed point selectionMath. MorphologyNCSM98.0% accuracy
[86]2013Morphological gradient watershedAdaptive medianMIAS & NMRMIAS: 95.3%, NMR: 94.0%
[71]2013OtsuMorphologicalDEMS95.06% accuracy
[86]2014Marker-controlled watershedSobelMIAS90.83% detection, ROC: 91.3%
[76]2014Wavelet + GAWienerMIAS & DDSM79.2 ± 8% (mean ± SD)
[87]2014Watershed transformationMSKESensitivity: 90.47%, Specificity: 75.0%, Accuracy: 84.85%
[88]2015WatershedMedianMini-MIAS96.18% accuracy
Table 8. Summary of reviewed works related to classical segmentation in mammogram images: threshold-based segmentation methods.
Table 8. Summary of reviewed works related to classical segmentation in mammogram images: threshold-based segmentation methods.
Ref.YearTechniqueFilterDatabaseEvaluation Metric
[89]2001Histogram thresholdingMorphologicalDDSM96.0% detection rate, 90.0% accuracy
[90]2001Kittler’s optimal thresholdingBCCCF92.0–95.0% Spearman, 6.9% avg. density
[91]2009Otsu + K-means clusteringCLAHEMIAS85.0% accuracy, 70.0% sensitivity
[92]2011Adaptive global + local thresholdMeteorologicalMIAS91.3% sensitivity, 0.71% false positive
[93]2012Otsu thresholdingMorphologicalMIASME1: 1.7188, ME2: 0.0083, MHD: 0.8702
[94]2012Histogram + edge detectionGaussianMIAS & EPICMIAS: 98.8%, EPIC: 91.5%
[95]2013OtsuMedianMIAS
[96]2014OtsuMedianMIAS92.86% accuracy, 4.97% error rate
[97]2015Threshold + evolutionaryAverageDDSM95.2% accuracy
[98]2016OtsuMedianMIAS96.55% accuracy, 96.97% sensitivity, 96.29% specificity
[99]2016Morphological thresholdMedianMIAS94.54% accuracy, 5.45% false ID
Table 9. Summary of reviewed works related to classical segmentation in mammogram images: edge-based segmentation methods.
Table 9. Summary of reviewed works related to classical segmentation in mammogram images: edge-based segmentation methods.
Ref.YearTechniqueFilterDatabaseEvaluation Metric
[100]2004Edge detectionMedianMIAS83.9% accuracy
[101]2011Active contourBinary homogeneityMIAS99.6% CM, 98.7% CR, 98.3% quality
[102]2012Sobel, Prewitt, LaplacianAdobe PhotoshopNCSM79.0% (Sobel), 72.0% (Prewitt), 71.0% (Laplacian)
[103]2014Energy minimization + contourMIAS90.0% accuracy, 92.27% precision
[104]2015Dynamic graph cutMIAS + DDSM98.88% sens., 98.89% spec., 93.0% (neg. values)
[105]2016Edge detection2-DMIAS92.5% accuracy, 93.0% sensitivity, 85.0% specificity
Table 10. Summary of reviewed works related to classical segmentation in ultrasound images.
Table 10. Summary of reviewed works related to classical segmentation in ultrasound images.
Ref.YearTechniqueFilterDatabaseEvaluation Metric
[115]2014Geodesic Active Contours (GAC)SRAD
[116]2015Frequency-domainGaussian 2-D + Adaptive Z-shaped function184 BUS imagesAPR: 99.39%, ARR: 29.29%
[117]2015Robust Graph-Based (RGB) + Active Contour Model (ACM)Total Variation (TV)46 BUS imagesSensitivity: 94.50%, Accuracy: 95.42%
[118]2015Support Vector Machine (SVM) + AdaBoostBandpass + non-linearPrecision rate 89.3%
[110]2016WatershedMorphological + 3D Sobel21 cases of whole breast ultrasoundAccuracy: 85.7%
[119]2018Exploding Seeds Method (ESM) + Distance Regularized Level Set Evolution (DRLSE)Gaussian180 BUS imagesAccuracy: 99.10%, Sensitivity: 95.76%
[120]2019GLCM + Morphological + HistogramMedian250 BUS images
[121]2020Walking ParticleCanny Edge + Histogram400 BUS imagesAccuracy: 97.12%, Sensitivity: 96.30%
[122]2020Superpixel-based Graph Cut + Fuzzy C-MeansAnisotropic Diffusion110 BUS imagesAccuracy: 98%
[123]2020Simple Linear Iterative Clustering (SLIC)Histogram + Bilateral + Pyramid mean shift320 BUS imagesF1-score: 89.87 ± 4.05%
Table 11. Summary of CNN-based literature reviewed.
Table 11. Summary of CNN-based literature reviewed.
StudyYearModalityModel/FrameworkKey Results
Castiglioni et al. [126]2021Multi-modalU-Net family (U-Net, 3D U-Net, Attention U-Net)-
Suzuki [127]2017Multi-modalDeep CNNs vs. classical ML/MTANN comparison-
Lundervold [128]2019MRICNN variants (U-Net, V-Net, QSMnet) across workflow-
Raghu et al. [129]2020Medical imaging (general)Transfer learning analysis (e.g., ResNet, VGG)-
Wang et al. [130]2021Mammography, Ultrasound, MRIFine-tuned ResNet-50, InceptionV3 for breast image classificationHigh sensitivity/specificity reported
Ben et al. [131]2024Breast MRI (DCE-MRI implied)CNN-based automated lesion segmentation (vs. SVM+wavelets)-
Rouhi et al. [137]2015MammographyRegion-growing + Cellular Neural Network + Genetic Algorithm feature selectionAccuracy = 96.47%
Chiao et al. [133]2019UltrasoundMask R-CNN (detection + segmentation + classification) m A P 0.75 , Accuracy = 85%
Singh et al. [135]2019MammographycGAN-generated mass masks + shape-aware CNNDice = 0.94, IoU = 0.87, Shape accuracy = 80%
El Adoui et al. [136]2019DCE-MRIU-Net vs. SegNet segmentation comparisonMean IoU = 76.14% (U-Net) vs. 68.88% (SegNet)
Abdelhafiz et al. [139]2020MammographyU-Net for automated mass segmentationDice = 0.951, IoU = 0.909
Nawaz et al. [140]2018HistopathologyFine-tuned DenseNet for subtype classificationAccuracy = 95.4%
Zuluaga-Gomez et al. [138]2021ThermographyCustom CNN with aggressive augmentationAccuracy = 92%
Salama & Aly [134]2021Mammography (DDSM)Modified U-Net + InceptionV3 end-to-end pipelineAccuracy = 98.87%, AUC  0.988
Hossain et al. [132]2023UltrasoundRKO-UNet + spatial/channel self-attention + CNN classifierAccuracy = 98.41%
Tsochatzidis et al. [141]2021Mammography (DDSM/CBIS-DDSM)ResNet-50 + segmentation mask concatenation + spatially-aware lossAUC 0.84 0.90
Heenaye-Mamode Khan et al. [142]2021MammographyResNet-50 + adaptive learning-rate strategyAccuracy = 81.5% → 88%
Islam et al. [144]2024UltrasoundU-Net + Ensemble CNN classifier (MobileNet + Xception) + Grad-CAM-
Madhu et al. [145]2024Ultrasound (implied)UCapsNet (Enhanced U-Net + Capsule classifier)-
Aumente-Maestro et al. [143]2025Ultrasound (BUSI discussed)Multi-task segmentation + classification + dataset integrity analysis-
Table 12. Summary of transformer-based literature reviewed for breast cancer imaging.
Table 12. Summary of transformer-based literature reviewed for breast cancer imaging.
StudyYearModel/FrameworkKey Results
Chen et al. [163]2022Multi-view Vision Transformer (MVT) with local (intra-view) and global (inter-view) transformer blocks for four-view mammogramsAUC = 0.818; outperforming multi-view CNN baseline AUC = 0.784
Kassis et al. [164]2024Swin Transformer for DBT tumour detection with slice-level processing and cross-slice contextual reasoningAUC = 0.934 (high-resolution setting); outperforming ResNet101 and vanilla ViT
Mahoro & Akhloufi [165]2024TransUNet-based segmentation + downstream classification pipeline for breast thermography (background removal prior to classification)Accuracy > 97% (healthy vs. sick vs. uncertain classification)
Khan et al. [166]2025Hybrid/ensemble framework combining ViT-L16 attention representations with ResNet50 and EfficientNetB1 using ProDense block and stack-ensemble strategyAccuracy = 98.08% on INbreast dataset
Chen et al. [167]2025PatchCascade-ViT: self-supervised Vision Transformer + cascade learning for BI-RADS mammography classification-
Abdallah et al. [168]2025HybMNet: hybrid CNN + Swin Transformer components with self-supervised pretraining for mammography classification-
Sriwastawa et al. [169]2024Comparative benchmarking of transformer architectures for breast histopathology classification (ViT, PiT, CvT, CrossFormer, CrossViT, NesT, MaxViT, SepViT)-
Carriero et al. [170]2024Review synthesis of deep learning trends in breast imaging (emphasis on vision transformers, deployment barriers)-
Table 13. Summary of datasets and evaluation metrics in reviewed deep learning studies.
Table 13. Summary of datasets and evaluation metrics in reviewed deep learning studies.
ModalityCommon DatasetsEvaluation MetricsDataset Assessment
MammographyDDSM, CBIS-DDSM, INbreast, MIASAccuracy, Sensitivity, Specificity, AUC, Dice, IoUPublic and widely used, enabling benchmarking; however, some are small (MIAS) and imbalanced.
UltrasoundBUSI, institutional datasetsAccuracy, F1-score, Sensitivity, Specificity, DiceUseful for lesion detection, but limited size and diversity; high variability across institutions.
MRIInstitutional DCE-MRI cohorts, small public setsDice, IoU, AHD, AUCHigh-resolution data, but scarce public datasets; single-institution bias reduces generalisability.
HistopathologyBreakHis, BACH2018, Bioimaging2015Accuracy, Precision, Recall, F1-score, AUCLarge number of images and magnifications; strong benchmarks, though annotation variability exists.
Other (Thermal, DBT)Institutional thermal datasets, DBT cohortsAccuracy, AUC, mAPEmerging modalities; datasets are small and lack standardisation, limiting robustness.
Table 14. Comparison of classical image processing methods and deep learning methods in breast cancer detection.
Table 14. Comparison of classical image processing methods and deep learning methods in breast cancer detection.
AspectClassical MethodsDeep Learning MethodsTrade-Off
Feature BasisHandcrafted features (thresholds, edges, textures, wavelets)Automatically learned features (CNNs, RNNs, Transformers)Interpretability vs. automation
SegmentationThresholding, region growing, watershed, active contoursU-Net, Mask R-CNN, cGANs, CNN–Transformer hybridsSimplicity vs. precision
Classificationk-NN, SVM, LDA trained on handcrafted featuresCNNs (ResNet, Inception), CNN–RNN hybrids, TransformersLow data needs vs. high accuracy
PerformanceAccuracy: 75–93%, Dice ≤ 0.94Accuracy: 95–99%, Dice ≥ 0.90, AUC up to 0.98Moderate accuracy vs. state-of-the-art
Data NeedsSmall datasets are sufficient for trainingRequire large annotated datasetsResource-efficiency vs. scalability
Computational CostLow; feasible on basic hardwareHigh; requires GPUs and large memoryFeasibility vs. performance
InterpretabilityTransparent and clinician-friendlyBlack-box, difficult to explainTrust vs. complexity
GeneralisationLimited cross-dataset transferabilityStrong adaptability, but domain-shift sensitiveStability vs. adaptability
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jin, W.; Asli, B.H.S. Advances in Medical Image Processing for Early Breast Cancer Detection: Classical Techniques and Deep Learning Perspectives. Electronics 2026, 15, 790. https://doi.org/10.3390/electronics15040790

AMA Style

Jin W, Asli BHS. Advances in Medical Image Processing for Early Breast Cancer Detection: Classical Techniques and Deep Learning Perspectives. Electronics. 2026; 15(4):790. https://doi.org/10.3390/electronics15040790

Chicago/Turabian Style

Jin, Wenxian, and Barmak Honarvar Shakibaei Asli. 2026. "Advances in Medical Image Processing for Early Breast Cancer Detection: Classical Techniques and Deep Learning Perspectives" Electronics 15, no. 4: 790. https://doi.org/10.3390/electronics15040790

APA Style

Jin, W., & Asli, B. H. S. (2026). Advances in Medical Image Processing for Early Breast Cancer Detection: Classical Techniques and Deep Learning Perspectives. Electronics, 15(4), 790. https://doi.org/10.3390/electronics15040790

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop