Next Article in Journal
Diagnostic Role of Native T1 Mapping Compared to Conventional Magnetic Resonance Techniques in Cardiac Disease in a Real-Life Cohort
Previous Article in Journal
Unusual Duodenal Ulcer: Metastatic Urothelial Carcinoma of the Renal Pelvis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

A Review of Machine Learning Techniques for the Classification and Detection of Breast Cancer from Medical Images

1
Maharaja Research Foundation, University of Mysore, Mysuru 570005, India
2
Department of Computer Science and Engineering, Maharaja Research Foundation, Maharaja Institute of Technology, Mysuru 570004, India
3
Biomaterial Bank Nord, Research Center Borstel Leibniz Lung Center, Parkallee 35, 23845 Borstel, Germany
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(14), 2460; https://doi.org/10.3390/diagnostics13142460
Submission received: 2 June 2023 / Revised: 17 July 2023 / Accepted: 18 July 2023 / Published: 24 July 2023

Abstract

:
Cancer is an incurable disease based on unregulated cell division. Breast cancer is the most prevalent cancer in women worldwide, and early detection can lower death rates. Medical images can be used to find important information for locating and diagnosing breast cancer. The best information for identifying and diagnosing breast cancer comes from medical pictures. This paper reviews the history of the discipline and examines how deep learning and machine learning are applied to detect breast cancer. The classification of breast cancer, using several medical imaging modalities, is covered in this paper. Numerous medical imaging modalities’ classification systems for tumors, non-tumors, and dense masses are thoroughly explained. The differences between various medical image types are initially examined using a variety of study datasets. Following that, numerous machine learning and deep learning methods exist for diagnosing and classifying breast cancer. Finally, this review addressed the challenges of categorization and detection and the best results of different approaches.

1. Introduction

The second-leading cause of mortality is breast cancer, which is the cancer type that is most frequently diagnosed. In 2020, more than 2 million new cases of breast cancer were discovered, making it the most diagnosed disease in the world, according to the World Health Organization (WHO). A total of 626,700 women lose their lives to cancer-related conditions every year. Breast cancer is the most common malignancy in women and the second leading cause of death, and if it is not caught early enough, it can be fatal [1]. If the cancer is found before it expands to a size of 10 mm, the patient has an 85% probability of going into complete remission. Cohort studies show that 30% of breast cancer cases are identified when the tumor is 30 mm. Breast cancer is usually detected during screening when the tumor is at least 20 mm in size [2]. Therefore, encouraging early diagnosis of breast cancer is crucial. A positive clinical breast examination (CBE) and breast self-examination (BSE) may warrant early intervention. Healthcare professionals conduct a CBE as part of routine medical examinations to look for breast lesions. BSE also includes a patient physical examination to look for any physical changes and breast appearance. Women may take charge of their health thanks to the BSE technique. The World Health Organization recommends that at-risk women learn more about BSE [3]. Screening procedures are used to create medical images of the breasts. Professionals with human skills, such as radiologists and doctors, typically interpret these images. Medical imaging’s poor diagnostic accuracy is caused by a lack of technological expertise and expertise in analyzing such images.
Breast cancer has two types, benign (not hazardous) and cancer (malignant) and falls into two groups: normal or aberrant. Benign conditions are thought to be non-cancerous or not life-threatening. But occasionally, it might progress to cancer status. Cancer begins with unnatural cell development and has the potential to spread or infiltrate neighboring tissue quickly. The nuclei of malignant tissue are typically substantially larger than those of normal tissue, which can be fatal in later stages. If malignancy is discovered, tissue is often collected from a surgical biopsy, and less frequently from a tiny needle or larger cardiac biopsy for microscopic examination. Examining current investigative data and extracting pertinent information from earlier data are the key components of an early and precise diagnosis of this disease. Machine learning (ML) algorithms and medical imaging support the procedure. Various deep learning and machine learning techniques have been created to help clinicians understand medical images. Over the past ten years, the development of machine and deep learning models has attracted much attention. Pre-trained networks can be used to install models because they are freely available.
Artificial intelligence (AI) has advanced quickly in recent years. AI aids and assists medical professionals in recognizing and forecasting illness risk more quickly and precisely, enabling early disease identification. Artificial intelligence (AI), a rapidly expanding phenomenon, may soon lead to significant developments in various industries, including medical education.
AI techniques have improved medical image processing, computer-aided diagnosis, image interpretation, fusion, registration, segmentation, image-guided therapy, image retrieval, and image analysis. These techniques enhance the ability of scientists and medical practitioners to comprehend how to recognize the genetic changes that will cause disease [4].

2. Artificial Intelligence and Machine Learning Techniques

To anticipate, characterize, or respond to an issue, models must be created using Machine Learning (ML), a subsection of Artificial Intelligence (AI) (Figure 1). Machine learning is a term used to describe a group of techniques that “detect patterns in data, use the uncovered patterns to predict future data or other outcomes of interest”. ML approaches learn directly from data, eliminating the need for numerous expert rules or precisely representing every environmental element. Due to its independence from typical ML extraction techniques and capacity to focus on a complicated hierarchy of image attributes, Deep Learning (DL) is a subclass of Machine Learning (ML) and AI. Computers read data from photos and apply deep learning methods to various previously used computer models to make significant gains. Many medical specialties, particularly radiology and pathology, have used these algorithms to fulfill tasks. Their performance has occasionally been on par with that of human professionals. According to [5], DL can extract information from medical images similar to human analysis and offer details on molecular status, prognosis, and therapy sensitivity. Convolutional neural networks, deep Boltzmann machines, and deep neural networks are popular DL methods.
It is possible to distinguish between supervised, unsupervised, and agent-based machine learning techniques.
Supervised learning: In supervised machine learning (ML), every issue may be viewed as the learning of a parametrized function, also referred to as a “model”, that maps inputs (i.e., predictor variables) to outputs (i.e., “target variables”). The purpose of supervised learning is to utilize an algorithm to extract the parameters of those functions from the given data. It is possible to think of supervised learning as using logistic and linear regressions. The majority of ML approaches fall within this category. SVM, DT, Clustering-NN, and K-means are examples of supervised machine learning methods.
Unsupervised learning: ML issues are often far more difficult to solve if the target variables are unavailable. Unsupervised learning uses the common dimensionality reduction and clustering tasks to find correlations or patterns in the data without providing any direction for the “correct” answer.
Agent-based learning: Between guided and unguided learning: It is a collection of machine learning techniques where learning occurs by replicating the activities and communications of a single autonomous agent or a group of autonomous agents. To effectively learn, carefully determine values (or preferences), and employ inquiry procedures, one must deal with problems that regularly arise in real life. It is necessary to develop generalizable models because these general unsupervised approaches rely on target obtain variables for which there is little information. Only by experimenting can one detect essential parts of the surroundings. In this context, a specific example of a problem with decision-making over time is reinforcement learning.
This Section provides a concise introduction to frequently employed ML approaches.

2.1. Decision Tree

Decision trees (DT) are frequently used in “divide and conquer” data classification strategies. With this method, the data can be shown as a tree, with the leaf nodes serving as the labels for the data sample labels and the inside nodes serving as the various qualities [6]. The appropriate data set is identified by moving up and down the tree from root to leaf. Decision trees (DTs) are a class of supervised learning algorithms, and DTs can be used to classify classification and regression problems. The most widely used DT algorithm is C4.5 [7]. Authors [8] compared C4.5 with other DT algorithms.

2.2. Random Forests

The most common use of a bagged DT is called a random forest (RF) [9]. It is an ensemble model made up of numerous separately trained DTs. The class with the most votes is chosen as the final classification for the input data by each component DT in an RF model, which makes classification decisions for each component DT individually. Regression models that use RFs can also produce results that are averaged over the results from individual trees. The RF algorithm’s fundamental idea is that a randomly chosen subset of features is selected at each node of each tree; the samples used to train each component tree are determined using bagging, which resamples (with replacement) the initial set of data points.

2.3. K-Nearest Neighbor

A data sample is compared to other data samples using a distance metric in the K-Nearest Neighbor (k-NN) algorithm. The distance between two identical data samples can be reduced using a distance metric, and the distance between two data samples can be increased (The typical Euclidean distance is typically used to calculate the separation between two data samples). An equation will provide the Euclidean distance between x and y. This approach is called the nearest k-method [6].

2.4. Support Vector Machine: (SVM)

The Support Vector Machine (SVM) is a popular machine-learning technique for issue regression and categorization. SVM was applied in various applications, including cheminformatics [10] and bioinformatics. The SVM classifier builds a model for the classification using training data. A later step is the classification of an unidentified sample. The main idea behind SVM is using hyperplanes to separate diverse groups. In cases where data can be separated linearly, SVM has attained significant precision. SVM output, however, is unable to separate separable data non-linearly. The data can be split linearly after being mapped to a new, high-dimensional space, utilizing kernel functions to tackle this problem. The correct kernel function selection and its parameters are two of the main issues with SVM [11]. The kernel function is a mathematical technique that enables Support Vector Machines (SVMs) to perform a classification of a set of initially one-dimensional data in a ‘two-dimensional’ manner. Generally, a kernel function facilitates the projection of data from a lower-dimensional space to a higher-dimensional space. Linear kernel function is commonly described as:
K (x, xi) = x·xT
Polynomial Kernel Function: The polynomial kernel function is directional, i.e., the output depends on the direction of the two vectors in low dimensional space. This is due to the dot product in the kernel. The magnitude of the output is also dependent on the magnitude of the vector xi [12].
K (x, x1) = (1+ x·Txi)) d, ‘d’ is degree of kernel function

2.5. Artificial Neural Networks (ANN) and Deep Learning (DL)

According to [13] the interconnected neuronal biologic network in the human brain is comparable to an artificial neural network (ANN). The most popular ANN for categorizing ANN issues using the backpropagation training strategy is feedback. In a direct-acting ANN, a single neuron’s basic structure is depicted in Figure 2 [14]. In an ANN, a single neuron obtains input from other neurons, multiplies it by the appropriate weight Wij, and then uses an activation function to create a weighted output f. (Xj). In this kind of network, the neurons are arranged into 180 different types of strata. One input layer, one output layer, and numerous hidden layers are common for direct-action ANNs. The training algorithm provides the technique for weight adjustments in Feedforward ANN during preparation for backpropagation. The output layer compares the actual output generated by the ANN to the desired outcome. The difference between the actual (determined) and anticipated (target) outputs is then used to calculate the error. Finally, during the following iteration, the network receives a new transmission of the mistake of changing the weights. “Deep learning” refers to convolutional neural networks and deep neural networks, which are ANNs with several hidden layers.

2.5.1. Convolutional Neural Network (CNN)

CNNs are the most widely used image algorithms. Convolutional neural networks are commonly used in image classification, image recognition, object detection, etc. [15,16]. The classification of visual images is the deep learning task where convolutional neural networks are most frequently used. In addition, CNN is a high-dimensional image dataset, and each image typically comprises thousands of pixels as traits that are complicated and large in volume. CNN is a feed-forward network that can identify an image’s topological properties. CNNs are multilayer perceptron-driven models. The three distinct layers CNNs use—which are more closely related to neural networks—are convolutional, pooling, and fully connected. Every layer performs a different purpose. The convolutional layer has been used as a feature extractor. The fully connected layer uses the extracted function to determine the current entry’s category. Reducing the number of feature maps and network parameters is the function of a pooling layer.
Convolutional Layer: Convolutional layers are utilized to produce feature maps using the weight distribution theory and the local connection concept. Local connectivity and weight distribution objectives are used to reduce the number of parameters while maximizing the advantages of the strongly connected local pixel neighborhood and location-independent local image statistics. The weight distribution model looks like this. Each unit (neuron) in a feature map only has a “local connection” to surrounding patches of the feature map at the previous stage thanks to a weight group called a “filter bank”. Each unit has a filter row they share on a feature map. Other feature maps employ different filter banks as well. The weighted sum of each unit serves as the input to the activation function, a nonlinear transformation function. The weighted total of each succeeding unit is sent to the activation function, a nonlinear transformation function. According to [17], the activation function enables the transmitted data to change nonlinearly for subsequent processing steps.
Pooling layer: The pooling layer combines (semantically) linked convolutional layer features into one using a subsampling technique. A unit within a pooling layer uses a local patch as input from a previous entity map (convolutional layer) to calculate the maximum or average patch value at the output. Reduced representation size and increased robustness lower the parameters needed in later stages.
Fully connected layer: According to a multilayer perceptron, a classic neural network, the units in this layer are fully connected to all the units in the layer above.
The organization of its physical space significantly influences CNN’s effectiveness and efficiency. How each layer is constructed, what materials are utilized, and how the layers are arranged impacts how quickly and precisely specific tasks may be completed. Ten CNN architectures will sum it up as follows:
  • LeNet-5: LeNet-5 [13] has two convolutional layers and three fully linked layers.
  • According to [15], AlexNet includes five convolutional and three fully linked layers.
  • VGG-16 [18] uses three fully connected layers and thirteen convolutional layers, taking the ReLU from Alex Net.
  • According to [19], Inception-v1 has a 22-layer architecture with 5 M parameters.
  • According to [20], Inception-v3 is a version of Inception-v1 with parameters of 24 M.
  • ResNet-50: A network with 50 layers [21].
  • Thirty-six convolutional layers, according to Xception [22].
  • Inception-V4: According to [23] Inception-V4 consists of a feature extractor and fully connected layers.
  • One hundred and sixty-four layers are deep in Inception-ResNet.
  • According to ResNeXt-50 [24], it comprises five phases, including an identification and convolution block in each stage. Every identity block contains three levels, and every convolution block has three layers [23].

2.5.2. Recurrent Neural Network (RNN)

The RNN is a specific type of neural network with a specific mathematical description. In the sequence-to-sequence problem, both the input and the output have sequential structures. Recurrent neural networks (RNN) can be utilized to address this problem. The RNN unit has a hidden state that allows it to handle sequential data of different sizes. The hidden state of RNN, which holds some information about a sequence, is its primary and most significant feature. The state, which recalls the previous input to the network, is also known as the memory state. It executes the same procedure on all of the inputs or hidden layers to produce the output, using identical variables for each input. Compared to other neural networks, this minimizes the complexity of the parameter set.
The input and output architecture of RNNs is identical to that of other deep neural network architectures (Figure 2). An RNN keeps track of every piece of data over time. Only the ability to remember past inputs makes it helpful for time series prediction, known as a long short-term memory. RNN is very useful for image recognition, face detection, time series forecasting, language modelling and generating text, and speech recognition machine translation. RNNs are formed as:
Bidirectional Neural Network (BiNN): A BiNN is a type of recurrent neural network in which data are input in both directions and output from both directions is combined to create the input. In cases such as NLP tasks and time-series analysis issues, where the context of the input is more crucial, BiNN is helpful.
Long Short-Term Memory (LSTM): based on the read–write–forget principle, which states that given an input of information, a network should read and write just the information that will be most valuable in forecasting the outcome and ignore the rest. Three new gates are added to the RNN to accomplish this. Only the chosen information is transmitted via the network in this way [25].

2.5.3. Deep Convolutional Neural Network

The most widely used type of deep learning model is the deep convolutional neural network, which is utilized for large-scale image recognition tasks, particularly in the study of medical imaging.
The DCNN consists of three main layers: convolutional layer, pooling layer, and fully connected layer.
(1)
Convolutional layer: The main component of the DCNN is the convolutional layer, filter or kernel weights represent the layer parameters. The output feature map is created by multiplying each of the receptive fields, which are the small areas formed by the input feature map. If the stride hyperparameter is smaller than the filter size, the convolution is performed in overlapping windows. The stride is the distance between the applications of filters.
(2)
Pooling layer: Down sampling the input’s spatial dimension is performed by the pooling layer. The primary goals of this layer type are the gradual reduction of the representation’s spatial dimension and the reduction of the parameters and calculations needed by the network. Although many different pooling functions are available, such as average pooling and L2-norm pooling, max pooling is the most popular since it computes the maximum in the input patch.
(3)
Fully connected layer: A conventional multi-layer perceptron with a SoftMax activation function in the output layer makes up the completely connected layer. Neuronal cells link to every activation in the preceding layer. To categorize the input image using high-level features extracted from convolutional and pooling layers is the goal of the fully connected layer [26].

3. Types of Breast Cancer Imaging

Medical imaging, which frequently employs a range of modalities, including MRI, CT, PET, mammography, radiographic ultrasonography, and duplex ultrasound, is the most efficient way to find breast cancer. Medical images help in disease diagnosis, pathological lesion detection, patient therapeutic care, and the assumption of many disorders. Medical image analysis is one application where ML and AI have been quite successful recently. By using image processing and machine learning techniques for the early identification and diagnosis of malignancies, it is now possible to increase the accuracy of breast cancer diagnosis.

3.1. Mammography Images

Mammography is a type of medical imaging that uses a low-dose X-ray system and is mainly used to test for breast cancer. It can be used to find malignant tumors inside the breast. Mammograms help diagnose breast cancer in persons with odd symptoms or breast nodules; even while screening, mammography helps determine the cancer risk in women without overt symptoms. According to the American Cancer Society, every woman over 40 should obtain a mammogram once a year. On a mammography, dense breast tissue may appear white or light grey. Mammograms of younger women may be easier to view since they appear to have larger breasts. They may spot added in situ lesions and reduce invasive tumors compared to MRI and ultrasounds. Mammography remains the gold standard for community breast cancer screening [27]. Mammography is the most reliable and accurate screening technique. Mammography is still used with MRI and ultrasound, especially with high breast tissue density, but it cannot replace mammography. Mammograms can be viewed in various ways to give more details before detection or diagnosis. The two most common mammography views are cardio-caudal (CC) and mediolateral oblique (MLO). The breast is squashed between two paddles in the CC view mammography, which is taken horizontally from an upward projection at a C-arm angle of e 0°. This reveals the glandular tissue, the surrounding fatty tissue, and the outermost edge of the chest muscle. The breast is squashed diagonally between the paddles in the MLO view of mammography, which is captured at a 45-degree angle from the side of the C-arm. As a result, more breast tissue is visible than in other views.

Mammography Datasets

Several datasets are available to the public, and they differ in terms of their size, resolution, image format, and the types of images they contain (such as Full-Field Digital Mammography (FFDM), Film Mammography (FM), or Screen-Film Mammography (SFM), and the kinds of abnormalities they contain). DDSM breasts, Mini-MIAS, the Curated Breast Imaging Subset of DDSM (CBIS-DDSM), and BCDR are a few examples of public datasets.
  • The digital database for screening mammography (DDSM) comprises 2620 mammograms scanned from film which were then separated into 43 volumes (Figure 3). For each example, there are four breast mammograms since the Mediolateral Oblique (MLO) and Cranio-Caudal projections were used to photograph each breast side. The dataset includes pixel-level annotations for the suspicious regions and the ground truth. The breast density score for each patient was calculated using the ACR BI-RADS (American College of Radiology Breast Imaging Reporting and Data System). The file for each case also contains information about the size and resolution of each scanned image. JPEG (Joint Photographic Experts Group) format, available in various formats and resolutions, was used for the images.
  • The Curated Breast Imaging Subset of the DDSM (CBIS-DDSM) is an upgraded version of the DDSM that includes bounding boxes for the region of interest (ROI), updated mass segmentation, and decompressed pictures. The data were picked and reviewed by mammographers with the necessary training, and the images are in the Digital Imaging and Communication in Medicine (DICOM) format. The collection is 163.6 GB in size and contains 6775 studies. There were 10,239 images in total, all mammography scans with associated mask images. CSV files are associated with the dataset that includes the patients’ pathological data. A mass training set, a mass testing set, a training set for calcification, and a testing set for calcification make up the dataset’s four CSV files. The mass testing set only includes images for 378 cancers, whereas the dataset consists of images of 1318 tumors. Images for 1622 calcifications are included in the calcification training set, whereas photos for 326 calcifications are included in the calcification testing set.
  • IN Breast: Breast consists of 410 images and 115 cases. In 90 of the 115 cases, there was malignancy in both breasts. The dataset represents the four types of breast illnesses: breast bulk, breast calcification, breast asymmetries, and breast distortions. Images of (CC) and (MLO) views, stored in DICOM format, are included in the dataset. The dataset also offers the breast density score from the Breast Imaging-Reporting and Data System (BI-RADS).
  • Mini-MIAS: The dataset includes ground truth indicators for potential abnormalities and 322 digital films. The collection contains five types of abnormalities: masses, architectural distortion, asymmetry, and normal. Ultimately, 1024 by 1024 pixels) were used as the final resolution for the images. The images are accessible to everyone on the University of Essex’s Pilot European Image Processing Archive (PEIPA).
  • BCD: The BCDR consists of two mammographic repositories:
    The BCDR-FM and the Film Mammography-based Repository.
    The BCDR-DM, or Full Field Digital Mammography-based Repository.
The BCDR repositories include normal and atypical breast cancer cases and the clinical information needed to treat them. The 1010 cases in the BCDR-FM are split between 998 females and 12 males. Additionally, it contains 104 detected lesions and 3703 mammographic pictures in the two perspectives, MLO and CC, from 1125 investigations.

3.2. Ultrasound Images

Another type of medical imaging technique for finding tumors is breast ultrasonography (Figure 4). The comprehensive medical images of the breast produced by ultrasound are created using sound wave techniques. This method is considered suitable and safe for pregnant women who cannot utilize X-rays or CT scans and nursing mothers. Additionally, pregnant women and nursing mothers who cannot use X-rays or CT scans are considered suitable and safe candidates for ultrasound.

Ultrasound Dataset

Grayscale images make up the majority of ultrasound (US) images stored in a DICOM format at Baheya Hospital. The three categories of the US dataset are benign, malignant, and normal. A total of 1100 images were initially acquired. After preprocessing, 780 of the dataset’s initial 1080 images remained. The LOGIQ E9 ultrasound system and the LOGIQ E9 Agile ultrasound system are used in the scanning procedure. High-quality radiology, cardiology, and vascular care imaging routinely use these approaches. They create images with a 1280 × 1024 resolution. ML6-15-D linear probe transducers operate at 1–5 MHz (Table 1).

3.3. Magnetic Resonance Imaging (MRI)

Along with ultrasound and mammography, magnetic resonance imaging (MRI) is an early cancer cell detection method. Magnetic fields are used in MRI to produce incredibly accurate three-dimensional (3D) transverse images. A considerable radiation dose is required for a human body MRI to provide precise 3D breast images. Therefore, the diseased area changes rather noticeably when we use an MRI, and no malignancy is found that cannot be seen in any other method. Breast MRI is important for the early detection of breast cancer, in part because it provides quick diffusion-weighted imaging and T2- and T1-weighted CE (contrast-enhanced) imaging, which can be used to characterize lesions further. Breast MRI is essential in identifying complications in women at high risk of breast cancer, among other factors.
  • High risk of developing breast cancer.
  • Evaluation of the staging period.
  • Neoadjuvant chemotherapy (NAC) follow-up.
  • Evaluation of an auxiliary lymph node region when mammography could not identify the primary location.
A breast MRI takes thirty to forty minutes, following best practices and benchmark procedures.

MRI Datasets

  • Breast–MRI–NACT–Pilot dataset: The database for this dataset is 19.5 GB in size and contains 99,058 MRI images for 64 patients.
  • Mouse–Mammary: This dataset has 23,487 Images, the database is 8.6 GB, and there are 32 patients.

3.4. Histopathological Images

The standard for diagnosing cancer has not altered, despite the fast advancements in medical technology. The tissues involved in disease are portrayed in microscopic detail in histopathological image analysis (Figure 5). The pathologists’ experience and factors such as fatigue and a decline in brain function impact the histopathological investigation’s lengthy and highly skilled process.

Dataset for Breast Cancer Histopathological Images

The majority of studies on BC histopathology image analysis, according to the literature, are based on small datasets that are often not shared with the scientific community. The Break His dataset is introduced in this review. At four distinct magnifications (40×, 100×, 200×, and 400×), 82 patients provided 7909 microscopic photos of breast tumor tissue that were clinically realistic. These images were collected for BreaKHis. It now has 2480 benign samples and 5429 cancerous ones. All information was made anonymous. Hematoxylin and eosin (HE)-stained breast tissue biopsy slides were used to create the samples. Pathologists from the P&D Lab obtained the samples through surgical (open) biopsy (SOB), prepared them for histological analysis, and labeled them.

3.5. Thermography Images

Thermal imaging (Figure 6) often called breast thermography, is created under a microscope. This makes it possible to examine how cells, tissues, and organs’ microscopic anatomy relates to their structural and functional makeup. Finding breast changes that might be symptoms of breast cancer is routinely performed using this painless, non-invasive method [28]. A thermal infrared camera that converts infrared light into electrical impulses and displays it as a thermogram can be used to diagnose breast cancer by identifying body sections that suggest an unusual temperature shift. Thermography is the name of this technique. Thermal imaging employs sensitive and high-resolution thermal cameras and is a promising early diagnosis technique. Thermal imaging, in conjunction with artificial intelligence (AI), is a successful technique for identifying early-stage breast cancer and is projected to have very high levels of prediction.

3.5.1. Thermal Camera

The Plank equation illustrates the connection between the body surface’s wavelength, temperature, and radiation. The discharge of electromagnetic waves occurs when a body’s temperature exceeds absolute zero. A gadget is required to detect this wavelength since infrared light has a range of invisible wavelengths to the human eye. One of the better methods for figuring out the wavelength range is to use a thermal camera. Any object warmer than absolute zero will emit infrared radiation, which can be detected using a thermal camera. The wavelengths of infrared light typically lie between the visible and microwave spectrum. The wavelength range of this infrared light is 0.75 Mm to 1000 mm. Infrared breast thermography may be used to identify breast cancer early to boost the survival probability of patients with the disease. Therefore, a slight asymmetry between the left and right breast temperature patterns may indicate a breast anomaly. The interpretation of the asymmetry in breast thermograms depends critically on several textural characteristics.

3.5.2. Thermography Datasets

The Mastology Research with Infrared Image (DMR) database is used for most IR-thermal image research. There are 287 people in the DMR-IR database, ranging in age from 23 to 120; 186 of them are healthy, and 48 have breast cancer.

3.6. Positron Emission Tomography (PET)

Breast PET, also known as Positron Emission Tomography, is a medical imaging method used to visualize and identify metabolic activity in breast tissue. This technique involves introducing a small amount of a radioactive tracer, typically a sugar-based compound, into the patient’s bloodstream. The tracer accumulates in regions with higher metabolic activity, such as cancerous cells, emitting positrons (positively charged particles). When these positrons interact with electrons in the tissue, they produce gamma rays, which are then detected by the PET scanner to generate images.
Machine learning plays a crucial role in breast PET imaging, particularly in improving image quality, enhancing cancer detection accuracy, and aiding in diagnosis. However, compared to other imaging techniques, nuclear medicine modalities such as PET or scintigraphy have been considered less effective in evaluating early-stage breast cancer. Additionally, publicly available breast PET datasets are not as abundant as datasets for other medical imaging modalities.
It is important to emphasize that the successful application of machine learning in breast PET heavily depends on access to high-quality and well-annotated datasets. Having access to such datasets is vital for training accurate and reliable machine learning models for breast PET imaging analysis (Figure 7) [29].

4. Review on Machine Learning in Breast Cancer

Using the information acquired from medical imaging, machine learning, and deep learning algorithms may recognize, categorize, and diagnose breast cancer. This section discusses how to spot breast cancer using a mammography, ultrasound, MRI, or a thermogram.

4.1. Machine Techniques for Mammogram Images

To determine the most accurate technique to categorize mammography pictures based on breast cancer concerns, the authors of most papers utilized different methods to classify machine learning. The articles for mammography images using several machine learning models (SVM classifier, ANN, K-NN, Fuzzy C-Means, and CNN) are gathered and summarized in Table 2.
The author [30] employed an SVM classifier to categorize the mammography images by the traits deduced from the Hough transformation to determine the mammograms’ traits. On 95 clinical pictures, they used the SVM approach, and the findings demonstrate that, without mentioning accuracy, the suggested method efficiently categorizes the problematic classes of mammograms.
With a 93.1% accuracy rate [31], classified tumors into benign or malignant types using SVM on the DDSM dataset.
The authors [32] conducted a simulation experiment using 44 mammography images from the MIAS database and the same premise as the earlier work. They stated that the accuracy of the mass categorization was as high as 95%. The DDSM database’s simulated mammography images demonstrated the method’s 93% accuracy.
K-means clustering technique is used in the K-SVM-based model for cancer diagnostics described by [33] to extract symbolic objects from tumors. K-SVM estimated an accuracy of 97.38% using the WDBC dataset.
In [34], the authors proposed a method for mass detection that combines deep CNN and SVM. The final fully connected layer data was utilized to build a high-level characteristic representation of the image for categorization on mammography spots in the CNN model.
An improved DenseNet neural network model, also known as the DenseNet neural network model, was suggested in a different study [35] for the accurate and reliable categorization of benign and malignant mammography. The mammography images from 2042 Shanxi Medical University hospital cases were utilized to create the dataset. According to the study’s findings, the DenseNet-II neural network model outperforms other models in categorization.
In [36] used the SVM classifier on the Mini-MIAS and INBreast datasets to categorize breast density and mass as abnormal or normal. With an AUC of 0.9325, the results and performance reviews are 99% correct.
The accuracy of the classification of the extracted regions as mass or non-mass by the authors [37] using DDSM was 98.88%.
The classification approach proposed by [38] uses an SVM classifier to distinguish between abnormalities (mass or microcalcification) and (benign or malignant). INBreast and MIAS were employed for this investigation. The outcomes indicated that the precision value was 99%. The AUC score was 0.9990.
Another researcher [39] also developed an SVM classifier to categorize anomalies utilizing fusion functions on MIAS data. According to the findings, the precision value was 93.17%.
With an accuracy of 91.25%, Ref. [40] suggested a method for categorizing cancer as normal and abnormal using SVM algorithms on the MIAS dataset (109 cases).
Using the MIAS dataset [41], employed an ANN classifier to determine whether something was normal or abnormal before classifying the abnormal into benign or malignant conditions. The results of their work are as follows: RBF accuracy (normal/abnormal): 93.98%; Sn value: 97.22%; RBF accuracy (benign/malignant): 94.29%; Sn value: 100%.
To classify tumors into benign or malignant tissue, authors [42] built Fuzzy C-Means (FCM) on DDSM and MIAS datasets, and they verified that accuracy is 87%, sensitivity value (90.47%), and specificity value (84.84%).
In [43] used an associative classifier with fuzzy–ANN for breast tissue and mass classification. According to the performance assessment, the accuracy for the DDSM dataset is 95.11%.
Authors [44] used the IRMA and MIAS datasets to deploy a K-NN classifier that correctly identified ROI as normal or abnormal with an accuracy of 92.881 ± 0.0099, a Sn value of 92.885 ± 0.0099, and an AUC value of 0.9713.
An updated Dense-Net neural network model, also known as the DensNet neural network model, was suggested by a different study [35] for the accurate and reliable categorization of benign and malignant mammography. The authors initially preprocessed the mammography images. The DenseNet neural network model was then enhanced, and a new DenseNet-II neural network model was produced by replacing the first convolutional layer with the startup structure. Finally, the preprocessed mammography datasets were added to the neural network models AlexNet, VGGNet, GoogLeNet, DenseNet, and DenseNet II. The average model’s accuracy is 94.55%.
The study [45] used an ANN classifier on the MIAS dataset (57 pictures, 37 benign, and 20 malignant tumors), and the accuracy of the results was 96.89%.
K-NN was employed by the authors [46] to classify 100 pictures from MIAS and BancoWeb as benign or malignant. The accuracy of the result was 96%.
Ref. [47] developed CNN with an end-to-end algorithm to obtain high-level attributes. The Henan Provincial People’s Hospital’s Department of Radiology provided the mammography data used in this investigation. The accuracy and AUC ratings for the suggested model are higher.
Convolution neural network (CNN) architectures such as Inception V3, ResNet50, Visual Geometry Group networks (VGG)-19, VGG-16, and Inception-V2 ResNet were applied to the MIAS dataset by the authors of [48]. The classification of mammogram breast images produced the following results: For the 80–20 method and the 10-fold cross-validation method, the overall accuracy, sensitivity, specificity, and precision were 98.96%, 97.83%, 99.13%, 97.35%, 97.66%, and 0.995, respectively. The author used techniques on the DDSM dataset and implemented CNN architecture with transfer learning, with results of 92.84% accuracy.
In addition, the authors [49] proposed K-NN on 252 images from Mini-MIAS and DDSM databases to explore the differences between normal and diseased breast tissues. The accuracy of the result was shown to be 91.2%.
The Fuzzy Gaussian Mixture Model (FGMM) is a method proposed by [50] for classifying 300 photos from DDSM as benign or malignant. The accuracy of the method was 93%.
Authors in [51] used a novel deep learning framework for the detection and classification of breast cancer. They classified images to benign and malignant. They are based on the CNN architecture of GoogleNet and VGGNet for classification. The GoogLeNet, VGGNet, and ResNet architecture individually give an average classification accuracy of 93.5%.
Table 2. A summary of the research publications for using deep learning and machine learning in mammogram images.
Table 2. A summary of the research publications for using deep learning and machine learning in mammogram images.
ReferencesDatasetML MethodResults
[36]Mini-MIAS INBreastSVM ClassifierAccuracy of 99%
AUC value is 0.933
[52]DDSMSVM ClassifierSn value is 82.4%
[37]DDSMSVM ClassifierAccuracy is 98.9%
[38]MIAS
INBreast
SVM ClassifierAccuracy is 99% ± 0.50
AUC value 0.99 ± 0.005
[39]MIASSVM ClassifierAccuracy 93.17%
[40]MIAS: 109 casesSVM ClassifierAccuracy value from 68% to 100%
[41]MIASRBFNN classifierRBF (normal/abnormal)
Accuracy is 93.9%
Sn value is 97.2%
RBF (benign/malignant)
Accuracy is 94.3%
Sn value is 100%
[53]Private-1896 casesGLCM
SFFS (sequential floating forward selection)
the bilateral CC and MLO view images
Sn-value is 68.8%
Sp value is 95.0%
The AUC value is 0.85 ± 0.046
[45]MIAS: 57 images
37 benign
and 20 malignant
CNN classifierAccuracy is 90.9%
AUC value is 96.9%
[46]MIAS
-BancoWeb: 100 images
CNN and hybrid of K-means aAccuracy 96%
[42]DDSMFuzzy C-Means
(FCM)
Accuracy is 87%
Sn value is 90 to 47%
Sp value is 84 to 84%
[49]252 images from
Mini-MIAS
-DDSM
KNNAbnormality detecting: Accuracy is 91.2%
AUC value is 0.98 Malignancy detecting:
Accuracy is 81.4%
AUC value is 0.84
[50]300 images from
DDSM
Fuzzy Gaussian
Mixture Model
(FGMM)
Accuracy is 93%
Sn value is 90%
Sp value is 96%
[44]IRMA-MIASk-NNAccuracy is 92.8% ± 0.009
Sn value is 92.85% ± 0.01
AUC value is 0.971
[54]DDSMCNN and transfer learningSensitivity of the mass 89.9%
[55]DDSM, MIASLS SVM, KNN, Random Forest, and Naive BayesAccuracy 92%
[56](Mini-MIAS)
DDSM
CNNThe accuracy of 0.936, 0.890, 0.871 on the DDSM, 0.944, 0.915, 0.892 on the Mini-MIAS for normal, benign, and malignant regions
[48]MIASCNN a pre-trained architecture such as Inception V3, ResNet50, Visual Geometry Group networks (VGG)-19, VGG-16, and Inception-V2 ResNetOverall Accuracy, Sn, Sp, precision, F-score, and AUC of 98.96%, 97.8%, 99.1%, 97.4%, 97.7%, and 0.995, respectively, for the 80–20 method and 98.87%, 97.3%, 98.2%, 98.84%, 98.04%, and 0.993 for the 10-fold cross-validation method, the TL of the VGG16 model is adequate for diagnosis.
[57]DDSMCNN Accuracy 71.4%

4.2. Machine Learning Techniques for Ultrasound Images

By analyzing 138 privately-owned instances, the researchers [58] suggested an SVM machine learning approach for discriminating benign and malignant tumors. The findings show an accuracy of 86.96%.
In 283 privately-owned cases, authors [59] employed numerous models (DT, SVM, RF, K-NN) to distinguish between benign and unsettling lesions accurately.
A data collection contained 8145 breast ultrasound images overall. The authors developed a deep convolutional approach with multi-scale kernels and skip connections [60]. This technique consists of two steps: the first is to identify solid nodules, and the second is to detect whether there are malignant tumors in the image.
Three convolutional layers and two linked layers were employed by the authors of [61] to classify 166 private tumors with 292 benign masses as breast cancer. The outcome for accuracy was 83%.
The authors [62] proposed CNN to detach skin, fibro glandular tissue, 3D mass, and adipose tissue from breast ultrasound images. Quantitative criteria used to evaluate segmentation effects, such as precision, recall, and F1 calculations, all topped 80%, confirming the suggested technique’s capacity to identify functional tissue in a breast ultrasound image.
To identify ultrasonic lesions, the authors [63] investigated using a patch-based LeNet, a U-Net, and a transfer learning method that employed a FCN (Fully Convolutional Network).
In [61] suggested a CNN system concentrating on transfer learning that produced a 93.6% AUC on 292 benign and 166 malignancies to distinguish between benign and malignant breast lesions. They set up the CNN VGG19 model, which had been trained on 882 images of ultrasonography breast masses from the ImageNet dataset. The findings were 83.0% accuracy and 82.4% sensitivity using the mean AUC.
On 46 privately-held images, the authors [64] utilized an SVM classifier to find the tumor location. The accuracy of this method was 95%.
Other studies, such as [65] achieved a 94.81% accuracy rate while utilizing SVM algorithms to distinguish between normal and abnormal instances using 169 privately-held cases.
Additionally, Ref. [66] provided SVM for detecting and diagnosing breast masses using 120 privately-obtained pictures, with findings demonstrating a 95.85% accuracy.
CNN is based on Google Net. The authors [67] trained and tested CNN on 7408 ultrasound images on 829 images. Results from the suggested model attained 90% accuracy and 86% sensitivity.
U-net was used by [68] to identify the bulk of 433 private ultrasound images and the results achived 84% accurately (Table 3).

4.3. Machine Learning Techniques (MLT) for Thermography

There are many publications that have utilized thermography to identify and categorize breast cancer. A list of thermography articles employing ML and DL methods is shown in Table 4.
According to varied results between 88.10% and 2.5%, some authors in various papers employed SVM to categorize their images [74]. However, the authors used SVM, K-Nearest Neighbor, and Naive Bayes with 40 owned images with an accuracy of 92.4% for breast cancer detection. In contrast, 90.48%, Sn value 87.6%, and Sp value 89.73% were the results of segmentation and detection using DWT for 306 images by [75].
Additionally, Ref. [76] developed a metaclassifier (ANN, DT, Bayesian, ELM, MLP) to classify 1052 thermogram images, and the accuracy of the findings was 76.01%.
Another method is the proposed ROI, ANN, for segmentation on the Mastology Research dataset with an accuracy of 90.17% [77].
Overall, 1056 thermography images were used by researchers in [78] to apply the CNN model for extracting breast features based on biodata, image processing, and image statistics. The accuracy was 98.95%.
Authors have suggested that Convolutional Neural Networks (CNN) [47] create a system that automatically takes thermographic images of the breast and categorizes them as normal or abnormal. It was used for 63 images (35 normal and 28 aberrant). These data were utilized for training and how to assess how well CNN performed in comparison to three classification methods: Bayes Network (BN), Tree Random Forest (TRF), and Multilayer Perception (MLP). The results were produced using a CNN classifier (100% TPR, SPC, and ACC). On the other hand, the authors used transfer learning with various deep learning pre-trained architectures, including AlexNet, Google Net, ResNet-50, ResNet-101, Inception V3, VGG-16, and VGG-19, to classify breast cancer using a fresh dataset. The dataset was randomly divided into 30% for validation and 70% for training, with images cropped to a fixed size of (224_224) or (227_227) pixels. With a sensitivity of 100%, specificity of 82.35%, and balanced accuracy of 91.18%, their findings showed that the VGG-16 Convolutional Neural Network surpassed the competition.
To diagnose breast cancer, Ref. [79] applied multiple models (CNN, SVM, and Random Forest) on more than 1000 images available on Kaggle. Additionally, to obtain these results, the researcher compared the accuracy of the three models (CNN model accuracy was 99.67%, SVM accuracy was 89.84, and RF accuracy was 90.55%).
Researchers [80] used the CNN (U-Net) model using the DMR-IR dataset to automatically separate and extract the breast area from other thermal image regions. They achieved an accuracy of 99.33%, a sensitivity of 100%, and a specificity of 98.67%.
On the other hand, Ref. [81] used various CNN architectures to detect breast cancer, including Resnet18, Resnet34, Resnet50, Resnet152, Vgg16, and Vgg19. On 5604 thermography images, they applied the proposed models, and 2411 obtained healthy images. They came to these conclusions because Resnet18, Resnet34, and Resnet50, the most stable design among these three, exhibited sound results in cancer classification using thermography.
Table 4. This table summarizes the research articles for using machine learning algorithms in thermography images.
Table 4. This table summarizes the research articles for using machine learning algorithms in thermography images.
ReferencesDatasetML MethodPerformance Evaluation
[82]50 breast imagesSVMAccuracy is 88.1%
Sn value is 85.7%
Sp value is 90.5%
[83]40 imagesSVM
Naive Bayes
K-Nearest
Neighbor
Accuracy is 92.5%
[75]306 imagesDWTAccuracy is 90.5%
Sn value is 87.6%
Sp value is 89.7%
[84]63 imagesFuzzy c-means
ROI
SVM
Accuracy is 100%
[85]63 thermography
Images
Bio-inspired Swarm
Techniques
Accuracy: 85.71%, 84.12%, 85.71%, and 96.83% for each swarm
[77]Mastology Research
-Dataset
ROI
ANN
Accuracy is 90.2%
Sn value is 89.34%
Sp value is 91%
[78]Mastology Research
Dataset
CNNAccuracy is 98.95%
[86]Mastology Research
Dataset
DWANSn value: 0.95
[87]63 thermographic (35 normal and 28 abnormal)CNN
TRF
MLP
BN
CNN presents better results than TRF, MLP, and BN
and the accuracy between (80–100% for CNN)
[88]DMR-IRANN
SVM
SVM sensitivity of 76% and specificity of 84%
ANN sensitivity of 92% and specificity of 88%
[79]Images of approximately 150 patients, either with or without breast cancer, totaling over 1000 (Kaggle available)CNN
SVM
Random forest
The accuracy that CNN acquired was 99.67%
SVM was 89.84%
The accuracy that RF obtained was 90.55%
[80]DMR_IRCNN
U_NET
Accuracy = 99.33%
Sensitivity = 100%
Specificity = 98.67%

5. Discussion

This section highlights the difficulties in combining ML with breast cancer and identifies many future research goals.

5.1. Datasets

Most researchers (48%) employed the mammography picture dataset, despite literature assessments of past studies in the classification and diagnosis of breast cancer. On various datasets, they applied a range of machine learning approaches. DDSM accomplished 32% of the work and 29% by MIAS. Following that, some used histological images (33%), several studies used open-access datasets (BreakHis), 11% of the researchers used ultrasound, and only 4% used thermography images (Figure 8 and Table 5).

5.2. Results

It is clear from earlier studies that researchers put a lot of effort into applying various machine learning models, including Support Vector Machine Learning (SVM), Probability Neural Networks (PNN), and K-Nearest Neighbors (KNN). They ran the models on various medical images then compared the outcomes.
Researchers have proposed convolutional neural networks (CNNs) for early breast cancer detection. Various CNN architectures, including -Resnet18, Resnet34, -Resnet50, -Resnet152, -vgg16, and vgg19, have been used, along with each architecture’s median and interquartile range. The best outcomes were from the resnet34 and resnet50 convolutional neural network designs, with 100% predicted accuracy in blind validation.
ML and DL methods for breast cancer still have significant limits and challenges that need to be addressed despite the positive findings of the reviewed literature.
These approaches offer outstanding outcomes in early breast cancer diagnosis and categorization. And as a result of the review, several important issues were found.

5.3. Challenges

The following discussion includes these issues, as well as internal aspects, potential future research areas, and issues:
DL needs considerable training data because the data set’s size and quality significantly impact the classifier’s effectiveness. But a lack of data is one of the biggest obstacles to using DL in medical imaging. Generating significant amounts of medical imaging data is challenging because eliminating human error takes a great deal of work from experts and one person. Large medical imaging data sets are difficult to construct because annotating the data takes a great deal of time and effort from a single expert and many experts to eliminate human error. The absence of substantial training datasets has made it challenging to construct deep-learning models for medical imaging, which was the first problem we saw in our studies. Most reviewed studies evaluated and assessed these using various datasets that cancer research organizations or clinics privately collected. The main issue with this method is that it is impossible to compare how well such models function across several investigations.
The absence of benchmarks provided a hurdle and highlighted a lack of flexibility.
Another issue with specific papers is using data expansion techniques rather than transferring learning to minimize overfitting.
Techniques for breast cancer categorization using unsupervised grouping: The supervised learning method was used to classify breast cancer in most of the selected primary papers. These strategies have provided superior results when labeled images are used throughout the training. However, finding breast cancer images with precise, medically labeled criteria might be difficult. There are frequently many unidentified medical images available. Despite being useful knowledge sources, many blank labels cannot be used for supervised learning. Therefore, there is a pressing need for a breast cancer categorization model that may be created using several grouping techniques without supervision.
Methodology of reinforcement learning for breast cancer classification: The fundamental issue is a lack of sufficient breast cancer image examples to depict all types of breast cancer. Creating a machine learning model that simultaneously learns from its surroundings can be difficult. Therefore, systems for identifying breast cancer from medical photos can perform and be more effective when employing a learning-based reinforcement model.
Reliability of data collection techniques: The robustness issue of various clinical and technical circumstances must be addressed to integrate new datasets gradually. Different image acquisition scanners, lighting configurations, sizes, and views across many picture modalities, and varying presentation aspects of the coloring and enlargement factors, are a few examples of these modifications.
Despite its significance in medical picture segmentation, the segmentation’s influence still falls short of what is required for practical use.

5.4. Future

In addition to what we have already discussed, there is another crucial point for the future. Thermal images or computed tomography (CT) images of breast cancer may be used in the future in place of the traditional image modalities (mammograms, ultrasound, MRI, and histological) to improve the accuracy of classification models for breast cancer. The identical patient must receive MRI or CT scans. Additionally, pictures of all different breast cancer cases will be collected. Boundary pictures should be labeled to classify multi-class breast cancer because they enable researchers to evaluate the efficacy of the recently created multi-class breast cancer classification model. Furthermore, emerging technologies and methods mentioned in this comprehensive review may be used for future medical idea cases to predict other cancer types such as cervical cancer or lung cancer, which also use any type of medical images.

6. Conclusions

This paper covers the most recent research on machine learning and deep learning methods for identifying and categorizing breast cancer. This study aimed to thoroughly understand contemporary breast cancer detection and diagnosis and identify and categorize breast cancer from various medical image types. This review concentrated on the most widely used machine learning approaches, including SVM, DT, Nearest Neighbor, Naive Bayesian Networks, ANN, and Convolutional Neural Networks. According to the papers analyzed, one interesting direction is the development of sophisticated systems that use artificial intelligence that can foretell serious medical issues and assist doctors and patients in preventing breast cancer.

Author Contributions

Conceptualization, R.J. and R.A.; methodology, R.J., H.K.C. and R.A.; formal analysis, R.J., H.K.C. and R.A.; investigation, R.A. and H.K.C.; resources, R.J. and R.A.; data curation, R.J. and R.A.; writing—original draft preparation, R.J., R.A. and H.K.C.; writing—review and editing, R.J. and R.A.; visualization, R.J. and R.A.; supervision, H.K.C. and R.A.; project administration, H.K.C.; funding acquisition, R.A. All authors have read and agreed to the published version of the manuscript.

Funding

Publication fund for this is supported by Research Center Borstel.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Arevalo, J.; Gonzalez, F.A.; Ramos-Pollan, R.; Oliveira, J.L.; Guevara Lopez, M.A. Representation learning for mammography mass lesion classification with convolutional neural networks. Comput. Methods Programs Biomed. 2016, 127, 248–257. [Google Scholar] [CrossRef] [PubMed]
  2. Roslidar, R.; Rahman, A.; Muharar, R.; Syahputra, M.R.; Arnia, F.; Syukri, M.; Pradhan, B.; Munadi, K. A Review on Recent Progress in Thermal Imaging and Deep Learning Approaches for Breast Cancer Detection. IEEE Access 2020, 8, 116176–116194. [Google Scholar] [CrossRef]
  3. Lancet, T. Breastfeeding: Achieving the new normal. Lancet 2016, 387, 404. [Google Scholar] [CrossRef]
  4. Razzak, M.I.; Naz, S.; Zaib, A. Deep learning for medical image processing: Overview, challenges and the future. In Classification in BioApps; Springer: Cham, Switzerland, 2018; pp. 323–350. [Google Scholar]
  5. Erickson, B.J.; Korfiatis, P.; Akkus, Z.; Kline, T.L. Machine learning for medical imaging. Radiographics 2017, 37, 505–515. [Google Scholar] [CrossRef] [Green Version]
  6. Witten, H.; Frank, E.; Kaufmann, M.; Geller, J. Data Mining: Practical Machine Learning Tools and Techniques with Java Implementations. ACM Sigmod Rec. 2000, 31, 76–77. [Google Scholar] [CrossRef]
  7. Amin, R.K.; Sibaroni, Y. Implementation of Decision Tree Using C4.5 Algorithm in Decision Making of Loan Application by Debtor (Case Study:Bank Pasar of Yogyakarta Special Region). In Proceedings of the 2015 3rd International Conference on Information and Communication Technology (ICoICT), Bali, Indonesia, 27–29 May 2015. [Google Scholar]
  8. Lim, T.-S.; Loh, W.-Y.; Shih, Y.-S. A Comparison of Prediction Accuracy, Complexity, and Training Time of Thirty-Three Old and New Classification Algorithms. Mach. Learn. 2000, 40, 203–228. [Google Scholar] [CrossRef]
  9. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  10. Vatsa, M.; Singh, R.; Noore, A. Improving biometric recognition accuracy and robustness using DWT and SVM watermarking. IEICE Electron. Express 2005, 2, 362–367. [Google Scholar] [CrossRef] [Green Version]
  11. Tharwat, A.; Hassanien, A.E.; Elnaghi, B.E. A BA-based algorithm for parameter optimization of Support Vector Machine. Pattern Recognit. Lett. 2017, 93, 13–22. [Google Scholar] [CrossRef]
  12. Deepak, P.A.; Chouhan, S. ICATE 2013 Paper Identification Number-102 SVM Kernel Functions for Classification. 2013. Available online: https://zero.sci-hub.se/2209/fd9e0a644e382bebb08d8cfa46399f2c/patle2013.pdf (accessed on 1 June 2023).
  13. Boser, B.; Denker, J.S.; Henderson, D.; Howard, R.E.; Hubbard, W.; Jackel, L.D.; Laboratories, H.Y.L.B.; Zhu, Z.; Cheng, J.; Zhao, Y.; et al. Backpropagation Applied to Handwritten Zip Code Recognition. Neural Comput. 1989, 1, 541–551. [Google Scholar]
  14. Shen, C. A Transdisciplinary Review of Deep Learning Research and Its Relevance for Water Resources Scientists. Water Resour. Res. 2018, 54, 8558–8593. [Google Scholar] [CrossRef]
  15. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
  16. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. 2015, 12, 211–252. [Google Scholar] [CrossRef] [Green Version]
  17. Guo, Y.; Liu, Y.; Oerlemans, A.; Lao, S.; Wu, S.; Lew, M.S. Deep learning for visual understanding: A review. Neurocomputing 2016, 187, 27–48. [Google Scholar] [CrossRef]
  18. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  19. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A.; Liu, W.; et al. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  20. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  21. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  22. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  23. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning. arXiv 2017, arXiv:160207261v2. [Google Scholar] [CrossRef]
  24. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5987–5995. [Google Scholar]
  25. Wang, F.; Tax, D.M.J. Survey on the attention based RNN model and its applications in computer vision. arXiv 2016, arXiv:1601.06823. [Google Scholar]
  26. Mechria, H.; Gouider, M.; Hassine, S.K. Breast cancer detection using deep convolutional neural network. In Proceedings of the ICAART 2019—Proceedings of the 11th International Conference on Agents and Artificial Intelligence, Prague, Czech Republic, 19–21 February 2019; Volume 2, pp. 655–660. [Google Scholar]
  27. Shen, L.; Margolies, L.R.; Rothstein, J.H.; Fluder, E.; McBride, R.; Sieh, W. Deep Learning to Improve Breast Cancer Detection on Screening Mammography. Sci. Rep. 2019, 9, 12495. [Google Scholar] [CrossRef] [Green Version]
  28. Hossam, A.; Harb, H.M.; Abd, H.M.; Kader, E. Automatic Image Segmentation Method for Breast Cancer Analysis Using Thermography. JES J. Eng. Sci. 2018, 46, 12–32. [Google Scholar] [CrossRef]
  29. Robert, M.; Miller, J. About Cancer. Available online: http://www.aboutcancer.com/breast_pet_scans.htm (accessed on 17 July 2023).
  30. Vijayarajeswari, R.; Parthasarathy, P.; Vivekanandan, S.; Basha, A.A. Classification of mammogram for early detection of breast cancer using SVM classifier and Hough transform. Measurement 2019, 146, 800–805. [Google Scholar] [CrossRef]
  31. Zemmal, N.; Azizi, N.; Sellami, M. CAD system for classification of mammographic abnormalities using transductive semi supervised learning algorithm and heterogeneous features. In Proceedings of the 2015 12th International Symposium on Programming and Systems (ISPS), Algiers, Algeria, 28–30 April 2015; pp. 1–9. [Google Scholar]
  32. Anitha, J.; Dinesh, P.J. A wavelet based morphological mass detection and classification in mammograms. In Proceedings of the 2012 International Conference on Machine Vision and Image Processing (MVIP), Tamil Nadu, India, 14–15 December 2012; pp. 25–28. [Google Scholar]
  33. Zheng, B.; Yoon, S.W.; Lam, S.S. Breast cancer diagnosis based on feature extraction using a hybrid of K-means and support vector machine algorithms. Expert Syst. Appl. 2014, 41, 1476–1482. [Google Scholar] [CrossRef]
  34. Wichakan, I.; Vateekul, P. Combining Deep Convolutional Networks SVMs for Mass Detection on Digital Mammograms. In Proceedings of the 2016 8th International Conference on Knowledge and Smart Technology (KST), Chiang Mai, Thailand, 3–6 February 2016; pp. 239–244. [Google Scholar]
  35. Li, H.; Zhuang, S.; Li, D.-A.; Zhao, J.; Ma, Y. Benign and malignant classification of mammogram images based on deep learning. Biomed. Signal Process. Control. 2019, 51, 347–354. [Google Scholar] [CrossRef]
  36. Abdel-Nasser, M.; Rashwan, H.A.; Puig, D.; Moreno, A. Analysis of Tissue Abnormality and Breast Density in Mammographic Images Using a Uniform Local Directional Pattern; Elsevier: Amsterdam, The Netherlands, 2015. [Google Scholar]
  37. Spanhol, F.A.; Oliveira, L.S.; Petitjean, C.; Heutte, L. A dataset for breast cancer histopathological image classification. IEEE Trans. Biomed. Eng. 2016, 63, 1455–1462. [Google Scholar] [CrossRef] [PubMed]
  38. Wajid, S.K.; Hussain, A. Local energy-based shape histogram feature extraction technique for breast cancer diagnosis. Expert Syst. Appl. 2015, 42, 6990–6999. [Google Scholar] [CrossRef] [Green Version]
  39. Phadke, A.C.; Rege, P.P. Fusion of local and global features for classification of abnormality in mammograms. Sādhanā 2016, 41, 385–395. [Google Scholar] [CrossRef] [Green Version]
  40. Khan, S.; Hussain, M.; Aboalsamh, H.; Bebis, G. A comparison of different Gabor feature extraction approaches for mass classification in mammography. Multimed. Tools Appl. 2017, 76, 33–57. [Google Scholar] [CrossRef] [Green Version]
  41. Pratiwi, M.; Alexander; Harefa, J.; Nanda, S. Mammograms Classification Using Gray-level Co-occurrence Matrix and Radial Basis Function Neural Network. Proc. Comput. Sci. 2015, 59, 83–91. [Google Scholar] [CrossRef] [Green Version]
  42. Hamoud, M.; Merouani, H.F.; Laimeche, L. The power laws: Zipf and inverse Zipf for automated segmentation and classification of masses within mammograms. Evol. Syst. 2015, 6, 209–227. [Google Scholar] [CrossRef]
  43. Abubacker, N.F.; Azman, A.; Doraisamy, S.; Murad, M.A.A. An integrated method of associative classification and neuro-fuzzy approach for effective mammographic classification. Neural Comput. Appl. 2017, 28, 3967–3980. [Google Scholar] [CrossRef]
  44. Gardezi, S.J.S.; Faye, I.; Bornot, J.M.S.; Kamel, N.; Hussain, M. Mammogram classification using dynamic time warping. Multimed. Tools Appl. 2018, 77, 3941–3962. [Google Scholar] [CrossRef]
  45. Rouhi, R.; Jafari, M. Classification of benign and malignant breast tumors based on hybrid level set segmentation. Expert Syst. Appl. 2016, 46, 45–59. [Google Scholar] [CrossRef]
  46. Peng, W.; Mayorga, R.; Hussein, E. An automated confirmatory system for analysis of mammograms. Comput. Methods Programs Biomed. 2016, 125, 134–144. [Google Scholar] [CrossRef] [PubMed]
  47. Hai, J.; Tan, H.; Chen, J.; Wu, M.; Qiao, K.; Xu, J.; Zeng, L.; Gao, F.; Shi, D.; Yan, B. Multi-level features combined end-to-end learning for automated pathological grading of breast cancer on digital mammograms. Comput. Med. Imaging Graph. 2019, 71, 58–66. [Google Scholar] [CrossRef] [PubMed]
  48. Saber, A.; Sakr, M.; Abo-Seida, O.M.; Keshk, A.; Chen, H. A Novel Deep-Learning Model for Automatic Detection and Classification of Breast Cancer Using the Transfer-Learning Technique. IEEE Access 2021, 9, 71194–71209. [Google Scholar] [CrossRef]
  49. Dhahbi, S.; Barhoumi, W.; Zagrouba, E. Breast cancer diagnosis in digitized mammograms using curvelet moments. Comput. Biol. Med. 2015, 64, 79–90. [Google Scholar] [CrossRef]
  50. Aminikhanghahi, S.; Shin, S.; Wang, W.; Jeon, S.I.; Son, S.H. A new fuzzy Gaussian mixture model (FGMM) based algorithm for mammography tumor image classification. Multimed. Tools Appl. 2017, 76, 10191–10205. [Google Scholar] [CrossRef]
  51. Khan, S.U.; Islam, N.; Jan, Z.; Ud Din, I.; Rodrigues, J.J. A novel deep learning based framework for the detection and classification of breast cancer using transfer learning. Pattern Recognit. Lett. 2019, 125, 1–6. [Google Scholar] [CrossRef]
  52. Liu, X.; Mei, M.; Liu, J.; Hu, W. Microcalcification detection in full-field digital mammograms with PFCM clustering and weighted SVM-based method. EURASIP J. Adv. Signal Process. 2015, 2015, 73. [Google Scholar] [CrossRef] [Green Version]
  53. Tan, M.; Qian, W.; Pu, J.; Liu, H.; Zheng, B. A new approach to develop computer-aided detection schemes of digital mammograms. Phys. Med. Biol. 2015, 60, 4413–4427. [Google Scholar] [CrossRef] [Green Version]
  54. Suzuki, S.; Zhang, X.; Homma, N.; Ichiji, K.; Sugita, N.; Kawasumi, Y.; Ishibashi, T.; Yoshizawa, M. Mass detection using deep convolutional neural networks for mammoghraphic computer-aided diagnosis. In Proceedings of the 55th Annual Conference of the Society of Intruments and Control Engineers of Japan (SICE), Tsukuba, Japan, 20–23 September 2016; pp. 1382–1386. [Google Scholar]
  55. Jasti, V.D.P.; Zamani, A.S.; Arumugam, K.; Naved, M.; Pallathadka, H.; Sammy, F.; Raghuvanshi, A.; Kaliyaperumal, K. Computational Technique Based on Machine Learning and Image Processing for Medical Image Analysis of Breast Cancer Diagnosis. Secur. Commun. Netw. 2022, 2022, 1918379. [Google Scholar] [CrossRef]
  56. Ranjbarzadeh, R.; Tataei Sarshar, N.; Jafarzadeh Ghoushchi, S.; Saleh Esfahani, M.; Parhizkar, M.; Pourasad, Y.; Anari, S.; Bendechache, M. MRFE-CNN: Multi-route feature extraction model for breast tumor segmentation in Mammograms using a convolutional neural network. Ann. Oper. Res. 2022, 11, 1–22. [Google Scholar] [CrossRef]
  57. Qiu, Y.; Wang, Y.; Yan, S.; Tan, M.; Cheng, S.; Liu, H.; Zheng, B. An initial investigation on developing a new method to predict short-term breast cancer risk based on deep learning technology. In Medical Imaging 2016: Computer-Aided Diagnosis; SPIE: Cergy, France, 2016; Volume 9785, p. 978521. [Google Scholar]
  58. Cai, L.; Wang, X.; Wang, Y.; Guo, Y.; Yu, J.; Wang, Y. Robust phase-based texture descriptor for classification of breast ultrasound images. Biomed. Eng. Online 2015, 14, 26. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  59. Shan, J.; Alam, S.K.; Garra, B.; Zhang, Y.; Ahmed, T. Computer-Aided Diagnosis for Breast Ultrasound Using Computerized BI-RADS Features and Machine Learning Methods. Ultrasound Med. Biol. 2016, 42, 980–988. [Google Scholar] [CrossRef] [PubMed]
  60. Qi, X.; Zhang, L.; Chen, Y.; Pi, Y.; Chen, Y.; Lv, Q.; Yi, Z. Automated diagnosis of breast ultrasonography images using deep neural networks. Med. Image Anal. 2019, 52, 185–198. [Google Scholar] [CrossRef]
  61. Byra, M.; Galperin, M.; Ojeda-Fournier, H.; Olson, L.; O’Boyle, M.; Comstock, C.; Andre, M. Breast mass classification in sonography with transfer learning using a deep convolutional neural network and color conversion. Med. Phys. 2019, 46, 746–755. [Google Scholar] [CrossRef]
  62. Xu, Y.; Wang, Y.; Yuan, J.; Cheng, Q.; Wang, X.; Carson, P.L. Medical breast ultrasound image segmentation by machine learning. Ultrasonics 2019, 91, 1–9. [Google Scholar] [CrossRef] [PubMed]
  63. Yap, M.H.; Pons, G.; Marti, J.; Ganau, S.; Sentis, M.; Zwiggelaar, R.; Davison, A.K.; Marti, R. Automated Breast Ultrasound Lesions Detection Using Convolutional Neural Networks. IEEE J. Biomed. Health Inform. 2018, 22, 1218–1226. [Google Scholar] [CrossRef] [Green Version]
  64. Huang, Q.; Yang, F.; Liu, L.; Li, X. Automatic segmentation of breast lesions for interaction in ultrasonic computer-aided diagnosis. Inf. Sci. 2015, 314, 293–310. [Google Scholar] [CrossRef]
  65. Moon, W.K.; Huang, Y.-S.; Lo, C.-M.; Huang, C.-S.; Bae, M.S.; Kim, W.H.; Chen, J.-H.; Chang, R.-F. Computer-aided diagnosis for distinguishing between triple-negative breast cancer and fibroadenomas based on ultrasound texture features. Med. Phys. 2015, 42, 3024–3035. [Google Scholar] [CrossRef]
  66. Prabusankarlal, K.M.; Thirumoorthy, P.; Manavalan, R. Assessment of combined textural and morphological features for diagnosis of breast masses in ultrasound. Hum.-Centric Comput. Inf. Sci. 2015, 5, 12. [Google Scholar] [CrossRef] [Green Version]
  67. Han, S.; Kang, H.-K.; Jeong, J.-Y.; Park, M.-H.; Kim, W.; Bang, W.-C.; Seong, Y.-K. A deep learning framework for supporting the classification of breast lesions in ultrasound images. Phys. Med. Biol. 2017, 62, 7714–7728. [Google Scholar] [CrossRef]
  68. Kumar, A.; Singh, S.K.; Saxena, S.; Lakshmanan, K.; Sangaiah, A.K.; Chauhan, H.; Shrivastava, S.; Singh, R.K. Deep feature learning for histopathological image classification of canine mammary tumors and human breast cancer. Inf. Sci. 2020, 508, 405–421. [Google Scholar] [CrossRef]
  69. Chmielewski, A.; Dufort, P.; Scaranelo, A.M. A Computerized System to Assess Axillary Lymph Node Malignancy from Sonographic Images. Ultrasound Med. Biol. 2015, 41, 2690–2699. [Google Scholar] [CrossRef]
  70. Shibusawa, M.; Nakayama, R.; Okanami, Y.; Kashikura, Y.; Imai, N.; Nakamura, T.; Kimura, H.; Yamashita, M.; Hanamura, N.; Ogawa, T. The usefulness of a computer-aided diagnosis scheme for improving the performance of clinicians to diagnose non-mass lesions on breast ultrasonographic images. J. Med. Ultrason. 2016, 43, 387–394. [Google Scholar] [CrossRef]
  71. Lo, C.-M.; Chan, S.-W.; Yang, Y.-W.; Chang, Y.-C.; Huang, C.-S.; Jou, Y.-S.; Chang, R.-F. Feasibility Testing: Three-dimensional Tumor Mapping in Different Orientations of Automated Breast Ultrasound. Ultrasound Med. Biol. 2016, 42, 1201–1210. [Google Scholar] [CrossRef] [PubMed]
  72. Abdel-Nasser, M.; Melendez, J.; Moreno, A.; Omer, O.A.; Puig, D. Breast tumor classification in ultrasound images using texture analysis and super-resolution methods. Eng. Appl. Artif. Intell. 2017, 59, 84–92. [Google Scholar] [CrossRef]
  73. Moon, W.K.; Chen, I.-L.; Chang, J.M.; Shin, S.U.; Lo, C.-M.; Chang, R.-F. The adaptive computer-aided diagnosis system based on tumor sizes for the classification of breast tumors detected at screening ultrasound. Ultrasonics 2017, 76, 70–77. [Google Scholar] [CrossRef]
  74. Ali, M.A.; Sayed, G.I.; Gaber, T.; Hassanien, A.E.; Snasel, V.; Silva, L.F. Detection of breast abnormalities of thermograms based on a new segmentation method. In Proceedings of the Federated Conference on Computer Science and Information Systems (FedCSIS), Lodz, Poland, 13–16 September 2015; pp. 255–261. [Google Scholar]
  75. Pramanik, S.; Bhattacharjee, D.; Nasipuri, M. Wavelet based thermogram analysis for breast cancer detection. In Proceedings of the 2015 International Symposium on Advanced Computing and Communication (ISACC), Silchar, India, 14–15 September 2015; pp. 205–212. [Google Scholar]
  76. Santana, M.; Pereira, J.; Silva, F.; Lima, N.; Sousa, F.; Arruda, G.; Lima, R.; Silva, W.; Santos, W. Breast cancer diagnosis based on mammary thermography and extreme learning machines. Int. J. Artif. Intell. Mach. Learn. 2018, 34, 45–53. [Google Scholar] [CrossRef] [Green Version]
  77. Sánchez-Ruiz, D.; Olmos-Pineda, I.; Olvera-López, J.A. Automatic region of interest segmentation for breast thermogram image classification. Pattern Recognit. Lett. 2020, 135, 72–81. [Google Scholar] [CrossRef]
  78. Ekici, S.; Jawzal, H. Breast cancer diagnosis using thermography and convolutional neural networks. Med. Hypotheses 2020, 137, 109542. [Google Scholar] [CrossRef]
  79. Reddy, A.V. Breast cancer detection based on thermographic images using machine learning and deep learning algorithms Healthcare View project Breast cancer detection based on thermographic images using machine learning and deep learning algorithms. Comput. Sci. 2022, 4, 49–56. [Google Scholar]
  80. Mohamed, E.A.; Rashed, E.A.; Gaber, T.; Karam, O. Deep learning model for fully automated breast cancer detection system from thermograms. PLoS ONE 2022, 17, e0262349. [Google Scholar] [CrossRef] [PubMed]
  81. Fernández-Ovies, F.; Alférez, S.; DeAndrés-Galiana, E.J.; Cernea, A.; Fernández-Muñiz, Z.; Fernández-Martínez, J.L. Detection of Breast Cancer Using Infrared Thermography and Deep Neural Networks. In Bioinformatics and Biomedical Engineering; Rojas, I., Valenzuela, O., Rojas, F., Ortuño, F., Eds.; Springer International Publishing: Cham, Switzerland, 2019; pp. 514–523. [Google Scholar]
  82. Mookiaha, M.R.K.; Acharyaa, U.R.; Ng, E.Y. Data mining technique for breast cancer detection in thermograms using hybrid feature extraction strategy. Quant. InfraRed Thermogr. J. 2012, 9, 151–165. [Google Scholar] [CrossRef]
  83. Milosevic, M.; Jankovic, D.; Peulic, A. Thermography based breast cancer detection using texture features and minimum variance quantization. EXCLI J. 2014, 13, 1204. [Google Scholar] [PubMed]
  84. Gaber, T.; Ismail, G.; Anter, A.; Soliman, M.; Ali, M.; Semary, N.; Hassanien, A.E.; Snasel, V. Thermogram breast cancer prediction approach based on Neutrosophic sets and fuzzy c-means algorithm. In Proceedings of the 2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Milan, Italy, 25–29 August 2015; pp. 4254–4257. [Google Scholar]
  85. Sayed, G.I.; Soliman, M.; Hassanien, A.E. Bio-Inspired Swarm Techniques for Thermogram Breast Cancer Detection; Springer: Berlin/Heidelberg, Germany, 2016. [Google Scholar]
  86. De Freitas Barbosa, V.A.; de Santana, M.A.; Andrade, M.K.S.; de Lima, R.d.C.F.; dos Santos, W.P. Deep-wavelet neural networks for breast cancer early diagnosis using mammary termographies. In Deep Learning for Data Analytics; Elsevier: Hoboken, NJ, USA, 2020; pp. 99–124. [Google Scholar]
  87. Tello-Mijares, S.; Woo, F.; Flores, F. Breast Cancer Identification via Thermography Image Segmentation with a Gradient Vector Flow and a Convolutional Neural Network. J. Health Eng. 2019, 2019, 9807619. [Google Scholar] [CrossRef] [Green Version]
  88. Al Husaini, M.A.S.; Habaebi, M.H.; Gunawan, T.S.; Islam, M.R.; Elsheikh, E.A.A.; Suliman, F.M. Thermal-based early breast cancer detection using inception V3, inception V4 and modified inception MV4. Neural Comput. Appl. 2022, 34, 333–348. [Google Scholar] [CrossRef]
Figure 1. Relationship between artificial intelligence, machine learning, and deep learning.
Figure 1. Relationship between artificial intelligence, machine learning, and deep learning.
Diagnostics 13 02460 g001
Figure 2. The input and output architecture of RNNs.
Figure 2. The input and output architecture of RNNs.
Diagnostics 13 02460 g002
Figure 3. Mammography images from the DDSM dataset from Kaggle.
Figure 3. Mammography images from the DDSM dataset from Kaggle.
Diagnostics 13 02460 g003
Figure 4. Different types of ultrasound images from Kaggle.
Figure 4. Different types of ultrasound images from Kaggle.
Diagnostics 13 02460 g004
Figure 5. Different types of Breast histopathology images.
Figure 5. Different types of Breast histopathology images.
Diagnostics 13 02460 g005
Figure 6. Thermography Images from Irthermo database.
Figure 6. Thermography Images from Irthermo database.
Diagnostics 13 02460 g006
Figure 7. Positron emission tomography scan images.
Figure 7. Positron emission tomography scan images.
Diagnostics 13 02460 g007
Figure 8. Different types of medical images used.
Figure 8. Different types of medical images used.
Diagnostics 13 02460 g008
Table 1. Table of three classes of ultrasound breast images.
Table 1. Table of three classes of ultrasound breast images.
CaseNumber of Images
Benign487
Malignant210
Normal133
Total780
Table 3. This table summarizes ultrasound papers using different (ML and DL) techniques.
Table 3. This table summarizes ultrasound papers using different (ML and DL) techniques.
ReferencesDatasetML MethodPerformance Evaluation
[58]138 private
cases
SVM classifierAccuracy, Sn value, and Sp value all 86.9%
AUC value is 0.89
[66]120 private
images
-(benign 70)
-(malignant 50)
SVM classifierAccuracy is 95.85%
Sn value is 96.0%
Sp value is 91.5%
The AUC value is 0.94
[69]105 private
images
SVM classifierSn value is 95%
Sp value is 90%
AUC value is 95%
[65]169 private casesSVM classifierAccuracy is 94.8%
Sn value is 94.1%
Sp value is 96.7%
[64]46 private
Images
SVM classifierAccuracy is 0.98 ± 0.013
Sn value is 0.97 ± 0.035
Sp value is 0.98 ± 0.019
AUC value is 0.997 ± 0.003
[70]97 private
images
K-NNSn value is 87.8%
Sp value is 89.5%
AUC value is 0.93
[71]18 private
cases
Binary-LRAccuracy is 80.4%
[72]59 private
images
RFAUC value is 99%
[73]156 owned casesLR
ANN
Accuracy is 81.8%
Sn Value is 85.4%
Sp Value is 77.8%
AUC value is 0.855
[59]283 owned casesDT
KNN
RF
SVM
SVM accuracy is 77.7%
AUC Value is 0.84
RF accuracy is 78.5%
AUC value is 0.83
[67]7408CNN based on
VGG19
Accuracy value is 91.2%
TP value is 84.3%
TN value is 96.1%
AUC value is 96.0%
[61]882CNN based on
VGG19
Acc value is 88.7%
TP value is 84.8%
TN value is 89.7%
AUC value is 93.6%
[63]306FCN–AlexNetTP value is 98%
[68]433U-NetTP value is 84%
Table 5. Different types of available public datasets.
Table 5. Different types of available public datasets.
DatasetImage TypeURL
MIASmammogramhttps://www.repository.cam.ac.uk/handle/1810/250394 (accessed on 1 June 2023)
DDSMmammogramhttp://marathon.csee.usf.edu/Mammography/Database.html (accessed on 1 June 2023)
mini-MIASmammogramhttp://peipa.essex.ac.uk/info/mias.html (accessed on 1 June 2023)
Break-His-histologicalhttps://web.inf.ufpr.br/vri/databases/breastcancer-histopathological-databasebreakhis/ (accessed on 1 June 2023)
DMR-IRThermographyhttp://visual.ic.uff.br/dmi (accessed on 1 June 2023)
BI-RADSmammogramhttps://radiopaedia.org/articles/breast-imaging-reporting-and-data-system-bi-rads (accessed on 1 June 2023)
INbreastmammogramhttp://dx.doi.org/10.17632/x7bvzv6cvr.1 (accessed on 1 June 2023)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jalloul, R.; Chethan, H.K.; Alkhatib, R. A Review of Machine Learning Techniques for the Classification and Detection of Breast Cancer from Medical Images. Diagnostics 2023, 13, 2460. https://doi.org/10.3390/diagnostics13142460

AMA Style

Jalloul R, Chethan HK, Alkhatib R. A Review of Machine Learning Techniques for the Classification and Detection of Breast Cancer from Medical Images. Diagnostics. 2023; 13(14):2460. https://doi.org/10.3390/diagnostics13142460

Chicago/Turabian Style

Jalloul, Reem, H. K. Chethan, and Ramez Alkhatib. 2023. "A Review of Machine Learning Techniques for the Classification and Detection of Breast Cancer from Medical Images" Diagnostics 13, no. 14: 2460. https://doi.org/10.3390/diagnostics13142460

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop