You are currently viewing a new version of our website. To view the old version click .
Diagnostics
  • Article
  • Open Access

16 October 2022

Deep Learning Model for the Detection of Real Time Breast Cancer Images Using Improved Dilation-Based Method

,
,
,
and
1
Applied College in Abqaiq, King Faisal University, P.O. Box 400, Al-Ahsa 31982, Saudi Arabia
2
School of Computing Science & Engineering, VIT Bhopal University, Bhopal 466114, India
3
Computer Science Department, King Faisal University, P.O. Box 400, Al-Ahsa 31982, Saudi Arabia
4
School of Computer Science, University of Petroleum & Energy Studies, Dehradun 248007, India
This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics

Abstract

Breast cancer can develop when breast cells replicate abnormally. It is now a worldwide issue that concerns people’s safety all around the world. Every day, women die from breast cancer, which is especially common in the United States. Mammography, CT, MRI, ultrasound, and biopsies may all be used to detect breast cancer. Histopathology (biopsy) is often carried out to examine the image and discover breast cancer. Breast cancer detection at an early stage saves lives. Deep and machine learning models aid in the detection of breast cancer. The aim of the research work is to encourage medical research and the development of technology by employing deep learning models to recognize cancer cells that are small in size. For histological annotation and diagnosis, the proposed technique makes use of the BreCaHAD dataset. Color divergence is caused by differences in slide scanners, staining procedures, and biopsy materials. To avoid overfitting, we used data augmentation with 19 factors, such as scale, rotation, and gamma. The proposed hybrid dilation deep learning model is of two sorts. It illustrates edges, curves, and colors, and it improves the key traits. It utilizes dilation convolution and max pooling for multi-scale information. The proposed dilated unit processes the image and sends the processed features to the Alexnet, and it can recognize minute objects and thin borders by using the dilated residual expanding kernel model. An AUC of 96.15 shows that the new strategy is better than the old one.

1. Introduction

Cancer is now the main cause of mortality in almost every country. Cellulitis is a condition characterized by abnormal cell development. Tissues are made up of cells that collaborate with one another. Diseased cells are replaced by healthy cells. Tumors and cancers are uncontrollable cell growths that have shifted from their native environment. These are the most prevalent cancer kinds [1]. Pathologists employ histology to assess the progression of cancer in organs [2] By putting together a histological tissue slide for microscopy. Histopathological specimens include a variety of tissues that surround cells and structures. It takes time to physically understand historical images. Many years of expertise in computer-aided image analysis have the potential to increase histopathological pictures’ analytical and predictive capacities. A trustworthy second opinion enables histopathologists to do more.
This may aid in the diagnosis of the problem. In 2012, cancer claimed the lives of 8.2 million individuals, and 8.8 million in 2015. Breast cancer claimed the lives of 5.7 lakh individuals in 2015. Between 2005 and 2015, the global cancer incidence increased by 33% [3]. We were able to give each of the images a uniform appearance by utilizing a basic H&E color normalizing procedure that we obtained online. The photographs were then scaled up to four distinct sizes, each with a different aspect ratio of one, a half, a third, and a quarter. These enlarged images were separated into four identical, non-overlapping halves. Individual patches were made without any augmentation, then stitched together by increasing each patch to a total of 224 by 224 pixels. We were able to obtain exactly what we needed. Our team created an enhanced patch-wise dataset on top of a concatenated subset of four non-overlapping neighboring patches. Pathologists are trained to analyze histological images from several perspectives, since staining and recording can take place in a variety of conditions. To mimic the actions of a pathologist during an examination and represent real-world variation, we used a variety of data augmentation techniques, such as horizontal and vertical flipping, rotation, shifting (in both dimensions), brightness, magnification, and blurring. The idea was to create a scenario equivalent to that of a pathologist employing these techniques to carry out their work. The dataset’s integrity will not be jeopardized in any way, even if its size increases as more data are uploaded to it. In a published study, the use of data augmentation and patching methods for histological classification was shown to be theoretically feasible. This argument is reinforced by recently released study data. Once an agreement was achieved, each of the newly created patches was given a name that matched the category of the original picture that had been utilized in the creation process. After training using image-net weights, the model compared the values it had given to patches of varied sizes. This was completed after the model had been trained.
A patch-wise classifier evaluates multiple unique parts of the image at various magnification levels, before selecting how to categorize the full image. The results of this processing will be combined with the results of processing the whole picture patches in the following step to produce an image-wise classification. This classification will be determined by the total number of image patches under development. To accurately categorize histology photographs into one of the various required classifications, first, one must find the key components of the images and then study these features. Only then will it be possible to properly label the pictures. The data supplied describe the nucleus characteristics, texture, and overall tissue architecture. The shape, color, and structure of the nucleus can be used to identify and distinguish cancer cells from healthy ones. Normal cells can be identified by the absence of specific characteristics. This stage is critical for distinguishing between benign and malignant cells (e.g., density or variability). A fundamental understanding of tissue architecture is required to distinguish between in situ and metastatic carcinomas. The categorization method is used to generalize learned traits over a wide range of spatial sizes, from glands to cell nuclei.
Cancer is the leading cause of mortality among Indians. In 2016, only one new instance of breast cancer was discovered and identified in India. Since more people have been diagnosed with cancer in the last 10 years, experts expect that the overall number of cancer patients in India will rise by 2025. Breast cancer is the leading cause of mortality for women under the age of 50, and it ranks second on the list of the world’s deadliest diseases. Early detection has been shown to have a high association with public awareness [4], and both variables can dramatically reduce death rates. If a patient’s prognosis is positive and they obtain an accurate diagnosis, they have a far greater chance of fully recovering. It is, therefore, critical to develop effective technologies that can increase the rate of early diagnosis and reduce death rates. However, histological examinations continue to be the most used imaging tool for detecting breast cancer [5]. Pathology diagnosis in its current condition is highly interpretative. Even in ideal settings, this operation is time-consuming and demanding; one can imagine how much more difficult it is when the pathologist is weary or under pressure at work. Tissue samples obtained during biopsies or surgical procedures are evaluated to determine whether the patient has cancer. These samples are inspected under a microscope. The operation is carried out by pathologists, who are highly experienced medical professionals. Histology, on the other hand, is a method that needs to be performed quickly, even though it is a difficult process and takes a long time. Following treatment, the tissue is sliced into small sections that are thin enough to enable light to pass through them, allowing light microscopy to be performed on the sections after they have been cut. Although frozen section histology can be utilized for intraoperative assessment, it is a labor-intensive and time-consuming technique. The tissue can only be studied using exceedingly small samples. The use of computerized image inspection and machine learning algorithms has recently become practical. A decade ago, digital pathology used microscopes with cameras to digitally scan complete tissue samples. Recent advances in image analysis and processing power have assisted in the incorporation of computer-aided diagnostic (CAD) systems into pathology lab operations. These tools were made to help pathologists find, diagnose, and predict illnesses. In terms of diagnosis and prognosis, DL-based CAD systems have made major contributions to the growth of the medical business. Data from radiology, pathology, cardiology, pharmacology, cancer, and genomes are analyzed by these systems. In recent decades, more sophisticated systems based on machine learning principles have been used to diagnose cancer. Immunohistochemistry is one such approach. There has been a substantial increase in the number of times this strategy has been deployed. Since it is beneficial in cancer prediction and prognosis, DL is generally recognized as one of these approaches. The higher diagnostic accuracy of DL can aid in the improvement of mammograms, ultrasounds, and digital breast tomosynthesis (DBT). Because it provides a better degree of precision than other approaches, DL is used in the treatment of breast cancer (BrC). There has been a flood of scientific articles pertaining to BrC during the last several years. Deep learning is used in several contexts in all these studies. Deep learning algorithms are able to automate the BrC diagnosis procedure, which is riddled with flaws and errors. Despite the large number of BrC classification review studies that have been conducted thus far, only a small number of them can offer guidance to future researchers. Even though these articles give a rather complete literature overview on BrC, they may benefit from including a few additional parts on deep learning. An example sentence is as follows: the great bulk of BrC literature evaluations have focused on either classic ML techniques with broad strokes. Both approaches involve feature extraction at various stages of the diagnostic process, which is comparable. Deep-learning architectures, such as generative adversarial networks, extreme learning machines, and similar systems, were unable to generate a BrC diagnosis. Recent BrC research has led to the development of several imaging modalities. However, most prior review studies ignored emerging imaging modalities, such as infrared thermal imaging, digital breast tomosynthesis, and computed tomography. There are review articles on digital breast tomosynthesis [6], but they do not include all imaging modalities that can be used to classify BrC. This is due to the article’s structure, which limits what can and cannot be included. They also failed to provide a clear explanation of the advantages and disadvantages of previous research, which impeded an otherwise fascinating analysis of deep learning-based approaches. They also did not give a full explanation of the positive or negative aspects of the previous investigations.
Automatic biological data analysis may help pathologists make early or on-the-spot diagnoses. These strategies assist pathologists in reducing their workload by isolating and screening out non-cancerous regions. They also assist pathologists in detecting breast cancer early, which reduces fatalities. An assessment of the research paradigm, strategy, or framework is required. In a validated system, performance measurements for segmentation and classifiers may be employed. As a result, two datasets must be used for training and testing. The system must be evaluated on a separate dataset to prevent memorization.

3. Proposed Model

It gathers multiscale contextual data, while maintaining image quality and without increasing the number of model parameters. Spatial resolution is reduced when convolutions are dilated. Gridding of dilated convolutional networks has already begun. In this section, the preprocessing and the working principal of the proposed model are given.

3.1. Data Augmentation

The model’s initial goal is to deal with overfitting concerns caused by the restricted number of photos used in the training phase. To address this, we employ data augmentation as the first step, so that we may make use of rotation, scaling, shift, and gamma correction with various parameters. Breast Cancer Histopathological Image Classification (BreakHis) includes 2480 benign photos and 5429 malignant images, whereas Breast Cancer Histopathological Annotation and Diagnosis (BreCaHAD) includes 162 breast cancer images. Techniques such as flipping, rotating, shifting, scaling, and gamma are all valuable. The parameters tested for augmentation are shown in Table 1. In the top row of Figure 5, a histology picture from the dataset is shown, followed by a strain normalized image and a rotational image. The second row shows a flipped image with gamma values between 0.3 and 1.2.
Table 1. Parameters of Data Augmentation.
Figure 5. Data Augmentation Results.
Data augmentation increases the number of images in a dataset, which eliminates overfitting issues caused by dataset size restrictions [21,22,23]. Figure 5 depicts the picture enhancement outcome. The histology picture in the dataset is followed by a strain normalized and rotational image. The second row has an inverted picture with gamma values of 0.2, 0.5, 0.8, 1.0, and 1.2. Image augmentation is the process of rotating, scaling, and flipping images. The photos have been scaled by 0.4, 0.7, and 1.3. Using the reflections, one can rotate the picture.

3.2. Hybrid Dilation Deep Learning Model

Histology images were utilized to evaluate the correctness of the pre-models. Using this technology, cancer cells are captured and differentiated. The proposed model decreases network overfitting, while improving accuracy, as shown in Figure 6. The missing pixels are then dilated and the Hadamard product is used to connect the input result to the entirely connected layers. The technology improves the accuracy, while broadening the receptive field. To extract features, dilated convolution is employed. Dropout and overfitting are reduced when data are normalized in bulk. Excellent results in terms of localization and mitotic centroid measurement are achieved. All cancer cells need high-resolution images. The suggested architecture makes use of fewer than half of the network layers shown in the picture. Batch normalization of a simplified symmetric skip network is carried out. The proposed model has the highest AUC (96.15). In addition, 3 × 3 convolution aided in the identification of micro-cells.
Figure 6. The Proposed Hybrid Dilation Deep Learning Model.
In this section, we present some of the most pressing issues and solutions. In the absence of visuals, convolutional neural networks overfit, impairing detection. The model provides two preprocessing processes for extending the dataset. First and foremost, it combines BreakHis and BreCaHAD. The dataset is then rotated, shifted, gamma-flipped, and scaled. While the degree of malignancy remains the same, the manifestation of the disease is altered. This method improves photographs by adjusting the contrast without affecting the image data.
We can look for minute variations in histology images to predict breast cancer. It is recommended that a deep learning breast cancer screening system based on photographs is used. Our key scientific contribution is the recognition of breast tumors in histopathology photos using deep learning. Breast cancer categorization considers overfitting and color imbalance. Data expansion is currently ongoing. This model can detect low-level edge, contour, and color characteristics. There is less duplication, and ghost models are more straightforward. It depicts the model’s performance without sacrificing visual information. Breast cancer may be discovered by mammography, PET, biopsies, ultrasound, and magnetic resonance imaging. Therefore, this therapy might be advantageous. A pathologist must study the photos to establish if they are benign or malignant. CNN enhances deep learning systems for identifying breast cancer by using a cancer cell of grade 3 and its prognosis. The CNN model’s multilayer architecture identifies the cancer [24,25,26]. Convolution captures the local characteristics, pooling reduces dimensionality, and a fully connected layer produces the desired outcome. Convolution can present the edges and local visual elements. Normalization, pooling, and convolutional layers are all types of terminology used to describe the normalization, pooling, and convolution processes. The picture of the CNN layer is 128 pixels wide and has 3 channels (128 × 128 × 3). Small image databases need less memory and compute power. Pixels have an influence on algorithms in the extraction of characteristics from histology pictures. As a result, we employ normalized and improved photos to find microscopic cells. The smallest cancer cells and thin cancer cell borders may be found via max pooling. It selects a value at random from each receptive field. Assume that one wants to apply a filter that creates a pooling layer on a 3 × 3 pixel square image. One-pixel rotations in both the horizontal and vertical directions have little influence on the outcome of the pooling layer. One must remember that this applies whether you apply it to the original layer or to a modified version of the image, such as the one created after applying a convolution layer to the original. The pooling layer comes after the convolution layer and is often the next stage in the process. Despite the convolution layer’s effectiveness in detecting a meaningful structure within a sample of the data, the pooling layer introduces more noise into the equation, making it more difficult to localize the structure to a precise place. This is because the convolution layer handles the first stage of structure recognition. By offering an abstracted form of the representation, max pooling is one strategy for preventing model over-fitting. The main motivation is to increase the model’s accuracy. This decreases the number of facts that must be memorized and, as a result, the amount of effort necessary. Furthermore, the internal representation is translationally invariant. Because the total number of parameters has been cut down, translation invariance can now be maintained. The preprocessing phases of strain normalization and data augmentation increase feature extraction [27,28,29]. Many histology images with the same label may have different pixel intensity levels. This has an adverse effect on performance. This problem has been solved by the proposed feature extraction module, which takes the input image and performs the convolution procedure. Those used in histopathology frequently have a slightly greater resolution and cover a significantly larger area than images used in other imaging disciplines. A typical pathology slide’s resolution can exceed 100,000 pixels wide and 100,000 pixels deep, when scanned at high magnification. This is possibly the most distinguishing feature of histology photos. Because clinical annotations, such as written reports, apply to the entire picture or collections of images rather than specific locations within the image, it may be difficult to precisely match a search “query” with the relevant section of the image. Therefore, it may also be difficult to locate the region of the image that matches the “query” that was entered in the search field. This is demonstrated by the fact that a tumor in a pathology imaging study may be just one hundred pixels wide, or around one millionth of the whole picture area. Even if a medical specialist, researcher, or medical student found the photos using a text-based search, physical inspection would still be required to diagnose the lesion. One must complete this stage before proceeding to any other form of analysis. Real-world pathology cases, such as those in many other professions, typically involve multiple photographs. The text labels that are given may not be specific enough for the illness subtype that is of interest.
Reverse image search, also known as content-based image retrieval, looks for images that are visually “similar”. This strategy might be applied to address concerns outside of medicine. A clinician may opt to search a database for identical lesions as part of the diagnostic procedure to confirm whether a given trait of interest is suggestive of a benign or malignant histologic mimic, such as basal cell carcinoma. This approach can be used to determine if a worrying discovery is a benign or malignant histologic mimic. The doctor may assess the affected location to determine whether the characteristic of concern is a benign or malignant histologic mimic. Examples of technology that may have applications outside of the medical area include “search by photo” for common images, visual search for retail products, and various tools for faces and art. Analogous research exists in the medical imaging-related subfields of pathology and radiology. Models were created for each application in early CBIR systems that relied on machine learning. So, in order to install these technologies, it was necessary to collect labeled data for each application, which took a significant amount of time. Dimensionality reduction has two components, Fx and Fy. Fx then goes through average pooling to obtain smaller Fx (avg) and max pooling Fy (max) values to obtain a better value, which is then combined to make the feature map.
Feature maps are dilated with a dilation value of 4 to increase the receptive field of the architecture. As the dilation value increases, the receptive field grows, and the dilation mechanism multi-scales the features. The features are then obtained by performing element-wise multiplication of the input image and the feature map. The feature is then aggregated with the features generated by global average pooling and maximum pooling, followed by fully connected layers and dilation, and the weighted sum of these two blocks is used to retain the features of both long and short regions, enhancing the essential elements, while suppressing the useless information. As a result, the pixels can be observed more clearly. Furthermore, the cancer cell may be detected more precisely.
Expansion convolution searches for items while minimizing complexity (Perone, Calabrese & Cohen-Adad, 2018). Dilation broadens the network rather than increases the size of the filters at all levels. The filter depth may be calculated using the formula (2d + 1) × (2d + 1). The four receptive fields for d = 1, 2, 3 are 3 × 3, 5 × 5, 7 × 7, and 9 × 9. In addition, (x, y) = I (x + s × I y + s × j) denotes the dilated convolution (7). D(x, y) is a sparse dilated convolution filter with the input parameter I(x, y). When s = 1, dilation and standard convolution provide identical results. Dilation widens the receptive field without increasing the number of parameters. It also makes the process easier because dilation only adds a convolution operation to the process.

4. Experimental Results

This section evaluates the deep learning model. We begin by comparing the model’s performance to other models.

4.1. Trial and Evaluation of Models

There are 70 and 30 training and testing datasets, respectively. In addition, we update both the training and enriched data in 80/20 splits. The experimental data sets are summarized in. Table 2 contains image information for both datasets. Table 3. They are used to configure the network and choose hyperparameters. The proposed model is used for object recognition. Convolution was applied after max pooling to increase the receptive field, as shown in Figure 7.
Table 2. The Images from Both Datasets, where A: Training Images, B: Testing Images, C: Augmentation, and D: Validation.
Table 3. The Proposed Model’s Evolution Metrics.
Figure 7. Receptive Field Growth Layer vs. Receptive Field Number: (a) Stride (S) = 1, dilation 1, 1, 2, 2, 2 (b) S = 2, dilation 1, 1, 2, 2, 2 (c) S = 3, dilation 1, 1, 2, 2, 2 (d) S = 4, dilation 1, 1, 2, 2, 2.

4.2. Testing

TensorFlow 2 and Keras 2.4 are used in this Python 3.9 model. It has two Nvidia GeForce GTX 2070 GPUs, as well as a 16 MB cache. A dilated convolution model with many scales is used for object identification. Convolution was applied after max pooling to increase the receptive field (Figure 7). Dilated convolutions are the most dominant. The proposed multi-scale extended convolution model is better at predicting pixels for both malignant and benign tumors than the previous model.
BreakHis comprises 7909 actual samples (photos) from 82 people, including 2480 benign and 5429 malignant photographs, as well as 162 BreCaHAD breast cancer images. There were 153,349 parameters in the 7909 BreakHis samples and 3078 BreCaHAD images. The sections that follow compare the results of binary and multiclass classification. This model excels at identifying small details. Convolution was used to enhance the receptive field after max pooling. Instead, dilated convolution is used in the suggested model. The suggested model differentiates normal and malignant cells. Other convolution layers, as well as stochastic units for downsampling and upsampling, are used in the new model.
Better specificity, accuracy, precision, recall, and F1-score are shown in Table 3. The proposed model has a 98.60 percent accuracy for BreCaHAD and a 98.41 percent accuracy for BrekHis in the model with enhanced images, where A is the original picture and B is the enhanced image.
The receiver operating characteristic curve of the BreaKHis dataset was enlarged 40 times (1–5). These cell numbers indicate that the cells are in good health and indicate 90/10, 80/20, and 70–30 percentiles. RESNet50 and ResNeXt-101 models were used. ResNeXt101 outperformed ResNet50 on the provided datasets (Table 4). The accuracy, sensitivity, and precision of the model are exceptional. The VGG19 and VGG16 demonstrated the top five testing error rates of 8.9 percent and 9.2 percent, respectively. Their performance is evaluated in this section. Units such as the ghost unit and stochastic upsampling were proposed. The proposed model, which is derived from a network, discovers missing and minor properties. It was brightened and contrasted using adaptive histogram equalization. By integrating the two datasets (BreCaHAD and BreakHis), overfitting was reduced (BreCaHAD and BreakHis). This model improves the precision of the tip, contour, and color. These tools were used to analyze BreakHis and BreCaHAD data. Its small dataset hinders further study; however, it improves model accuracy. Small deviations are identified without the need for additional complexity. The detection of small cells improves network performance. The attributes are extracted algorithmically. Extra variables are not needed to distinguish malignant from normal cells and Figure 8 shows the malignant image segmented by units. Our suggested technique is ideal for detecting breast cancer. It can properly identify breast cancer at a low processing cost. Our method can also identify a variety of disorders.
Table 4. The Proposed Model’s Evolution Metrics.
Figure 8. Detection of Breast Cancers. (a) Input H&E Image (b) image after gaussian smoothing (c) It is the gradient magnitude it is determining in which direction the change in intensity is pointing (d) Gradient in the X direction (e) It is the non-maximum suppression for the detection of edges of cancerous cells (f) Image with minor contours and features.

5. Conclusions

Every day, women die from breast cancer in the United States. The aim of the research work is to encourage medical research and the development of technology by employing deep learning. The proposed hybrid dilation deep learning model is of two sorts. It utilizes dilation convolution and max pooling for multi-scale information. Cancer is now the main cause of mortality in almost every country. Between 2005 and 2015, the global cancer incidence increased by 33%. Breast cancer is the world’s sixth most deadly illness. Early identification and public awareness have the potential to significantly decrease mortality. Advances in image analysis and processing power have assisted in the incorporation of CAD systems into pathology lab operations. They also assist pathologists in detecting breast cancer early, which reduces fatalities. Image augmentation increases the number of pictures in a dataset, removing overfitting difficulties caused by dataset size constraints. Flipping, rotating, shifting, scaling, and gamma are all useful techniques. Our main contribution is the use of deep learning to recognize breast tumors in histopathology photographs and the proposed model achieved this with better accuracy. A pathologist must examine the photographs to determine if they are benign or cancerous. The layered design of the CNN model detects cancer and convolution searches for objects, while keeping the complexity to a minimum. Dilation broadens the network rather than enlarging the filters. Adding a convolution operation to the equation makes the process easier, which also makes the process faster. We can evaluate the findings of this study using multi-classification breast cancer detection tasks, rather than the more common binary assessments. It is critical to use primary data while carrying out additional research to better validate the efficacy of BC detection. The application of the CNN technique to the analysis of a wide range of data, such as gene expression data and image data from MRI scans, is a key outcome of this study. The suggested model cannot be used in any of these scenarios because it lacks the capacity to account for mixed-type data. In terms of output quality, these models are typically superior to their counterparts. There is little doubt that CNN would benefit from more research and may employ even more hybrid algorithms to attract the attention of academics.

Author Contributions

Conceptualization, T.H.H.A., R.N., E.A., H.A. and D.K.; methodology, TT.H.H.A., R.N. and D.K.; software, T.H.H.A., R.N. and D.K.; validation, E.A. and H.A.; formal analysis, T.H.H.A., R.N., H.A. and D.K.; investigation H.A., D.K. and E.A.; resources, T.H.H.A., R.N., H.A. and D.K.; data curation, T.H.H.A., R.N., H.A. and D.K.; original draft, T.H.H.A., R.N., E.A., H.A. and D.K.; preparation, T.H.H.A., R.N., E.A., H.A. and D.K.; writing—review and editing, T.H.H.A., R.N., E.A., H.A. and D.K.; visualization, T.H.H.A., R.N., H.A. and D.K.; supervision, D.K. project administration, T.H.H.A., R.N., H.A. and D.K.; funding acquisition, T.H.H.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research and the APC were funded by the Deanship of Scientific Research, Vice Presidency for Graduate Studies and Scientific Research, King Faisal University, Saudi Arabia (Project No. GRANT628).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Acknowledgments

This work was supported by the Deanship of Scientific Research, Vice Presidency for Graduate Studies and Scientific Research, King Faisal University, Saudi Arabia (Project No. GRANT628).

Conflicts of Interest

The authors declare that they have no competing interests.

References

  1. Pacilè, S.; Lopez, J.; Chone, P.; Bertinotti, T.; Grouin, J.M.; Fillard, P. Improving breast cancer detection accuracy of mammography with the concurrent use of an artificial intelligence tool. Radiol. Artif. Intell. 2020, 2, e190208. [Google Scholar] [CrossRef] [PubMed]
  2. Yari, Y.; Nguyen, T.V.; Nguyen, H.T. Deep learning applied for histological diagnosis of breast cancer. IEEE Access 2020, 8, 162432–162448. [Google Scholar] [CrossRef]
  3. IHME. Cancer’s Deadly Toll Grows in Less Developed Countries as New Cases Increase Globally; Institute for Health Metrics and Evaluation: Seattle, WA, USA, 2016; Available online: https://www.healthdata.org/news-release/cancer%E2%80%99s-deadly-toll-grows-less-developed-countries-new-cases-increase-globally (accessed on 25 September 2022).
  4. Mammary Duct Ectasia-Symptoms and Causes; Mayo Clinic: Rochester, MN, USA, 2022; Available online: https://www.mayoclinic.org/diseases-conditions/mammary-duct-ectasia/symptoms-causes/syc-20374801 (accessed on 18 July 2022).
  5. Suzuki, T. Research on analysis of final diagnosis and prognostic factors, and development of new therapeutic drugs for malignant tumors (especially malignant pediatric tumors). Yakugaku Zasshi 2020, 140, 229–271. [Google Scholar] [CrossRef] [PubMed]
  6. Spanhol, F.A.; Oliveira, L.S.; Petitjean, C.; Heutte, L. A Dataset for Breast Cancer Histopathological Image Classification. IEEE Trans. Biomed. Eng. 2015, 63, 1455–1462. [Google Scholar] [CrossRef]
  7. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  8. Szymak, P.; Gasiorowski, M. Using pretrained alexnet deep learning neural network for recognition of underwater objects. NAŠE MORE Znanstveni Časopis za More i Pomorstvo 2020, 67, 9–13. [Google Scholar] [CrossRef]
  9. Starke, S.; Leger, S.; Zwanenburg, A.; Leger, K.; Lohaus, F.; Linge, A.; Schreiber, A.; Kalinauskaite, G.; Tinhofer, I.; Guberina, N.; et al. 2D and 3D convolutional neural networks for outcome modelling of locally advanced head and neck squamous cell carcinoma. Sci. Rep. 2020, 10, 15625. [Google Scholar] [CrossRef]
  10. Kashyap, R. Breast Cancer Histopathological Image Classification Using Stochastic Dilated Residual Ghost Model. Int. J. Inf. Retr. Res. 2022, 12, 1–24. [Google Scholar] [CrossRef]
  11. Aldhyani, T.H.H.; Verma, A.; Al-Adhaileh, M.H.; Koundal, D. Multi-Class Skin Lesion Classification Using a Lightweight Dynamic Kernel Deep-Learning-Based Convolutional Neural Network. Diagnostics 2022, 12, 2048. [Google Scholar] [CrossRef]
  12. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale visual recognition. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2020. [Google Scholar]
  13. Shen, L.; Margolies, L.R.; Rothstein, J.H.; Fluder, E.; McBride, R.; Sieh, W. Deep learning to improve breast cancer detection on screening mammography. Sci. Rep. 2019, 9, 12495. [Google Scholar] [CrossRef]
  14. Koné, I.; Boulmane, L. Hierarchical ResNeXt models for breast cancer histology image classification. In Lecture Notes in Computer Science; INESC TEC: Porto, Portugal, 2018; pp. 796–803. [Google Scholar] [CrossRef]
  15. Gardezi, S.J.S.; Elazab, A.; Lei, B.; Wang, T. Breast cancer detection and diagnosis using mammographic data: Systematic review. J. Med. Internet Res. 2019, 21, e14464. [Google Scholar] [CrossRef] [PubMed]
  16. Aldhyani, T.H.H.; Alkahtani, H. A Bidirectional Long Short-Term Memory Model Algorithm for Predicting COVID-19 in Gulf Countries. Life 2021, 11, 1118. [Google Scholar] [CrossRef] [PubMed]
  17. Nair, R.; Bhagat, A. Healthcare information exchange through blockchain-based approaches. Transform. Bus. Bitcoin Min. Blockchain Appl. 2020, 2, 234–246. [Google Scholar] [CrossRef]
  18. Rajit, N.; Vishwakarma, S.; Soni, M.; Patel, T.; Joshi, S. Detection of COVID-19 cases through X-ray images using hybrid deep neural network. World J. Eng. 2021, 19, 33–39. [Google Scholar]
  19. Anghel, A.; Stanisavljevic, M.; Andani, S.; Papandreou, N.; Rüschoff, J.; Wild, P.; Gabrani, M.; Pozidis, H.A. High-performance system for robust stain normalization of whole-slide images in histopathology. Front. Med. 2019, 6, 193. [Google Scholar] [CrossRef]
  20. Kim, J.; Song, J.; Lim, D. CT image denoising using inception model. J. Korean Data Inf. Sci. Soc. 2020, 31, 487–501. [Google Scholar]
  21. Abdallah, T.; Elleuch, I.; Guermazi, R. Student behavior recognition in classroom using deep transfer learning with VGG-16. Procedia Comput. Sci. 2021, 12, 951–960. [Google Scholar] [CrossRef]
  22. Kashyap, R. Evolution of histopathological breast cancer images classification using stochastic dilated residual ghost model. Turk. J. Electr. Eng. Comput. Sci. 2021, 29, 2758–2779. [Google Scholar] [CrossRef]
  23. Lin, Y.; Wu, J. A novel multichannel dilated convolution neural network for human activity recognition. Math. Probl. Eng. 2020, 2020, 5426532. [Google Scholar] [CrossRef]
  24. Han, Z.; Wei, B.; Zheng, Y.; Yin, Y.; Li, K.; Li, S. Breast cancer multi-classification from histopathological images with structured deep learning model. Sci. Rep. 2017, 7, 4172. [Google Scholar] [CrossRef]
  25. Wang, B.; Zhang, X.; Zhou, X.; Li, J. A gated dilated convolution with attention model for clinical cloze-style reading comprehension. Int. J. Environ. Res. Public Health 2020, 17, 1323. [Google Scholar] [CrossRef] [PubMed]
  26. Zhang, J.; Lin, S.; Ding, L.; Bruzzone, L. Multi-scale context aggregation for semantic segmentation of remote sensing images. Remote Sens. 2020, 12, 701. [Google Scholar] [CrossRef]
  27. Li, X.; Shen, X.; Zhou, Y.; Wang, X.; Li, T. Classification of breast cancer histopathological images using interleaved DenseNet with SENet (IDSNet). PLoS ONE 2020, 15, e0232127. [Google Scholar] [CrossRef] [PubMed]
  28. Zhang, W.; Lu, G.; Zhang, S.; Li, Y. Self-paced hybrid dilated convolutional neural networks. Multimed. Tools Appl. 2020, 81, 34169–34181. [Google Scholar] [CrossRef]
  29. Lu, Z.; Bai, Y.; Chen, Y.; Su, C.; Lu, S.; Zhan, T. The classification of gliomas based on a pyramid dilated convolution resent model. Pattern Recognit. Lett. 2020, 133, 173–179. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.