Abstract
Ovarian cancer ranks as the fifth leading cause of cancer-related mortality in women. Late-stage diagnosis (stages III and IV) is a major challenge due to the often vague and inconsistent initial symptoms. Current diagnostic methods, such as biomarkers, biopsy, and imaging tests, face limitations, including subjectivity, inter-observer variability, and extended testing times. This study proposes a novel convolutional neural network (CNN) algorithm for predicting and diagnosing ovarian cancer, addressing these limitations. In this paper, CNN was trained on a histopathological image dataset, divided into training and validation subsets and augmented before training. The model achieved a remarkable accuracy of 94%, with 95.12% of cancerous cases correctly identified and 93.02% of healthy cells accurately classified. The significance of this study lies in overcoming the challenges associated with the human expert examination, such as higher misclassification rates, inter-observer variability, and extended analysis times. This study presents a more accurate, efficient, and reliable approach to predicting and diagnosing ovarian cancer. Future research should explore recent advances in this field to enhance the effectiveness of the proposed method further.
1. Introduction
The yearly mortality for ovarian cancer is 151,900, making it the deadliest cancer globally [1]. According to Miller, it is women’s fifth highest cause of death. The most common kind of gynaecological carcinoma is ovarian cancer, which emanates from epithelial tissue, and 90% of the cases are due to this type. The five histologic carcinomas are Mucinous-Ovarian Cancer (M-OC), High-Grade Serous-Ovarian Cancer (H-GS-OC), Low-Grade Serous-Ovarian Cancer (L-GS-OC), Clear-Cell Ovarian Cancer (C-COC), and Endometrioid-Ovarian-Cancer (E-O-C), with poor prognosis at an advanced stage [2].
Early Identification boosts survival from 3% in Stage IV to about 90% in Stage I [3]. Van Haaften-Day et al. (2001) [4] found that Carbohydrate Antigen 125 (CA125) has been used for over four decades, but that its accuracy is not acceptable as it has not improved survivability. 50% of early-stage tumours, primarily type I ovarian cancers, and 92% of advanced-stage tumours, primarily type II ovarian cancers, have increased serum CA125 levels. Skates et al. [5] found that physiological variables influence normal CA125 serum concentrations and the presence of menopause impacts CA125 levels. Sopik et al. [6] found that benign illnesses also had increased CA125 levels, causing false positives. Only 20% of ovarian tumors express CA125. A screening’s anticipated sensitivity is around 80%. Moss et al. [7] confirmed that relying on the biomarker alone is misleading. Akinwunmi et al. [8] found that 1% of healthy people had increased serum levels, including 5% of patients with benign illnesses, e.g., endometriosis.
HE4, WFDC2, is overexpressed in Endometrioid ovarian cancer and less in epithelial tissues of the respiratory system and reproductive organs [9]. Yana-ranop et al. [10] found that the specificity of HE4 was 86% compared to that of CA125, and the AUC of it was 0.893 compared to that of CA125 0.865 [11]. HE4 levels vary in smokers (30%) and non-smokers (20%) [12]. Contraceptives impact HE4 levels. Ferraro et al. [13] found lower HE4 levels in oral contraceptive users (p = 0.008). Biopsy, imaging (US, CT, MRI, PET), and algorithms for learning (deep) Convolutional Neural networks (CNN) can predict and diagnose epithelial ovarian cancer (serous) accurately.
The challenge is that there is no effective screening method; hence, ovarian cancer is diagnosed when it has already advanced to Stage III or IV. Radiologists manually analyse and interpret medical images of a suspected cancer patient for cancer subtyping and staging. This results in the misclassification of the cancer subtypes, inter-observer variations, subjectivity, and time consumption. To address this, a deep CNN model was developed to predict and diagnose cancers.
Expert pathologists interpret Cellular morphology, defining OC categories and directing treatments [14]. Inter-observer variations cause inaccurate, suboptimal treatments, poor prognosis, and reduced life quality [15]. This shows the need for computational methods to accurately predict cancer class and diagnose cancer subtype.
Accurate prediction and diagnosis of the cell tumours are vital as they lead to proper prognosis and treatment, increasing survivability. Deep learning merits include:
- Processing huge data and producing highly accurate predictions, reducing incorrect diagnoses.
- Permitting early detection of ovarian carcinoma, increasing treatment success.
- Permitting personalised treatment. Deep learning can predict how treatments affect different women, enabling personalised, efficient care.
Deep learning can improve patient outcomes and reduce mortality through early detection and personalised treatment. The algorithm could predict and diagnose the images in under 5 s with an F1-score of 0.94. The order of the paper is as follows: related work, the materials and methods used, followed by results and discussion, and, finally, the conclusion.
2. Related Work
Various deep-learning techniques have been studied to classify ovarian carcinoma based on the cell type. Personalized treatment plans for ovarian cancer patients depend on accurately identifying the type of ovarian cancer. In the past decade, various studies have aimed to enhance cancer screening outcomes in the preclinical stage, focusing on utilizing histopathological images and biomarkers, such as CA-125 and HE-4. However, biomarker-based detection can be slow and prone to missing detection [16,17]. CA-125 has limited accuracy and specificity in detecting early-stage ovarian cancer, and imaging methods such as CT, US, and MRI are commonly used for locating and detecting features of masses. However, the interpretation of images by expert radiographers can be time-consuming and prone to inter-observer variability [18].
Recent studies have focused on using machine learning algorithms for early prediction and diagnosis of ovarian cancer. Different methods have been proposed to extract features from ultrasound images and classify ovarian tumors. Some techniques include SVM, shallow neural network classifiers, and wavelet transform filters. The features extracted from the images include textural and pathological features, wavelet coefficients, homogeneousness, histogram, and grey difference multi-scaling. These features are then integrated with a support-vector-machine (SVM) classifier to classify the types of cysts fully [19,20,21,22,23,24]. El-Bendary and Belal [25] developed a combined approach for subtyping cancer stage and ovarian cancer classification using gene expression and clinical datasets. Boosting and Ensemble SVM achieved a classification accuracy of 80%, while other ML classifiers yielded 70.77%. The proposed method also had a high recall, specificity, precision, F measure, and AUC.
The authors in [26] developed a CAD method for diagnosing borderline ovarian cancer by analyzing S-HG images. They used a k-NN classifier, and an optimization tool called T-POT, which resulted in an average accuracy between 0.976 and 0.96. The T-POT classifier achieved AUROC values of 0.97, 1, 0.98, and 0.99 for normal, malignant, borderline, and benign cases, respectively.
Deep learning techniques offer the advantage of intricate feature learning directly from raw data, making it possible to define a system without manually creating features that are necessary for other machine learning methods. This property has drawn interest in examining the benefits of using deep learning in medical image processing [27,28]. Multiphoton Microscopy (MPM) imaging and wide-field fluorescence imaging (WFI) has been proposed for ovarian imaging to capture specific biomarkers that cannot be detected by MPM alone while providing high-resolution images of the entire ovary. Using the Linear Discriminant Analysis classification algorithm, Sawyer et al. [29] achieved an accuracy of 66.66%, 87.50%, and 62.5% for genotype, age, and treatment, respectively [30,31].
The article highlights the different methods researchers propose for detecting and classifying various types of cancer using deep learning algorithms. Lu et al. [32] proposed a Tumour Origin and Assessment method via Deep Learning (TOAD), yielding an AUC-ROC of 0.988. Booma et al. [33] introduced a method using max-pooling enhanced with ML algorithms, which achieved an accuracy of 89%. Wen et al. utilized a custom collection of 3D filters achieving accuracy with AU-ROC ranging from 83% to 90% [34] (Wen et al., 2016). Huttunen et al. [35] used deep neural networks to classify unstained tissue multiphoton microscopy images achieving 95% sensitivity and 97% specificity. Wang et al. [36] proposed a two-level deep transfer learning approach, achieving an accuracy of 87.54%. Lastly, Yu et al. [37] proposed a radiomics model based on a convolutional neural network (CNN), which achieved an AUC of 0.955 to 0.975. They suggested that a CNN-based radiomics method for predicting endometrial cancer and distinguishing histopathology slides with cancer cells, showing potential clinical value in detecting tumour cells and predicting chemotherapy response.
Sengupta et al. [38] used DHL with lamin protein distribution to diagnose ovarian cancer based on the morphology of the nucleus, showing potential as a useful marker in prognosis and histology. Zhang et al. [39] used radiomics and a CNN-based model to predict Endometrioid Cancer with promising clinical potential, although with a limited sample size. Liao et al. [40] proposed the MTDL method for addressing high-dimensional feature space issues and improving classification accuracy, with the potential for further application to additional dataset categories. Various studies have proposed promising methods for diagnosing and predicting gynecological cancers, including ovarian and endometrial cancers, using advanced techniques such as MPM, WFI, DHL, radiomics, and CNN-based models, which hold potential for improving clinical outcomes.
Guo et al. [41] used deep and machine-learning techniques for clustering and classification in ovarian cancer subtyping using gene expression features. Kasture et al. [42] proposed a DL approach for predicting and classifying ovarian cancer subtypes using histopathological images. The authors in [43] used F-R-CNN to classify ovarian images. Ghoniem et al. [44] proposed a hybrid DL using multimodal data, combining gene and histopathology images with ALO-optimized LSTM and CNN networks. The proposed study of Ghoniem et al. suggests a novel approach to diagnosing ovarian cancer using a hybrid evolutionary deep-learning model that integrates multiple types of data and various assessment indicators. By comparing the performance of the proposed model with nine other multi-modal fusion models, the study demonstrates that the proposed model achieves higher precision and accuracy in ovarian cancer diagnoses.
A visual example of early detection through conventional methods as shown in Figure 1.
Figure 1.
A visual example of early detection through conventional methods.
Finally, Xao et al. [45] used multi-omics to identify biomarkers for early OC diagnosis. We requested readers to check out in details the figure [46], which shows women with genetic mutations, or a family cancer history are at high risk. Tissue and liquid biopsy were analyzed using omics technology, combined with ML technology for biomarker development, and combined with the machine learning algorithm to form omics [47]. In addition, the authors refer to figure [47], which relates to using multi-omics to identify biomarkers, which speeds up the process of discovering new biomarkers. The process is displayed in Figure 2.
Figure 2.
Tissue and liquid biopsy were analysed using omics technology, combined with ML technology for biomarker development, combined with the machine learning algorithm to form omics.
In their seminal work, Arezzo et al. investigated the application of radiomics in the qualitative and quantitative analysis of images obtained from various imaging techniques, such as magnetic resonance imaging (MRI) and ultrasound [48]. The primary objective of the radiomics approach is to achieve personalized ovarian cancer medicine. However, they concluded that radiomics has not yet been widely adopted in clinical settings due to several challenges, including:
- The absence of a benchmark protocol for feature extraction in each method involved.
- Bias is introduced by differences in images acquired from various instruments.
- A lack of prospective external validation of algorithmic models on large datasets, resulting in insufficient studies addressing the performance of radiomics [48]. Arezzo et al. emphasized that images cannot only show the extent of the disease but also aid in its diagnosis. They suggested that deep learning could potentially overcome some of the limitations of radiomics in image analysis.
In a subsequent study, Arezzo et al. [48] revealed that peritoneal tuberculosis (TBP) exhibits similar signs and symptoms to advanced ovarian cancer. TBP accounts for approximately 1% to 2% of tuberculosis cases. The similarity between TBP and ovarian carcinoma presents challenges in the differential diagnosis. Biopsy remains the definitive method for confirming the presence of the disease, while computed tomography (CT) and ultrasound can be employed for visualizing nodules [49]. The authors proposed that a deep learning framework could effectively enhance the accuracy of image analysis, providing more objective diagnoses when using these imaging techniques.
Reilly et al. [50] outline the development and validation process of MIA3G, a deep neural network algorithm that aims to detect ovarian cancer. The algorithm was trained on a dataset of 1067 specimens and validated on a separate set of 2000 specimens. The results show that MIA3G achieves a sensitivity of 89.8% and a specificity of 84.02% in detecting ovarian cancer.
Machine learning integrates and analyses high-throughput molecular experiments to create new biomarkers for understanding illness. Integrated data analysis through machine learning can reveal biologically meaningful biomarker candidates despite changes at each omics level. Artificial intelligence studies are predicted to improve precision medicine for OC significantly. Elyan et al. [51] reviewed advances in using deep learning to analyse images and compared ML and DL algorithms. Figure 3 compares machine learning techniques that require manual feature extraction and deep learning, which do the extraction automatically.
Figure 3.
The approach of Earlier C-V techniques (top) against DL (bottom). (Gumbs et al.) [52].
Machine learning (ML) [53] enables computers to learn from past data without creating algorithms to account for the unlimited number of possible feature combinations. DL techniques differ from traditional approaches that use manual labelling. DL techniques learn the features of incoming photos from start to end without feature extraction or engineering. Convolutional neural networks (C-NNs) can capture the underlying representation of images by using partially linked layers and weight sharing, as shown in Figure 4.
Figure 4.
CNN model schematic (Sarvamangala & Kulkarni, 2022) [53].
The studies highlighted the potential of deep learning and machine learning techniques in ovarian cancer subtyping and diagnosis using various data modalities, including gene expression, histopathological images, and ultrasound images. These methods have shown promising results in improving accuracy and identifying subtypes at the molecular level. However, the small sample sizes and the need for larger datasets remain a knowledge gap.
Table 1 compares machine learning methods used for diagnosing and subtyping ovarian cancer. The table lists various researchers’ studies, methodologies, and performance metrics, such as accuracy, sensitivity, specificity, and AUC-ROC.
Table 1.
Comparative Analysis of Machine Learning Methods for Diagnosis and Subtyping of Ovarian Cancer.
3. Materials and Methods
This section discusses the methodology used in the research, including the dataset preparation and the proposed Convolutional Neural Network (CNN) architecture. The dataset used in this project consists of 200 images, with 100 images each of serous ovarian cancer and non-cancerous samples. The original dataset was obtained from The Cancer Genome Atlas TCGA repository and was augmented to 11,040 images for both classes to be effectively used in the deep learning architecture (Kasture et al., 2021) [42]. This dataset was split into 80% for the training set and 20% for validation. The proposed CNN architecture is shown in Figure 5.
Figure 5.
The proposed architecture of the model.
The preprocessed dataset was then fed for convolution operation. The preprocessing conducted on the images was the elimination of the images that had errors, such as the ones which were encoded badly and did not have the jpeg extension. The data was cleaned to remove the images that were distorted. Convolution is a special line function used to extract a feature using some feature detectors, also called a kernel, which are applied to all inputs, and a series of numbers called a tensor. The product-wise element between each kernel element and the input tensor is calculated in each tensor area and summarized to find the output value in the corresponding outgoing tensor area called the feature map. With striding and padding included, the formula for determining the output image dimension was as follows:
where
- n = Number of input pixels
- f = Number of pixels of filter
- p = Padding
- s = Stride
To mimic the real-world representation mathematically of the behavior of neurons, the result of convolution was passed on to the non-linear activation function, ReLU, defined mathematically as The activation function choice includes the fact that it does not engage all of the neurons at the same time and does not stimulate all of the neurons at the same time. Hence during backpropagation, not all the neurons are activated. Next, the data was passed onto the pooling layer, which provided a standard downtime that reduced the size of the feature map element to introduce translation flexibility in tiny shifts and distortions, and that reduced the number of subsequent parameters that could be learned. The research used Maxpooling. The data was then fed into the flattening layer and converted into a one-dimensional array or (vector) numbers for inputting into the fully connected layer. After that, the data is fed to the fully connected layer of feedforward neural networks. Every node in the previous layer is linked to the next layer’s nodes by a learnable weight. The features from the pooling layers are mapped to the network’s outputs. Lastly, the result was fed to the output layer, which used SoftMax for classification. The Loss function used was cross entropy and was given by the following formula for binary classification:
where:
- l = Loss function
- p = The predicted probability
- y = 0 or 1 in binary classification
In conclusion, the proposed CNN architecture and dataset preparation for ovarian cancer classification has been discussed. The algorithm used to train the augmented image dataset has been summarized, and Table 2 summarizes the hyperparameters used in the CNN architecture.
Table 2.
Hyperparameters and Configuration Settings of CNN for Image Classification.
4. Results
The model was fed with 11,040 images––half healthy cells, and the other half infected with serous cancer subtype. The number of epochs was chosen for hyperparameter selection and incremented in tens. The training and validation accuracy were recorded.
Table 3 shows a steady increase in the training and validation accuracy as the number of epochs increased from 10 to 50, reaching a training accuracy of 99.52% and a validation accuracy of 99.91% after 50 epochs.
Table 3.
The number of epochs and the corresponding training and validation accuracies.
After the training process, the testing process was performed by uploading an image from the test dataset, and the algorithm would give the percentage of it being either serous or healthy cells. This result demonstrates the superiority of the Xception network, which achieves high training and validation accuracy with a small number of epochs. Compared to conventional generic convolutional neural networks, the model does not perform channel-wise convolution, reducing the number of connections and making the model lighter. As a result, excellent accuracy can be achieved with just 10 epochs (99.03% training accuracy and 94.43% validation accuracy).
The model accuracy shows that the percentage for training and validation accuracy rose as shown in Figure 6. This graph shows accuracy from 98.56% and 99.73% to 99.52% and 99.91% for 10 to 50 epochs, respectively.
Figure 6.
Model Accuracy for both training and validation after 50 epochs.
The percentage loss (shown in Figure 7) for training and validation stood at 0.0445 and 0.0083 and 0.0147, and 0.0020 after 10 and 50 epochs, respectively.
Figure 7.
Model loss for both training and validation after 50 epochs.
Figure 8 shows the 93.02% of the images were correctly classified as health, and 95.12% were classified as serous cells. From the confusion matrix, valuable performance metrics can be derived, such as:
Figure 8.
Confusion Matrix of the proposed model.
The sensitivity of the model is 0.9502, which represents the model’s ability to predict and classify the health cell images correctly as health when the label class is health.
The specificity represents the model’s ability to correctly predict and classify the infected cells as infected cells when the class label is serous.
Recall measures how well the model can detect positive events, as in the health images class detection.
Precision measures how well the model has assigned the positive events to the positive class.
The F-score is the harmonic mean of the model’s precision and recall, and it measures the model’s accuracy on the dataset of health and serous.
Table 4 shows that the precision for health is higher than that of serous cells by 2 units, and the recall for both is equal at 0.93. The F1 score is 0.94 for all classes. The classification report summarizes the performance of the Xception network on the medical images for the entire dataset.
Table 4.
Classification Report for Health and Serous Classifications.
Table 5 shows the accuracy score for different classification models; deep hybrid learning, convolutional neural network, GoogleNet (v3), and linear discriminant analysis classification. The GoogleNet v3 achieved the highest accuracy of 0.925, and the linear discriminant achieved the lowest of 0.666. As a comparison, the accuracy of GoogleNet v3 can be attributed to the transfer learning process, whereby the features are input to a pre-trained architecture. The hyperparameters would have been tuned already. The method that was proposed outperformed the ones by other researchers. A classification accuracy of 0.94 can be attributed to the efficient use of the parameters. The network is built because it has batch normalization after every convolution, which helps fight the model’s overfitting.
Table 5.
Comparison of Accuracy scores for different models.
Table 5 shows that a “Deep Hybrid Learning” model achieved an accuracy score of 0.907. This model combines deep learning and traditional machine learning techniques, such as decision trees or linear regression, to improve performance. The specific structure of the network used in the study is not provided. Still, it likely involves multiple layers of deep neural networks and incorporates techniques such as batch normalization to combat overfitting.
5. Discussion
In this study, we aimed to develop a deep-learning model to classify healthy and serous ovarian cancer cells accurately. Our model achieved a high accuracy of 94.43%, a sensitivity of 0.9502, and a specificity of 0.9316, demonstrating its potential to assist physicians in diagnosing and predicting ovarian cancer prognosis, ultimately improving patient outcomes.
Our findings build upon previous research in the field, with our model surpassing the performance of other models reported in the literature. The superior performance can be attributed to the depth-wise separable convolution, which reduces model complexity, and the absence of activation functions in intermediate layers and residual connections. These innovations enabled our model to achieve higher accuracy than the 80% reported by El-Nabawy et al. [25], the 66.66–87.5% by Sawyer et al. [29], and the 83.6% by Lu et al. [32].
However, the study has some limitations. A public TCGA dataset may not fully represent local patient populations, and the retrospective design does not account for disease progression. Additionally, the lack of clinical data, such as patient genetics, may affect model robustness. To address these issues, future research should involve partnerships with hospitals and research centers to validate the model on real-world data and incorporate clinical datasets.
Alternative explanations for the findings should also be considered. Model interpretation techniques could provide insights into the underlying factors driving the model’s predictions, and the model’s design can be further refined to improve performance.
The practical implications of our findings include the potential to support ovarian cancer diagnosis and monitoring in clinical settings. Our model also offers recommendations for future research, such as addressing limitations, improving model interpretability, and validating the model on diverse populations.
6. Conclusions
In summary, this study explored the potential of a deep learning model for predicting and diagnosing ovarian cancer. Our model achieved an accuracy of 94%, highlighting its promise for early prediction and diagnosis. The study contributes to the existing literature by demonstrating the potential of deep learning in preclinical cancer screening and personalized management of ovarian cancer.
The broader implications of our research include the potential impact on related fields, such as radiology, and real-world applications in ovarian cancer diagnosis and treatment optimization. Despite some limitations, our study offers suggestions for future research to build upon our findings, such as improving data quality and quantity, evaluating models on new independent datasets, and enhancing generalizability through diverse population studies.
Overall, this research emphasizes the potential of advanced computing techniques in the timely detection of ovarian cancer. By addressing current challenges and refining these models, we can develop more reliable and broadly applicable tools that can be adopted for practical medical purposes, ultimately improving cancer diagnosis, outcome prediction, and personalized medical care.
Author Contributions
Conceptualization, Q.A., M.A. (Mubarak Albathan) and A.H.; Data curation, A.Y., Q.A., M.B. and A.H.; Formal analysis, A.Y., T.M., M.U.T., M.B., M.A. (Muhammad Asim) and A.H.; Funding acquisition, M.A. (Mubarak Albathan); Investigation, B.Z., A.Y., T.M., M.U.T., M.A. (Mubarak Albathan), S.J. and A.H.; Methodology, B.Z., A.Y., T.M., M.U.T., S.J., M.B. and M.A. (Muhammad Asim); Project administration, M.A. (Mubarak Albathan); Resources, B.Z., T.M., M.U.T., S.J., M.B. and M.A. (Muhammad Asim); Software, B.Z., A.Y., T.M., M.U.T., Q.A., M.A. (Mubarak Albathan), M.B. and A.H.; Supervision, M.A. (Muhammad Asim); Validation, T.M., M.U.T., M.A. (Mubarak Albathan) and S.J.; Visualization, Q.A., M.A. (Mubarak Albathan) and M.A. (Muhammad Asim); Writing—original draft, B.Z., Q.A., S.J., M.B. and A.H.; Writing—review and editing, Q.A. and M.A. (Muhammad Asim). All authors have read and agreed to the published version of the manuscript.
Funding
The authors extend their appreciation to the Deanship of Scientific Research at Imam Mohammad Ibn Saud Islamic University (IMSIU) for funding and supporting this work through Research Partnership Program no. RP-21-07-11.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Dataset will be available upon request.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Torre, L.A.; Bray, F.; Siegel, R.L.; Ferlay, J.; Lortet-Tieulent, J.; Jemal, A. Global cancer statistics, 2012. CA Cancer J. Clin. 2015, 65, 87–108. [Google Scholar] [CrossRef] [PubMed]
- Reid, B.M.; Permuth, J.B.; Sellers, T.A. Epidemiology of ovarian cancer: A review. Cancer Biol. Med. 2017, 14, 9–32. [Google Scholar] [CrossRef] [PubMed]
- Vázquez, M.A.; Mariño, I.P.; Blyuss, O.; Ryan, A.; Gentry-Maharaj, A.; Kalsi, J.; Manchanda, R.; Jacobs, I.; Menon, U.; Zaikin, A. A quantitative performance study of two automatic methods for the diagnosis of ovarian cancer. Biomed. Signal Process. Control. 2018, 46, 86–93. [Google Scholar] [CrossRef] [PubMed]
- Van Haaften-Day, C.; Shen, Y.; Xu, F.; Yu, Y.; Berchuck, A.; Havrilesky, L.J.; De Bruijn HW, A.; Van Der Zee AG, J.; Bast, R.C.; Hacker, N.F. OVX1, Macrophage-Colony Stimulating Factor, and CA-125-II as Tumor Markers for Epithelial Ovarian Carcinoma A Critical Appraisal. Cancer Interdisciplin. Int. J. Am. Cancer Soc. 2001, 92, 2837–2844. [Google Scholar] [CrossRef]
- Skates, S.J.; Mai, P.; Horick, N.K.; Piedmonte, M.; Drescher, C.W.; Isaacs, C.; Armstrong, D.K.; Buys, S.S.; Rodriguez, G.C.; Horowitz, I.R.; et al. Large Prospective Study of Ovarian Cancer Screening in High-Risk Women: CA125 Cut-Point Defined by Menopausal Status. Cancer Prev. Res. 2011, 4, 1401–1408. [Google Scholar] [CrossRef]
- Sopik, V.; Rosen, B.; Giannakeas, V.; Narod, S.A. Why have ovarian cancer mortality rates declined? Part III. Prospects for the future. Gynecol. Oncol. 2015, 138, 757–761. [Google Scholar] [CrossRef]
- Moss, E.L.; Hollingworth, J.; Reynolds, T.M. The role of CA125 in clinical practice. J. Clin. Pathol. 2005, 58, 308–312. [Google Scholar] [CrossRef]
- Akinwunmi, B.O.; Babic, A.; Vitonis, A.F.; Cramer, D.W.; Titus, L.; Tworoger, S.S.; Terry, K.L. Chronic Medical Conditions and CA125 Levels among Women without Ovarian Cancer. Cancer Epidemiol. Biomark. Prev. 2018, 27, 1483–1490. [Google Scholar] [CrossRef]
- Drapkin, R.; von Horsten, H.H.; Lin, Y.; Mok, S.C.; Crum, C.P.; Welch, W.R.; Hecht, J.L. Human Epididymis Protein 4 (HE4) Is a Secreted Glycoprotein that Is Overexpressed by Serous and Endometrioid Ovarian Carcinomas. Cancer Res. 2005, 65, 2162–2169. [Google Scholar] [CrossRef]
- Yanaranop, M.; Anakrat, V.; Siricharoenthai, S.; Nakrangsee, S.; Thinkhamrop, B. Is the Risk of Ovarian Malignancy Algorithm Better Than Other Tests for Predicting Ovarian Malignancy in Women with Pelvic Masses? Gynecol. Obstet. Investig. 2017, 82, 47–53. [Google Scholar] [CrossRef]
- Wu, C.; Wang, Y.; Wang, F. Deep Learning for Ovarian Tumor Classification with Ultrasound Images. In Advances in Multimedia Information Processing–PCM 2018: 19th Pacific-Rim Conference on Multimedia, Hefei, China, 21–22 September 2018; Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer International Publishing: Berlin/Heidelberg, Germany, 2018; pp. 395–406. [Google Scholar] [CrossRef]
- Bolstad, N.; Øijordsbakken, M.; Nustad, K.; Bjerner, J. Human epididymis protein 4 reference limits and natural variation in a Nordic reference population. Tumor Biol. 2011, 33, 141–148. [Google Scholar] [CrossRef]
- Ferraro, S.; Schiumarini, D.; Panteghini, M. Human epididymis protein 4: Factors of variation. Clin. Chim. Acta 2015, 438, 171–177. [Google Scholar] [CrossRef] [PubMed]
- Jayson, G.C.; Kohn, E.C.; Kitchener, H.C.; Ledermann, J.A. Ovarian cancer. Lancet 2014, 384, 1376–1388. [Google Scholar] [CrossRef] [PubMed]
- Kommoss, S.; Pfisterer, J.; Reuss, A.; Diebold, J.; Hauptmann, S.; Schmidt, C.; Du Bois, A.; Schmidt, D.; Kommoss, F. Specialized Pathology Review in Patients with Ovarian Cancer. Int. J. Gynecol. Cancer 2013, 23, 1376–1382. [Google Scholar] [CrossRef]
- Yoshida-Court, K.; Karpinets, T.V.; Mitra, A.; Solley, T.N.; Dorta-Estremera, S.; Sims, T.T.; Delgado Medrano, A.Y.; El Alam, M.B.; Ahmed-Kaddar, M.; Lynn, E.J.; et al. Immune environment and antigen specificity of the T cell receptor repertoire of malignant ascites in ovarian cancer. PLoS ONE 2023, 18, e0279590. [Google Scholar] [CrossRef] [PubMed]
- de Leon, A.; Perera, R.; Nittayacharn, P.; Cooley, M.; Jung, O.; Exner, A.A. Ultrasound Contrast Agents and Delivery Systems in Cancer Detection and Therapy. Adv. Cancer Res. 2018, 139, 57–84. [Google Scholar] [CrossRef] [PubMed]
- Lusk, J.F.; Miranda, C.; Howell, M.; Chrest, M.; Eshima, J.; Smith, B.S. Photoacoustic Flow System for the Detection of Ovarian Circulating Tumor Cells Utilizing Copper Sulfide Nanoparticles. ACS Biomater. Sci. Eng. 2019, 5, 1553–1560. [Google Scholar] [CrossRef]
- Danaee, P.; Ghaeini, R.; Hendrix, D.A. A deep learning approach for cancer detection and relevant gene identification. In Proceedings of the 22nd Pacific Symposium on Biocomputing (PSB), Kohala Coast, HI, USA, 4–8 January 2017; pp. 219–229. [Google Scholar]
- Chen, S.-J.; Chang, C.-Y.; Chang, K.-Y.; Tzeng, J.-E.; Chen, Y.-T.; Lin, C.-W.; Hsu, W.-C.; Wei, C.-K. Classification of the Thyroid Nodules Based on Characteristic Sonographic Textural Feature and Correlated Histopathology Using Hierarchical Support Vector Machines. Ultrasound Med. Biol. 2010, 36, 2018–2026. [Google Scholar] [CrossRef]
- Acharya, U.R.; Sree, S.V.; Swapna, G.; Gupta, S.; Molinari, F.; Garberoglio, R.; Witkowska, A.; Suri, J.S. Effect of complex wavelet transform filter on thyroid tumor classification in three-dimensional ultrasound. Proc. Inst. Mech. Eng. Part H J. Eng. Med. 2013, 227, 284–292. [Google Scholar] [CrossRef]
- Katsigiannis, S.; Keramidas, E.G.; Maroulis, D. A Contourlet Transform Feature Extraction Scheme for Ultrasound Thyroid Texture Classification. Int. J. Eng. Intell. Syst. Electr. Eng. Commun. 2010, 18, 171. [Google Scholar]
- Chang, C.-Y.; Chen, S.-J.; Tsai, M.-F. Application of support-vector-machine-based method for feature selection and classification of thyroid nodules in ultrasound images. Pattern Recognit. 2010, 43, 3494–3506. [Google Scholar] [CrossRef]
- Chang, C.-Y.; Liu, H.-Y.; Tseng, C.-H.; Shih, S.-R. Computer-aided diagnosis for thyroid graves’ disease in ultrasound images. Biomed. Eng. Appl. Basis Commun. 2012, 22, 91–99. [Google Scholar] [CrossRef]
- El-Nabawy, A.; El-Bendary, N.; Belal, N.A. Epithelial Ovarian Cancer Stage Subtype Classification using Clinical and Gene Expression Integrative Approach. Procedia Comput. Sci. 2018, 131, 23–30. [Google Scholar] [CrossRef]
- Wang, G.; Sun, Y.; Jiang, S.; Wu, G.; Liao, W.; Chen, Y.; Lin, Z.; Liu, Z.; Zhuo, S. Machine learning-based rapid diagnosis of human borderline ovarian cancer on second-harmonic generation images. Biomed. Opt. Express 2021, 12, 5658–5669. [Google Scholar] [CrossRef] [PubMed]
- Ma, S.; Sigal, L.; Sclaroff, S. Learning Activity Progression in LSTMs for Activity Detection and Early Detection. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1942–1950. [Google Scholar] [CrossRef]
- Aliamiri, A.; Shen, Y. Deep learning based atrial fibrillation detection using wearable photoplethysmography sensor. In Proceedings of the 2018 IEEE EMBS International Conference on Biomedical Health Informatics (BHI), Las Vegas, NV, USA, 4–7 March 2018; pp. 442–445. [Google Scholar]
- Sawyer, T.W.; Koevary, J.W.; Rice, F.P.S.; Howard, C.C.; Austin, O.J.; Connolly, D.C.; Cai, K.Q.; Barton, J.K. Quantification of multiphoton and fluorescence images of reproductive tissues from a mouse ovarian cancer model shows promise for early disease detection. J. Biomed. Opt. 2021, 24, 096010. [Google Scholar] [CrossRef]
- Liang, Q.; Wendelhag, I.; Wikstrand, J.; Gustavsson, T. A multiscale dynamic programming procedure for boundary detection in ultrasonic artery images. IEEE Trans. Med. Imaging 2000, 19, 127–142. [Google Scholar] [CrossRef]
- LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
- Lu, M.Y.; Chen, T.Y.; Williamson, D.F.K.; Zhao, M.; Shady, M.; Lipkova, J.; Mahmood, F. AI-based pathology predicts origins for cancers of unknown primary. Nature 2021, 594, 106–110. [Google Scholar] [CrossRef]
- Booma, P.M.; Thiruchelvam, V.; Ting, J.; Ho, S. Max pooling technique to detect and classify medical image for max pooling technique to detect and classify medical image for ovarian cancer diagnosis. Test Eng. Manag. J. 2020, 82, 8423–8442. [Google Scholar]
- Wen, B.; Campbell, K.R.; Tilbury, K.; Nadiarnykh, O.; Brewer, M.A.; Patankar, M.; Singh, V.; Eliceiri, K.W.; Campagnola, P.J. 3D texture analysis for classification of second harmonic generation images of human ovarian cancer. Sci. Rep. 2016, 6, 35734. [Google Scholar] [CrossRef]
- Huttunen, M.J.; Hassan, A.; McCloskey, C.W.; Fasih, S.; Upham, J.; Vanderhyden, B.C.; Boyd, R.W.; Murugkar, S. Automated classification of multiphoton microscopy images of ovarian tissue using deep learning. J. Biomed. Opt. 2018, 23, 66002. [Google Scholar] [CrossRef] [PubMed]
- Wang, C.-W.; Lee, Y.-C.; Chang, C.-C.; Lin, Y.-J.; Liou, Y.-A.; Hsu, P.-C.; Chang, C.-C.; Sai, A.-K.-O.; Wang, C.-H.; Chao, T.-K. A Weakly Supervised Deep Learning Method for Guiding Ovarian Cancer Treatment and Identifying an Effective Biomarker. Cancers 2022, 14, 1651. [Google Scholar] [CrossRef] [PubMed]
- Yu, K.-H.; Hu, V.; Wang, F.; Matulonis, U.A.; Mutter, G.L.; Golden, J.A.; Kohane, I.S. Deciphering serous ovarian carcinoma histopathology and platinum response by convolutional neural networks. BMC Med. 2020, 18, 236. [Google Scholar] [CrossRef] [PubMed]
- Sengupta, D.; Ali, S.N.; Bhattacharya, A.; Mustafi, J.; Mukhopadhyay, A.; Sengupta, K. Nuclear Morphology Optimized Deep Hybrid Learning (NUMODRIL): A novel architecture for accurate diagnosis/prognosis of Ovarian Cancer. bioRxiv 2020. [Google Scholar] [CrossRef]
- Zhang, Y.; Gong, C.; Zheng, L.; Li, X.; Yang, X. Deep Learning for Intelligent Recognition and Prediction of Endometrial Cancer. J. Health Eng. 2021, 2021, 1148309. [Google Scholar] [CrossRef] [PubMed]
- Liao, Q.; Ding, Y.; Jiang, Z.L.; Wang, X.; Zhang, C.; Zhang, Q. Multi-task deep convolutional neural network for cancer diagnosis. Neurocomputing 2019, 348, 66–73. [Google Scholar] [CrossRef]
- Guo, L.-Y.; Wu, A.-H.; Wang, Y.-X.; Zhang, L.-P.; Chai, H.; Liang, X.-F. Deep learning-based ovarian cancer subtypes identification using multi-omics data. BioData Min. 2020, 13, 10. [Google Scholar] [CrossRef]
- Kasture, K.R.; Shah, D.D.; Matte, P.N. Research Article A New Deep Learning method for Automatic Ovarian Cancer Prediction & Subtype classification. Turk. J. Comput. Math. Educ. Res. Artic. 2021, 12, 1233–1242. [Google Scholar]
- Kavitha, S.; Vidyaathulasiraman. Identification and classification of early stage Ovarian cancer using convolutional neural network. Ilkogr. Online-Elem. Educ. Online 2021, 20, 1908–1924. [Google Scholar] [CrossRef]
- Ghoniem, R.M.; Algarni, A.D.; Refky, B.; Ewees, A.A. Multi-Modal Evolutionary Deep Learning Model for Ovarian Cancer Diagnosis. Symmetry 2021, 13, 643. [Google Scholar] [CrossRef]
- Xiao, Y.; Bi, M.; Guo, H.; Li, M. Multi-omics approaches for biomarker discovery in early ovarian cancer diagnosis. Ebiomedicine 2022, 79, 104001. [Google Scholar] [CrossRef] [PubMed]
- Woman Ovarian Cancer Figure. Available online: https://ars.els-cdn.com/content/image/1-s2.0-S2352396422001852-gr1_lrg.jpg (accessed on 13 April 2023).
- Machine Learning Technology for Biomarker Development Figure. Available online: https://ars.els-cdn.com/content/image/1-s2.0-S2352396422001852-gr2.jpg (accessed on 13 April 2023).
- Arezzo, F.; Loizzi, V.; La Forgia, D.; Moschetta, M.; Tagliafico, A.S.; Cataldo, V.; Kawosha, A.A.; Venerito, V.; Cazzato, G.; Ingravallo, G.; et al. Radiomics Analysis in Ovarian Cancer: A Narrative Review. Appl. Sci. 2021, 11, 7833. [Google Scholar] [CrossRef]
- Arezzo, F.; Cormio, G.; La Forgia, D.; Kawosha, A.A.; Mongelli, M.; Putino, C.; Silvestris, E.; Oreste, D.; Lombardi, C.; Cazzato, G.; et al. The Application of Sonovaginography for Implementing Ultrasound Assessment of Endometriosis and Other Gynaecological Diseases. Diagnostics 2022, 12, 820. [Google Scholar] [CrossRef]
- Reilly, G.P.; Dunton, C.J.; Bullock, R.G.; Ure, D.R.; Fritsche, H.; Ghosh, S.; Pappas, T.C.; Phan, R.T. Validation of a deep neural network-based algorithm supporting clinical management of adnexal mass. Front. Med. 2023, 10, 1–11. [Google Scholar] [CrossRef] [PubMed]
- Elyan, E.; Vuttipittayamongkol, P.; Johnston, P.; Martin, K.; McPherson, K.; Moreno-García, C.F.; Jayne, C.; Sarker, M.K. Computer vision and machine learning for medical image analysis: Recent advances, challenges, and way forward. Artif. Intell. Surg. 2022, 2, 24–45. [Google Scholar] [CrossRef]
- Gumbs, A.A.; Frigerio, I.; Spolverato, G.; Croner, R.; Illanes, A.; Chouillard, E.; Elyan, E. Artificial Intelligence Surgery: How Do We Get to Autonomous Actions in Surgery? Sensors 2021, 21, 5526. [Google Scholar] [CrossRef]
- Sarvamangala, D.R.; Kulkarni, R.V. Convolutional neural networks in medical image understanding: A survey. Evol. Intell. 2022, 15, 1–22. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).