Next Article in Journal
Serous Carcinoma of the Endometrium with Mesonephric-Like Differentiation Initially Misdiagnosed as Uterine Mesonephric-Like Adenocarcinoma: A Case Report with Emphasis on the Immunostaining and the Identification of Splice Site TP53 Mutation
Previous Article in Journal
Magnetic Resonance-Guided High-Intensity Focused Ultrasound Ablation of Uterine Fibroids—Efficiency Assessment with the Use of Dynamic Contrast-Enhanced Magnetic Resonance Imaging and the Potential Role of the Administration of Uterotonic Drugs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using Dynamic Features for Automatic Cervical Precancer Detection

by
Roser Viñals
1,*,
Pierre Vassilakos
2,*,
Mohammad Saeed Rad
1,
Manuela Undurraga
2,
Patrick Petignat
2 and
Jean-Philippe Thiran
1
1
Signal Processing Laboratory (LTS5), École Polytechnique Fédérale de Lausanne (EPFL), 1015 Lausanne, Switzerland
2
Department of Pediatrics, Gynecology and Obstetrics, Geneva University Hospitals, Boulevard de la Cluse 30, 1205 Geneva, Switzerland
*
Authors to whom correspondence should be addressed.
Diagnostics 2021, 11(4), 716; https://doi.org/10.3390/diagnostics11040716
Submission received: 15 March 2021 / Revised: 7 April 2021 / Accepted: 15 April 2021 / Published: 17 April 2021
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
Cervical cancer remains a major public health concern in developing countries due to financial and human resource constraints. Visual inspection with acetic acid (VIA) of the cervix was widely promoted and routinely used as a low-cost primary screening test in low- and middle-income countries. It can be performed by a variety of health workers and the result is immediate. VIA provides a transient whitening effect which appears and disappears differently in precancerous and cancerous lesions, as compared to benign conditions. Colposcopes are often used during VIA to magnify the view of the cervix and allow clinicians to visually assess it. However, this assessment is generally subjective and unreliable even for experienced clinicians. Computer-aided techniques may improve the accuracy of VIA diagnosis and be an important determinant in the promotion of cervical cancer screening. This work proposes a smartphone-based solution that automatically detects cervical precancer from the dynamic features extracted from videos taken during VIA. The proposed solution achieves a sensitivity and specificity of 0.9 and 0.87 respectively, and could be a solution for screening in countries that suffer from the lack of expensive tools such as colposcopes and well-trained clinicians.

1. Introduction

In 2020, cervical cancer, a largely avoidable disease, affected approximately 600,000 women worldwide and resulted in more than 340,000 deaths [1]. The World Health Organization (WHO) emphasized the importance of acting immediately to combat cervical cancer through a triple intervention strategy that should be reached by the year 2030, including (i) 90% of girls fully vaccinated by 15 years of age, (ii) 70% of women screened with a high-performance test twice in their lifetime, and (iii) 90% of women identified with cervical disease receive treatment and care.
Cervical cancer is caused by persistent high-risk human papillomavirus (HPV) infection and develops through precursor precancerous lesions named cervical intraepithelial neoplasia grade 2 (CIN2) and grade 3 (CIN3). As far as it concerns screening, WHO recommended, for low- and middle-income countries (LMICs), HPV testing as a primary screening test [2]. Nevertheless, screening with HPV testing alone could result in unnecessary treatment due to its limited positive predictive value and specificity. Therefore, to identify women requiring treatment, visual inspection with acetic acid (VIA) of HPV-positive women is recommended by WHO. VIA has been used for several years in LMICs as a primary screening test, whereby the appearance of aceto-white areas after the application of a 3–5% acetic acid solution helps to define the pathological areas of the cervix and permits their immediate treatment. For countries where HPV testing is not affordable, VIA is still recommended by WHO as a stand-alone screening modality in its “see and treat” strategy [2]. However, VIA is a highly subjective procedure with a diagnostic accuracy for detecting precancer and cancer, varying from setting to setting [3,4,5,6]. Consequently, tools for a reliable noninvasive detection and characterization of neoplastic lesions based on quantitative diagnostic algorithms are desirable to assist front-line providers when performing VIA, especially in low-resource settings.
There are several techniques for computer-aided diagnosis of cervical cancer [7,8,9,10]. Most of them use cytological images [11,12,13,14,15,16,17]. However, an increasing number of studies are developing methods for automatic classification of images captured during VIA, and often, adding images taken during the visual inspection with Lugol’s iodine (VILI) or with the green lens. Particularly, the authors of [18,19,20,21] apply convolutional neural networks (CNN) to images captured with colposcopes.
Nevertheless, standard colposcopes are rarely available for screening in low-income countries due to limited financial resources, health manpower, and facilities. Consequently, recent studies are using other acquisition devices. For instance, a low-cost and portable colposcope has been developed and used to acquire images during VIA and VILI [22,23]. From these images, textural-based features are extracted and used in a support vector machine model. Another example is a colposcope based on a smartphone with optical lenses attached that incorporates an Artificial Intelligence (AI) classifier [24,25]. The smartphone-based system was evaluated in [26], after highly selecting the images and using gynecologic oncologists’ impressions as a reference standard, showing a large variable diagnosis among experts and demonstrating the potential of smartphones as an aiding tool for VIA. In [27], the authors used images taken during VIA with a fixed-focus camera. Finally, authors in [28] developed a digital video colposcope with an optical head that evaluates the effect of aceto-whitening based on the evolution of the diffuse reflectance with respect to time. As a result, it provides a dynamic spectral imaging (DSI) map that has proven to be useful to assist biopsy of suspicious lesions and distinguish high-grade CIN [29,30,31,32].
The approaches described are based on the analysis of a single static image taken during VIA and, often, combining additional images taken during VILI or with the green lens. By contrast, authors in [33] used five colposcopic images taken during VIA and used a graph convolutional network to classify them. Similarly, authors in [34] used multiple images taken during VIA, one image during VILI, and one taken with the green colposcope lens.
In this project, instead of using a colposcope and static images, we use dynamic images recorded via a smartphone during VIA [35]. It is our conviction that the different evolution of the color in the images of neoplastic and healthy tissue after application of acetic acid is a key element for detecting cervical precancer and cancer. Furthermore, the minimal training required for using our diagnosis system, and the increasing penetration rate of smartphones in LMICs, makes it ideal for countries with limited healthcare resources, where it can be easily integrated into a single-visit approach [36].
The rest of the paper is organized as follows. Section 2 presents the dataset and the proposed classification algorithm. In Section 3, the performance of the proposed scheme is analyzed by comparing it with the histology results, the colposcopists’ diagnosis, and the colposcopists’ annotations after VIA. Finally, Section 4 presents an evaluation of the results and Section 5 concludes the paper.

2. Materials and Methods

2.1. Dynamic Image Dataset

The appearance of aceto-white areas after the application of an acetic acid solution helps to define the pathological areas of the cervix. The CIN2 and CIN3 precancerous lesions (collectively referred to as CIN2+) tend to become whiter just after the application of the acetic acid, and then the whiteness decreases smoothly. By contrast, in the non-neoplastic tissue, the whiteness remains more constant. Most lesions can be distinguished one minute after the application of acetic acid, although it is reasonable to do VIA up to 3 min [37]. In this work, the temporal evolution of neoplastic and non-neoplastic areas will be the key feature used for its classification.

2.1.1. Smartphone as an Acquisition Device

To acquire the dynamic images, we use a Samsung Galaxy S5 (Samsung, Seoul, South Korea) with a camera of 16 megapixels, that record 2 min videos during VIA at 1 frame per second (fps). Patients are placed in the lithotomy position and a speculum is inserted and adjusted to provide more visibility. The smartphone is fixed to a tripod to minimize the movements during recording and placed around 15 cm of the vaginal vestibule. The recording is started immediately after the application of 3% acetic acid.
Our smartphone-based artificial intelligence classifier is intended to be introduced in LMICs using primary HPV screening in a “screen and treat” approach. The mobile app will be used as a triaging method to refer women testing positive for HPV for treatment. Therefore, our dataset used to train the algorithm is derived from a selected group of HPV-positive women aged between 30 to 49 years, which is the age group recommended by WHO for screening asymptomatic women in LMICs [2].
Video sequences were obtained after informed and signed consent from 44 asymptomatic HPV-positive women at the colposcopy consultation in Maternity of Geneva (Geneva University Hospitals), Switzerland, and at the Gynecologic Department of the District Hospital of Dschang, Cameroon. The cases from Geneva were referred for colposcopy after positive cytology and HPV testing, whereas Cameroonian cases were women screened with HPV testing in a “test-triage-and-treat program”. All HPV-positive women received VIA for triage and biopsies, endocervical brushing (ECB), and cytology as a quality control measure [36].
Cervical tissue obtained from biopsies, ECB, or conization served as the reference standard to evaluate the accuracy of the algorithm in this study. Adjudicated histopathology diagnosis was made after preparation of slides in the Division of Pathology (Geneva University Hospitals). Guided biopsies were performed on all visible lesions. If no lesion was seen, a random biopsy at 6 o’clock within the transformation zone and near the squamocolumnar junction was obtained. ECB was performed on all women, with or without visible lesions. The adjudicated diagnoses of 44 cases revealed 15 of them as negative (including 3 CIN1 cases) and 29 as positive (11 CIN2 and 18 CIN3 cases).
Three expert colposcopists (gynecologic oncologists) from the Geneva University Hospitals analyzed the 44 recorded videos, blinded from the histopathologic diagnoses, and were asked to (i) classify them as positive (CIN2+) or negative and (ii) draw the neoplastic lesions on the last frames of the videos.
Note that more sequences had been acquired but some of them were discarded due to severe movement and blurriness. The dataset specifications are shown in Table 1. Due to the limited amount of data, the classification is performed pixel-wise, classifying each pixel independently. Leave-one-out cross-validation was applied at the patient level, i.e., the video’s pixels of the patient being diagnosed were not used to train the model for that patient.

2.1.2. Labels

Each acquisition is labelled negative or positive based on histologic assessment, and, when positive, the neoplastic area (CIN2+) is delineated manually in the last frame of the dynamic image by an expert colposcopist. Two challenges are faced here. First, the selected area may be inaccurate for the pixels close to the boundaries. Second, movement during acquisitions causes variations of the acetowhite region’s position from frame to frame.
To palliate these two issues, the ground truth labels are eroded and dilated using a disk as a morphological operator whose size is manually set based on the size of the positive region. Pixels for training the neural network are randomly selected from the modified labels: positive pixels from the eroded label and negative pixels from the dilated, ensuring that none of the pixels selected are close to the boundaries between neoplastic (CIN2+) and non-neoplastic regions and the image borders. Figure 1 shows a label drawn by a colposcopist after histological confirmation of a CIN2+ lesion, and the resulting eroded and dilated labels. Note that the area where negative pixels are chosen was obtained not only by dilating the original label but also masking it to avoid regions far from the center of the cervix. From the 44 recorded videos, 21,851 positive pixels and 93,725 negative pixels were selected for training.

2.2. Classifier

As mentioned previously, this work is based on the analysis of dynamic image sequences of the cervix under a contrast agent to observe the reaction of cells with acetic acid (VIA) during a defined time window, and regarding the color evolution of each region, deciding whether that region is suspicious or not. Therefore, the intensity curves of the pixels of the video are the key elements of our prediction.
To do so, as shown in Figure 2, we propose a pipeline to process each dynamic image. The different components of the pipeline are described in this section.

2.2.1. Preprocessing

From each patient, a two minute Red Green Blue (RGB) video was captured, and it was sampled to obtain 120 frames. Motion compensation based on the point feature matching technique [38] was applied, ensuring the method to be robust to small movements of the camera. The frames are cropped to center the cervix and spatially resized to have a resolution of 150 × 150 pixels. An example of the effect of motion compensation is shown in Figure 3 where the ground truth labels, drawn by the expert colposcopist in the last frame of the video, are plotted in the first, middle, and last frames of the recording. When considering the frames recorded without applying motion compensation, the region bounded by the colposcopist’s label changes considerably from frame-to-frame, while when stabilizing the video, the defined region remains more constant.
After stabilizing the video, each color channel was scaled to the range [−0.5, 0.5]. To reduce the complexity of the problem, we used principal component analysis (PCA) [39]. PCA is defined as an orthogonal linear transformation that transforms the data, an RGB vector in this case, to a new coordinate system with a reduced dimensionality. The RGB values of a given positive pixel for each frame of the video create an elongated cluster which can be represented by a line. This line, called principal axis and represented in Figure 4a, is estimated to keep the variance at its maximum. Once this axis is obtained, the input data is projected to convert a three-channel pixel to a scalar, while keeping the color evolution still visible (Figure 4b). Note that the principal axis is not the same for every sequence. However, as their variances are not significant, we used the average line of principal axes of our training set.
Once the RGB pixel values are converted to a scalar, the first ten frames are discarded since they are the most affected by movement. The sequence is then downsampled to 11 time points using cubic spline interpolation, which are the final input to the neural net. Recall that the classification is performed pixel-wise. Thus, each pixel with its corresponding 11 time points is an independent input to the neural network.
For the training data, the positive and negative pixels were selected, as explained in Section 2.1.1. Two additional steps were applied to the pixels selected for training. First, pixels that contain reflections were identified by thresholding the amplitude of the maximum of its time evolution. Figure 5c shows the detected reflections for a specific sequence. Second, data augmentation by scaling was applied to positive pixels, ensuring an equal number of positive and negative pixels for training.

2.2.2. Neural Network Architecture

The classification was performed by using an artificial neural network (ANN), which has been proven to have the capability to model complex relationships between inputs and outputs to find patterns in data. We trained our network in a supervised manner, and we attempted to learn a function that maps an input to an output based on example input–output pairs, with the input being the intensity curve of a pixel. The neural network applied contains three layers: one input layer, a hidden layer with 15 nodes, and an output layer.
Leave-one-out cross-validation at the patient level was applied: when testing the algorithm for a specific patient, its pixels are removed from the training set and are not used for training or validation.

2.2.3. Postprocessing

As an output of the ANN, the probability of each pixel of being precancerous is given, which can be combined as shown in Figure 5b, resulting in a probability map. Reflections were detected as in the preprocessing of training pixels (Figure 5c), and the probability was set to 0.5 (Figure 5d). The probability map was then multiplied by a distance map with maximum and minimum of 1 and 0.5 (Figure 5e). The distance map applied intends to emphasize that precancerous regions tend to appear in the central regions of the cervix. The final probability map is shown Figure 5f.
From the resulting probability map, region growing segmentation was applied as follows:
  • The first seed is randomly selected from the highest values of the modified probability map and added to the region.
  • The pixel’s four adjacent neighbors are separately analyzed. Any neighbor is considered to lie on the affected region if they satisfy two criteria:
    • Homogeneity criterion: the difference between the seed probability and the neighbor should be less than a fixed threshold.
    • Minimum probability criterion: the probability of the neighbor should be above a threshold.
  • The newly added pixels are compared to their own neighbors under the same criteria.
  • Steps 2 and 3 are repeated until the criteria are not met for any neighbor of the pixels lying on the region.
  • A new seed is selected such that it has the highest probability and has not been identified as part of the region before.
  • The procedure is repeated until the predefined maximum number of seeds has been distributed.
Following the region growing segmentation, a closing operation was performed, which is dilation followed by erosion with the same kernel, and the contours of the found regions were obtained. An example of the results of the region growing segmentation and closing operation is shown in Figure 5g,h, respectively. Finally, a threshold was applied to the size of the lesions detected, deciding whether they are neoplastic (positive) or not.

3. Results

In this section, we examine the results and performance of our classifier. Since the number of sequences is low, leave-one-out cross-validation was applied at the patient level. To guarantee class balance, the same number of negative pixels as total positive pixels available for training were randomly selected. The parameters used are specified in Table 2.
The classifier results will be compared to (i) histologic assessment and colposcopists’ diagnosis, and (ii) colposcopists’ annotations after VIA. Figure 6 shows the algorithm results for five examples of the dataset.

3.1. Comparision of the Final Sequence Classification, Colposcopists’ Classification, and Histology Results

In this section, the results of our model are compared to the output of the histologic assessment (negative or positive for CIN2+), considered as the reference standard. From the 44 sequences classified, 39 were classified correctly as positive or negative, giving rise to an accuracy of 0.89. The precision, sensitivity, and specificity achieved were 0.93, 0.90, and 0.87, respectively (Table 3).
The 44 sequences of the database have been shown to three colposcopists and classified as positive or negative, giving rise to the results shown in Table 3. None of the colposcopists reached a better performance than the algorithm, achieving, on average, an accuracy, precision, sensitivity, and specificity of 0.71, 0.85, 0.68, and 0.78, respectively. The Cohen’s kappa coefficient between the three colposcopists and the proposed method has been computed, which varies within the range from 0.25 to 0.62. Consequently, there is a fair to substantial agreement between the algorithm and the three expert colposcopists.
The 5 sequences wrongly classified by the algorithm are shown in Figure 7. The first two are false positives: (a) an ectropion associated with inflammation and (b) an area of metaplasia. Case (a) has also been misclassified by one of the colposopists. The last three are false negative, (c), (d), and (e). For these cases, the transformation zone (TZ), i.e., the area that almost all manifestations of cervical carcinogenesis occur, is not fully visible and extends into the endocervical canal. In a case where precancerous lesions accompany TZ into the endocervical canal, visual inspection may be negative. For this reason, (c), (d), and (e) have also been misclassified by the colposcopists. Nevertheless, note that our model has found lesions similar to the ground truth in (d) and (e), but due to their small size, it has discarded them as negative.

3.2. Comparision of the Predicted Lesions by the Algorithm and Annotations by Colposcopists during VIA

VIA presents a high observer variability among colposcopists. Therefore, the annotations drawn in the last frames of our sequences are subjective, a fact highlighted in Figure 8, where the annotations given by the three colposcopists for five positive patients and the resulting predictions are represented. Note that in Figure 8a–d, colposcopists even disagree if lesions are placed in the two sides of the cervix or not, and in all cases, the annotations are significantly different.
We have extracted the positive and negative pixels for training the annotations given by the first colposcopist (represented in green) after erosion and dilation, but we have estimated the performance of the lesion positioning of our model by comparing the results to three available annotations. Note that the inaccuracy of the annotations is aggravated by the movement, especially present in the borders of the lesions.
Each preprocessed dynamic image contains 150 × 150 pixels which are independently classified. On the one hand, the corresponding annotation indicates us as to whether a pixel lies on a neoplastic lesion or not. On the other hand, the proposed method, after the region growing and closing operation, classifies every pixel as belonging to a neoplastic lesion or not. By comparing both, we can compute for each image the accuracy, precision, sensitivity, and specificity achieved by the classification of its pixels. Table 4 presents the results after averaging among images and comparing them to the three available annotations. The algorithm achieved, averaging the results for the three colposcopists’ annotations, a mean accuracy, precision, sensitivity, and specificity of 0.92, 0.53, 0.48, and 0.96, respectively. To evaluate the overlapping between the ground truth labels and the predicted ones, the Intersection over Union (IoU) has been computed, which is, on average, 0.32.

4. Discussion

In this study, we have developed and validated a smartphone-based solution that automatically detects cervical precancer from videos taken during VIA, intended to be used in sub-Saharan Africa as a triage method of HPV-positive women. We have measured the diagnostic accuracy of the model for detecting CIN2+ lesions and compared with the diagnostic accuracy of gynecologic oncologists. The proposed method has achieved an accuracy, sensitivity, and specificity of 0.89, 0.9, and 0.87, using histologic assessment as a reference standard.
The proposed algorithm results in a significant improvement compared to the reported sensitivities and specificities of VIA, which vary between 0.25 and 0.80, and between 0.40 and 0.90 among studies [3,4,40,41]. Furthermore, in our study, we asked three expert colposcopists blinded from the histopathologic diagnoses to classify the dynamic images from our database. The algorithm outperformed all of three of the experts, achieving a mean improvement of accuracy, sensitivity, and specificity of 17.42%, 21.84%, and 8.89%. The Cohen’s kappa coefficient, averaging among colposcopists, is 0.47, which implies that there is a moderate agreement between the proposed algorithm and the experts.
Tools using AI for cervical cancer screening have shown impressive effectiveness in automatic classification of cervical neoplastic lesions, as shown in Table 5, where some of the latest results in the literature are presented. It is important to emphasize that our algorithm presents comparable results to these state-of-the-art computer-aided diagnosis techniques, but with a much simpler setting: only VIA images are used and acquired with a smartphone, without any additional accessory. The other approaches in Table 5 require the use of colposcopes, which are rarely available in low-income countries, and some of them require a more sophisticated setup, acquiring images not only during VIA, but also during VILI and with the green lens.
Furthermore, in this work, we have been severely limited by the number of available sequences. To overcome this limitation, the inputs to the neural network are pixels’ time evolutions and leave-one-out cross-validation, at the patient level, has been applied. Nevertheless, the pixel classification approach poses several challenges. First, the selection of positive and negative pixels relies on very subjective colposcopists’ annotations and they are affected by movement. The subjectivity of the annotations has been demonstrated by comparing the annotations of the three colposcopists. We have palliated these issues by (i) applying motion compensation, (ii) selecting the pixels after eroding and dilating the annotations, and (iii) post-processing together the neural network’s output of all the pixels of one dynamic image’s pixels, ensuring homogeneity between neighboring pixels. Second, while the patient classification can be objectively evaluated by comparing it to the histological assessment, the colposcopists’ annotations are not an objective ground truth of the positions of the lesions. To have an estimation of the performance of the positioning, the predicted lesions have been compared with three colposcopists’ annotations. Third, specular reflections during recording hinder the classification of the pixels. For training, our method discards the pixels containing reflections, which are automatically detected. When predicting an image, the probability of the pixels containing reflections is set to 0.5, before starting the postprocessing of the image.
Our model failed to classify 5 sequences out of 44: in 3 of them, TZ was not fully visible and extended into the endocervical canal. The vast majority of precancers occur in the TZ. Therefore, if TZ lies in the endocervical canal, precancerous lesions may not be visible. For this reason, these three cases have been misclassified by colposcopists and by the algorithm. A cervical ectropion associated with inflammation and a case of metaplasia were misclassified as positive. Benign lesions such as inflammations, metaplasia, leukoplakia, condyloma, and CIN1 will be added as an additional class in our model in the near future. Our goal is implementing an algorithm able to distinguish between neoplastic lesions, benign lesions, and normal epithelium.
Among the different advantages of our screening tool, the minimal training required for using it and the increasing penetration rate of smartphones in LMICs make it ideal for countries with limited healthcare resources, where it can be easily integrated into a single-visit approach with the aim of reducing the preventable burden of morbidity, mortality, disability, and of reducing inequality.
Because this is a proof-of-principle investigation, it has limitations. Our selected dataset of video sequences contains a large number of positive cases, which does not correspond to the CIN2+ prevalence of a real-world setting in sub-Saharan Africa. Although these results are promising, further work is needed to improve the performance of the algorithm. First, we will increase the amount of training data and secondly, the upgraded algorithm will be validated in Cameroon in a real-life diagnosis process.

5. Conclusions

This work presented a smartphone-based algorithm to detect precancerous lesions from videos taken during VIA. Using the histological assessment as a gold standard, the proposed model achieved an accuracy, sensitivity, and specificity of 0.89, 0.9, and 0.87. Furthermore, the algorithm was compared to three gynecologic oncologists, outperforming all of them. The proposed model could create a new opportunity to facilitate the triage of HPV-positive women in LMICs.

Author Contributions

Conceptualization, R.V., P.V., M.S.R., M.U., P.P. and J.-P.T.; methodology, R.V., M.S.R. and J.-P.T.; software, R.V. and M.S.R.; validation, R.V., P.V., M.S.R., M.U., P.P. and J.-P.T.; formal analysis, P.V. and R.V.; investigation, R.V., P.V., M.S.R., M.U., P.P. and J.-P.T.; resources, R.V., P.V., M.S.R., M.U., P.P. and J.-P.T.; writing—original draft preparation, R.V., P.V., M.S.R., M.U., P.P. and J.-P.T.; writing—review and editing, R.V., P.V., M.S.R., M.U., P.P. and J.-P.T.; supervision, P.V., P.P. and J.-P.T.; project administration, P.V., P.P. and J.-P.T.; funding acquisition, P.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research has been supported by the University Hospitals of Geneva (Commission des Affaires Humanitaires (CAH)), Solidarité Internationale Genève, ESTHER Switzerland Partnership Project Grant (17G1), District Hospital of Dschang, and École Polytechnique Fédérale de Lausanne.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki (DoH), the Essentials of Good Epidemiological Practice (EGEP) issued by Public Health Switzerland, the International Conference on Harmonisation for Good Clinical Practice (ICH-GCP), as well as the Swiss law and regulatory authorities’ requirements, as applicable. The study has been nested in a large prospective trial named 3T-study in West Cameroon, registered in ClinicalTrials.gov (NCT03757299, October 2018). The 3T-study in Cameroon also obtained approval from the Cantonal Ethics Board of Geneva, Switzerland (Commission cantonale d’éthique de la recherche, CCER, N°2017-01110), and the National Ethics Committee for Research on Human Health (Comité national d’éthique de la recherche pour la santé humaine, CNERSH, N°2018/07/1083/CE/CNERSH/SP). The use of video sequences after VIA obtained from patients at the HUG colposcopy clinic has also been approved by the CCER (N°2020-00868).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study. Written informed consent has been obtained from the patient(s) to publish this paper.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy restrictions.

Acknowledgments

The authors would like to thank all the health staff who helped us to record the videos in Geneva and in Cameroon and the Plateforme de recherche pédiatrique of the University Hospitals of Geneva.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Sung, H.; Ferlay, J.; Siegel, R.L.; Laversanne, M.; Soerjomataram, I.; Jemal, A.; Bray, F. Global cancer statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J. Clin. 2021, caac.21660. [Google Scholar] [CrossRef] [PubMed]
  2. World Health Organization. WHO Guidelines for Screening and Treatment of Precancerous Lesions for Cervical Cancer Prevention; WHO: Geneva, Switzerland, 2013. [Google Scholar]
  3. Bigoni, J.; Gundar, M.; Tebeu, P.M.; Bongoe, A.; Schäfer, S.; Fokom-Domgue, J.; Catarino, R.; Tincho, E.F.; Bougel, S.; Vassilakos, P.; et al. Cervical cancer screening in sub-Saharan Africa: A randomized trial of VIA versus cytology for triage of HPV-positive women. Int. J. Cancer 2015, 137, 127–134. [Google Scholar] [CrossRef] [PubMed]
  4. Gravitt, P.E.; Paul, P.; Katki, H.A.; Vendantham, H.; Ramakrishna, G.; Sudula, M.; Kalpana, B.; Ronnett, B.M.; Vijayaraghavan, K.; Shah, K.S. Effectiveness of VIA, pap, and HPV DNA testing in a cervical cancer screening program in a Peri-Urban community in Andhra Pradesh, India. PLoS ONE 2010, 5. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Ajenifuja, K.O.; Gage, J.C.; Adepiti, A.C.; Wentzensen, N.; Eklund, C.; Reilly, M.; Hutchinson, M.; Burk, R.D.; Schiffman, M. A population-based study of visual inspection with acetic acid (VIA) for cervical screening in rural nigeria. Int. J. Gynecol. Cancer 2013, 23, 507–512. [Google Scholar] [CrossRef] [Green Version]
  6. Etherington, I.J.; Luesley, D.M.; Shafi, M.I.; Dunn, J.; Hiller, L.; Jordan, J.A. Observer variability among colposcopists from the West Midlands region. BJOG Int. J. Obstet. Gynaecol. 1997, 104, 1380–1384. [Google Scholar] [CrossRef]
  7. Devi, M.A.; Ravi, S.; Vaishnavi, J.; Punitha, S. Classification of Cervical Cancer Using Artificial Neural Networks. Procedia Comput. Sci. 2016, 89, 465–472. [Google Scholar] [CrossRef] [Green Version]
  8. Fernandes, K.; Cardoso, J.S.; Fernandes, J. Automated Methods for the Decision Support of Cervical Cancer Screening Using Digital Colposcopies. IEEE Access 2018, 6, 33910–33927. [Google Scholar] [CrossRef]
  9. Jusman, Y.; Ng, S.C.; Abu Osman, N.A. Intelligent screening systems for cervical cancer. Sci. World J. 2014, 2014. [Google Scholar] [CrossRef] [Green Version]
  10. Chitra, B.; Kumar, S.S. Recent advancement in cervical cancer diagnosis for automated screening: A detailed review. J. Ambient. Intell. Humaniz. Comput. 2021, 1, 3. [Google Scholar] [CrossRef]
  11. Taha, B.; Dias, J.; Werghi, N. Classification of cervical-cancer using pap-smear images: A convolutional neural network approach. In Proceedings of the Medical Image Understanding and Analysis (MIUA), Edinburgh, UK, 11–13 July 2017. [Google Scholar]
  12. Mariarputham, E.J.; Stephen, A. Nominated texture based cervical cancer classification. Comput. Math. Methods Med. 2015, 2015. [Google Scholar] [CrossRef]
  13. Sokouti, B.; Haghipour, S.; Tabrizi, A.D. A framework for diagnosing cervical cancer disease based on feedforward MLP neural network and ThinPrep histopathological cell image features. Neural Comput. Appl. 2014, 24, 221–232. [Google Scholar] [CrossRef]
  14. Wu, M.; Yan, C.; Liu, H.; Liu, Q.; Yin, Y. Automatic classification of cervical cancer from cytological images by using convolutional neural network. Biosci. Rep. 2018, 38, 20181769. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Hyeon, J.; Choi, H.J.; Lee, B.D.; Lee, K.N. Diagnosing cervical cell images using pre-trained convolutional neural network as feature extractor. In Proceedings of the 2017 IEEE International Conference on Big Data and Smart Computing (BigComp), Jeju, Korea, 13–16 February 2017. [Google Scholar]
  16. Almubarak, H.A.; Stanley, R.J.; Long, R.; Antani, S.; Thoma, G.; Zuna, R.; Frazier, S.R. Convolutional Neural Network Based Localized Classification of Uterine Cervical Cancer Digital Histology Images. Procedia Comput. Sci. 2017, 114, 281–287. [Google Scholar] [CrossRef]
  17. Xiang, Y.; Sun, W.; Pan, C.; Yan, M.; Yin, Z.; Liang, Y. A novel automation-assisted cervical cancer reading method based on convolutional neural network. Biocybern. Biomed. Eng. 2020, 40, 611–623. [Google Scholar] [CrossRef] [Green Version]
  18. Zhang, B.; Zhang, Q.; Zhou, H.; Xia, C.; Wang, J. Automated Prediction of Cervical Precancer Based on Deep Learning. In Proceedings of the Chinese Intelligent Systems Conference 2020, Shenzhen, China, 24–25 October 2020. [Google Scholar]
  19. Miyagi, Y.; Takehara, K.; Miyake, T. Application of deep learning to the classification of uterine cervical squamous epithelial lesion from colposcopy images. Mol. Clin. Oncol. 2019, 11, 583–589. [Google Scholar] [CrossRef] [PubMed]
  20. World Health Organization; Buiu, C.; Dănăilă, V.R.; Răduţă, C.N.; Guo, P.; Singh, S.; Xue, Z.; Long, R.; Antani, S.; Das, A. World Health Organization; Buiu, C.; Dănăilă, V.R.; Răduţă, C.N.; Guo, P.; Singh, S.; Xue, Z.; Long, R.; Antani, S.; Das, A.; et al. Preprocessing for automating early detection of cervical cancer. In Proceedings of the 2019 IEEE EMBS International Conference on Biomedical and Health Informatics, Chicago, IL, USA, 19–22 May 2019; Volume 8, pp. 597–600. [Google Scholar]
  21. Buiu, C.; Dănăilă, V.R.; Răduţă, C.N. MobileNetV2 ensemble for cervical precancerous lesions classification. Processes 2020, 8, 595. [Google Scholar] [CrossRef]
  22. Asiedu, M.N.; Simhal, A.; Chaudhary, U.; Mueller, J.L.; Lam, C.T.; Schmitt, J.W.; Venegas, G.; Sapiro, G.; Ramanujam, N. Development of Algorithms for Automated Detection of Cervical Pre-Cancers with a Low-Cost, Point-of-Care, Pocket Colposcope. IEEE Trans. Biomed. Eng. 2019, 66, 2306–2318. [Google Scholar] [CrossRef]
  23. Asiedu, M.N.; Skerrett, E.; Sapiro, G.; Ramanujam, N. Combining multiple contrasts for improving machine learning-based classification of cervical cancers with a low-cost point-of-care Pocket colposcope. In Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society 2020, Montreal, QC, Canada, 20–24 July 2020; pp. 1148–1151. [Google Scholar]
  24. Matti, R.; Gupta, V.; Sa, D.; Sebag, C.; Peterson, C.W.; Levitz, D. Introduction of mobile colposcopy as a primary screening tool for different socioeconomic populations in urban India. Pan Asian J. Obs. Gyn. 2019, 2, 4–11. [Google Scholar]
  25. Peterson, C.; Rose, D.; Mink, J.; Levitz, D. Real-Time Monitoring and Evaluation of a Visual-Based Cervical Cancer Screening Program Using a Decision Support Job Aid. Diagnostics 2016, 6, 20. [Google Scholar] [CrossRef] [Green Version]
  26. Xue, Z.; Novetsky, A.P.; Einstein, M.H.; Marcus, J.Z.; Befano, B.; Guo, P.; Demarco, M.; Wentzensen, N.; Long, L.R.; Schiffman, M.; et al. A demonstration of automated visual evaluation of cervical images taken with a smartphone camera. Int. J. Cancer 2020, 147, 2416–2423. [Google Scholar] [CrossRef]
  27. Hu, L.; Bell, D.; Antani, S.; Xue, Z.; Yu, K.; Horning, M.P.; Gachuhi, N.; Wilson, B.; Jaiswal, M.S.; Befano, B.; et al. An Observational Study of Deep Learning and Automated Evaluation of Cervical Images for Cancer Screening. J. Natl. Cancer Inst. 2019, 111, 923–932. [Google Scholar] [CrossRef]
  28. Soutter, W.P.; Diakomanolis, E.; Lyons, D.; Ghaem-Maghami, S.; Ajala, T.; Haidopoulos, D.; Doumplis, D.; Kalpaktsoglou, C.; Sakellaropoulos, G.; Soliman, S.; et al. Dynamic spectral imaging: Improving colposcopy. Clin. Cancer Res. 2009, 15, 1814–1820. [Google Scholar] [CrossRef] [Green Version]
  29. Harris, K.E.; Lavin, P.T.; Akin, M.D.; Papagiannakis, E.; Denardis, S. Rate of detecting CIN3+ among patients with ASC-US using digital colposcopy and dynamic spectral imaging. Oncol. Lett. 2020, 20, 17. [Google Scholar] [CrossRef] [PubMed]
  30. Kaufmann, A.; Founta, C.; Papagiannakis, E.; Naik, R.; Fisher, A. Standardized Digital Colposcopy with Dynamic Spectral Imaging for Conservative Patient Management. Case Rep. Obstet. Gynecol. 2017, 2017, 1–5. [Google Scholar] [CrossRef] [Green Version]
  31. Budithi, S.; Peevor, R.; Pugh, D.; Papagiannakis, E.; Durman, A.; Banu, N.; Alalade, A.; Leeson, S. Evaluating Colposcopy with Dynamic Spectral Imaging During Routine Practice at Five Colposcopy Clinics in Wales: Clinical Performance. Gynecol. Obstet. Investig. 2018, 83, 234–240. [Google Scholar] [CrossRef] [PubMed]
  32. Cholkeri-Singh, A.; Lavin, P.T.; Olson, C.G.; Papagiannakis, E.; Weinberg, L. Digital Colposcopy with Dynamic Spectral Imaging for Detection of Cervical Intraepithelial Neoplasia 2+ in Low-Grade Referrals: The IMPROVE-COLPO Study. J. Low. Genit. Tract Dis. 2018, 22, 21–26. [Google Scholar] [CrossRef] [PubMed]
  33. Li, Y.; Chen, J.; Xue, P.; Tang, C.; Chang, J.; Chu, C.; Ma, K.; Li, Q.; Zheng, Y.; Qiao, Y. Computer-Aided Cervical Cancer Diagnosis Using Time-Lapsed Colposcopic Images. IEEE Trans. Med. Imaging 2020, 39, 3403–3415. [Google Scholar] [CrossRef]
  34. Yue, Z.; Ding, S.; Zhao, W.; Wang, H.; Ma, J.; Zhang, Y.; Zhang, Y. Automatic CIN Grades Prediction of Sequential Cervigram Image Using LSTM with Multistate CNN Features. IEEE J. Biomed. Health Inform. 2020, 24, 844–854. [Google Scholar] [CrossRef]
  35. Gallay, C.; Girardet, A.; Viviano, M.; Catarino, R.; Benski, A.C.; Tran, P.L.; Ecabert, C.; Thiran, J.P.; Vassilakos, P.; Petignat, P. Cervical cancer screening in low-resource settings: A smartphone image application as an alternative to colposcopy. Int. J. Womens. Health 2017, 9, 455–461. [Google Scholar] [CrossRef] [Green Version]
  36. Levy, J.; de Preux, M.; Kenfack, B.; Sormani, J.; Catarino, R.; Tincho, E.F.; Frund, C.; Fouogue, J.T.; Vassilakos, P.; Petignat, P. Implementing the 3T-approach for cervical cancer screening in Cameroon: Preliminary results on program performance. Cancer Med. 2020, 9, 7293–7300. [Google Scholar] [CrossRef]
  37. Hilal, Z.; Tempfer, C.B.; Burgard, L.; Rehman, S.; Rezniczek, G.A. How long is too long? Application of acetic acid during colposcopy: A prospective study. Am. J. Obstet. Gynecol. 2020, 223, 101.e1–101.e8. [Google Scholar] [CrossRef] [PubMed]
  38. Abdullah, L.M.; Tahir, N.M.; Samad, M. Video stabilization based on point feature matching technique. In Proceedings of the 2012 IEEE Control and System Graduate Research Colloquium (ICSGRC), Shah Alam, Malaysia, 16–17 July 2012; pp. 303–307. [Google Scholar]
  39. Wold, S.; Esbensen, K.; Geladi, P. Principal component analysis. Chemom. Intell. Lab. Syst. 1987, 2, 37–52. [Google Scholar] [CrossRef]
  40. Tebeu, P.M.; Fokom-Domgue, J.; Crofts, V.; Flahaut, E.; Catarino, R.; Untiet, S.; Vassilakos, P.; Petignat, P. Effectiveness of a two-stage strategy with HPV testing followed by visual inspection with acetic acid for cervical cancer screening in a low-income setting. Int. J. Cancer 2015, 136, E743–E750. [Google Scholar] [CrossRef] [PubMed]
  41. Kunckler, M.; Schumacher, F.; Kenfack, B.; Catarino, R.; Viviano, M.; Tincho, E.; Tebeu, P.M.; Temogne, L.; Vassilakos, P.; Petignat, P. Cervical cancer screening in a low-resource setting: A pilot study on an HPV-based screen-and-treat approach. Cancer Med. 2017, 6, 1752–1761. [Google Scholar] [CrossRef]
  42. Kudva, V.; Prasad, K.; Guruvare, S. Hybrid Transfer Learning for Classification of Uterine Cervix Images for Cervical Cancer Screening. J. Digit. Imaging 2020, 33, 619–631. [Google Scholar] [CrossRef] [PubMed]
  43. Zhang, T.; Luo, Y.M.; Li, P.; Liu, P.Z.; Du, Y.Z.; Sun, P.; Dong, B.H.; Xue, H. Cervical precancerous lesions classification using pre-trained densely connected convolutional networks with colposcopy images. Biomed. Signal Process. Control 2020, 55, 101566. [Google Scholar] [CrossRef]
Figure 1. For the supervised learning, positive and negative pixels are chosen from area far from ground truth contours, to avoid mislabeled training data due to inaccurate labeling. From left to right: (a) Label drawn by the colposcopist, validated by biopsy, (b) eroded label, where positive pixels are chosen from (green region), (c) dilated and masked label, where negative pixels are chosen from (green region).
Figure 1. For the supervised learning, positive and negative pixels are chosen from area far from ground truth contours, to avoid mislabeled training data due to inaccurate labeling. From left to right: (a) Label drawn by the colposcopist, validated by biopsy, (b) eroded label, where positive pixels are chosen from (green region), (c) dilated and masked label, where negative pixels are chosen from (green region).
Diagnostics 11 00716 g001
Figure 2. Pipeline of the proposed approach. The features mentioned on the intensity curve, such as how fast it is becoming fast and level of whiteness, are only human interpretation of the curves and are not pre-computed; the proposed network only uses the raw data at its input.
Figure 2. Pipeline of the proposed approach. The features mentioned on the intensity curve, such as how fast it is becoming fast and level of whiteness, are only human interpretation of the curves and are not pre-computed; the proposed network only uses the raw data at its input.
Diagnostics 11 00716 g002
Figure 3. (a) Ground truth labels drawn on the first, one-minute, and last frames from a recording before motion compensation. (b) Ground truth labels drawn on the first, one-minute, and last frames from a recording after motion compensation.
Figure 3. (a) Ground truth labels drawn on the first, one-minute, and last frames from a recording before motion compensation. (b) Ground truth labels drawn on the first, one-minute, and last frames from a recording after motion compensation.
Diagnostics 11 00716 g003
Figure 4. (a) An example of Red Green Blue (RGB) values of a positive pixel, projected on the principal axis, to reduce the dimensionality of the input data. (b) Resulting video frame after combining the three color channels.
Figure 4. (a) An example of Red Green Blue (RGB) values of a positive pixel, projected on the principal axis, to reduce the dimensionality of the input data. (b) Resulting video frame after combining the three color channels.
Diagnostics 11 00716 g004
Figure 5. (a) Ground truth label drawn in the image corresponding to one minute, (b) probability map, (c) detected reflections, (d) probability map after identifying reflections, (e) distance map, (f) modified probability map after multiplying by the distance map, (g) region growing segmentation results, (h) resulting prediction after closing operation drawn in the image corresponding to one minute.
Figure 5. (a) Ground truth label drawn in the image corresponding to one minute, (b) probability map, (c) detected reflections, (d) probability map after identifying reflections, (e) distance map, (f) modified probability map after multiplying by the distance map, (g) region growing segmentation results, (h) resulting prediction after closing operation drawn in the image corresponding to one minute.
Diagnostics 11 00716 g005
Figure 6. Frames corresponding to one minute after acetic acid application with the ground truth labels, histology results, predicted labels by the algorithm, resulting classification, and colposcopists’ diagnosis.
Figure 6. Frames corresponding to one minute after acetic acid application with the ground truth labels, histology results, predicted labels by the algorithm, resulting classification, and colposcopists’ diagnosis.
Diagnostics 11 00716 g006
Figure 7. Frames corresponding to one minute after acetic acid application from dynamic images that have misclassified, with the ground truth labels, histology results, predicted labels, resulting classification, and colposcopists’ diagnosis.
Figure 7. Frames corresponding to one minute after acetic acid application from dynamic images that have misclassified, with the ground truth labels, histology results, predicted labels, resulting classification, and colposcopists’ diagnosis.
Diagnostics 11 00716 g007
Figure 8. Colposcopists’ annotations and algorithm’s predicted labels drawn in one-minute frames.
Figure 8. Colposcopists’ annotations and algorithm’s predicted labels drawn in one-minute frames.
Diagnostics 11 00716 g008
Table 1. Dataset specifications.
Table 1. Dataset specifications.
Dataset Specifications
Number of sequences available68
Total number of discarded sequences24
Missing histology results and VIA annotations2
Severe movement16
Blurriness2
Blood flow that prevents the visualization of the cervix tissue3
Excess of mucus that prevents the visualization of the cervix tissue1
Total number of sequences used44
Number of positive sequences29
CIN318
CIN211
Negative sequences15
CIN13
Negative12
Video length (seconds)120
Number of frames120
Number of selected positive pixels21,851
Number of selected negative pixels93,725
Table 2. Algorithm parameters.
Table 2. Algorithm parameters.
Preprocessing Parameters
Mean principal axis[0.3609, 0.5941, 0.7074]
Discarded number of framesFirst 10
Downsampling factor0.1
Number of input features ANN11
Range scaling coefficients used for data augmentation[0.9, 1.15]
Reflections’ threshold0.25
ANN Parameters
Number of nodes15
Number of hidden layers1
Postprocessing Parameters
Homogeneity criteria probability threshold0.27
Probability threshold0.50
Reflections’ threshold 0.25
Number of seeds5
Closing cluster kernel[7, 7]
Final size threshold450
Table 3. Algorithm and colposcopists’ classification results based on histologic assessment.
Table 3. Algorithm and colposcopists’ classification results based on histologic assessment.
AlgorithmColposcopist 1Colposcopist 2Colposcopist 3
True Positives26231422
True Negatives13121013
False Positives2352
False Negatives36157
Accuracy0.890.800.550.80
Precision0.930.880.740.92
Sensitivity0.900.790.480.76
Specificity0.870.800.670.87
Cohen’s kappa-0.620.250.53
Table 4. Results when comparing predicted lesions by the algorithm and visual inspection with acetic acid (VIA) annotations by colposcopists.
Table 4. Results when comparing predicted lesions by the algorithm and visual inspection with acetic acid (VIA) annotations by colposcopists.
Colposcopist 1Colposcopist 2Colposcopist 3
Accuracy0.92 ± 0.100.92 ± 0.080.91 ± 0.15
Precision0.57 ± 0.280.49 ± 0.290.55 ± 0.3
Sensitivity0.46 ± 0.250.49 ± 0.290.5 ± 0.3
Specificity0.97 ± 0.040.96 ± 0.050.96 ± 0.04
Intersection over Union (IoU)0.32 ± 0.190.32 ± 0.200.33 ± 0.19
Table 5. Comparison of the proposed and state-of-the-art methods published between 2019 and 2021.
Table 5. Comparison of the proposed and state-of-the-art methods published between 2019 and 2021.
AccuracySensitivitySpecificityAcquisition DeviceNumber of Images Captured during VIAAdditional Images UsedNumber of Patients
Asiedu et al. [22]0.80.810.786Portable colposcope1VILI134
Miyagi et al. [19]0.820.800.88Colposcope1None330
Yue et al. [34]0.960.950.98Colposcope5VILI and green lens679
Li et al. [33]0.780.78Not specifiedColposcope4Pre-acetic acid7668
Kudva et al. [42]0.940.910.89Colposcope1None2198
Zhang et al. [43]0.760.440.88Colposcope1None229
Proposed algorithm0.890.90.87Smartphone120None44
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Viñals, R.; Vassilakos, P.; Rad, M.S.; Undurraga, M.; Petignat, P.; Thiran, J.-P. Using Dynamic Features for Automatic Cervical Precancer Detection. Diagnostics 2021, 11, 716. https://doi.org/10.3390/diagnostics11040716

AMA Style

Viñals R, Vassilakos P, Rad MS, Undurraga M, Petignat P, Thiran J-P. Using Dynamic Features for Automatic Cervical Precancer Detection. Diagnostics. 2021; 11(4):716. https://doi.org/10.3390/diagnostics11040716

Chicago/Turabian Style

Viñals, Roser, Pierre Vassilakos, Mohammad Saeed Rad, Manuela Undurraga, Patrick Petignat, and Jean-Philippe Thiran. 2021. "Using Dynamic Features for Automatic Cervical Precancer Detection" Diagnostics 11, no. 4: 716. https://doi.org/10.3390/diagnostics11040716

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop