Next Article in Journal / Special Issue
Mathematical Model for Chemical Reactions in Electrolytes Applied to Cytochrome c Oxidase: An Electro-Osmotic Approach
Previous Article in Journal
Exploring Polygonal Number Sieves through Computational Triangulation
Previous Article in Special Issue
Global Dynamics of a Within-Host Model for Usutu Virus
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Reinforcement Learning for Efficient Digital Pap Smear Analysis

by
Carlos Macancela
1,*,
Manuel Eugenio Morocho-Cayamcela
1 and
Oscar Chang
2
1
School of Mathematical and Computational Sciences, Yachay Tech University, Ibarra 100115, Ecuador
2
Electronics and Control Department, School of Electrical Engineering, Faculty of Engineering Central University of Venezuela, Los Chaguaramos, Caracas 1050, Venezuela
*
Author to whom correspondence should be addressed.
Computation 2023, 11(12), 252; https://doi.org/10.3390/computation11120252
Submission received: 18 October 2023 / Revised: 19 November 2023 / Accepted: 1 December 2023 / Published: 10 December 2023
(This article belongs to the Special Issue 10th Anniversary of Computation—Computational Biology)

Abstract

:
In August 2020, the World Health Assembly launched a global initiative to eliminate cervical cancer by 2030, setting three primary targets. One key goal is to achieve a 70% screening coverage rate for cervical cancer, primarily relying on the precise analysis of Papanicolaou (Pap) or digital Pap smears. However, the responsibility of reviewing Pap smear samples to identify potentially cancerous cells primarily falls on pathologists—a task known to be exceptionally challenging and time-consuming. This paper proposes a solution to address the shortage of pathologists for cervical cancer screening. It leverages the OpenAI-GYM API to create a deep reinforcement learning environment utilizing liquid-based Pap smear images. By employing the Proximal Policy Optimization algorithm, autonomous agents navigate Pap smear images, identifying cells with the aid of rewards, penalties, and accumulated experiences. Furthermore, the use of a pre-trained convolutional neuronal network like Res-Net50 enhances the classification of detected cells based on their potential for malignancy. The ultimate goal of this study is to develop a highly efficient, automated Papanicolaou analysis system, ultimately reducing the need for human intervention in regions with limited pathologists.

1. Introduction

Cervical cancer is a dangerous disease affecting women worldwide, with over 500 million diagnosed cases and 300 million deaths each year [1]. The primary cause of cervical cancer is the human papillomavirus (HPV), which is transmitted through sexual contact. However, there are some preventative measures including HPV vaccination and methods for screening for cervical lesions including Pap smears and HPV tests [2]. The HPV test is designed to identify the presence of the HPV virus, whereas the Pap smear test involves collecting cells from the cervix and searching for abnormal cells that have the potential to develop into cervical cancer if left untreated. Although there is treatment, the percentage of deaths is not decreasing because most deaths occur in low- and middle-income countries such as those in Latin America, the Caribbean, or Africa [3,4]. According to a study carried out in Ecuador, only thirty-six percent of women have received coverage through the HPV vaccination program by 2020, and as of 2019, five out of ten women have been screened for cervical cancer within the past five years [5]. These less developed regions experiment with several social, economic, and cultural barriers that prevent women from participating in preventive programs, leading to a high risk of cervical cancer. Moreover, the lack of access to preventative measures is caused by health system barriers such as shortages of personnel, deficient health services, long waiting times, and lack of adequate instruments and equipment. Also, cultural barriers such as lack of sexual education and knowledge barriers such as the lack of information about preventive treatments contribute to low/medium coverage in screening for cervical cancer [6].
A survey conducted in 2009 among 81 women in both urban and rural areas of Ecuador highlighted dissatisfaction with the extended waiting periods for test results. This delay is due to a lack of appropriate instruments and a limited number of trained personnel and specialists, especially in rural areas [7]. Cervix screening analysis can only be carried out by a few specialists, including anatomic pathologists, cytologists, and clinical pathologists. This limited number of specialists makes it difficult to evaluate a large number of tests, as evidenced by a 2021 World Health Organization (WHO) report on Ecuador, which found a paucity of medical staff per 10,000 cancer patients, including 0 radiation oncologists, only 3 medical physicists, 154 radiologists, and 4 nuclear medicine physicians [5].
This paper focuses on overcoming barriers in the healthcare system, specifically the long waiting times that patients face when waiting for their test results. To effectively tackle and minimize these delays, the integration of artificial intelligence (AI) techniques becomes imperative. This paper covers the implementation of a deep reinforcement learning (DRL) approach to identify the cells in digital Pap smear tests [8]. To enhance the precision of cell classification in Pap smear tests, this research employs deep learning (DL) and transfer learning (TL) techniques, implementing a ResNet model. The proposed methodology involves the creation of a DRL environment designed to train an agent by utilizing a robust algorithm like Proximal Policy Optimization (PPO) [9]. The agent, representing a defined region of interest (ROI), navigates through digital Pap smear tests to locate and classify cells effectively. By successfully implementing this technology, there is a substantial potential to alleviate patient waiting times and enhance the overall efficiency of the healthcare system. This innovative approach not only addresses a specific pain point in the current system but also showcases the transformative power of AI in improving healthcare processes.

2. Methods and Materials

2.1. Proposed Approach

This study presents an automatic system for detecting cervical cancer, leveraging DRL and DL mechanisms, and a TL technique. The methodology involves a systematic workflow, see Figure 1. The initial step is to select the public dataset of LBC images of Pap smears from the Mendeley data repository. The environment’s architecture is established using the OpenAI-GYM API and the Pap smear images. To facilitate the learning process, a customized CNN is implemented to define the rewards and penalties within the environment. The training of agents is executed through the PPO algorithm, unfolding across three distinct stages. Only the most adept agents progress to subsequent stages, ensuring refined learning and heightened competence. To classify the identified cells in the Pap smear, a pre-trained CNN, specifically ResNet50, is employed. The training and validation of the model were performed considering 80% of the dataset and the remaining 20% for testing, with different metrics to evaluate the performance of the model and seamlessly integrate it into the final cervical cancer detection system.

2.2. Data Acquisition

This paper used two datasets. The first dataset includes 963 images of Pap tests that were collected from the Mendeley data repository. These samples were taken in 2019 from 460 patients at three Indian institutions: Babina Diagnostic Pvt. Ltd. (Mumbai, India), Dr. B. Borooah Cancer Research Institute (Guwahati, India), and Gauhati Medical College and Hospital (Guwahati, India). To achieve higher-quality images with clearer backgrounds, liquid-based cytology was used instead of the conventional method. The images were captured using a Leica DM 750 microscope that was connected to a specialized high-definition camera ICC50 HD, and the official computer software was utilized. The microscope’s magnification was set to 400×, resulting in high-definition JPG images with a resolution of 2048 × 1536 pixels [10].
Each image in the dataset was categorized by a pathologist into one of four categories, leading to the following classification, see Figure 2.
The second dataset was obtained from the bachelor’s thesis of Topapanta [11], who cropped the images from the first dataset to extract images with a size of 150 by 150 pixels that contains single cell and Pap smear background images. The dataset consists of 100 images that are classified into two categories: “Cells” and “No Cells”, see Figure 3 and Figure 4.

3. Agent Design

In reinforcement learning an agent refers to an intelligent entity that interacts with an environment in order to learn and optimize its decision-making process. It is capable of perceiving the state of the environment through observations and receives feedback in the form of rewards or penalties based on its actions [8]. This paper proposes an ROI that functions as an agent, see Figure 5. The ROI agent interacts with the environment by moving through the digital Pap smear performing some actions, see Table 1, and collecting a large amount of cells during an episode. The proposed ROI agent extends the capabilities of traditional digital Pap smear test analysis by introducing an intelligent entity that actively participates in the cell collection process. Through its interactions with the digital environment, the ROI agent aims to enhance the efficiency and effectiveness of the diagnostic procedure, ultimately leading to improved accuracy and reducing long waiting times for results.

4. Environment Design

The environment defines the rules, dynamics, and feedback mechanisms that govern the agent’s behavior and learning process. The environment consists of states, actions, rewards, and a transition model. States describe the current situation or configuration of the environment, while actions are the choices available to the agent to modify or interact with the environment. Additionally, the environment provides feedback to the agent in the form of rewards or penalties based on its actions. Rewards indicate the desirability or quality of the agent’s behavior, guiding it toward maximizing cumulative rewards over time. This provides the necessary stimuli and feedback for the agent to explore, learn, and improve its decision-making capabilities. The environment in this project follows the specifications outlined by GYM, which is a framework developed by OpenAI for creating and simulating interactive environments used in reinforcement learning (RL) tasks. It provides a standardized platform for designing RL experiments and allows researchers to define observation spaces, action spaces, rewards, episode termination conditions, and other environment-specific configurations [12]. Finally, the entire digital Pap smear acts as the environment in which the agent can perform actions to navigate around the image and locate the cells presented in the sample, see Figure 6.
The observation space refers to a key feature in reinforcement learning environments. It represents the state or observation of the environment at any given time, providing information that the agent can perceive. The observation space defines the type and structure of the data that the agent receives as input. It could be in the form of raw sensory data, such as images, see Figure 5. By accessing the observation space, the agent can gather information about the environment’s features, conditions, or objects necessary for decision-making. It allows the agent to perceive and understand the current state of the environment, which enables it to select appropriate actions to achieve its objectives [13].

4.1. Reward Signal

The reward signal is a crucial component of the learning process. It represents the feedback provided to an agent based on its actions in an environment. The reward signal serves as a measure of the desirability or quality of the agent’s behavior, guiding it toward learning optimal or desirable policies. Also, it is typically a numerical scalar value that the agent receives from the environment after each action it takes. Furthermore, the reward signal can be explicitly provided by the environment, or it can be implicitly defined through a reward function, which maps states and actions to rewards. Finally, the reward function encapsulates the designer’s intentions and objectives, shaping the agent’s learning process. The reward signal can be written as R and it was defined as follows.
R = r 1 + r 2 + r 3 = r 1 = 0.5 , if agent is searching for cells r 2 = 10 × n , if agent detect a cell r 3 = 10 , if agent detect the same cell multiple times
where n is the number of found cells.

4.2. Cell Recognition Model

The reward signal in the environment consists of predefined rewards and penalties that are determined based on whether the agent detects a cell or not. To ascertain that the agent is located over on a cell, a convolutional neural network (CNN) is employed. The CNN needs to provide a balance between small size and high accuracy, as computational efficiency is crucial during training. The neural network architecture begins with a convolutional layer that takes a 50 × 50 pixel image as input and applies 10 filters sized 2 × 2 . This is followed by a second convolutional layer with 32 filters of the same size. The final convolutional layer also employs 64 filters of size 2 × 2 . The last convolutional layer is connected to a flatten layer, which is then followed by a dense layer with 16 neurons. In order to prevent overfitting, a dropout layer with a rate of 0.3 is implemented. Finally, a dense layer with two neurons determines whether the image corresponds to a cell or not, see Figure 7.

4.3. Pseudocode

The environment has been developed using the object-oriented programming paradigm, allowing for efficient extraction of all its features through separate code segments. By adopting this approach, various aspects of the environment can be accessed and manipulated independently, see Algorithm 1.
Algorithm 1 Environment feature extraction
 Initialize agent and its weights;
while train do Reset the environment and gather initial observation S;
     while episode not completed do
         for time step t = 0 , 1 , , T  do
            Let agent choose action A based on state S;
            Update environment according to action A;
            Get new image (State S ) from environment;
            Calculate reward R;
            Calculate advantage A ^ ;
            Check if episode completed;
             S S ;
         end for
         Update weights with PPO;
     end while
end while

4.4. Cell Classifier Model

A pre-trained CNN was used in order to classify the cells. Residual Network (ResNet) is a deep convolutional neural network architecture introduced in 2015 by Kaiming He et al. [14] in the paper “Deep Residual Learning for Image Recognition”, which mentions that the main idea behind ResNet is to address the issue of vanishing gradients in deep neural networks. Deep networks with many layers can suffer from vanishing gradients, where the magnitude of the gradients used to update the weights during training becomes very small, leading to slow convergence or even non-convergence. ResNet solves this problem by adding residual connections to the network, which allow for the gradients to bypass one or more layers and flow more easily through the network. A residual block in ResNet consists of several convolutional layers, with the output of each layer being added to the input of the next layer (the “residual connection”). This allows the network to learn residual functions or the difference between the desired output and the input, instead of trying to learn the whole mapping from scratch [15], see Figure 8.
ResNet also introduced the concept of deep supervision, where multiple loss functions are used for different parts of the network to help improve convergence.
The ResNet architecture achieved state-of-the-art results on several computer vision benchmarks and has been widely used in many subsequent deep learning models and systems. It is also one of the most popular models for transfer learning, where a pre-trained ResNet model is fine-tuned for a new task [15]. Residual Network (ResNet) is capable of categorizing cells into the following categories mentioned in the first dataset, see Figure 2.
In this study, ResNet-50 was used, where the number represents the total number of layers in the CNN. Nevertheless, it is necessary to emphasize that as the number of layers increases the computational cost too. This architecture was trained by using pre-default weights. This is useful for increasing the efficiency and precision of the model. As a consequence, it helps the network to converge faster, having an accurate model in less time. ResNet-50 is trained with a set of data from ImageNet that can be used to initialize the network before being trained with different data, see Figure 2. This network already has knowledge about how to recognize and extract image features. To avoid over-fitting, a dropout is added in the last layer of ResNet-50. The dropout is a hyper-parameter, and regularization technique where the main idea is forcing some neurons to not activate during the training [16]. For this architecture, the dropout was added in the last convolutional layer with a rate of 0.3.

4.5. Training Process

In reinforcement learning, there are many different algorithms that may be used to train agents. Choosing the right algorithm is essential in order to converge to a good behavior. The agent uses PPO for training since it has been shown to be reliable and effective in many applications [9,17,18]. PPO ensures that the changes made to the policy distribution remain within a specified range. This constraint is vital for maintaining stability during the learning process, as it prevents extreme policy updates that may result in divergent behavior or sub-optimal policies. Furthermore, PPO incorporates a clipping mechanism within the surrogate objective function, limiting the policy update to a certain threshold. This clipping helps to stabilize the learning process by preventing overly large policy updates, which can introduce instability and hinder convergence. The ROI uses this on-policy algorithm to train and learn, see Algorithm 2, how to move in the environment, and what features are important to solve the task.
Algorithm 2 PPO Clip
 Initialize θ ;
for iteration i = 0 , 1 ,  do
     for time step t = 0 , 1 , , T  do
          Sample time step with policy π θ , o l d ;
          Calculate advantage A ^ t ;
     end for
     for epoch k = 0 , 1 , , K  do
          Optimize L C L I P + V F + S with respect to θ ;
          Update θ ;
     end for
end for

4.6. Experiments

A total of eight agents will be trained using different reward signals, see Table 2. The PPO algorithm hyperparameters are tuned in order to test the best configuration. The first four agents (A–D) have an n_step value set to 512. Conversely, the remaining four agents (E–H) employ identical hyperparameters, but with an n_step value of 1024, see Table 3.
Finally, the selection of agents for further consideration relies on their performance throughout the training process. To facilitate this selection, a multi-stage approach is implemented, involving three distinct stages for rejecting agents and not well-defined reward signals. This process aims to identify the most promising agents from each stage and subject them to retraining in subsequent stages, see Figure 9.

5. Results and Discussion

5.1. First Stage

In the initial phase, we observed the initial training progress of the agents, see Figure 10 and Figure 11. It became apparent that the two agents failed to construct a viable policy. This outcome arose due to the inadequately defined reward signal, which imposed a significant penalty on the agents for cell searching. On the other hand, the remaining six agents exhibited diverse mean rewards throughout their training. Notably, agents C, E, and F outperformed the others by executing more effective actions and detecting a greater number of cells. Consequently, these agents achieved higher mean rewards throughout the three-million-time training period and demonstrated stability in their performance values.
The variations in mean rewards can be attributed to the random selection of environments or digital Pap smears from the initial dataset. Each environment contained a different number of cells, which introduced variability into the agents’ experiences.
Furthermore, the initial peak observed in all the agents’ performance can be attributed to the fact that the initial policies were not yet well-trained. These early policies often generated a large number of random actions, which resulted in the agents covering a larger area of the environment. However, since these actions were chosen randomly without any prior experience or knowledge, the agents did not gain meaningful insights or improve their performance significantly during this peak.

5.2. Second Stage

In this stage, agents C, E, and F were selected based on their superior performance, as they consistently achieved higher mean rewards compared to the other agents. These selected agents underwent an additional two million time steps of retraining, leading to diverse outcomes, as depicted in Figure 12. Agent E emerged as the standout learner during this stage, exhibiting positive mean rewards and surpassing the performance of the other agents. However, it is worth mentioning that in one particular environment, agent E obtained lower values compared to the other environments. After that lowest value, the agent began to recover its policy and exhibit improved results. On the other hand, agents C and F consistently obtained negative mean rewards in the majority of environments, leading to their exclusion from the final stage. The negative mean rewards obtained by agents C and F indicated a lack of effective policy and performance in various environments. These agents were therefore considered unsuitable to advance to the final training level.

5.3. Third Stage

In this stage, agent E underwent parallel retraining for 6 million additional time steps, carried out four times, see Figure 13. Surprisingly, as depicted in Figure 5 and Figure 6, despite employing identical hyperparameters and a shared pre-trained model, the learning process exhibited variability. This variation can be attributed to the use of a stochastic algorithm, which generates a policy based on action probabilities rather than deterministic choices. The stochastic nature of the algorithm introduces randomness into the learning process, resulting in divergent training trajectories. As a result, each retrained agent may exhibit unique patterns of improvement and performance. From the four retrained agents, it is possible to select one agent that achieves superior results compared to the others. This selection is based on the agent’s ability to optimize its policy and achieve higher performance scores. This highlights the importance of evaluating and identifying the most successful agent among multiple retraining iterations.

Behavior Testing

Now, it is interesting to see the agents’ movement patterns within the environment. To visualize their trajectories and action frequencies, a heatmap technique was employed. Figure 14 showcases this representation, highlighting the agents’ positions and the frequency of actions taken in each location. The heatmap provides valuable insights into how the agents navigate and interact with the environment. By observing the intensity of color in different areas, we can identify regions where the agents spend more time or take repetitive actions. This visualization aids in understanding the agents’ exploration patterns and potential areas of focus.
To assess the behavior of the agents, three distinct environments were chosen for policy testing. Each environment featured a different quantity of cells distributed throughout various sections of the test area, see Figure 15. By defining these diverse environments, we aimed to evaluate how the agents are adapting and performing under different conditions. The varying distribution of cells provided a challenge for the agents to locate and identify them effectively. This approach allowed us to analyze the agents’ performance across different scenarios, gaining insights into their capabilities and strategies in different parts of the test.
In the first environment, see Figure 16, it is interesting to observe that agent E1 displayed no behavior in the first environment, despite achieving higher mean rewards in training compared to the other agents in multiple environments. By analyzing the heatmap plots of agents E2 and E4, we can observe that these agents actively search for cells but eventually become trapped in a repetitive loop of actions. On the other hand, the heatmap plot of agent E3 shows a notable performance, as it covers a larger area compared to the other agents and successfully detects nearly all the cells present in the sample. Additionally, agent E3 exhibits a maximum of 20 passes through the same position, indicating efficient exploration and reduced redundancy in actions. Overall, these observations suggest that agent E3 demonstrates superior performance in terms of coverage, cell detection, and avoidance of repetitive behavior in the given environment.
In the second environment, see Figure 17, we observe that agent E1 remains trapped in the initial position throughout the entire episode without taking any actions. This indicates a lack of exploration or an inability to navigate effectively in this particular environment. Conversely, agents E2 and E4 exhibit more active behavior and successfully identify a significant number of cell positions. However, they eventually fall into repetitive action patterns, suggesting a limitation in their ability to adapt and explore further. Additionally, agent E3 once again stands out with impressive results. It achieves a high level of visual coverage, capturing almost all the cells, while demonstrating a maximum of 20 repetitions of passing through the same position. Furthermore, agent E3 covers a larger area compared to the other agents, indicating a more comprehensive exploration. In sum, these findings highlight the superior performance of agent E3 in terms of cell detection, reduced repetitive actions, and broader area coverage, making it the most promising agent in the second environment.
In the third environment, see Figure 18, we observe that agent E1 actively performs actions and successfully collects a significant number of cells compared to the other environments. This behavior can be attributed to the fact that more training steps are required for agent E1 to overcome the difficulty of becoming stuck in the starting position. After visual analysis of the other agents, we can infer that they exhibit a searching behavior, trying to maximize the number of found cells in the environment. Each agent has unique search patterns as they explore. However, a common issue arises: the agents tend to fall into repetitive action loops. The agents would benefit from further retraining to prevent this repetitive behavior and optimize their search strategies. However, despite this limitation, the agents still display a strong inclination towards exploring and gathering cells.

5.4. Cell Classifier Model

Training a model using a ResNet50 architecture with pre-trained weights offers numerous advantages over training from scratch. Pre-trained weights allow the model to take advantage of the knowledge acquired from a large dataset like ImageNet, which contains millions of labeled images. These pre-trained weights encapsulate general image features, including edges, textures, and object parts, which prove advantageous for TL. This initialization provides the model with a robust starting point, enabling faster convergence and superior performance compared to training from zero, see Figure 19 and Figure 20.
After training the model for classifying cervix cells in digital Pap smears, the results showed a training loss and accuracy of 26.40% and 89.89%, respectively. When the model was tested on new data, it achieved a validation loss and accuracy of 21.30% and 91.25%, respectively. Furthermore, Table 4 highlights the excellent classification performance of the model across different cervix cell types. It achieves an accuracy of 89.88% in identifying HSIL, demonstrating its capability to accurately detect this severe abnormality. Additionally, the model exhibits a high accuracy of 92.27% for LSIL, effectively distinguishing this less severe abnormality. Notably, It shows excellent performance in classifying NILM cases, achieving an accuracy of 98.51%. Moreover, the model achieves a respectable accuracy of 83.40% for SCC, crucial for the early detection and treatment of this malignant condition.

Comparison with Other Studies

In this comparative analysis, ResNet50V2* and ResNet101V2* from [19] emerge as leading contenders with accuracy rates of 0.97 and 0.95, respectively, showcasing their robust performance. However, our proposed ResNet50 model holds its ground with a commendable accuracy of 0.91, suggesting competitive efficacy. Notably, ResNetXt29_264d and ResNetXt29_464d from [20] demonstrate comparable accuracy, see Table 5. We can mention that [19,20] these findings collectively contribute valuable insights to the optimization of convolutional neural network models.

5.5. Final System

By integrating a cell classifier model into the environment and leveraging the intelligence of trained agents, we can create a highly advanced system for detecting and classifying located cells. In the final environment, the trained agents operate autonomously, utilizing cell recognition to scan their surroundings and identify cells in real-time. Once a cell is detected, the agents employ the cell classifier model to classify it according to predefined categories, see Figure 21.

5.5.1. Hyperspectral and Multispectral Systems Discussion

In the context of Pap smear analysis, the traditional approach involves manual examination by cytotechnologists or pathologists using stained cells under a microscope, while hyperspectral [21] and multispectral [22] imaging technologies offer significant capabilities in various medical imaging applications, their integration into Pap smear analysis faces challenges. The current standard practices have well-established staining techniques, and any transition to new imaging technologies requires thorough validation, standardization, and integration into routine clinical workflows. Moreover, challenges related to sample preparation, the complexity and cost of implementing these imaging systems, regulatory approval processes, and the need for extensive clinical validation must be addressed.
However, our research paper introduces a promising alternative perspective by combining DRL and DL techniques for autonomous analysis using standardized methods. This innovative approach offers a potential solution to the challenges associated with manual examination. By employing DRL, our system can autonomously navigate through smear images, providing a rapid and efficient means of analysis. Unlike the complexities of hyperspectral and multispectral imaging, our proposed framework focuses on streamlining the analysis process using readily available images. This not only addresses some of the challenges mentioned earlier but also opens up the possibility of swift implementation into routine clinical practice. Our research thus contributes to the ongoing exploration of modern, autonomous perspectives in Pap smear analysis, offering new possibilities for enhanced diagnostics and screening processes.

5.5.2. Faced Limitations

Addressing some limitations of the system, it is essential to note that the use of the PPO algorithm introduces a stochastic element due to its implementation of a stochastic policy, relying on a probability distribution to guide the agent’s actions [9]. This inherent randomness implies that agents trained under identical hyperparameters and conditions may exhibit diverse behaviors. To accurately assess an agent’s efficiency or performance, it becomes necessary to train the same agent multiple times. In this study, eight different agents were trained, Figure 9, and their behaviors were examined at various stages of training. Additionally, it is crucial to consider that the number of samples required for achieving robust results can be substantial and varies depending on the context. For instance, control tasks may necessitate hundreds of thousands to millions of time steps for an agent to learn effectively [23], and Atari games often demand millions of time steps [24]. Given that the dataset comprises 1080 × 1080 pixel images, the training process duration is influenced by the image size, and the scale of the environment is notably larger compared to other reinforcement learning studies such as those mentioned before. Therefore, the agents were trained for 11 million time steps, a duration that, while demonstrating promising results and observable searching behavior in the last three agents, may still be subject to further optimization. In the future, the complexity of the environment could be mitigated through the implementation of various techniques, including the application of thresholds, Gaussian blur, and a grayscale transformation [25]. These methods have the potential to simplify the environment by reducing unnecessary details and emphasizing key features, thus facilitating more streamlined processing and analysis. This strategic simplification could contribute to improved efficiency and performance, making the system more adaptable and responsive to the nuances of the environment it interacts with, while the ResNet50 model exhibited a high accuracy percentage and favorable metrics across different dataset categories, it is prudent to acknowledge that there is room for improvement in future models. Achieving greater precision and accuracy remains a prospective avenue for enhancement such as [19].

6. Conclusions

In conclusion, the trained agents have demonstrated exceptional learning capabilities, and they have successfully developed a searching behavior that allows them to efficiently locate cells within digital Pap smears. These results indicate the effectiveness of the deep reinforcement learning approach in navigating complex visual patterns and identifying cells of interest.
Additionally, a CNN model using ResNet-50 architecture was trained to classify the detected cells into four distinct categories including being negative for intraepithelial lesion or malignancy, low-grade intraepithelial lesions, high-grade intraepithelial lesions, and squamous cell carcinoma. This classification capability adds another layer of sophistication to the proposed system, enabling accurate identification and categorization of cells based on their shape and structure.
The successful training of both the agents and the CNN model shows the potential of this research to revolutionize the field of cervical cancer screening. By combining the searching behavior of the agents with the classification capabilities of the CNN model, an integrated and highly efficient system for automated analysis of Papanicolaou can be realized.
These results have important implications for clinical practice, as they open the door for the development of an automatic Papanicolaou analyzing machine that can analyze samples without human intervention. The utilization of deep reinforcement learning and CNN-based classification techniques has the potential to solve the lack of pathologists, particularly in low-middle incoming countries, and improve the accuracy and availability of cervical cancer screening services.
In conclusion, our research has successfully trained four agents over 11 million time steps, resulting in the development of a searching behavior for locating cells in digital Pap smear images. Furthermore, our CNN model, utilizing the ResNet-50 architecture, accurately classifies the detected cells into four categories. These achievements demonstrate the great capabilities of deep reinforcement learning and CNN-based classification techniques in advancing the field of cervical cancer screening. The proposed automated system could improve the efficiency, accessibility, and accuracy of Papanicolaou analysis, ultimately contributing to the global efforts of eliminating cervical cancer and improving women’s healthcare worldwide.

Author Contributions

Conceptualization, O.C.; Formal analysis, C.M.; Resources, O.C.; Software, C.M. and M.E.M.-C.; Supervision, M.E.M.-C. and O.C.; Validation, M.E.M.-C. and O.C.; Writing—original draft, C.M.; Writing—review and editing, M.E.M.-C. and O.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cohen, P.A.; Jhingran, A.; Oaknin, A.; Denny, L. Cervical cancer. Lancet 2019, 393, 169–182. [Google Scholar] [CrossRef] [PubMed]
  2. Hasenleithner, S.O.; Speicher, M.R. How to detect cancer early using cell-free DNA. Cancer Cell 2022, 40, 1464–1466. [Google Scholar] [CrossRef] [PubMed]
  3. Nuche-Berenguer, B.; Sakellariou, D. Socioeconomic determinants of cancer screening utilisation in Latin America: A systematic review. PLoS ONE 2019, 14, e0225667. [Google Scholar] [CrossRef] [PubMed]
  4. Davies-Oliveira, J.; Smith, M.; Grover, S.; Canfell, K.; Crosbie, E. Eliminating Cervical Cancer: Progress and Challenges for High-income Countries. Clin. Oncol. 2021, 33, 550–559. [Google Scholar] [CrossRef] [PubMed]
  5. World Health Organization. 2021. Available online: www.who.int/publications/m/item/cervical-cancer-ecu-country-profile-2021 (accessed on 17 October 2023).
  6. Liebermann, E.J.; VanDevanter, N.; Hammer, M.J.; Fu, M.R. Social and Cultural Barriers to Women’s Participation in Pap Smear Screening Programs in Low- and Middle-Income Latin American and Caribbean Countries: An Integrative Review. J. Transcult. Nurs. 2018, 29, 591–602. [Google Scholar] [CrossRef] [PubMed]
  7. Strasser-Weippl, K.; Chavarri-Guerra, Y.; Villarreal-Garza, C.; Bychkovsky, B.L.; Debiasi, M.; Liedke, P.E.R.; de Celis, E.S.P.; Dizon, D.; Cazap, E.; de Lima Lopes, G.; et al. Progress and remaining challenges for cancer control in Latin America and the Caribbean. Lancet Oncol. 2015, 16, 1405–1438. [Google Scholar] [CrossRef] [PubMed]
  8. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction, 2nd ed.; The MIT Press: Cambridge, UK, 2018. [Google Scholar]
  9. Schulman, J.; Wolski, F.; Dhariwal, P.; Radford, A.; Klimov, O. Proximal Policy Optimization Algorithms. arXiv 2017, arXiv:1707.06347. [Google Scholar]
  10. Hussain, E.; Mahanta, L.B.; Borah, H.; Das, C.R. Liquid based-cytology Pap smear dataset for automated multi-class diagnosis of pre-cancerous and cervical cancer lesions. Data Brief 2020, 30, 105589. [Google Scholar] [CrossRef] [PubMed]
  11. Chang, O.G.; Toapanta, B.O. Automatic High-Resolution Ananlysis of Pap Test Cells. Bachelor’s Thesis, Universidad de Investigación de Tecnología Experimental Yachay, San Miguel de Urcuquí, Ecuador, 2021. [Google Scholar]
  12. Brockman, G.; Cheung, V.; Pettersson, L.; Schneider, J.; Schulman, J.; Tang, J.; Zaremba, W. OpenAI Gym. In Proceedings of the International Conference on Learning Representations, San Juan, Puerto Rico, 2–4 May 2016. [Google Scholar]
  13. Plappert, M.; Houthooft, R.; Dhariwal, P.; Sidor, S.; Chen, R.Y.; Asfour, T.; Abbeel, P.; Andrychowicz, M. Multi-Goal Reinforcement Learning: Challenging Robotics Environments and Request for Research. In Proceedings of the Conference on Robot Learning, Zurich, Switzerland, 29–31 October 2018. [Google Scholar]
  14. Targ, S.; Almeida, D.; Lyman, K. Resnet in resnet: Generalizing residual architectures. arXiv 2016, arXiv:1603.08029. [Google Scholar]
  15. He, F.; Liu, T.; Tao, D. Why resnet works? residuals generalize. IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 5349–5362. [Google Scholar] [CrossRef] [PubMed]
  16. Chen, L.; Gautier, P.; Aydore, S. DropCluster: A structured dropout for convolutional networks. arXiv 2020, arXiv:2002.02997. [Google Scholar]
  17. Berner, C.; Brockman, G.; Chan, B.; Cheung, V.; Debiak, P.; Dennison, C.; Farhi, D.; Fischer, Q.; Hashme, S.; Hesse, C.; et al. Dota 2 with Large Scale Deep Reinforcement Learning. arXiv 2019, arXiv:1912.06680. [Google Scholar]
  18. Akkaya, I.; Andrychowicz, M.; Chociej, M.; Litwin, M.; McGrew, B.; Petron, A.; Paino, A.; Plappert, M.; Powell, G.; Ribas, R.; et al. Solving Rubik’s Cube with a Robot Hand. arXiv 2019, arXiv:1910.07113. [Google Scholar]
  19. Wong, L.; Ccopa, A.; Diaz, E.; Valcarcel, S.; Mauricio, D.; Villoslada, V. Deep Learning and Transfer Learning Methods to Effectively Diagnose Cervical Cancer from Liquid-Based Cytology Pap Smear Images. Int. J. Online Biomed. Eng. (iJOE) 2023, 19, 7–93. [Google Scholar] [CrossRef]
  20. Zhao, S.; He, Y.; Qin, J.; Wang, Z. A Semi-supervised Deep Learning Method for Cervical Cell Classification. Anal. Cell. Pathol. 2022, 2022, 1–12. [Google Scholar] [CrossRef] [PubMed]
  21. Zhu, H.; Luo, J.; Liao, J.; He, S. High-accuracy Rapid Identification and Classification of Mixed Bacteria Using Hyperspectral Transmission Microscopic Imaging and Machine Learning. Prog. Electromagn. Res. 2023, 178, 49–62. [Google Scholar] [CrossRef]
  22. Shen, F.; Deng, H.; Yu, L.; Cai, F. Open-source mobile multispectral imaging system and its applications in biological sample sensing. Spectrochim. Acta Part Mol. Biomol. Spectrosc. 2022, 280, 121504. [Google Scholar] [CrossRef] [PubMed]
  23. Lillicrap, T.; Hunt, J.; Pritzel, A.; Heess, N.; Erez, T.; Tassa, Y.; Silver, D.; Wierstra, D. Continuous control with deep reinforcement learning. In Proceedings of the 4th International Conference on Learning Representations, ICLR 2016, San Juan, Puerto Rico, 2–4 May 2016. [Google Scholar]
  24. Moreno-Vera, F. Performing Deep Recurrent Double Q-Learning for Atari Games. In Proceedings of the 2019 IEEE Latin American Conference on Computational Intelligence (LA-CCI), Guayaquil, Ecuador, 11–15 November 2019. [Google Scholar] [CrossRef]
  25. Gao, R.; Lu, H.; Al-Azzawi, A.; Li, Y.; Zhao, C. DRL-FVRestore: An Adaptive Selection and Restoration Method for Finger Vein Images Based on Deep Reinforcement. Appl. Sci. 2023, 13, 699. [Google Scholar] [CrossRef]
Figure 1. Conceptualization of the proposed approach.
Figure 1. Conceptualization of the proposed approach.
Computation 11 00252 g001
Figure 2. Pap smear samples: (a) Negative for intraepithelial lesion or malignancy (NILM). (b) Low-grade intraepithelial lesions (LSIL). (c) High-grade intraepithelial lesions (HSIL). (d) Squamous cell carcinoma (SCC).
Figure 2. Pap smear samples: (a) Negative for intraepithelial lesion or malignancy (NILM). (b) Low-grade intraepithelial lesions (LSIL). (c) High-grade intraepithelial lesions (HSIL). (d) Squamous cell carcinoma (SCC).
Computation 11 00252 g002
Figure 3. Examples of extracted “Cells”.
Figure 3. Examples of extracted “Cells”.
Computation 11 00252 g003
Figure 4. Examples of extracted “No cells”.
Figure 4. Examples of extracted “No cells”.
Computation 11 00252 g004
Figure 5. The figure illustrates a stack of frames, visually representing the environment in which an agent interacts. The agent is characterized by two black squares and the motion it exhibits within the frames.
Figure 5. The figure illustrates a stack of frames, visually representing the environment in which an agent interacts. The agent is characterized by two black squares and the motion it exhibits within the frames.
Computation 11 00252 g005
Figure 6. Graphical representation of two environments. (A) It is the environment where the agent searches for cells, and the captured image by the ROI is displayed in the right window. (B) The cells that have been detected.
Figure 6. Graphical representation of two environments. (A) It is the environment where the agent searches for cells, and the captured image by the ROI is displayed in the right window. (B) The cells that have been detected.
Computation 11 00252 g006
Figure 7. A visual depiction of how data moves through the CNN.
Figure 7. A visual depiction of how data moves through the CNN.
Computation 11 00252 g007
Figure 8. Residual block architecture of ResNet model.
Figure 8. Residual block architecture of ResNet model.
Computation 11 00252 g008
Figure 9. Graphic representation of the stages during the training process.
Figure 9. Graphic representation of the stages during the training process.
Computation 11 00252 g009
Figure 10. The left plot shows all the first set of agents’ mean rewards, permitting the ability to visually compare their training process. The right plot shows only three agents, agent D was not considered because it has the lowest values.
Figure 10. The left plot shows all the first set of agents’ mean rewards, permitting the ability to visually compare their training process. The right plot shows only three agents, agent D was not considered because it has the lowest values.
Computation 11 00252 g010
Figure 11. The left plot shows all the second set of agents’ scores, permitting the ability to visually compare their training process. The right plot shows only three agents, agent H was not considered because it has the lowest values.
Figure 11. The left plot shows all the second set of agents’ scores, permitting the ability to visually compare their training process. The right plot shows only three agents, agent H was not considered because it has the lowest values.
Computation 11 00252 g011
Figure 12. The graph shows the retraining results of the agents C, E, and F.
Figure 12. The graph shows the retraining results of the agents C, E, and F.
Computation 11 00252 g012
Figure 13. The graph shows the retraining results of the four extensively trained agents.
Figure 13. The graph shows the retraining results of the four extensively trained agents.
Computation 11 00252 g013
Figure 14. The figure illustrates the behavior of an untrained agent. Red dots represent discovered cells, while the blue bar indicates repetitive visits to the same position.
Figure 14. The figure illustrates the behavior of an untrained agent. Red dots represent discovered cells, while the blue bar indicates repetitive visits to the same position.
Computation 11 00252 g014
Figure 15. The figure shows three distinct environments utilized for testing the agents during the second and third stages of experiments.
Figure 15. The figure shows three distinct environments utilized for testing the agents during the second and third stages of experiments.
Computation 11 00252 g015
Figure 16. Tracking results of the agents E1, E2, E3, and E4 from the last stage in the first environment.
Figure 16. Tracking results of the agents E1, E2, E3, and E4 from the last stage in the first environment.
Computation 11 00252 g016
Figure 17. Tracking results of the agents E1, E2, E3, and E4 from the last stage in the second environment.
Figure 17. Tracking results of the agents E1, E2, E3, and E4 from the last stage in the second environment.
Computation 11 00252 g017
Figure 18. Tracking results of the agents E1, E2, E3, and E4 from the last stage in the third environment.
Figure 18. Tracking results of the agents E1, E2, E3, and E4 from the last stage in the third environment.
Computation 11 00252 g018
Figure 19. ResNet-50 behavior in training and validation accuracy during 100 epochs.
Figure 19. ResNet-50 behavior in training and validation accuracy during 100 epochs.
Computation 11 00252 g019
Figure 20. ResNet-50 behavior in training and validation loss during 100 epochs.
Figure 20. ResNet-50 behavior in training and validation loss during 100 epochs.
Computation 11 00252 g020
Figure 21. The figure demonstrates three instances in which the agent successfully detects cells and accurately classifies them. The agent detected LSHL, HSIL, and SCC cells.
Figure 21. The figure demonstrates three instances in which the agent successfully detects cells and accurately classifies them. The agent detected LSHL, HSIL, and SCC cells.
Computation 11 00252 g021
Table 1. The actions are numbered from 1 to 4, with each number representing a specific direction. The environment interprets these numbers to facilitate the movement of the ROI.
Table 1. The actions are numbered from 1 to 4, with each number representing a specific direction. The environment interprets these numbers to facilitate the movement of the ROI.
Actions NumberAction
1Right
2Left
3Up
4Down
Table 2. Brief description of the agents.
Table 2. Brief description of the agents.
AgentsDescription
A, EUsing the reward signal without changes.
B, FNo penalty for detecting the same cell multiple times
C, GNo penalty while searching for a cell
D, HHigh penalty while searching for a cell, r 1 = 10
Table 3. The hyperparameters included in the PPO Stable Baselines 3 implementation, with descriptions and values. The symbol following a hyperparameter name refers to the coefficient.
Table 3. The hyperparameters included in the PPO Stable Baselines 3 implementation, with descriptions and values. The symbol following a hyperparameter name refers to the coefficient.
HyperparametersValuesDescription
learning_rate ( α ) 0.0003 Progress remaining, which ranges from 1 to 0.
n_steps512; 1024Steps per parameters update.
batch_size128Images processed by the network at once
n_epochs10Updates for the policy using the same trajectory
gamma ( γ ) 0.99 Discount factor
gae_lambda ( λ ) 0.95 Bias vs. variance trade-off
clip_range ( ϵ ) 0.2 Range of clipping
vf_coef c 1 0.5 Value function coefficient
ent_coef c 2 0.0 Entropy coefficient
max_grad_norm 0.5 Clips gradient if it becomes too large
Table 4. Summary of metrics obtained with ResNet50 by category.
Table 4. Summary of metrics obtained with ResNet50 by category.
CategoriesPrecisionRecallF1-ScoreSupport
NILM1.000.970.98200
LSIL0.940.920.93200
HSIL0.780.920.84200
SCC0.910.860.85200
Table 5. Comparison with other studies ordered by accuracy.
Table 5. Comparison with other studies ordered by accuracy.
AI MethodsAccuracyClasses
ResNet50V2* [19]0.974
ResNet101V2* [19]0.954
ResNet500.914
ResNetXt29_2*64d [20]0.9110
ResNetXt29_4*64d [20]0.9110
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Macancela, C.; Morocho-Cayamcela, M.E.; Chang, O. Deep Reinforcement Learning for Efficient Digital Pap Smear Analysis. Computation 2023, 11, 252. https://doi.org/10.3390/computation11120252

AMA Style

Macancela C, Morocho-Cayamcela ME, Chang O. Deep Reinforcement Learning for Efficient Digital Pap Smear Analysis. Computation. 2023; 11(12):252. https://doi.org/10.3390/computation11120252

Chicago/Turabian Style

Macancela, Carlos, Manuel Eugenio Morocho-Cayamcela, and Oscar Chang. 2023. "Deep Reinforcement Learning for Efficient Digital Pap Smear Analysis" Computation 11, no. 12: 252. https://doi.org/10.3390/computation11120252

APA Style

Macancela, C., Morocho-Cayamcela, M. E., & Chang, O. (2023). Deep Reinforcement Learning for Efficient Digital Pap Smear Analysis. Computation, 11(12), 252. https://doi.org/10.3390/computation11120252

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop