Next Article in Journal
A Review of Emerging Immunotherapeutic Strategies for IDH-Mutant Glioma
Previous Article in Journal
Clinical Predictors and Recurrence Characteristics Following Radiotherapy for Primary Central Nervous System Lymphoma: A Retrospective Cohort Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SPP-SegNet and SE-DenseNet201: A Dual-Model Approach for Cervical Cell Segmentation and Classification

by
Betelhem Zewdu Wubineh
*,
Andrzej Rusiecki
and
Krzysztof Halawa
Faculty of Information and Communication Technology, Wroclaw University of Science and Technology, 50-370 Wroclaw, Poland
*
Author to whom correspondence should be addressed.
Cancers 2025, 17(13), 2177; https://doi.org/10.3390/cancers17132177
Submission received: 3 June 2025 / Revised: 26 June 2025 / Accepted: 27 June 2025 / Published: 27 June 2025
(This article belongs to the Section Methods and Technologies Development)

Simple Summary

Cervical cancer is one of the most common cancers in women, and detecting it early is critical for successful treatment. Currently, doctors examine Pap smear images manually, which takes a lot of time and can lead to mistakes. To help improve this process, we have developed two advanced computer models. One model accurately separates cervical cells from the background in images, while the other classifies whether the cells are normal or abnormal. We tested our method on two well-known image datasets and found that it works better than existing systems in both identifying and classifying cervical cells. These results show that our system could support medical professionals by making cervical cancer screening faster, more accurate, and more consistent. This approach may ultimately help detect cancer earlier and improve patient outcomes.

Abstract

Background/Objectives: Cervical cancer, the fourth most common malignancy in women worldwide, continues to pose a significant threat to global health. Manual examination of the Pap smear image is time-consuming, labor-intensive, and prone to human error due to the large number of slides and subjective judgment. This study proposes a novel SegNet-based spatial pyramid pooling (SPP-SegNet) deep learning model for segmentation and a Squeeze-and-Excitation-based (SE-DenseNet201) model for classification, aimed at improving the accuracy of cervical cancer detection. Methods: The model incorporates the SPP bottleneck and atrous convolution in the SegNet framework, allowing for the extraction of multiscale spatial features and improving segmentation performance. The segmentation output is used as input for the classification task. The proposed method is evaluated on the Pomeranian and SIPaKMeD datasets. Results: Segmentation results show that SPP-SegNet achieves 98.53% accuracy on the Pomeranian data set, exceeding standard SegNet, 97.86%. It also achieves 94.15% accuracy on the SIPaKMeD dataset, outperforming the standard SegNet, which is 90.95%. For classification, SE-DenseNet201 achieves 93% and 99% accuracy for the Pomeranian and SIPaKMeD binary classification, respectively, using the bounding box input. Conclusions: These results show that SPP-SegNet and SE-DenseNet201 can potentially automate cervical cell segmentation and classification, facilitating the early detection and diagnosis of cervical cancer.

1. Introduction

Cervical cancer is a global health concern, ranking as the fourth most prevalent cancer among women worldwide [1]. In 2018, approximately 570,000 new cases and 311,000 deaths were attributed to cervical cancer, highlighting the urgent need for effective prevention and early detection strategies [2]. The Pap smear test, a widely adopted screening method, involves collecting and microscopically examining cervical cells to detect precancerous or cancerous changes. This test has been instrumental in reducing the incidence and mortality rates of cervical cancer in many developed countries [3,4]. However, manual screening of Pap smear images is labor-intensive, time-consuming, and susceptible to human error due to subjective judgment and the large volume of slides that require analysis [5,6,7].
Automating the Pap smear screening process through image processing and analysis offers a promising solution to overcome these challenges. Image segmentation, a fundamental step in this automation, involves partitioning an image into distinct regions of interest, such as the nucleus and cytoplasm of cervical cells [4]. Accurate segmentation is crucial for subsequent feature extraction and classification tasks, as it facilitates the identification of abnormal cells that may indicate precancerous or cancerous conditions [1,8]. However, segmenting cervical cells in Pap smear images is inherently challenging due to factors such as overlapping cells, variations in stain quality, and the presence of artifacts [4].
Deep learning (DL), a subfield of machine learning, has been widely used in medical image analysis, demonstrating remarkable capabilities in tasks such as image segmentation and classification [2,9]. By learning complex patterns and representations from large datasets, DL models improve the accuracy and efficiency of medical image interpretation [4]. Convolutional neural networks (CNNs), in particular, automatically learn features from images [10], eliminating the need for manual feature engineering, and are especially effective for both segmentation and classification tasks [11,12].
SegNet is a DL architecture specifically designed for image segmentation [13]. It consists of an encoder network that extracts features from the input image and a decoder network that upsamples these features to produce a segmented output [14]. To enhance the performance of SegNet in cervical cell segmentation, we propose a novel approach that integrates a spatial pyramid pooling (SPP) bottleneck and incorporates an atrous convolution (dilation rate) in the encoder. SPP enables the extraction of multiscale features, improving robustness to variations in cell size and shape [15]. The SPP bottleneck is placed between the SegNet encoder and decoder networks to aggregate features at different scales before upsampling. For classification tasks, the squeeze-and-excitation block is incorporated into the DenseNet201 (SE-DenseNet201) model to improve feature recalibration by emphasizing informative features and suppressing less relevant ones.
This study aims to develop a novel framework for cervical cell segmentation and classification using SPP-SegNet and SE-DenseNet201. The contributions are as follows:
  • Introduced SPP-SegNet, incorporating spatial pyramidal grouping and atrous convolutions to improve cervical cell segmentation performance.
  • Developed SE-DenseNet201, integrating squeeze-and-excitation (SE) blocks to improve feature recalibration and boost classification accuracy.
The remainder of the paper is organized as follows. Section 2 discusses related work, Section 3 describes the materials and methods used in this study, Section 4 presents the results and discusses the findings, and Section 5 provides concluding remarks.

2. Related Works

Several research studies have been conducted on automating cervical cancer screening by segmenting and analyzing Pap smear images. Xiong et al. [16] conducted research on automated segmentation methods for the early detection of cervical cancer using DL-based magnetic resonance imaging (MRI). The study focused on segmenting the tumors using a single-shot multi-box detector (SSD), a squeeze-and-excitation block, and a U-Net attention gate (SE-ATT-Unet), achieving a remarkable precision of 92.32%. Chankong et al. [3] proposed a method for automatic cervical cell segmentation and classification using fuzzy C-means (FCMs) clustering and classifiers such as artificial neural networks (ANNs). The study scored a precision value of 85% using the Herlev dataset and 87% using the LCH datasets of nucleus segmentation. Zhu and Du [17] researched the segmentation of overlapping cervical cells based on boundary tracking. The study used the U-Net architecture to segment the image and used the center of each nucleus to identify the boundary point at a different angle. The model scored a precision value of 95.04%.
Furthermore, Beatriz et al. [18] proposed a U-Net and SegNet method to reduce mortality rates due to the implementation of Pap smears and issues related to the hindrance of false positive and negative results in the detection of cervical cancer. In relation to the performance of U-Net and SegNet neural networks for nuclei segmentation, these studies reported significant success, with SegNet achieving an accuracy of 95.52% in real cytological datasets (CRIC Cervix-Seg), demonstrating its ability to handle overlapping structures and artifacts effectively. Furthermore, Song et al. [19] researched overlapping cytoplasm segmentation via constrained multi-shape evolution for cervical cancer screening. The paper proposed an a priori-based shape model to address challenges in overlapping cytoplasm by combining local and global shapes. It yielded a Dice similarity coefficient of 83% in the pap smear image.
SegNet’s encoder-decoder architecture has been adapted for various segmentation tasks. Vianna et al. [20] demonstrated U-Net and SegNet performance on lesion segmentation of breast ultrasound images, achieving a Dice coefficient of 81.1%. Badrinarayanan et al. [21] proposed a SegNet deep convolutional encoder–decoder architecture for pixel-wise semantic image segmentation. The paper highlighted its memory efficiency and competitive performance compared to architectures such as FCN and DeepLab. The model scored an mIoU of 60.1%. Moreover, Brahmbhatt and Rajan [22] worked on skin lesion segmentation using SegNet with binary cross-entropy. The study achieved an intersection over union (IoU) value of 92% with minimal pre- and post-processing using the PH2 dataset. Goker [23] worked on the detection of cervical cancer from images of the uterine cervix using transfer learning. The model achieved an accuracy of 95.09% in the Herlev dataset using DenseNet20.
Although previous studies, such as those of [18], have taken advantage of U-Net and SegNet architectures for cervical cell segmentation, handling overlapping structure challenges, such as limited multiscale feature extraction, have been encountered [17,19]. To address these issues, this study proposes SPP-SegNet and SE-DenseNet201 to achieve superior segmentation and classification accuracy in the Pomeranian and SIPaKMeD datasets compared to the standard SegNet and DenseNet201 methods.

3. Materials and Methods

In this section, we describe the dataset, the pre-processing technique, the proposed method, the implementation details, and the performance metrics.

3.1. Dataset and Preprocessing Technique

The dataset was obtained from the Pomeranian Medical University in Szczecin, Poland. It comprised 419 images categorized into three classes: high squamous intraepithelial lesions (HSIL), low squamous intraepithelial lesions (LSIL), and normal squamous intraepithelial lesions (NSIL) [24]. The images were originally in BMP format, most of which had a resolution of 1130 × 1130 pixels. Furthermore, we used the publicly available SIPaKMeD dataset, which contains 4049 single-cell images classified into five categories: three representing abnormal cells and two representing normal cells. A summary of the details of the datasets is provided in Table 1.
During pre-processing, the images were resized to 224 × 224 pixels to ensure uniformity and optimize computational efficiency. The intensity values of the pixels were normalized to the range [−1, 1] to improve the stability of the model during training. Additionally, we created masks to segment Pap smear images using OpenCV, where all cells were represented as white and the background as black.

3.2. Proposed Method

For segmenting whole cells from the background, we employed the SegNet architecture, a well-known pixel-wise semantic segmentation method [25]. Built on the VGG-16 framework, SegNet removes fully connected layers and adopts a symmetric encoder–decoder structure [26]. Its ability to capture both low-level and high-level features, combined with its efficient upsampling strategy, makes it suitable for delineating fine-grained structures in medical images [18]. The standard SegNet model is illustrated in Figure 1.
However, the standard SegNet architecture struggles to capture global or multi-scale context effectively, which is crucial for segmenting objects or cells of different sizes and shapes. Thus, this study proposed a novel approach, SPP-SegNet, which improves the accuracy and efficiency of cervical cell segmentation by combining SegNet with spatial pyramid pooling (SPP) and atrous convolutions to increase the receptive field without increasing the number of parameters. The SPP block was designed to extract multiscale spatial information, improve the receptive field, and preserve fine-grained features for decoding.
In the encoding stage, the model comprised five convolutional blocks, each using atrous convolutions with progressively increasing dilation rates. This configuration expands the receptive field without increasing computational complexity or loss of resolution. The dilation rates were 1 × 1, 2 × 2, 4 × 4, 8 × 8, and 16 × 16 in the last layers, capturing broader contextual information. Batch normalization and ReLU activation were applied after each convolution to ensure stable learning and nonlinear transformations. The MaxPooling layers in each block progressively reduced the spatial resolution, generating feature maps with high-level semantics. The bottleneck integrated an SPP block to capture multi-scale features by pooling the feature maps at different scales.
The decoder mirrored the encoder structure but included upsampling and transposed convolutions to restore spatial resolution. Upsampling layers are applied in each block to upscale feature maps back to the original input resolution. Transposed convolutions replaced standard convolutions in the decoding stage, reconstructing spatial details. The final layer generated a single-channel output representing the segmented image, using a sigmoid activation function to produce probabilities in the shape of 224 × 224. The proposed model is depicted in Figure 2.
As shown in Figure 2, segmentation and classification tasks were performed using DL architecture. For the classification task, a novel SE-DenseNet201 model was employed, where SE blocks were integrated after every dense block to recalibrate channel-wise feature responses, improving the model’s ability to distinguish subtle variations in cervical cell features and enhance classification accuracy. First, we performed segmentation, followed by classification. For classification, we used two approaches: ROI-based (segmentation output multiplied by the original image) and bounding box (BBox) (determining bounded boxes containing the segmented fragment and cutting out the regions bounded by them from the original images). In the ROI image, the background was black, and the entire cell was extracted in RGB color. In contrast, the BBox approach cropped the RGB color from the original image based on the coordinates derived from the segmentation result.

3.3. Training Details and Performance Metrics

This study used Python 3.11 with TensorFlow and Keras on a Windows-based system with an NVIDIA GeForce RTX 3090 GPU. The model was trained using the Adam optimizer with a learning rate of 0.001 for 100 epochs to perform segmentation and classification tasks. A batch size of 18 was selected to balance computational efficiency and model stability, allowing frequent weight updates while optimizing memory usage. We used accuracy, precision, recall, and intersection over union (IoU) as performance metrics to segment all cells from the background. We used accuracy, precision, recall, and F1 score for classification as performance metrics.
The computational cost of the proposed SPP-SegNet method is summarized as follows. It contained approximately 21.9 million parameters, with training conducted over 100 epochs. The average training time per epoch was approximately 5 s, and the inference time per image was around 350 ms. Memory usage remained within acceptable limits for modern GPU-equipped systems. These results suggest that our model is efficient and feasible for deployment in moderately resource-constrained environments.

4. Results and Discussions

The following are the experimental setups implemented in this study. We used a novel SPP-SegNet DL architecture to separate entire cells from the background and SE-DenseNet201 to classify cervical cells. To evaluate the proposed model, we used the Pomeranian and SIPaKMeD datasets, which were divided into training, validation, and testing sets. Experiments were conducted using both the standard SegNet and the novel SPP-SegNet models for segmentation. The results of the segmentation for both the standard SegNet and the proposed SPP-SegNet models are presented in Table 2.
As shown in Table 2, the proposed SPP-SegNet model outperforms the standard SegNet across all evaluated datasets. On the Pomeranian dataset, SPP-SegNet achieved 98.53% accuracy, surpassing 97.86% for SegNet, with improved precision (96.41% vs. 91.9%), recall (91.32% vs. 90.31%), and IoU (95.75% vs. 95.55%). These results demonstrate the superior ability of SPP-SegNet to segment and detect relevant features while reducing false positives. Similarly, on the SIPaKMeD dataset, SPP-SegNet achieved 94.15% accuracy, significantly higher than the 90.95% of SegNet. Precision (93.87% vs. 92.40%), recall (94.94% vs. 90.05%), and IoU (95.08% vs. 92.89%) also show considerable improvements, confirming the effectiveness of SPP in enhancing multiscale feature extraction and segmentation performance. On Herlev, SPP-SegNet also demonstrated improved performance, achieving 87.58% accuracy compared to 86.74% for SegNet, along with gains in precision (91.95% vs. 91.05%), recall (91.67% vs. 90.35%), and IoU (91.13% vs. 90.33%). These consistent improvements across three distinct datasets reinforce the robustness and general applicability of the SPP-SegNet model for cervical cell segmentation under varying imaging and morphological conditions.

Ablation Study

We conducted an ablation study for the classification task to identify the best method for integrating classification with segmentation. In this ablation study, the classification task used the original image (without segmentation), the extracted ROI as input, and the BBox extracted using ResNet50V2, DenseNet121, and DenseNet201, our proposed SE-DenseNet201. The output of the segmentation was used as input for classification, as shown in Table 3.
Table 3 shows the classification performance of three different input methods: the original image, the extracted Region of Interest (ROI), and the BBox method. For the Pomeranian dataset, the BBox method achieved the highest accuracy, outperforming the ROI-based methods by 3% (88% accuracy compared to 85% for ROI). Similarly, for the SIPaKMeD and the Herlev dataset, the BBox method achieved an accuracy of 91% and 88%, respectively, which are comparable to the classification-only result for multiclass and binary classification. These results suggest that BBox extraction provides the most accurate classification results, particularly for the Pomeranian dataset. We then compare these results with the classification results using the proposed SE-DenseNet201 method, as shown in Table 4.
In Table 4, we can observe that the SE-DenseNet201 method with the BBox consistently outperforms the methods presented in Table 3, demonstrating a notable improvement in classification performance. For the Pomeranian dataset, the SE-DenseNet201 using the BBox method achieved 93% accuracy, significantly improving from results obtained with the standard method (ResNet50V2, DenseNet121, DenseNet201). For the SIPaKMeD multiclass dataset, the SE-DenseNet201 BBox method achieved an accuracy of 96%, outperforming the standard methods in Table 3 by 5%. Similarly, for binary classification, the SE-DenseNet201 BBox model showed an impressive 99% accuracy, precision, recall, and F1 score, surpassing all other methods. In the Herlev dataset, the proposed method achieves an accuracy of 90% using the BBox. These results highlight that the SE-DenseNet201 model, particularly with the BBox input, offers a significant enhancement in classification tasks compared to traditional methods, as shown by the improved metrics in both datasets. This reinforces the effectiveness of the proposed SE-DenseNet201 model, especially when using BBox output from segmentation, which is crucial for improving classification accuracy. In conclusion, the ablation study shows that the SPP-SegNet method, when combined with the proposed SE-DenseNet201, yields the best performance, particularly in terms of classification accuracy. The segmentation output for both the Pomeranian and SIPaKMeD datasets is shown in Figure 3.
As shown in Figure 3, there are two rows for the Pomeranian and SIPaKMeD datasets, respectively. The first column displays the actual image, and the second column shows the ground truth corresponding to each image. In the third column, the predicted outputs are displayed, and in the fourth column, an overlay of the predicted output is shown. The fifth column presents the segmented output using the proposed SPP-SegNet, the sixth column shows the segmented output using the standard SegNet, and the last column shows the BBox of the SPP-SegNet. The predicted output and the overlays are from the proposed SPP-SegNet. It can be noted that the proposed SPP-SegNet outperforms the standard SegNet, as the segmented result is much closer to the predicted output compared to the SegNet architecture. The confusion matrix of the proposed method is depicted in Figure 4.
As shown in Figure 4, six images were misclassified in the Pomeranian dataset, and fifteen images were misclassified in the SIPaKMeD binary classification. In this paper, we used both saliency maps and Grad-CAM to visualize and understand the regions of the image that influence the model’s decisions. A very popular technique for deriving per-decision explanations is through the generation of a saliency map, which visualizes the importance of different regions in the input image that contributed to the final decision. Saliency maps highlight the key image areas identified by the deep learning model [27]. Grad-CAM computes the weighted sum of feature maps based on their gradients, followed by the application of a ReLU activation function to generate the class activation map [7]. This map highlights the regions in the image that contributed most significantly to the model’s classification decision. By focusing on these high-impact areas, Grad-CAM provides valuable insights into the model’s attention and the reasoning behind its predictions. The visualization using saliency and Grad-CAM is depicted in Figure 5.
As shown in Figure 5, the first column displays the original cervical cell images, the second column shows the saliency maps, and the third column presents the Grad-CAM visualizations for both datasets—Pomeranian (left side) and SIPaKMeD (right side). The Pomeranian dataset includes NSIL samples in the first and second rows, and HSIL samples in the third and fourth rows. In the SIPaKMeD samples, the top two rows represent abnormal cells, while the bottom two show normal cells. The saliency maps and Grad-CAM visualizations consistently highlight regions with key morphological features. For NSIL samples (Pomeranian, rows 1 and 2), the model focuses primarily on the cytoplasmic region with less emphasis on the nucleus. In contrast, for HSIL samples (rows 3 and 4), there is strong model attention on the nuclear region, particularly areas showing nuclear enlargement and irregular contours, which are hallmarks of high-grade lesions. Similarly, in the SIPaKMeD dataset, the top two abnormal samples show heatmaps concentrated on the enlarged and hyperchromatic nuclei, consistent with pathological expectations. The normal cells in the bottom rows exhibit diffuse attention, indicating that the model is focusing more uniformly across the cytoplasm. These results demonstrate that the model is learning to prioritize diagnostically significant regions, particularly the nuclear morphology, when making predictions. This alignment with pathologist-defined criteria supports the model’s interpretability and reliability. It also validates that the visual explanation methods (saliency and Grad-CAM) reflect meaningful features, offering confidence in their use for model evaluation and refinement.
Table 5 presents a comparative analysis of our classification results with prior studies on the SIPaKMeD dataset. As shown in Table 5, our proposed SPP-SegNet model demonstrates superior performance in cervical cell segmentation compared to several existing methods. Traditional techniques such as FCM [3] achieved 85% precision on the Herlev dataset, while more advanced models like SE-ATT-Unet [16] and U-Net [17] reported precision scores of 92.32% and 95.04%, respectively. Although these models contribute valuable improvements, they often lack mechanisms for effective multiscale context aggregation. In contrast, our SPP-SegNet integrates spatial pyramid pooling and atrous convolution to enhance multiscale feature extraction, leading to higher accuracy (98.53% on Pomeranian and 94.15% on SIPaKMeD) and IoU scores (95.75% and 95.08%, respectively). These results confirm the model’s effectiveness and robustness in handling varying cervical cell morphologies and imaging challenges.
For classification, several deep learning architectures have been employed for the five-class task on the SIPaKMeD dataset. ResNet50 [28], a deep residual network designed to mitigate the vanishing gradient problem through skip connections, achieved 95% accuracy, demonstrating strong feature extraction capabilities. A basic CNN [29] reached 93% accuracy but lacked additional optimization techniques. To enhance feature selection, a CNN with principal component analysis (PCA) [30] was employed, improving accuracy slightly to 94% by reducing dimensionality and focusing on the most significant features. Another approach, CNN with a convolutional block attention module (CBAM) [31] incorporated attention mechanisms to refine spatial and channel feature selection, resulting in 92.8% accuracy with improved precision and recall. Additionally, for the two-class classification task, DenseNet121 [7], which leverages dense connectivity to strengthen gradient flow, achieved 95% in all performance metrics. In comparison, our proposed SE-DenseNet201 for two-class classification significantly outperforms previous methods, achieving 99% accuracy, precision, recall, and F1-score. Likewise, our approach for five-class classification attains 96% across all metrics, surpassing the existing methods and demonstrating the effectiveness of our model in capturing complex feature representations. Overall, the comparison demonstrates that our proposed models outperform existing methods, particularly in both binary and multi-class classification.
As a result, the proposed SPP-SegNet is effective at segmenting the entire cell from the background. Furthermore, the SE-DenseNet201 model achieves good results, particularly with the BBox method. This result may be helpful to cytotechnicians in identifying abnormal cells, allowing for a more accurate and efficient analysis. Improved segmentation can potentially aid in early detection and better diagnosis of cellular abnormalities.

5. Conclusions

In conclusion, the SPP-SegNet architecture significantly improves cervical cell segmentation in a single-cell Pap smear image, surpassing the performance of the standard SegNet model. By incorporating SPP and atrous convolution techniques, SPP-SegNet enhances the extraction of multiscale features, leading to more accurate and robust segmentation. Additionally, SE-DenseNet201 achieves better results in classifying cervical cells, with the segmentation output based on the BBox serving as input for classification. The results from both the Pomeranian and SIPaKMeD datasets demonstrate that the proposed method outperforms the standard method in all performance metrics. This improved segmentation accuracy is critical for the early detection of cervical cancer, as it helps to identify abnormal cells with greater precision. The proposed model shows great potential to automate Pap smear screening, reduce the workload of cytotechnicians, and minimize human error in diagnosing cervical abnormalities. Overall, SPP-SegNet and SE-DenseNet201 represent a valuable advance in the field of medical image analysis, offering a promising solution for the detection of cervical cancer. A limitation of the current study is the use of isolated single-cell images, which may not fully capture the complexity of real clinical samples. Future work will focus on extending the proposed method to multi-cell and whole-slide images, where challenges such as overlapping cells, heterogeneous staining, and large-scale image size must be addressed for effective clinical application. Additionally, clinical validation involving pathologists or cytotechnicians is needed to evaluate the system’s real-world performance and impact on diagnostic accuracy and workflow efficiency.

Author Contributions

Conceptualization, B.Z.W., A.R., and K.H.; methodology, B.Z.W., A.R., and K.H.; software, B.Z.W., A.R., and K.H.; validation, B.Z.W., A.R., and K.H.; investigation, B.Z.W., A.R., and K.H.; writing—original draft preparation, B.Z.W.; writing—review and editing, B.Z.W., A.R., and K.H.; visualization, B.Z.W.; supervision, A.R. and K.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The SIPaKMeD dataset is publicly available.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kurnianingsih; Allehaibi, K.H.S.; Nugroho, L.E.; Widyawan; Lazuardi, L.; Prabuwono, A.S.; Mantoro, T. Segmentation and classification of cervical cells using deep learning. IEEE Access 2019, 7, 116925–116941. [Google Scholar] [CrossRef]
  2. Li, J.; Dou, Q.; Yang, H.; Liu, J.; Fu, L.; Zhang, Y.; Zheng, L.; Zhang, D. Cervical cell multi-classification algorithm using global context information and attention mechanism. Tissue Cell 2022, 74, 101677. [Google Scholar] [CrossRef] [PubMed]
  3. Chankong, T.; Theera-Umpon, N.; Auephanwiriyakul, S. Automatic cervical cell segmentation and classification in Pap smears. Comput. Methods Programs Biomed. 2014, 113, 539–556. [Google Scholar] [CrossRef] [PubMed]
  4. Hussain, E.; Mahanta, L.B.; Das, C.R.; Choudhury, M.; Chowdhury, M. A shape context fully convolutional neural network for segmentation and classification of cervical nuclei in Pap smear images. Artif. Intell. Med. 2020, 107, 101897. [Google Scholar] [CrossRef]
  5. Bandyopadhyay, H.; Nasipuri, M. Segmentation of Pap Smear Images for Cervical Cancer Detection. In Proceedings of the 2020 IEEE Calcutta Conference CALCON 2020, Kolkata, India, 28–29 February 2020; pp. 30–33. [Google Scholar] [CrossRef]
  6. Desiani, A.; Suprihatin, B.; Yahdin, S.; Putri, A.I.; Husein, F.R. Bi-path Architecture of CNN Segmentation and Classification Method for Cervical Cancer Disorders Based on Pap-smear Images. IAENG Int. J. Comput. Sci. 2021, 48, 1–9. [Google Scholar]
  7. Wubineh, B.Z.; Rusiecki, A.; Halawa, K. Classification of cervical cells from the Pap smear image using the RES_DCGAN data augmentation and ResNet50V2 with self-attention architecture. Neural Comput. Appl. 2024, 36, 21801–21815. [Google Scholar] [CrossRef]
  8. Wubineh, B.Z.; Rusiecki, A.; Halawa, K. Segmentation and Classification Techniques for Pap Smear Images in Detecting Cervical Cancer: A Systematic Review. IEEE Access 2024, 12, 118195–118213. [Google Scholar] [CrossRef]
  9. Wubineh, B.Z.; Rusiecki, A.; Halawa, K. Segmentation of Cytology Images to Detect Cervical Cancer Using Deep Learning Techniques. In International Conference on Computational Science; Springer Nature: Cham, Switzerland, 2024; pp. 270–278. [Google Scholar]
  10. Wilhelmi, M.; Rusiecki, A. Simple CNN as an alternative for large pretrained models for medical image classification—MedMNIST case study. Procedia Comput. Sci. 2024, 239, 1298–1303. [Google Scholar] [CrossRef]
  11. Chlap, P.; Min, H.; Vandenberg, N.; Dowling, J.; Holloway, L.; Haworth, A. A review of medical image data augmentation techniques for deep learning applications. J. Med. Imaging Radiat. Oncol. 2021, 65, 545–563. [Google Scholar] [CrossRef]
  12. Wubineh, B.Z.; Rusiecki, A.; Halawa, K. Data Augmentation Techniques to Detect Cervical Cancer Using Deep Learning: A Systematic Review. In International Conference on Dependability of Computer Systems; Springer Nature: Cham, Switzerland, 2024; pp. 325–336. [Google Scholar]
  13. Zhang, C.; Lu, W.; Wu, J.; Ni, C.; Wang, H. SegNet Network Architecture for Deep Learning Image Segmentation and Its Integrated Applications and Prospects. Acad. J. Sci. Technol. 2024, 9, 224–229. [Google Scholar] [CrossRef]
  14. Afify, H.M.; Mohammed, K.K.; Hassanien, A.E. An improved framework for polyp image segmentation based on SegNet architecture. Int. J. Imaging Syst. Technol. 2021, 31, 1741–1751. [Google Scholar] [CrossRef]
  15. Elizar, E.; Zulkifley, M.A.; Muharar, R.; Zaman, M.H.M.; Mustaza, S.M. A Review on Multiscale-Deep-Learning Applications. Sensors 2022, 22, 7384. [Google Scholar] [CrossRef] [PubMed]
  16. Xiong, L.; Chen, C.; Lin, Y.; Song, Z.; Su, J. A Three-Step Automated Segmentation Method for Early Cervical Cancer MRI Images Based on Deep Learning. Int. J. Imaging Syst. Technol. 2024, 35, e23207. [Google Scholar] [CrossRef]
  17. Zhu, L.; Du, H. Instance Segmentation of Overlapping Cervical Cells Based on Boundary Tracking. In Proceedings of the Chinese Automation Congress CAC 2021, Beijing, China, 22–24 October 2021; pp. 7600–7604. [Google Scholar] [CrossRef]
  18. Teixeira, J.B.A.; Rezende, M.T.; Diniz, D.N.; Carneiro, C.M.; Luz, E.J.d.S.; Souza, M.J.F.; Ushizima, D.M.; de Medeiros, F.N.S.; Bianchi, A.G.C. Segmentation of cervical nuclei using convolutional neural network for conventional cytology. Comput. Methods Biomech. Biomed. Eng. Imaging Vis. 2023, 11, 1876–1888. [Google Scholar] [CrossRef]
  19. Song, Y.; Zhang, A.; Zhou, J.; Luo, Y.; Lin, Z.; Zhou, T. Overlapping cytoplasms segmentation via constrained multi-shape evolution for cervical cancer screening. Artif. Intell. Med. 2024, 148, 102756. [Google Scholar] [CrossRef]
  20. Vianna, P.; Farias, R.; Pereira, W.C.d.A. U-Net and SegNet performances on lesion segmentation of breast ultrasonography images. Res. Biomed. Eng. 2021, 37, 171–179. [Google Scholar] [CrossRef]
  21. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  22. Brahmbhatt, P. Skin Lesion Segmentation Using SegNet. 2019, pp. 14–15. Available online: https://www.kaggle.com/code/hashbanger/skin-lesion-segmentation-using-segnet#The-Model (accessed on 12 March 2025).
  23. Göker, H. Detection of Cervical Cancer From Uterine Cervix Images Using Transfer Learning Architectures. Eskişehir Tech. Univ. J. Sci. Technol. A-Appl. Sci. Eng. 2024, 25, 222–239. [Google Scholar] [CrossRef]
  24. Jeleń, Ł.; Stankiewicz-Antosz, I.; Chosia, M.; Jeleń, M. Optimizing Cervical Cancer Diagnosis with Feature Selection and Deep Learning. Appl. Sci. 2025, 15, 1458. [Google Scholar] [CrossRef]
  25. Wang, X.; Jing, S.; Dai, H.; Shi, A. High-resolution remote sensing images semantic segmentation using improved UNet and SegNet. Comput. Electr. Eng. 2023, 108, 108734. [Google Scholar] [CrossRef]
  26. Tang, D.; Li, C.; Du, T.; Jiang, H.; Ma, D.; Ma, Z.; Grzegorzek, M.; Jiang, T.; Sun, H. ECPC-IDS: A benchmark endometrial cancer PET/CT image dataset for evaluation of semantic segmentation and detection of hypermetabolic regions. Comput. Biol. Med. 2024, 171, 108217. [Google Scholar] [CrossRef] [PubMed]
  27. Hsu, C.-Y.; Li, W. Explainable GeoAI: Can saliency maps help interpret artificial intelligence’s learning process? An empirical study on natural feature detection. Int. J. Geogr. Inf. Sci. 2023, 37, 963–987. [Google Scholar] [CrossRef]
  28. Ozcan, C.; Akyıldız, B. Deep Learning Based Cervical Cells Classification from Pap Smear Images. In Proceedings of the 2nd International Conference on Information Technologies and Their Applications (ITTA 2024), Baku, Azerbaijan, 23–25 April 2024; Available online: https://www.researchgate.net/publication/386986535 (accessed on 12 March 2025).
  29. Arya, M.; Yadav, P.K.; Jain, A.; Gupta, M.; Agarwal, S.; Rajpoot, A.S. Neural Network to detect Cervical Cancer early Using Pap smears Images. In Proceedings of the International Conference on Contemporary Computing and Informatics (IC3I-2024), Greater Noida, India, 18–20 September 2024; Volume 7, pp. 324–329. [Google Scholar] [CrossRef]
  30. Al-Asbaily, S.A.; Almoshity, S.; Younus, S.; Bozed, K. Classification of Cervical Cancer using Convolutional Neural Networks. In Proceedings of the 2024 IEEE 4th International Maghreb Meeting of the Conference on Sciences and Techniques of Automatic Control and Computer Engineering (MI-STA), Tripoli city, Libya, 19–21 May 2024; pp. 735–739. [Google Scholar] [CrossRef]
  31. Cömert, Z.; Efil, F.; Türkoğlu, M. Convolutional Block Attention Module and Parallel Branch Architectures for Cervical Cell Classification. Int. J. Imaging Syst. Technol. 2025, 35, e70048. [Google Scholar] [CrossRef]
Figure 1. The standard SegNet architecture.
Figure 1. The standard SegNet architecture.
Cancers 17 02177 g001
Figure 2. The proposed segmentation and classification method.
Figure 2. The proposed segmentation and classification method.
Cancers 17 02177 g002
Figure 3. Segmentation results for the Pomeranian and SIPaKMeD dataset.
Figure 3. Segmentation results for the Pomeranian and SIPaKMeD dataset.
Cancers 17 02177 g003
Figure 4. Confusion matrix: (a) Pomeranian and (b) SIPaKMeD.
Figure 4. Confusion matrix: (a) Pomeranian and (b) SIPaKMeD.
Cancers 17 02177 g004
Figure 5. Saliency map and Grad-CAM visualization: (a) Pomeranian and (b) SIPaKMeD dataset.
Figure 5. Saliency map and Grad-CAM visualization: (a) Pomeranian and (b) SIPaKMeD dataset.
Cancers 17 02177 g005
Table 1. Details of the datasets used in this study.
Table 1. Details of the datasets used in this study.
DatasetCategoryNo. of ImgCategoriesImage SizeTotal Image
PomeranianHSIL124 1130 × 1130419
LSIL61
NSIL234
SIPaKMeDDyskeratotic813Abnormaldifferent4049
Koilocytotic825
Metaplastic793
Parabasal787Normal
Superficial–Intermediate831
Table 2. Segmentation results using the Pomeranian and SIPaKMeD dataset.
Table 2. Segmentation results using the Pomeranian and SIPaKMeD dataset.
DatasetMethodAccuracyPrecisionRecallIoU
PomeranianStandard SegNet97.86%91.9%90.31%94.55%
SPP-SegNet98.53%96.41%91.32%95.75%
SIPakMeDStandard SegNet90.95%92.40%90.05%92.89%
SPP-SegNet94.15%93.87%94.94%95.08%
HerlevStandard SegNet86.74%91.05%90.35%90.33%
SPP-SegNet87.58%91.95%91.67%91.13%
Table 3. Classification results using the Pomeranian and SIPaKMeD dataset in the ablation study.
Table 3. Classification results using the Pomeranian and SIPaKMeD dataset in the ablation study.
DatasetMethodResent50V2Densenet121Densenet201
AccPreRecF1AccPreRecF1AccPreRecF1
PomeranianOnly classification81%77%81%79%87%87%87%85%90%90%90%90%
ROI83%87%83%84%85%84%85%84%86%86%86%83%
BBox85%85%85%85%88%88%88%88%88%88%88%88%
SIPaKMeD MulticellOnly classification90%91%90%90%92%91%91%91%91%91%91%91%
ROI87%86%87%86%90%90%90%90%90%90%90%90%
BBox89%89%89%89%91%91%91%91%91%91%91%91%
SIPaKMeD BinaryOnly classification96%96%96%96%97%97%97%97%98%98%98%98%
ROI93%93%93%93%94%94%94%94%94%94%94%94%
BBox94%94%94%94%96%96%96%96%96%96%96%96%
Herlev BinaryOnly classification87%88%87%87%88%88%88%88%88%88%88%88%
ROI84%84%84%84%85%85%85%85%86%86%86%86%
BBox86%86%86%86%85%85%85%85%87%87%87%87%
Table 4. Classification results using the Pomeranian and SIPaKMeD dataset.
Table 4. Classification results using the Pomeranian and SIPaKMeD dataset.
DatasetMethodsAccuracyPrecisionRecallF1 Score
PomeranianSE-DesneNet201 ROI87%87%87%87%
SE-DesneNet201 BBox93%93%93%93%
SIPaKMeD
Multiclass
SE-DesneNet201 ROI94%94%94%94%
SE-DesneNet201 BBox96%96%96%96%
SIPaKMeD
Binary
SE-DesneNet201 ROI97%97%97%97%
SE-DesneNet201 BBox99%99%99%99%
Herlev BinarySE-DesneNet201 ROI89%90%89%88%
SE-DesneNet201 BBox90%91%90%90%
Table 5. Comparison of our results from the previous study using the SIPaKMeD dataset.
Table 5. Comparison of our results from the previous study using the SIPaKMeD dataset.
RefDatasetTaskClassMethodsAccuracyPrecisionRecallF1 ScoreIoU
[3]HerlevSegmentation-FCM-85%---
[16]Private dataSegmentation-SE-ATT-Unet-92.32%---
[17]TCT imageSegmentation-U-Net-95.04%96.19%-91.60%
[28]SIPaKMeDClassification5ResNet5095%----
[29]SIPaKMeDClassification5CNN93%----
[30]SIPaKMeDClassification5CNN + PCA94%----
[31]SIPaKMeDClassification5CNN_CBAM92.8%93%92.8%92.8%-
[7]SIPaKMeDClassification2DenseNet12195%95%95%95%-
OursSIPaKMeDClassification5SE-DenseNet20196%96%96%96%-
OursSIPaKMeDClassification2SE-DenseNet20199%99%99%99%-
OursPomeranianSegmentation-SPP-SegNet98.53%96.41%91.32%-95.75%
OursSIPaKMeDSegmentation-SPP-SegNet94.15%93.87%94.94%-95.08%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wubineh, B.Z.; Rusiecki, A.; Halawa, K. SPP-SegNet and SE-DenseNet201: A Dual-Model Approach for Cervical Cell Segmentation and Classification. Cancers 2025, 17, 2177. https://doi.org/10.3390/cancers17132177

AMA Style

Wubineh BZ, Rusiecki A, Halawa K. SPP-SegNet and SE-DenseNet201: A Dual-Model Approach for Cervical Cell Segmentation and Classification. Cancers. 2025; 17(13):2177. https://doi.org/10.3390/cancers17132177

Chicago/Turabian Style

Wubineh, Betelhem Zewdu, Andrzej Rusiecki, and Krzysztof Halawa. 2025. "SPP-SegNet and SE-DenseNet201: A Dual-Model Approach for Cervical Cell Segmentation and Classification" Cancers 17, no. 13: 2177. https://doi.org/10.3390/cancers17132177

APA Style

Wubineh, B. Z., Rusiecki, A., & Halawa, K. (2025). SPP-SegNet and SE-DenseNet201: A Dual-Model Approach for Cervical Cell Segmentation and Classification. Cancers, 17(13), 2177. https://doi.org/10.3390/cancers17132177

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop