Next Article in Journal
Machine Learning Recognition and Phase Velocity Estimation of Atmospheric Gravity Waves from OI 557.7 nm All-Sky Airglow Images
Previous Article in Journal
From Illusion to Insight: A Taxonomic Survey of Hallucination Mitigation Techniques in LLMs
Previous Article in Special Issue
An Explainable AI Framework for Stroke Classification Based on CT Brain Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Color Space Comparison of Isolated Cervix Cells for Morphology Classification

by
Irari Jiménez-López
,
José E. Valdez-Rodríguez
and
Marco A. Moreno-Armendáriz
*
Center for Computing Research, Instituto Politécnico Nacional, Mexico City 07738, Mexico
*
Author to whom correspondence should be addressed.
AI 2025, 6(10), 261; https://doi.org/10.3390/ai6100261
Submission received: 26 August 2025 / Revised: 2 October 2025 / Accepted: 3 October 2025 / Published: 7 October 2025
(This article belongs to the Special Issue AI in Bio and Healthcare Informatics)

Abstract

Cervical cytology processing involves the morphological analysis of cervical cells to detect abnormalities. In recent years, machine learning and deep learning algorithms have been explored to automate this process. This study investigates the use of color space transformations as a preprocessing technique to reorganize visual information and improve classification performance using isolated cell images. Twelve color space transformations were compared, including RGB, CMYK, HSV, Grayscale, CIELAB, YUV, the individual RGB channels, and combinations of these channels (RG, RB, and GB). Two classification strategies were employed: binary classification (normal vs. abnormal) and five-class classification. The SIPaKMeD dataset was used, with images resized to 256 × 256 pixels via zero-padding. Data augmentation included random flipping and ±10° rotations applied with a 50% probability, followed by normalization. A custom CNN architecture was developed, comprising four convolutional layers followed by two fully connected layers and an output layer. The model achieved average precision, recall, and F1-score values of 91.39%, 91.34%, and 91.31% for the five-class case, respectively, and 99.69%, 96.68%, and 96.89% for the binary classification, respectively; these results were compared with a VGG-16 network. Furthermore, CMYK, HSV, and the RG channel combination consistently outperformed other color spaces, highlighting their potential to enhance classification accuracy.

1. Introduction

Cervical cancer is one of the most dangerous cancers in women around the world and is the fourth most common cancer after breast, colorectal, and lung cancer. According to the International Agency for Research on Cancer of the World Health Organization (WHO), it is estimated that 660,000 people were diagnosed in 2022 [1]. In Mexico, cervical cancer is the second most common cancer and cause of death, only behind breast cancer, according to the National Institute of Statistics and Geography (INEGI) [2].
This type of cancer originates in the cervix, which is divided into the ectocervix, composed of squamous epithelial cells, and the endocervix, composed of columnar epithelial cells. The area where the ectocervix and endocervix meet is called the “transformation zone” and is the most likely site for malignant cells to develop from a persistent infection with the Human Papillomavirus (HPV). This disease is the most common cause of cervical cancer and causes cells to undergo morphological changes, such as increased nuclear size and alterations in cytoplasmic shape [3].
However, the progression from abnormal cells to an invasive carcinoma might occur 10 to 20 years after the first infection. Therefore, early detection helps stop the disease and prevent cancer proliferation; regular medical exams are essential in the battle against cervical cancer. Colposcopy and cytology are the most common tests for diagnosing abnormalities in cervical cells. The first procedure is a visual inspection in which a colposcope magnifies the cervix to examine it after applying an acetic acid solution [4].
Cervical cytology is an examination that detects abnormal cells under a microscope. Dr. George Papanicolaou was able to discern between normal and malignant cells in swabs smeared under microscopic slides in 1920. In 1943, he and Dr. Herbert Traut wrote that it is possible to classify normal and abnormal cells of the vagina and cervix; since then, this study has become the standard in cervical cancer screening. In Mexico, it is still the first screening test for the detection of cervical cancer [5]. However, the high accuracy of this exam has some technical limitations, as it requires an expert with extensive experience in detecting potential precancerous cells among thousands of cells to yield accurate results. It is a time-consuming test and a heavy workload. In this context, various scientific groups worldwide have investigated the automation of classification using Machine Learning (ML) and Deep Learning (DL) algorithms to create auxiliary tools for this analysis [6,7,8].
The automatic classification of cervical cytology samples according to their morphology through image processing involves several steps; however, the exact procedure may vary depending on the methodology and image type [9,10]. A first step is to highlight the characteristics that improve classification performance; these can be achieved through filtering, histogram operations, geometric transformations, and converting to different color spaces, among others.
From these techniques, switching between color spaces can be a powerful tool for analyzing images. A color space or color model is, in a simplified way, a vector field where each point corresponds to a specific color. When an image is transformed from one color space to another, it involves a change in the arrangement of such information. This transformation could lead to an inherent highlight of specific characteristics or attenuate some others, and it may allow Artificial Intelligence (AI) models to capture features that may not be as apparent in a specific space color as RGB, thereby contributing to achieving better performance metrics in detection and classification tasks. The most common and easiest to represent is the RGB color space, which has each of the additive primary colors (red, green, and blue) as the unit vectors.
Once this information has been remarked or at the very least highlighted, the next step may include feature extraction methods, usually calculus and measurements, to obtain specific information such as nucleus and cell size, ratio, perimeter, euclidean distances between extreme points for nucleus and cell pointed out, the smoothness or irregularity of the cell, position, and many others [11]. Then, ML or statistical methods are applied to sort the data.
Another feature extraction strategy is using DL models, which can capture abstract representations of the data and handle large-scale datasets. Typically, these models are trained or fine-tuned for classification or regression tasks. One of the advantages of DL techniques is their capacity to automatically extract features which can be complemented with classical ML or statistical algorithms, such as MLP, SVM, or RF, to categorize the data into classes [12]. Finally, an evaluation step must be applied to validate the results.
Convolutional Neural Networks (CNNs) are the most common network architecture in the field of DL. The main benefit is the automatic identification of relevant features without any human supervision. This characteristic has been applied to various fields, including medical imaging for detection, localization, segmentation, and classification in diseases such as breast, lung, liver, pancreas, brain, skin, cervical, and prostate cancers, among others. There is an extensive list of different architectures for sorting information in a medical context; however, AlexNet, VGG-16, ResNet, Inception, and EfficientNetB are among the most common models used with images in the RGB color space as input. In most cases, they are pre-trained on a large dataset, such as ImageNet, then a fine-tuning process is applied to complete the adjustment [8,13,14].
Automatic cytology image classification is a challenging task due to the high class variability of cell morphology, making it difficult for both traditional ML and DL models to generalize well. While most previous studies have relied mainly on the RGB color space, there is no clear evidence that this representation is the most suitable to highlight the relevant morphological or textural features of cytology cells. Exploring alternative color spaces may provide representations that emphasize discriminative features or reduce irrelevant variability, potentially leading to more robust and accurate classification. This work aims to determine whether specific transformations can enhance performance compared to the conventional RGB approach. A comparison of six color-space transformations, three RGB channel combinations, and the use of individual RGB channels for classifying cervical cell images into two and five classes was made, with a proposed CNN architecture and VGG-16 based on the SIPakMeD database. The key contributions of this study are as follows:
1
A comparative study of color spaces (RGB, CMYK, HSV, YUV, and CIELAB) was conducted to evaluate their impact on the classification performance of cervical cytology images.
2
The comparison included the combination of RGB channels (RG, RB, GB) and isolated channels (R, G, B) as input for the model.
3
We implemented a lightweight CNN with only four convolutional layers to classify the images into two and five categories, offering a computationally efficient alternative. For comparative purposes, the results were evaluated against those obtained with a VGG-16 model.
4
The discussion of the comparison focuses on analyzing which channel or channels provide the most helpful information for the classification.
Our implemented methodology can be found at the following link: https://github.com/IrariJL/Color-space-comparison-in-cervical-cytology-images (accessed on 25 August 2025).

2. Related Work

Numerous studies have explored the application of image processing and ML techniques for cervical cytology analysis in tasks such as detection, segmentation, and classification. In most cases, the CNN has been used due to its ability to extract features automatically; meanwhile, color transformation has received comparatively less attention. This section provides an overview of the literature related to cervical cytology analysis, with a focus on color transformation and its application to the tasks mentioned earlier.
Gowda et al. explored the importance of color transformation for image classification using the CIFAR-10, CIFAR-100, SVHN, and ImageNet datasets. The authors built an architecture comprising seven DenseNets, one for each color model (RGB, LAB, HSV, YUV, YCbCr, HED, and YIQ), with a final dense layer to predict the classes. They found that certain classes are better represented in some color spaces [15]. There are other studies related to the importance of choosing the color space to improve the performance of image colorization [16] and image compression [17]. Within the medical literature, Tellez et al. quantified the effects of stain color augmentation to classify histopathology images from four organs. Using the HSV color space, they implemented data augmentation techniques involving random modifications of the hue and saturation channels to simulate different color distributions; they concluded that color augmentation, either in the HSV or HED color space, should always be used to improve model robustness [18]. A comparative medical study was conducted by Valdez-Rodríguez et al. to detect glaucoma by estimating depth information and analyzing RGB channel combinations. They implemented methodologies using RGB, isolated channels, and channel combinations, which served as the input for the CNN, achieving good results with green and blue RGB channel combinations [19].
Aside from the relevance of color space transformations, isolated channels within any color space may contribute different amounts of information to the classification task. Lateef et al. investigated the role of individual color channels in classifying histopathological images of prostate and breast cancer. Using multiple color spaces, texture feature extraction, SVM-RFE for feature selection, and SVM for classification, they found that isolated channels in each color space had a distinct impact on classification accuracy [20]. Furthermore, some authors have suggested combining channels to capture the most relevant information. Rahmanimotlagh et al. evaluated the BACH breast cancer dataset using VGG, ResNet, Xception, and InceptionResNet for binary and four-class classification. They reported that fusing features from different convolutional layers improved results across nearly all models and color spaces tested (RGB, HSV, YUV). Across all experiments, a fully connected layer as the classifier outperformed an external SVM with an RBF kernel [21].
Nucleus and cytoplasm segmentation is a crucial analysis in cervical image processing, as it helps determine whether the cell is normal or abnormal. Alquran et al. proposed a method to segment the nucleus in the Herlev database using RGB and HSV color space channels. Histogram equalization was applied to each channel to enhance contrast and intensities. Following this, the images were binarized, yielding the best results with the green and value channels [22]. Kangkana Bora et al. employed a methodology with three color models (RGB, HSV, and YCbCr), three multi-resolution transform-based techniques for feature extraction (DWT, RT Type-I, and NSCT), two feature representation schemes, and two classifiers (LSSVM and MLP) to detect cervical dysplasia with an owner and Herlev database. They demonstrated that using YCbCr, NSCT (Non-subsampled Contourlet Transform), and MLP (Multilayer Perceptron) improved the classification results [23].
Marina et al. presented a dataset, SIPakMeD, consisting of 4049 labeled isolated cell images. These cells are classified into five classes depending on their appearance and morphology [24]. They also performed image classification using isolated cells in the RGB color space. A VGG-19 architecture was adapted for this purpose, and data augmentation techniques were implemented for training. The average accuracy was 95.35%; the classification rate was 96.8% for dyskeratotic cells, 98.32% for superficial–intermediate cells, 89.82% for koilocytotic cells, 94.07% for metaplastic cells, and 97.84% for parabasal cells.
Nazmul et al. conducted a study using the previously mentioned database, SIPaKMED, to classify cells into five classes. The first approach involved converting the images from RGB to HSV color space, as this separation of data into saturation and intensity of the color, based on its hue value, could improve the model’s performance. An original lightweight CNN model was proposed; images were processed in both RGB and HSV color spaces, reporting RGB F1-scores for each class of 96% for dyskeratotic, 93% for koilocytotic, 96% for metaplastic, 100% for parabasal, and 97% for superficial/intermediate. For the HSV, the F1-score for each class was 96% for dyskeratotic, 94% for koilocytotic, 98% for metaplastic, 100% for parabasal, and 98% for superficial/intermediate [25].
Prinyanka and Prabhpreet presented a study to develop a model to identify cells, combining pre-trained DL models and ML classifiers using the SiPakMED dataset. The first stage involved segmenting the region of interest by converting the color space from RGB to CIELab, utilizing the L channel, which contains the lightness information of the images. Subsequently, the K-means clustering method was employed for classification into three distinct clusters. To classify the information, a pre-trained ResNet-50 model was used, where the input images were re-scaled to different sizes: 64 × 64 , 224 × 224 , 512 × 512 stage 1, 512 × 512 stage 2, and 1024 × 1024 pixels. The following F1-scores were obtained: for 64 × 64 —dyskeratotic 86%, koilocytotic 89%, metaplastic 85%, parabasal 96%, and superficial/intermediate 96%; for 224 × 224 —dyskeratotic 95%, koilocytotic 96%, metaplastic 94%, parabasal 94%, and superficial/intermediate 98%; for 512 × 512 stage 1—dyskeratotic 95%, koilocytotic 96%, metaplastic 96%, parabasal 97%, and superficial/intermediate 98%; for 512 × 512 stage 2—dyskeratotic 95%, koilocytotic 96%, metaplastic 95%, parabasal 100%, and superficial/intermediate 98%; for 1024 × 1024 —dyskeratotic 97%, koilocytotic 95%, metaplastic 96%, parabasal 100%, and superficial/intermediate 98% [26].
Haryanto et al. conducted a comparison between a zero-padding implementation and its absence in various CNN architectures, utilizing the SIPakMeD dataset and the AlexNet architecture. The number of epochs increased from 500 to 5000 after the padding scheme was implemented in the images. The average accuracy of non-padding was 85% and 87% for the padding scheme; the classification rates using zero padding were 90% for dyskeratotic cells, 100% for superficial–intermediate cells, 98% for koilocytotic cells, 54% for metaplastic cells, and 95% for parabasal cells [27]. A fusion strategy was used by Manna et al. to classify the SIPakMeD dataset through ensemble learning [28]. They combined three CNN models: Inception v3, Xception, and DenseNet-169; then, a fuzzy ranking-based approach was employed to assign class probabilities. The average accuracies were 98.55% for binary classification and 95.43% for five sorted classes.

3. Methodology

Figure 1 illustrates the work’s block diagram. The proposed research leverages the re-accommodation of pixels after a color space transformation, individual channel selection, and combinations to enhance classification metrics. Yet, it focuses on the cases where abnormal images were sorted as normal, since misclassifying a sick cell image could be potentially more dangerous. The SIPakMeD database was selected for this research because it allows for classification based on cellular appearance and morphological characteristics.
The first stage is image preprocessing, where the images are resized to work in a specific format and size. Every image was resized into 256 × 256 pixels. Then, six strategies of color space transformation were applied, converting the images from RGB color space to HSV, CMYK, CIELAB, YUV, and grayscale models. RGB channel combinations were implemented, resulting in two-channel matrices corresponding to the red and green (RG), red and blue (RB), and green and blue (GB) channels. Additionally, individual RGB channels were used as input for a classification model.
The training model block corresponds to the proposed CNN or VGG-16 architectures that categorize the images into either two or five classes. To evaluate which color space transformation provided the most relevant information for the classification task, performance metrics such as the F1-score were calculated for each model. Additionally, Gradient-weighted Class Activation Mapping (Grad-CAM) was applied to visualize the regions of interest emphasized by the CNN during prediction.

3.1. SIPakMeD Dataset

The SIPakMeD dataset [24] consists of 4049 images manually cropped from isolated cell images obtained from 966 clusters of cells on Pap smear slides. The images were captured using a CCD camera (Infinity 1, Lumenera) adapted for use with an optical microscope (OLYMPUS BX53F).
Among the collected images, there are five types of cells: dyskeratotic, considered an abnormal cell, with the database containing 813 images of this class; koilocytotic, considered an abnormal cell, with the database containing 825 images of this class; parabasal cells, considered normal cells, with the database containing 787 images of this class; superficial intermediate, considered a normal cell, with the database containing 831 images of this class; metaplastic, considered a benign cell, with the database containing 793 images of this class; however, the presence of these types of cells is associated with precancerous lesions.
Under this context, there is a total of 1618 normal cell images and 2431 abnormal cell images. These will be used to train and test the proposed CNN for classification. It is worth noting that 26 of the 4049 images lacked information regarding the identification of the nucleus and cytoplasm (18 koilocytic, 1 metaplastic, and 7 superficial intermediate), resulting in only 4023 valuable images in the collection.

3.2. Image Preprocessing

The SIPakMeD dataset provides images with isolated cells of different sizes. The minimum and maximum sizes corresponded to dyskeratotic cells, with a size of 40 × 90 pixels, and superficial-intermediate cells, with a size of 553 × 388 pixels, respectively, with an average size of 300 × 239 pixels. The final size selected was 256 × 256 pixels, taking into account the average size, computational efficiency, and model comparability. A smaller input reduces memory usage, and this size allows for training of well-established DL architectures such as VGG or ResNet, facilitating benchmarking and future work. Resizing was performed using the bilinear interpolation method included in the OpenCV library [29], with different scales depending on the image size. Images larger than the selected size were reduced by a scale factor of less than one. While the images below are of the selected size, they were increased using a scale factor greater than one. To avoid distortion of the nucleus, zero padding was added to fit the images to a 256 × 256 × 3 matrix when they were close enough to the selected standard size. Therefore, resizing the image to the selected size did not require proportion changes. This should ensure that the rescaling process did not alter the distribution of pixel intensities, a finding later confirmed by the preservation of the histogram shape. The main effect was an increase or decrease in the total number of pixels. Figure A1 and Figure A2 in Appendix A show two histogram distributions of an original and resized sample for both cases.

3.3. Data Augmentation

One of the issues in DL models is the requirement for a large dataset to be trained. Medical image tasks can be challenging due to patient privacy concerns, data scarcity, and imbalanced data collection. Under this scenario, data augmentation is often employed to increase data, diversify it, reduce overfitting, and improve the model’s generalization [30]. In this study, data augmentation was applied before training the model through a random flip and a ±10° rotation, both of which were applied with a 50% probability. After this process, a normalization operation was carried out on every cell image by scaling pixel values from the original [0–255] range to [0–1] using the ToTensor function of PyTorch version 2.8, which also converts the image format from [H, W, C] to [C, H, W], where H, W, are the height and width of the images, respectively.

3.4. Color Space Transformations

The colors humans perceive in an object are the light reflected from the object. This light corresponds to the visible portion of the electromagnetic spectrum, which spans wavelengths from approximately 400 nm (blue) to 700 nm (red). For example, a yellow object reflects light with wavelengths in the range of 570 to 580 nm. Human color vision exists because of the 6 to 7 million cone cells in the retina, which are sensitive to three wavelength ranges corresponding to the colors red, green, and blue. As a result, people see color as the combination of these primary colors.
To distinguish among different colors, there are three principal properties: brightness, hue, and saturation. The brightness refers to the intensity of the color; hue is associated with the position in the visible spectrum; and the saturation refers to the amount of white light mixed with a hue. Color spaces were created to standardize and represent color components. Color space models are mathematical models that are able to represent colors as points in a coordinate space. They are widely used in multiple applications across fields such as computer vision, image processing, and computer graphics, as their properties can facilitate image processing tasks [31,32].

3.4.1. HSV Model

The HSV model is a cylindrical color model that represents colors based on hue (H, type of color), saturation (S, intensity), and value (V, brightness). The hue component goes from 0° to 360° to represent all colors. Saturation (S) represents the colorfulness. If a hue is further from the cylinder center, the color will be saturated, while if it is near the center, it will appear as shades of gray. The value (V) defines the brightness of the hue, causing the colors to change from black to white. The HSV model is intuitive for color selection in graphics and image editing because it provides a practical human interpretation compared to the RGB and CMYK models. However, its transformation is less intuitive than the CMYK model. Additionally, since information about the visible spectrum is contained in a single channel, the remaining channels provide unique details not available in other color spaces.
The procedure for converting an RGB image to HSV color space begins by normalizing the image range from [0, 255] to [0, 1]. Then, the maximum ( C m a x ) and minimum ( C m i n ) values from each RGB channel must be found, and their difference must be calculated ( Δ = C m a x C m i n ). The conversion formulas for H, S, and V channels are given in Equation (1), Equation (2), and Equation (3), respectively [33].
H = 0 , C max = 0 , 60 × G B Δ , C max = R , 60 × B R Δ + 120 , C max = G , 60 × R G Δ + 240 , C max = B .
S = Δ C m a x , C max = 0 , 0 , C max = 0 .
V = C m a x .

3.4.2. CMYK Model

The CMYK model is a subtractive color model in which the principal colors are the complementary pigments: cyan, magenta, yellow, and black (key). It works by absorbing light, with black (K) added to improve depth and contrast. The transformation from RGB to CMYK involves subtracting each component (R, G, and B) from white to create the C, M, and Y channels (that is, C = (1-R), M = (1-G), Y = (1-B)); the K channel is formed with minimum values of each C, M, and Y channel. Equation (4) shows the mathematical transformation for them. It is worth mentioning that the image must be normalized to the [0–1] range.
C = ( C K ) ( 1 k ) , M = ( M K ) ( 1 k ) , Y = ( Y K ) ( 1 k )
Figure 2 displays the decomposed channels of each color space, showing how each transformation encodes the chromatic and luminance information. This visualization is adapted to the screen it will be shown on or the paper it will be printed on, since the color space and output device cannot be separated.

3.5. Classification Model

CNNs have been the most used DL technique in medical image analysis. They can capture spatial hierarchies and extract features from the input through convolutional and pooling layers. However, a CNN architecture consists of more building blocks that lend training stability and generalization, such as nonlinear activation function layers, dropout, batch normalization, and fully connected layers. The combination of these operations enables the extraction and classification of complex features [34,35,36].
In this work, we designed a CNN model to analyze the dataset for both binary and five-class classification tasks. The LeNet model [37] served as the base for our proposed CNN; the parameters were determined through an exhaustive grid search, using the RGB space model as input. The network architecture consisted of four convolutional layers, followed by two fully connected layers and an output layer. The first convolutional layer used a 5 × 5 kernel with a stride of one. The subsequent convolutional layers employed a stride of two; the second layer used a 3 × 3 kernel, the third a 5 × 5 kernel, and the final convolutional layer also used a 3 × 3 kernel. Batch normalization was applied after each convolutional layer, and all layers employed the ReLU activation function. Max-pooling operations with a 3 × 3 window and a stride of one were used after each convolutional layer, except for the third max-pooling layer, which used a stride of two. The interleaved structure of large and small kernels was intended to capture both coarse features through larger receptive fields and fine-grained details through smaller ones, thereby enhancing feature representation across spatial scales. Mingxing Tan and colleagues studied the impact of using different kernel sizes, such as 3 × 3 , 5 × 5 , and 7 × 7 , and reported the improvement in metrics for using a different kernel in a single convolutional layer [38]. The use of big kernels had also been demonstrated as a powerful paradigm in contrast to the use of small kernels [39].
The previous layers perform the feature extraction from the input images. After this process, the resulting feature maps were concatenated to begin the classification technique. Two fully connected layers were used to sort the data, each with a ReLU activation function. The output layer consisted of either two or five neurons, depending on the strategy (binary or five-class classification) and the use of the Softmax activation function. Figure 3 shows the whole model, the kernel size, the number of neurons in each fully connected layer, and the feature map size output of each convolutional layer.

VGG-16

To compare and validate our proposal, we utilize the VGG-16 architecture [40], which consists of a deep CNN with small convolution filters in all layers. The model achieved a test data accuracy of 92.77% using ImageNet. They highlighted the importance of a deep model in visual representations.

4. Experiments and Results

All the experiments for our CNN architecture were run on a computer equipped with an Intel Core i7-13th generation processor and an NVIDIA RTX 4070 GPU with 8GB of RAM. The experiments using the VGG-16 model were implemented on a computer with an Intel Core i7-8th generation processor and two GPUs, 1080Ti, with 12 GB of RAM each. The algorithms were coded in Python 3.11.5 using the Pytorch framework. Table 1 summarizes the hyperparameters for both architectures.
A 5-fold cross-validation scheme was applied in all experiments. The data augmentation and color space transformations were applied online to save memory and automate the experiments.

4.1. Binary Classification Results

For the binary classification task, both the abnormal and normal classes consisted of 1611 images each. These datasets were used to evaluate 12 different color space transformations and RGB channel combinations, tested with both the proposed CNN and the VGG-16 architecture. For each color model and architecture, ten confusion matrices were generated. All results are available in our repository.
The best average F1-scores per fold obtained with the proposed CNN are as follows: fold 0 for RGB (0.9736), fold 3 for CMYK (0.9891), fold 0 for HSV (0.9891), fold 1 for CIELAB (0.9814), fold 3 for YUV (0.9674), fold 1 for grayscale (0.9549), fold 2 for RG (0.9627), fold 0 for RB (0.9720), fold 1 for GB (0.9798), fold 4 for the R channel (0.9519), fold 3 for the G channel (0.9564), and fold 4 for the B channel (0.9487).
For the VGG-16 model, the best average F1-scores per fold are as follows: fold 0 for RGB (0.9752), fold 0 for CMYK (0.9519), fold 0 for HSV (0.9674), fold 0 for CIELAB (0.9658), fold 1 for YUV (0.9550), fold 2 for grayscale (0.9518), fold 0 for RG (0.9566), fold 2 for RB (0.9534), fold 2 for GB (0.9658), fold 1 for the R channel (0.9083), fold 0 for the G channel (0.9347), and fold 2 for the B channel (0.9392).

4.2. Classification of the Five Classes

The results of the classification of the five types of cells are shown in Figure 4, where the best fold from RGB, CMYK, and HSV experiments corresponds to (a) fold 1, (b) fold 2, and (c) fold 2, respectively.
Figure 5 displays the best fold from CIELAB, YUV, and grayscale experiments, where (a) corresponds to fold 4, (b) to fold 3, and (c) to fold 0, respectively.
Figure 6 presents the best fold from the combinations of RGB color space channels from RG, RB, and GB experiments, where (a) corresponds to fold 1, (b) to fold 2, and (c) to fold 3, respectively.
Finally, Figure 7 shows the best fold from each RGB channel, R, G, and B experiments, where (a) corresponds to fold 1, (b) to fold 2, and (c) to fold 4, respectively.
The results of the VGG-16 models are shown in Figure 8, which presents the best results from the RGB, CMYK, and HSV color transformations, where (a) corresponds to fold 4, (b) to fold 2 and (c) to fold 3, respectively.

Grad-CAM

Additionally, the Grad-CAM algorithm [41] was used to select the color space based on the characteristics that each model chose to inform its decision. In Figure 9, an example of the relationship between the RGB, HSV, and CMYK color spaces for both architectures can be observed. Warmer colors (red/yellow) indicate areas with higher contributions to the decision, while cooler colors (blue) represent less relevant regions.

5. Discussion

To evaluate our methodology, three metrics were calculated: precision, recall, and F1-score. Table 2 reports the final metrics using the above confusion matrices, where A, B, C, D, and E correspond to dyskeratotic, koilocytotic, metaplastic, parabasal, and superficial–intermediate samples, respectively. CMYK and HSV had all their F1-scores above 90%, being consistent with the binary case. RGB, CIELAB, YUV, and RB combinations had four scores above 90%; in all of them, koilocytotic samples had the lowest value. However, when analyzing the values independently per class, there was no clear ranking or consistent pattern that suggested one color space outperformed the others across all classes.
The current analysis suggests that the metrics may not be a definitive criterion for defining the best color model. It is necessary to consider the misclassifications observed in the confusion matrices. The focus was oriented to False Positive (FP) results, that is, abnormal cells misclassified as normal samples, given that this confusion is more critical than a False Negative result. In this case, the upper-right quadrant of the confusion matrix is particularly relevant, where either dyskeratotic, koilocytic, or metaplastic cells are wrongly classified as parabasal or superficial–intermediate. To quantify this, two parameters were considered: (1) the total number of misclassifications in this quadrant and (2) the number of zero-valued cells (categories with no errors). Using the proposed CNN, HSV, and CMYK exhibited the best performance, followed by GB. For the VGG-16 model, HSV and CMYK also had the best results along with CIELAB. Both cases reported alternative color spaces that performed better than the original RGB of the images. Additionally, in both cases, HSV and CMYK were among the top three in performance.
For the binary case, a similar analysis was conducted. Through the proposed CNN model, HSV, CMYK, and GB combinations produced the lowest number of FP. In comparison, the VGG-16 architecture had the HSV, RG, and RB combinations with the lowest number of FP. This reinforces the idea that switching to alternative color spaces can enhance classification. Additionally, in both the binary and multiclass cases, HSV performed better than the RGB color space. However, since the ranking was not consistent in the one-on-one order comparison, it is suggested to consider not only the HSV color space but also the CMYK as viable options to differentiate between normal and abnormal cells.
Aside from differentiating between normal and abnormal cells, it is also essential to be able to identify specific abnormal cell types, as the presence of certain cells may indicate a particular condition. Focusing on the upper-left quadrant, the classification of these cell types in the HSV, CMYK, and CIELAB color spaces resulted in fewer misclassification cases for both models.
When comparing our results with those reported in [25] using the HSV color space, the F1-scores are similar, with a maximum difference of 3%. This suggests that the use of alternative color spaces is a practical approach for classifying cervical cytology. However, when considering the RGB color space, our metrics are comparatively lower. It could be attributed to the image resolution: using smaller image sizes might reduce the number of faded or noisy pixels, potentially improving classification metrics. Even so, the maximum difference is only 5%, which is not substantial enough to invalidate our approach. These findings support the relevance and robustness of our methodology for future research.
The Grad-CAM analysis highlights the specific area that the model focuses on for class differentiation; in Figure 9, the three color spaces show a main focus on the nucleus of the cell. However, in the proposed model, the RGB model also has a relatively high gradient in the cytoplasm. This dispersion of attention may be another factor contributing to its lower performance. Additionally, when comparing our model to the VGG-16 architecture, the attention focus is slightly displaced to the left in the VGG-16, emerging from the nucleus, whereas the proposed model focuses on the nucleus’s center.
Lastly, the computational time and cost of transforming color spaces were also taken into consideration. Among all, the RGB to CIELAB conversion required significantly more processing time than the rest. Given this, and its lower performance in comparison to false positives, CIELAB may not be suitable for practical implementation, despite its theoretical potential.

6. Conclusions

The proposed model’s obtained results are comparable to those reviewed in the state of the art. Having a simpler model with positive results may lead to a future portable application. As the results point out and the discussion synthesizes, changing the information contained in an image to a different color space can improve the classification results using DL models. In both cases, the binary and multiclass classification, the HSV color space obtained the highest performance, yet CMYK and the GB channel combination had equivalent results when differentiating normal and abnormal cells. In this context, it is advisable not only to work with the RGB color space; using alternative representations, such as HSV or CMYK, can emphasize features in the images that improve classification performance and provide more robust evaluation metrics. Color spaces such as CIELAB, YUV, and Grayscale also had notable performance compared to the original RGB; however, these did not match the performance of the ones mentioned before. The analysis of the independent channels of the RGB color space and the two-channel combinations of such showed that each channel from a color space may provide more or less relevant information for the classification task. Furthermore, the information provided to the DL model does not need to encompass the entire space, as is the case in this study, where only portions of it are considered.
A future study may involve combining channels from different spaces to maximize the valuable information obtained, both for differentiating between normal and abnormal cells and for specifically identifying abnormal cells. Additionally, the current analysis and biological background indicate that the nucleus is the primary characteristic that differentiates a normal cell from an abnormal one. Exploring the bases for implementing binary masks for classification may be beneficial.

Author Contributions

Conceptualization, I.J.-L. and J.E.V.-R.; methodology, J.E.V.-R.; software, I.J.-L.; validation, M.A.M.-A.; formal analysis, M.A.M.-A. and J.E.V.-R.; investigation, I.J.-L.; resources, I.J.-L. and J.E.V.-R.; data curation, I.J.-L.; writing—original draft preparation, I.J.-L.; writing—review and editing, I.J.-L., J.E.V.-R. and M.A.M.-A.; visualization, I.J.-L.; supervision, J.E.V.-R. and M.A.M.-A.; project administration, J.E.V.-R.; funding acquisition, M.A.M.-A. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Instituto Politécnico Nacional (IPN) through Secretaría de Investigación y Posgrado (SIP-IPN) research grants SIP-20250071 and SIP-20251352; Comisión de Operación y Fomento de Actividades Académicas del IPN (IPN-COFAA) and Programa de Estímulos al Desempeño de los Investigadores (IPN-EDI) and Secretaría de Ciencia, Humanidades, Tecnología e Innovación, Sistema Nacional de Investigadores (SECIHTI-SNII).

Institutional Review Board Statement

This study did not require ethical approval from the Ethics Committee of the IPN because we did not conduct medical experiments involving humans.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used in this study are available from the proprietary authors at: https://www.cs.uoi.gr/~marina/sipakmed.html (accessed on 25 August 2025).

Acknowledgments

We are grateful to Instituto Politécnico Nacional (IPN) and Secretaría de Ciencia, Humanidades, Tecnología e Innovación, for the support and resources provided to complete this research.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Histogram Shape

In this appendix, we present two histograms corresponding to a sample that increases its size and an image that decreases its size. Figure A1 shows the histogram of a dyskeratotic sample that increases the size by a 1.2 coefficient, and Figure A2 presents a superficial–intermediate sample that was decreased in size by a 0.5 coefficient. It can be observed that the overall shape of the intensity distribution, in both cases, remains essentially unchanged. This may indicate that the resizing process only increases the number of pixels but does not alter the underlying statistical properties of the image. However, the ratios between nucleus and cytoplasm of the original and re-sized images were calculated; if the image is not experiencing any distortion or significant deformation, these measurements should remain similar through the resizing process. To calculate the area, the contourArea function from the OpenCV library was used. Table A1 shows the results for this evaluation, where the differences between the measurements of the original and resized image are 6.30 % and 0.52 % for the dyskeratotic and superficial–intermediate samples, respectively.
Figure A1. Histogram distribution for each RGB channel of a dyskeratotic sample, (a) RGB image, (b) Red channel histogram comparison between the resized and original image, (c) Green channel histogram comparison between the resized and original image, (d) Blue channel histogram comparison between the resized and original image.
Figure A1. Histogram distribution for each RGB channel of a dyskeratotic sample, (a) RGB image, (b) Red channel histogram comparison between the resized and original image, (c) Green channel histogram comparison between the resized and original image, (d) Blue channel histogram comparison between the resized and original image.
Ai 06 00261 g0a1
Figure A2. Histogram distribution for each RGB channel of a superficial–intermediate sample, (a) RGB image, (b) Red channel histogram comparison between the resized and original image, (c) Green channel histogram comparison between the resized and original image, (d) Blue channel histogram comparison between the resized and original image.
Figure A2. Histogram distribution for each RGB channel of a superficial–intermediate sample, (a) RGB image, (b) Red channel histogram comparison between the resized and original image, (c) Green channel histogram comparison between the resized and original image, (d) Blue channel histogram comparison between the resized and original image.
Ai 06 00261 g0a2
Table A1. Measurements.
Table A1. Measurements.
DyskeratoticSuperficial–Intermediate
OriginalResizedOriginal Resized
Nucleus area662.510571765444.5
Cytoplasm area47127043.599,922.525,032.5
Nucleus/cytoplasm0.14060.15010.01770.0178

References

  1. Ferlay, J.; Ervik, M.; Lam, F.; Laversanne, M.; Colombet, M.; Mery, L.; Piñeros, M.; Znaor, A.; Soerjomataram, I.; Bray, F. Global Cancer Observatory: Cancer Today. 2024. Available online: https://gco.iarc.who.int/today (accessed on 7 July 2025).
  2. INEGI. Comunicado de Prensa 39/25: Estadísticas a Propósito del Día Mundial Contra el Cáncer. 2025. Available online: https://www.inegi.org.mx/contenidos/saladeprensa/aproposito/2025/EAP_DMvsCancer25.pdf (accessed on 7 July 2025).
  3. Burmeister, C.A.; Khan, S.F.; Schäfer, G.; Mbatani, N.; Adams, T.; Moodley, J.; Prince, S. Cervical cancer therapies: Current challenges and future perspectives. Tumour Virus Res. 2022, 13, 200238. [Google Scholar] [CrossRef]
  4. Bedell, S.L.; Goldstein, L.S.; Goldstein, A.R.; Goldstein, A.T. Cervical cancer screening: Past, present, and future. Sex. Med. Rev. 2020, 8, 28–37. [Google Scholar] [CrossRef] [PubMed]
  5. León-Maldonado, L.S.; López-Olmedo, N.; Murillo, R.; Hurtado-Salgado, E.; Allen-Leigh, B.; Armengol-Alonso, A.; Torres-Mejía, G.; Bautista-Arredondo, S.; Barrientos-Gutiérrez, T.; Lazcano-Ponce, E. Tamizaje del cáncer cervical. Salud Pública México 2024, 66, 549–555. [Google Scholar] [CrossRef] [PubMed]
  6. Debelee, T.G.; Kebede, S.R.; Schwenker, F.; Shewarega, Z.M. Deep Learning in Selected Cancers’ Image Analysis—A Survey. J. Imaging 2020, 6, 121. [Google Scholar] [CrossRef]
  7. Sambyal, D.; Sarwar, A. Recent developments in cervical cancer diagnosis using deep learning on whole slide images: An Overview of models, techniques, challenges and future directions. Micron 2023, 173, 103520. [Google Scholar] [CrossRef] [PubMed]
  8. Sarhangi, H.A.; Beigifard, D.; Farmani, E.; Bolhasani, H. Deep learning techniques for cervical cancer diagnosis based on pathology and colposcopy images. Inform. Med. Unlocked 2024, 47, 101503. [Google Scholar] [CrossRef]
  9. Kieu, S.T.H.; Bade, A.; Hijazi, M.H.A.; Kolivand, H. A Survey of Deep Learning for Lung Disease Detection on Medical Images: State-of-the-Art, Taxonomy, Issues and Future Directions. J. Imaging 2020, 6, 131. [Google Scholar] [CrossRef]
  10. He, W.; Liu, T.; Han, Y.; Ming, W.; Du, J.; Liu, Y.; Yang, Y.; Wang, L.; Jiang, Z.; Wang, Y.; et al. A review: The detection of cancer cells in histopathology based on machine vision. Comput. Biol. Med. 2022, 146, 105636. [Google Scholar] [CrossRef]
  11. Jeleń, Ł.; Stankiewicz-Antosz, I.; Chosia, M.; Jeleń, M. Optimizing Cervical Cancer Diagnosis with Feature Selection and Deep Learning. Appl. Sci. 2025, 15, 1458. [Google Scholar] [CrossRef]
  12. Alzubaidi, L.; Zhang, J.; Humaidi, A.J.; Al-Dujaili, A.; Duan, Y.; Al-Shamma, O.; Santamaría, J.; Fadhel, M.A.; Al-Amidie, M.; Farhan, L. Review of deep learning: Concepts, CNN architectures, challenges, applications, future directions. J. Big Data 2021, 8, 53. [Google Scholar] [CrossRef]
  13. Gao, J.; Jiang, Q.; Zhou, B.; Chen, D. Convolutional neural networks for computer-aided detection or diagnosis in medical image analysis: An overview. Math. Biosci. Eng. 2019, 16, 6536–6561. [Google Scholar] [CrossRef]
  14. Painuli, D.; Bhardwaj, S.; köse, U. Recent advancement in cancer diagnosis using machine learning and deep learning techniques: A comprehensive review. Comput. Biol. Med. 2022, 146, 105580. [Google Scholar] [CrossRef] [PubMed]
  15. Gowda, S.N.; Yuan, C. ColorNet: Investigating the importance of color spaces for image classification. In Proceedings of the Asian Conference on Computer Vision, Perth, Australia, 2–6 December 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 581–596. [Google Scholar]
  16. Ballester, C.; Bugeau, A.; Carrillo, H.; Clément, M.; Giraud, R.; Raad, L.; Vitoria, P. Influence of Color Spaces for Deep Learning Image Colorization. arXiv 2022, arXiv:2204.02850. [Google Scholar] [CrossRef]
  17. Prativadibhayankaram, S.; Panda, M.P.; Seiler, J.; Richter, T.; Sparenberg, H.; Fößel, S.; Kaup, A. A Study on the Effect of Color Spaces in Learned Image Compression. arXiv 2024, arXiv:2406.13709. [Google Scholar] [CrossRef]
  18. Tellez, D.; Litjens, G.; Bándi, P.; Bulten, W.; Bokhorst, J.M.; Ciompi, F.; van der Laak, J. Quantifying the effects of data augmentation and stain color normalization in convolutional neural networks for computational pathology. Med. Image Anal. 2019, 58, 101544. [Google Scholar] [CrossRef] [PubMed]
  19. Valdez-Rodríguez, J.E.; Felipe-Riverón, E.M.; Calvo, H. Optic Disc Preprocessing for Reliable Glaucoma Detection in Small Datasets. Mathematics 2021, 9, 2237. [Google Scholar] [CrossRef]
  20. Lateefa, A.K.; Sahran, S.; Hamzah, A.E.; Qasim, H.H.; Albashih, D.; Sapiee, N.M. Comparative Analysis of Color Space in Histopathology Image Classification. J. Kejuruter. 2025, 37, 617–634. [Google Scholar] [CrossRef]
  21. Rahmanimotlagh, S.; Rahmanimotlagh, F.; Khalilzadeh, M.M. Multi-Dimensional Color Space Analysis with Deep Neural Network Architectures for Precision Breast Cancer Diagnosis. Future Res. AI IoT 2025, 38–51. [Google Scholar]
  22. Alquran, H.; Badarneh, A.; Alsalatie, M.; Mustafa, W.A.; Alquraan, B. Nucleus region detection of cervical cytology using image channel conversion techniques. In Proceedings of the AIP Conference Proceedings, Bali, Indonesia, 5–6 October 2023; AIP Publishing LLC: Melville, NY, USA, 2024; Volume 3135, p. 020013. [Google Scholar]
  23. Bora, K.; Mahanta, L.B.; Borah, K.; Chyrmang, G.; Barua, B.; Mallik, S.; Das, H.S.; Zhao, Z. Machine Learning Based Approach for Automated Cervical Dysplasia Detection Using Multi-Resolution Transform Domain Features. Mathematics 2022, 10, 4126. [Google Scholar] [CrossRef]
  24. Plissiti, M.E.; Dimitrakopoulos, P.; Sfikas, G.; Nikou, C.; Krikoni, O.; Charchanti, A. Sipakmed: A new dataset for feature and image based classification of normal and pathological cervical cells in pap smear images. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; IEEE: Piscataway Township, NJ, USA, 2018; pp. 3144–3148. [Google Scholar]
  25. Islam, M.N.; Faruk, M.F.; Srizon, A.Y.; Hasan, S.M.M.; Mamun, M.A.; Hossain, M.A.; Islam, M.A. HSV-Driven Light-Weight CNN Architecture: Balancing Computational Complexity and Diagnostic Precision in Cervical Cancer Detection. In Proceedings of the 2024 27th International Conference on Computer and Information Technology (ICCIT), Cox’s Bazar, Bangladesh, 20–22 December 2024; pp. 1182–1187. [Google Scholar] [CrossRef]
  26. Mahajan, P.; Kaur, P. Improving cervical cancer classification in PAP smear images with enhanced segmentation and deep progressive learning-based techniques. Diagn. Cytopathol. 2024, 52, 313–324. [Google Scholar] [CrossRef]
  27. Haryanto, T.; Sitanggang, I.S.; Agmalaro, M.A.; Rulaningtyas, R. The utilization of padding scheme on convolutional neural network for cervical cell images classification. In Proceedings of the 2020 International Conference on Computer Engineering, Network, and Intelligent Multimedia (CENIM), Surabaya, Indonesia, 17–18 November 2020; IEEE: Piscataway Township, NJ, USA, 2020; pp. 34–38. [Google Scholar]
  28. Manna, A.; Kundu, R.; Kaplun, D.; Sinitca, A.; Sarkar, R. A fuzzy rank-based ensemble of CNN models for classification of cervical cytology. Sci. Rep. 2021, 11, 14538. [Google Scholar] [CrossRef]
  29. Bradski, G. The OpenCV Library. Dr. Dobb’s J. Softw. Tools 2000, 25, 120–123. [Google Scholar]
  30. Wubineh, B.; Rusiecki, A.; Halawa, K. A systematic review of effective data augmentation in cervical cancer detection. Int. J. Electron. Telecommun. 2025, 71, 369–377. [Google Scholar] [CrossRef]
  31. Gonzalez, R.C.; Woods, R.E. Digital Image Processing, 3rd ed.; Prentice-Hall, Inc.: Saddle River, NJ, USA, 2006. [Google Scholar]
  32. Ibraheem, N.A.; Hasan, M.M.; Khan, R.Z.; Mishra, P.K. Understanding color models: A review. ARPN J. Sci. Technol. 2012, 2, 265–275. [Google Scholar]
  33. James, I.S.P. Face Image retrieval with HSV color space using clustering techniques. SIJ Trans. Comput. Sci. Eng. Appl. 2013, 1, 17–20. [Google Scholar] [CrossRef]
  34. Shobayo, O.; Saatchi, R. Developments in Deep Learning Artificial Neural Network Techniques for Medical Image Analysis and Interpretation. Diagnostics 2025, 15, 1072. [Google Scholar] [CrossRef]
  35. Mienye, I.D.; Swart, T.G.; Obaido, G.; Jordan, M.; Ilono, P. Deep Convolutional Neural Networks in Medical Image Analysis: A Review. Information 2025, 16, 195. [Google Scholar] [CrossRef]
  36. Matsuzaka, Y.; Yashiro, R. The Diagnostic Classification of the Pathological Image Using Computer Vision. Algorithms 2025, 18, 96. [Google Scholar] [CrossRef]
  37. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-Based Learning Applied to Document Recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  38. Tan, M.; Le, Q.V. Mixconv: Mixed depthwise convolutional kernels. arXiv 2019, arXiv:1907.09595. [Google Scholar] [CrossRef]
  39. Ding, X.; Zhang, X.; Han, J.; Ding, G. Scaling up your kernels to 31 × 31: Revisiting large kernel design in cnns. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11963–11975. [Google Scholar]
  40. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  41. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
Figure 1. Block diagram of methodology.
Figure 1. Block diagram of methodology.
Ai 06 00261 g001
Figure 2. Channel representation for each color space.
Figure 2. Channel representation for each color space.
Ai 06 00261 g002
Figure 3. CNN architecture to classify two or five categories.
Figure 3. CNN architecture to classify two or five categories.
Ai 06 00261 g003
Figure 4. Fold confusion matrices for (a) RGB, (b) CMYK, and (c) HSV color spaces.
Figure 4. Fold confusion matrices for (a) RGB, (b) CMYK, and (c) HSV color spaces.
Ai 06 00261 g004
Figure 5. Fold confusion matrices for (a) LAB, (b) YUV, and (c) grayscale color spaces.
Figure 5. Fold confusion matrices for (a) LAB, (b) YUV, and (c) grayscale color spaces.
Ai 06 00261 g005
Figure 6. Fold confusion matrices for (a) RG, (b) RB, and (c) GB combinations.
Figure 6. Fold confusion matrices for (a) RG, (b) RB, and (c) GB combinations.
Ai 06 00261 g006
Figure 7. Fold confusion matrices for (a) R, (b) G, and (c) B channels.
Figure 7. Fold confusion matrices for (a) R, (b) G, and (c) B channels.
Ai 06 00261 g007
Figure 8. Fold confusion matrices for (a) RGB, (b) CMYK, and (c) HSV using VGG-16 model.
Figure 8. Fold confusion matrices for (a) RGB, (b) CMYK, and (c) HSV using VGG-16 model.
Ai 06 00261 g008
Figure 9. Heatmaps for RGB, HSV, and CMYK color spaces using the VGG-16 model and the proposed model.
Figure 9. Heatmaps for RGB, HSV, and CMYK color spaces using the VGG-16 model and the proposed model.
Ai 06 00261 g009
Table 1. Hyperparameters for both architectures.
Table 1. Hyperparameters for both architectures.
Hyperparameters
Learning Rate0.001
Batch size32
OptimizerADAM
Loss functionCross-entropy
Epochs200
Table 2. Metrics of multiclass problem.
Table 2. Metrics of multiclass problem.
Own CNNVGG-16
Color Space Class Precision Recall F1-Score Precision Recall F1-Score
RGBA0.90630.92950.91770.92640.92640.9264
B0.89380.87200.88270.85540.89870.8765
C0.94380.93210.93790.95300.89310.9221
D0.95070.95740.95410.98760.95780.9725
E0.96170.96700.96440.93940.98100.9598
CMYKA0.96320.96320.96320.93680.97600.9560
B0.91030.93620.92310.83910.90680.8716
C0.95390.95390.95390.92250.87920.9003
D0.98800.97060.97920.99260.88240.9343
E0.99440.98880.99160.94410.96570.9548
HSVA0.94090.97770.95890.95540.96150.9585
B0.92210.90450.91320.89030.87900.8846
C0.97280.95330.96300.91330.96480.9384
D0.99370.98750.99060.97590.96430.9701
E0.96860.96860.96860.97730.95030.9636
CIELABA0.92590.95540.94040.93980.90170.9204
B0.89610.87900.88750.78160.87180.8242
C0.93460.91080.92260.87260.88390.8782
D0.98290.95560.96900.97450.94440.9592
E0.95000.99350.97120.96690.91820.9419
YUVA0.89360.93330.91300.86960.91430.8914
B0.90670.80950.85530.81930.80470.8119
C0.89740.91500.90610.80370.90970.8534
D0.94290.96350.95310.91280.89470.9037
E0.96470.98800.97620.96670.84300.9006
GrayscaleA0.84560.78770.81560.80810.80350.8058
B0.88820.78950.83590.69180.82090.7509
C0.74730.91280.82180.79560.75170.7730
D0.95480.94870.95180.98190.87170.9235
E0.95560.93990.94770.92940.95760.9433
RGA0.93080.91360.92210.90860.96360.9353
B0.84570.88830.86650.84760.82740.8373
C0.91470.90080.90770.80700.88460.8440
D0.96180.97420.96790.97140.90670.9379
E0.96510.93260.94860.96750.90300.9342
RBA0.95560.92140.93820.89940.93250.9157
B0.87100.88240.87660.77380.82800.8000
C0.93790.89880.91790.82660.85630.8412
D0.96530.97660.97090.95540.89820.9259
E0.94480.98840.96610.94160.86000.8990
GBA0.90000.88820.89400.89570.84880.8716
B0.82820.88240.85440.74430.80370.7729
C0.89510.87350.88410.77500.83220.8026
D1.00000.97500.98730.93040.91880.9245
E0.95380.95380.95380.97300.89440.9320
RA0.87790.87120.87450.82520.76130.7919
B0.84420.77840.81000.75000.69770.7229
C0.83630.88270.85890.79290.86450.8272
D0.90360.94940.92590.82350.90320.8615
E0.96720.95160.95930.95090.92260.9366
GA0.77500.82670.80000.82420.86710.8451
B0.86670.74710.80250.74840.74360.7460
C0.76110.82040.78960.75350.66880.7086
D0.95830.92620.94200.89870.94670.9221
E0.91810.95150.93450.92260.93370.9281
BA0.87260.81550.84310.78920.85880.8225
B0.73290.77120.75160.73680.62920.6788
C0.82930.81440.82180.71150.70700.7093
D0.93620.93620.93620.86490.91430.8889
E0.93920.97140.95510.92640.94970.9379
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jiménez-López, I.; Valdez-Rodríguez, J.E.; Moreno-Armendáriz, M.A. Color Space Comparison of Isolated Cervix Cells for Morphology Classification. AI 2025, 6, 261. https://doi.org/10.3390/ai6100261

AMA Style

Jiménez-López I, Valdez-Rodríguez JE, Moreno-Armendáriz MA. Color Space Comparison of Isolated Cervix Cells for Morphology Classification. AI. 2025; 6(10):261. https://doi.org/10.3390/ai6100261

Chicago/Turabian Style

Jiménez-López, Irari, José E. Valdez-Rodríguez, and Marco A. Moreno-Armendáriz. 2025. "Color Space Comparison of Isolated Cervix Cells for Morphology Classification" AI 6, no. 10: 261. https://doi.org/10.3390/ai6100261

APA Style

Jiménez-López, I., Valdez-Rodríguez, J. E., & Moreno-Armendáriz, M. A. (2025). Color Space Comparison of Isolated Cervix Cells for Morphology Classification. AI, 6(10), 261. https://doi.org/10.3390/ai6100261

Article Metrics

Back to TopTop