Next Article in Journal
Bone Mineral Density (BMD) Assessment Using Dual-Energy CT with Different Base Material Pairs (BMPs)
Previous Article in Journal
Implantation of an Artificial Intelligence Denoising Algorithm Using SubtlePET™ with Various Radiotracers: 18F-FDG, 68Ga PSMA-11 and 18F-FDOPA, Impact on the Technologist Radiation Doses
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MST-AI: Skin Color Estimation in Skin Cancer Datasets

1
Electrical and Computer Engineering Department, Temple University, Philadelphia, PA 19122, USA
2
Cancer Epigenetics Institute, Nuclear Dynamics and Cancer Program, Fox Chase Cancer Center, Philadelphia, PA 19111, USA
3
Department of Medical Genetics and Molecular Biochemistry, Lewis Katz School of Medicine (LKSOM), Temple University Health System, Philadelphia, PA 19140, USA
4
Department of Cancer and Cellular Biology, Lewis Katz School of Medicine (LKSOM), Temple University Health System, Philadelphia, PA 19140, USA
5
Cancer Prevention and Control Program, Fox Chase Cancer Center, Philadelphia, PA 19111, USA
6
Division of Dermatology, Melanoma and Skin Cancer Program, Fox Chase Cancer Center, Philadelphia, PA 19111, USA
7
Department of Sociology, Harvard University, Cambridge, MA 02138, USA
*
Author to whom correspondence should be addressed.
J. Imaging 2025, 11(7), 235; https://doi.org/10.3390/jimaging11070235
Submission received: 28 May 2025 / Revised: 4 July 2025 / Accepted: 9 July 2025 / Published: 13 July 2025
(This article belongs to the Section AI in Imaging)

Abstract

The absence of skin color information in skin cancer datasets poses a significant challenge for accurate diagnosis using artificial intelligence models, particularly for non-white populations. In this paper, based on the Monk Skin Tone (MST) scale, which is less biased than the Fitzpatrick scale, we propose MST-AI, a novel method for detecting skin color in images of large datasets, such as the International Skin Imaging Collaboration (ISIC) archive. The approach includes automatic frame, lesion removal, and lesion segmentation using convolutional neural networks, and modeling normal skin tones with a Variational Bayesian Gaussian Mixture Model (VB-GMM). The distribution of skin color predictions was compared with MST scale probability distribution functions (PDFs) using the Kullback-Leibler Divergence (KLD) metric. Validation against manual annotations and comparison with K-means clustering of image and skin mean RGBs demonstrated the superior performance of the MST-AI, with Kendall’s Tau, Spearman’s Rho, and Normalized Discounted Cumulative Gain (NDGC) of 0.68, 0.69, and 1.00, respectively. This research lays the groundwork for developing unbiased AI models for early skin cancer diagnosis by addressing skin color imbalances in large datasets.

1. Introduction

Skin cancer is the fifth most prevalent form of cancer in the United States, with current estimates suggesting that one in five Americans will develop it in their lifetime [1,2]. Among the various types of skin cancer, melanoma is of particular concern due to its potential for metastasis and poor prognosis in certain populations. Although melanoma can affect individuals of all skin tones, its incidence is notably higher among white populations than among other racial and ethnic groups. For instance, the lifetime risk of developing melanoma is approximately 3% (1 in 33) for White individuals, compared to just 0.1% (1 in 1000) for Black individuals and 0.5% (1 in 200) for Hispanic individuals. Furthermore, melanoma incidence rates are 1.3 per 80,000 for Asian/Pacific Islanders and 10.7 per 80,000 for Native Americans [3,4,5,6].
Despite the lower incidence in individuals with darker skin tones, skin cancer, including melanoma, can still occur in these populations, and the prognosis is often worse due to delayed diagnosis [7]. Research has shown that patients with darker skin tones are more likely to be diagnosed at later stages of melanoma, making treatment more challenging and less effective [8]. In total, 12.6% of non-Hispanic Whites (NHWs) presented with regional or distant disease, compared to 21.0% of Hispanics and 34.1% of non-Hispanic Blacks (NHBs), who were diagnosed at more advanced stages [9]. In fact, 22% of melanoma cases in African−American patients are diagnosed when the cancer has spread to nearby lymph nodes, and 14% are diagnosed when it has metastasized to distant sites [10]. Moreover, individuals with darker skin tones are prone to melanoma in areas that are less commonly exposed to the sun, such as the palms of the hands, soles of the feet, groin, inside the mouth, and under the nails [11,12,13,14]. This disparity underscores the importance of improving early detection, awareness, and targeted interventions for skin cancer in all populations, regardless of skin type.
Recent research on skin cancer diagnosis has explored advanced imaging and AI techniques to improve detection across diverse skin types. Huang et al. [15] introduced a hyperspectral imaging approach combined with YOLOv5 [16] for skin cancer classification, demonstrating improved accuracy through detailed spectral information. Lin et al. [17,18] assessed the performance of the Spectrum-Aided Vision Enhancer (SAVE) system in detecting various melanoma subtypes, highlighting its potential utility across different lesion presentations. However, these studies primarily focused on technology efficacy without explicitly addressing the influence of skin color variability on diagnostic outcomes. Understanding and mitigating the impact of skin color is critical for developing equitable skin cancer detection tools that perform reliably across all populations.
Skin cancer detection from skin samples is a valid test for clinical diagnosis within the dermatology society. Unfortunately, many available medical datasets are not fair. This means that they usually have biases and tendencies toward one or some of the populations because of the geographical and temporal sources of data. These datasets are usually obtained from one, two, or three institutes in a single area or country with a fixed date range. Hence, it cannot be generalized to a wider range of patients. Moreover, patient data privacy enforces anonymization, and data such as race and ethnicity must be removed before distribution. This biased dataset affects the diagnosis process for medical students in training, as it was sourced from a single institution and contained a significant imbalance between data from White and Black patients [19]. In contrast, state-of-the-art artificial intelligence models are data-oriented, trained, validated, and deployed based on the available data. Unfairness in data creates unfair AI models. Unfair AI means that it can diagnose a disease for a specific population group better than other groups, such as skin cancer in white people compared to people of color [20,21].
To overcome skin color disparities, Daneshjou et al. [22] developed the publicly available, curated, and pathologically confirmed Stanford Diverse Dermatology Images (DDI) dataset. DDI dataset spans Fitzpatrick Skin Types (FST) I–VI [23,24] and was curated from Stanford Clinics between 2010 and 2020, and contains 656 images from 570 patients. It covers a wide range of skin tones, with 208 images from FST I–II, 241 from FST III–IV, and 207 images from FST V–VI. This allows for a direct comparison between lighter and darker skin tones while controlling for variables like age, sex, and diagnostic category. Although DDI is a valuable dataset, the number of samples is not sufficient to train and validate an AI system for clinical applications. Moreover, the Fitzpatrick Skin Type Scale is not an extensive indicator of human skin color [25,26,27].
To address the skin color disparity in skin cancer datasets, we leveraged the Monk skin tone (MST) scale [28,29], which offers reduced bias compared to the Fitzpatrick scale. We developed MST-AI models, a skin color estimation method based on the MST scale, and applied our models to the skin color of samples in a large publicly available skin cancer dataset, the International Skin Imaging Collaboration (ISIC) dataset. MST-AI automatically (1) removes image frames using a classical image processing algorithm, (2) segments and removes lesions using convolutional neural networks, and (3) models normal skin tones using a Variational Bayesian Gaussian Mixture Model (VB-GMM). We compared the resulting skin color distributions with the MST scale’s probability distribution functions (PDFs) using the Kullback-Leibler Divergence (KLD) metric. Validation was performed using manually annotated reference labels and comparisons with K-means clustering of image mean RGBs and skin mean RGBs. Our thorough validation demonstrates that MST-AI outperforms alternatives according to established ranking metrics such as Kendall’s Tau.
Bevan and Atapour-Abarghouei [28] made a valuable contribution by developing an automated skin tone detection method using the Fitzpatrick scale and applying bias unlearning techniques to enhance fairness in melanoma classification. Their approach did not include preprocessing steps, such as lesion or frame removal, which means that lesion and border colors could influence skin tone estimation. They also assigned a single skin tone per image, which may not fully capture the color variation due to overlaps within the Fitzpatrick scale. In comparison, our study uses the MST scale and incorporates preprocessing to focus on skin regions, offering an alternative perspective on automatic skin tone assessment in skin cancer diagnosis.
The remainder of this paper is organized as follows: Section 2 introduces the ISIC dataset and commonly used skin color scales and illustrates the proposed skin color estimator, detailing each step, including frame and lesion removal, the requirements for the skin color PDF, and the use of Kullback-Leibler Divergence (KLD). Section 3 presents the estimation results and compares them with the manual annotations. Section 4 provides a comparison and statistical analysis of the results. Finally, Section 5 concludes the paper with some comments and outlines directions for future work.

2. Materials and Methods

2.1. Dataset

The ISIC archive is one of the largest and most widely used repositories of dermoscopic images in melanoma research. With over 80,000 images, the collection encompasses a wide array of skin conditions, including benign and malignant lesions. Although all samples in the archive can be classified into two broad categories, benign and malignant, only a subset of the dataset is accompanied by detailed annotations. These annotations include lesion segmentation masks, which outline the precise boundaries of skin lesions, as well as essential metadata such as patient age, sex, and diagnosis [29]. The number of samples for the classification and segmentation tasks for each year is shown in Table 1 and Table 2, respectively.
We used the ISIC datasets spanning several years, which incorporated dermoscopic images collected from multiple international institutions. The multi-institutional and multi-year nature of the dataset ensured a broad representation of patient populations, imaging equipment, and clinical settings. The geographic diversity of the contributing centers introduces spatial variability, while the inclusion of data collected across different years adds temporal diversity. This combination enhances the heterogeneity of the dataset, supporting the development and evaluation of models that are more likely to generalize well across diverse, real-world conditions. Consequently, the use of the ISIC dataset provides a strong foundation for building robust and widely applicable algorithms, reducing the risk of dataset bias and overfitting to specific clinical settings. This rich combination of annotated images and metadata makes the ISIC archive an invaluable resource for a variety of research and clinical applications. It supports tasks such as (1) melanoma detection, where machine learning algorithms can be trained to distinguish between benign and malignant lesions, (2) lesion segmentation, which involves identifying and isolating skin lesions in images, and (3) diagnostic predictions, where predictive models can utilize patient metadata and image data to forecast the likelihood of various skin conditions. Thus, the breadth and depth of the dataset make it a versatile and essential tool for advancing melanoma research and improving diagnostic accuracy in dermatology [30,31,32].
Although the ISIC datasets contain many samples, which is one of the largest publicly available datasets, they lack information on race, ethnicity, and skin color groups or labels, which can lead to inaccurate predictions of skin melanoma. This research investigates the development of an AI-based method for estimating skin color from skin lesion image samples.

2.2. Skin Color Scales

Two well-known scales for skin color or tone are the Fitzpatrick [23] and Monk [33] skin tone (MST) scales. Both the Fitzpatrick and Monk scales assess skin color types but serve different purposes and have distinct methodologies. The Fitzpatrick scale, introduced by Dr. Thomas Fitzpatrick in 1975 [23], categorizes skin into six types based on its response to UV radiation, particularly focusing on whether the skin tends to burn or tan. These categories range from Type I, which always burns and never tans, to Type VI, which rarely burns and always tans. This scale is widely used in dermatology to evaluate the risk of sun damage, the likelihood of developing skin cancer, and to determine appropriate sun protection strategies for different individuals [23,24].
The Monk skin tone (MST) scale, developed by Dr. Ellis P Monk [33], takes a broader approach to classifying skin tone types. It is designed to be more generalizable across different populations and regions, as it focuses on a wider range of skin characteristics that are less influenced by UV exposure. The MST scale seeks to categorize skin color groups based on broader environmental and genetic factors, making it more versatile across varying climates and ethnicities, as shown in Figure 1. It is not specifically tied to sunburn or tanning but looks at overall skin characteristics that can vary more widely between individuals. While the Fitzpatrick scale is primarily concerned with sun-related skin reactions, the MST scale offers a more holistic view that can be applied across diverse demographics, providing insights into skin sensitivity and adaptability beyond just UV response [25,26,27].

2.3. Skin Color Estimator

An ISIC sample can be divided into three regions: the frame, lesion, and normal skin. The frame is typically a dark (or occasionally white) area surrounding the image, while the lesion is the suspicious region, which may be benign or malignant. Although not all images in the dataset contain these regions, normal skin usually remains after the removal of the frame and lesion. Given that the colors of both the frame and lesion differ significantly from those of normal skin, it is crucial to detect these regions to accurately estimate the skin color. We developed a model to estimate the skin tone of samples in the ISIC dataset based on the MST scale. A schematic of the proposed model is shown in Figure 2. The source code of the proposed system is publicly available at the GitHub repository (https://github.com/comp-epi-lee-lab/skin_color_estimator; released version: v1.0.0; accessed on 12 July 2025). Each step is described in detail below.

2.3.1. Frame Detection and Removal

Manually annotating medical data is time-consuming, erroneous, and labor-intensive, making unsupervised methods an attractive alternative for saving time and money. Certain characteristics can aid in developing algorithms for frame detection. Frames typically appear as solid black or white areas at the four corners of the sample images. Based on these traits, we created an algorithm to detect the frames. The algorithm is outlined in Algorithm 1, and a flowchart of the algorithm is shown in Figure 3a.
Similar to other unsupervised algorithms, certain parameters in this frame detection method must be manually or heuristically determined and tuned according to the dataset. The input to the algorithm is a color image in RGB format, and the output is a binary mask where ‘1’ indicates the frame region. The four key parameters, namely, white, black, object, and hole thresholds, were heuristically set to 95%, 5%, 64 pixels, and 64 pixels, respectively. The white and black thresholds define the upper and lower percentiles of the colors for non-frame areas. The object and hole thresholds specify the minimum size (in pixels) for the frame within the image and for the non-frame regions within the frame, respectively. After defining these thresholds, the algorithm begins by converting the extreme white regions into extreme black regions (Step 1). Next, the grayscale image is analyzed by taking the maximum value of the RGB channels, rather than their mean value, to enhance the contrast (Step 2). This approach functions similarly to a high-pass filter because it highlights high-contrast areas while filtering out low-contrast regions. The grayscale image is then normalized to the range [0, 1] (Step 3) and binarized using the black threshold (Step 4). The output mask from this step highlights the extremely dark regions, which are expected to correspond to the frame, although very dark spots (e.g., lesions) within the image may also resemble the frame color. Therefore, both the frame and dark spots are detected in Step 4. However, the frame area is not entirely dark, and some brighter spots within the frame region may remain undetected. These bright spots inside the frame are classified as “holes,” while the dark spots are referred to as “objects.” Step 5, which is divided into three sub-steps, removes small objects that are likely not related to the frame. The algorithm identifies objects using eight-neighbor connectivity, computes the area of each object by summing the binary pixel values, and removes objects with an area smaller than the object threshold. Step 6 mirrors Step 5 but operates on a negative binary image. To improve the estimation of the skin region (including both normal skin and lesions), a convex hull estimation [35] is applied (Step 7). The final output of the algorithm is a region of interest (ROI) representing the skin region, which includes normal skin and lesion areas. The frame detection algorithm consists of seven steps, with the outputs from each step shown in Figure 3b for an example image from the ISIC dataset.
Algorithm 1: Frame segmentation and removal.
Input
  • Three color channels (red, green, blue) of input image: I = <R, G, B>
  • Each R, G, and B channel is a matrix with r rows and c columns.
Output
  • The binary mask of the frame
Initialization
  • The white threshold is manually determined by W t h = 0.95
  • The black threshold is manually determined by B t h = 0.05
  • The object threshold is manually determined by O t h = 64
  • The hole threshold is manually determined by H t h = 64
Steps
  • Convert extreme whites into black
    I 1 = I , i f   1 3 R + G + B < 256 × W t h 0 , i f   1 3 R + G + B 256 × W t h  
  • Gray scale version of the image is computed as
    I 2 = max ( R ,   G ,   B )
  • Float scale normalization of the gray scale image is computed as
    I 3 = ( I 2 I 2 m i n   ) I 2 m a x I 2 m i n
    Here matrix division is an element-wise operation.
  • The black regions are separated by manual thresholding. The output is a binary matrix.
    I 4 = 0 , i f   I 3 < B t h 1 , i f   I 3 B t h  
  • Small objects are removed
    5.1.
    Label each region by finding the eight-neighbor connectivity in all pixels. Each region is assumed to be an object. Object labels are 0 ,   1 ,   2 ,   ,   n o b j 1 which n o b j is the total number of objects (regions) on the binary image.
    5.2.
    Objects areas are computed by
    A k = i = 0 r 1 j = 0 c 1 1 , p i j = k 0 , o t h   ,                           k 0 ,   1 ,   2 ,   ,   n o b j 1
    where p i j is the pixel value at i ,   j in a binary image I 4 .
    5.3.
    Objects smaller than the given threshold are converted to background (removed):
    I 5 = k = 0 n o b j 1 0 , i f   A k < O t h 1 , i f   A k O t h  
  • Small holes are removed. The algorithm is exactly similar to step 5, but instead of labeling pixels with value 1, pixels with value 0 will be labeled. Also, the threshold for holes H t h is used.
  • Convex Hull is computed for smooth segmentation. The convex hull is the set of pixels included in the smallest convex polygon that surrounds all white pixels in the input image. The area inside this convex hull will be considered as the region of interest (ROI) and anything outside as the frame.

2.3.2. Lesion Segmentation

Following frame detection, it is essential to identify and segment the lesion region, regardless of whether it is benign or malignant. This is because the color of the lesion typically differs significantly from that of normal skin, and including it in the skin color estimation would lead to substantial errors.
In contrast to frame detection, unsupervised lesion detection is highly challenging due to the wide variety of patterns and shapes that lesions can exhibit. Fortunately, subsets of the ISIC dataset from 2016, 2017, and 2018 included lesion masks. We used these subsets to develop a fully convolutional neural network (FCNN) model for the automatic semantic segmentation of lesions [36]. A transfer learning approach on PyTorch/v2.3.1 [37] pre-trained FCN-50 [36] is re-trained on training datasets of ISIC 2016 to 2018 and tested on their test cohorts. The images are batched into 32 samples and augmented with random horizontal/vertical flips, random translation, rotation, scale, shearing, and perspective, and then resized to 256 × 256 pixels and normalized to [0, 1]. Adam optimizer with a learning rate of 10 4 tuned the model to the data in 20 epochs. Jaccard loss is minimized as the optimization metric with the following definition:
J a c c a r d   L o s s = l o g ( c r X . Y c r X + c r Y c r X . Y )
where X , Y , Σ c , and Σ r are actual binary mask, detected mask (after sigmoid), element-wise product, summation over columns, and summation over rows, respectively. Jaccard Loss is an extension to the intersection over union match of two binary images that is rescaled to a logarithmic range for better convergence with stochastic gradient optimizers, such as Adam.
The training and testing results are presented in Table 3. In addition to Jaccard Loss, the DICE score is included as a well-known metric for better comparison. DICE score is defined as:
D I C E = 2 c r X . Y c r X + c r Y
The high DICE score for lesion detection in unseen test data confirmed that the FCN-50 model effectively excluded a substantial portion of the lesion, allowing the remaining areas to be used for color estimation of normal skin. Three examples of normal skin segmentation are shown in Figure 4. It is important to note that the frame and lesion removal steps can be performed in either order, as they are independent of one another.

2.3.3. Color Density Estimation

The central question is whether it is possible to assign a single MST scale value to the skin color in dermoscopic images. To address this, we can examine the MST scales in a 3D plot, as shown in Figure 5. This plot is generated by normalizing the RGB values of the MST scale depicted in Figure 1. As observed in Figure 5, the MST scale exhibits a significant overlap, indicating that neither an MST nor a skin color can be represented by a single RGB point in space. While the MST scales are associated with a primary color that varies in shade, this is not the case for skin. Normal skin may consist of several base colors combined with different tones. Therefore, both the MST and skin color represent a distribution of points (pixels) in the 3D RGB space. This distribution can be modeled using a PDF.
Probability density functions (PDFs) can be estimated using various methods, with the kernel method being one of the most commonly used approaches [38]. This method assumes that the distribution can be approximated using a weighted sum of kernels. The Gaussian kernel is the most frequently used due to its ability to satisfy key kernel conditions, including non-negativity, symmetry, and smoothness.
Visual examination of the 3D plots of the color distributions suggests that a single Gaussian distribution is insufficient for accurately estimating the skin color PDF. However, a more precise representation can be obtained using multiple Gaussians or a Gaussian Mixture Model (GMM). The GMM is a parametric approach that utilizes the kernel method to estimate the density function. It is assumed that the data are generated from a mixture of several Gaussian distributions, with each component characterized by its own mean and variance (or covariance in the multivariate case), as shown in the following equation [38]:
f x = k = 1 K w k N x | μ k , Σ k
where K , w k , and N x | μ k , Σ k are the number of components, the weight of the k-th component, and the Gaussian distribution with a mean of μ k and covariance of Σ k , respectively.
Although GMM is an effective method for density function estimation, it carries the risk of overfitting or underfitting because the number of components must be predetermined. A more robust alternative is the Variational Bayesian Gaussian Mixture Model (VB-GMM), which addresses these concerns. VB-GMM treats model parameters as random variables, assigns prior distributions to them, and aims to compute a posterior distribution over the parameters rather than just providing point estimates. The regularization introduced in the VB-GMM reduces the need to fix the exact number of components. In our approach, we used eight components to estimate the PDFs. Although we observed that half of these components overlapped with others, the computational overhead was an acceptable trade-off for minimizing underfitting and estimation errors. Two examples are shown in Figure 6 in 2D (each RGB component separately) for better visualization, as they are originally in 3D (RGB).
Our approach fits a separate VB-GMM to each of the 1618 image samples individually, estimating the RGB probability distribution in the 3D color space per sample. Because each model is computed independently, the global imbalance of skin tones in the dataset does not influence the performance of the VB-GMM.
After estimating the PDF, we generated random samples with similar coordinates (RGB values) to measure the differences and similarities in the next step. The red, green, and blue values were integers ranging from 0 to 255. We normalized these values to a range between 0 and 1, knowing that there are only 256 possible states for each variable. Consequently, we expect that a similar spatial sampling rate will be sufficient for an accurate PDF estimation. In practice, we selected 100 samples per axis, resulting in a total of 1003 = 1,000,000 points. We found that increasing the sample count to 2563 = 16,777,216 did not yield significant improvements in the results, but it was computationally much more expensive.

2.3.4. Kullback-Leibler Divergence and Membership Scores

In the previous section, we have estimated the PDFs of all MST scales and the skin sample. We now want to determine how similar or different the two PDFs are. A well-known metric is Kullback-Leibler divergence (KLD), defined as
D K L P Q = x χ P x log P x Q x
where P x and Q x are actual and detected PDFs, respectively, which are defined on the domain χ . D K L P Q measures the information loss by estimating the P x by Q x . If we assign a skin color to a MST scale in an image, it must represent the MST scale in the best possible form. It means that the skin PDF must represent the MST PDF, so P x is the MST PDF and Q x is the skin color PDF.
Since all MST scale PDFs and a skin color sample have overlap theoretically, ten KLDs are calculated per skin image sample. These KLDs (or other distance metrics such as Euclidean distance) are deployed to define the membership score, as shown below:
M P , Q m = s o f t m a x 1 D K L P Q m max m D K L P Q m
where P and Q m are PDFs of the skin color and m -th MST scale, respectively. The softmax function is defined as
s o f t m a x x m = e x m m e x m
M P , Q m is a normalized score between [0,1], where higher values indicate higher similarity.
Both GMM and KLD involve operations that are computationally sensitive to underflows and overflows, as GMM is exponential, and KLD includes a logarithmic term. Using standard floating-point calculations, such as float32 or float64, can lead to frequent computational failures when calculating the KLD, rendering the membership score computation infeasible. To address this, we developed a computationally efficient method that combines the KLD and GMM to prevent underflows and overflows. This method is described in Appendix A.

2.4. Test Cohort Annotation

To assess the validity of our membership definition in the ISIC datasets, we used human annotation as the gold standard. Our human annotator was tasked with segmenting an area of normal skin by drawing a rectangle around it, with everything inside the rectangle representing the expected skin color. We validated our estimator by calculating the statistical correlation between the manually segmented areas and the skin color estimator. For each image sample, we obtained ten membership measurements for each MST scale and annotated area, as well as ten membership measurements for each MST scale, and the skin area was automatically generated by our model in a semi-supervised manner. We then computed the ranked correlations between these ten values to evaluate how closely our model aligns with human judgment.
The ISIC dataset contains approximately 80,000 images, making it impractical to have an annotator manually label all of them. To address this, we randomly selected a subset of images that adequately represented color diversity. While random selection typically reflects the sample prevalence in the dataset, we recognize that the ISIC datasets inherently exhibit an imbalance in skin color and fairness. To mitigate this issue, we pre-classified skin color using an automatic approach. Although this pre-classification is not highly precise, it is sufficient for a weak categorization of skin colors across the MST scale. In this weak classifier, the mean RGB values of all pixels in the segmented normal skin region were used to represent the skin color. Similarly, the MST scale representation points were determined using the mean RGB values of all pixels in the MST scale. A nearest-neighbor approach based on the Euclidean distance between the RGB means was then used to assign a skin color label. We pre-classified the skin colors into ten MST scales and randomly selected 200 images from each scale, resulting in 2000 images prepared for manual annotation. Examples are shown in Figure 7.
During the annotation process, we identified two categories of images that should be excluded from the computation: those in which the normal skin area is too small to annotate and those in which body hair or other artifacts obscure almost the entire normal skin region. To address this, we divided the images in each MST scale into two groups: “included,” which contains manually annotated images, and “excluded,” which does not contain annotations. Figure 8 shows examples of these annotations. Of the 2000 images, 1618 were included for annotation, while 382 were excluded.
Although the pre-classifier was designed to create a weakly balanced dataset by selecting an approximately equal number of samples across different MST scales, we observed a higher exclusion rate of images in darker MST categories. This was primarily due to the presence of image artifacts that disproportionately affected the samples.
Addressing this imbalance requires a human-in-the-loop annotation pipeline to manually curate and correct the dataset, an approach that is both costly and time-intensive. Given these constraints, we acknowledge the remaining imbalances in the dataset. However, it represents a substantial improvement over the original ISIC dataset, which exhibited a more pronounced imbalance across skin tone categories.

3. Results

Three skin color categorization methods were compared using 1616 image samples. The methods are Image K-means, Skin K-means, and MST-AI, the proposed skin color estimator.
In both the Image K-means and Skin K-means methodologies, each sample is represented as a point in a three-dimensional RGB color space, where the coordinates correspond to the mean red, green, and blue intensity values of the associated pixel set. In the case of Image K-means, this pixel set comprises all pixels within the image, including those belonging to the frame and lesion regions. Conversely, Skin K-means uses only the pixels identified within the skin region, explicitly excluding both the frame and lesion pixels. Identical RGB-based coordinate representations are also assigned to each reference color in the MST scale [34]. For each sample, the Euclidean distance to every MST scale reference coordinate is computed and subsequently employed as the distance metric for determining the membership scores, as defined in Equation (5).
In the Skin Color Estimator framework, the dissimilarity between the samples and MST scale references is quantified using KLD. Specifically, the KLD is calculated between the VB-GMM estimated PDF of the sample and those of the MST scale reference distribution. These KLD values are then used to derive the membership scores in accordance with the formulation presented in Equation (5).
Table 4 indicates the ranks of the MST scales, ordered from the best to the worst match (only the top four out of ten values are shown for brevity). As shown, MST-AI provides the closest and most comprehensive match to the annotated regions. Although the other two methods perform relatively well for lighter skin tones, they struggle to accurately capture the diversity of darker skin tones. This limitation arises because these methods tend to make assumptions based on fewer characteristics or a narrower range of skin tones, which causes them to underperform when the skin color is darker or more complex. In these cases, Image and Skin K-means are particularly challenged when dealing with skin that contains multiple tones or variations in MST scales. These methods are less effective in identifying the subtleties of such diversity, resulting in less accurate matches for complex skin colors.
In contrast, MST-AI adopts a broader approach to classification by considering a wider selection of possible matches across different MST scales. This allows MST-AI to be more adaptable and better suited to handle a variety of skin tones, including those with greater complexity or diversity. This flexibility is particularly important in the decision-making process, as it allows the MST-AI to rely not only on the top-ranked match but also to take into account the second, third, or even fourth-ranked options if the top-ranked match does not fit well. This diversity in selection makes the MST-AI more robust, as alternative options can still offer a good fit, improving the overall reliability of the results.
In comparison, the other two methods do not exhibit the same flexibility. Their second, third-, and fourth-ranked matches are essentially extensions of the top-ranked match, meaning that they are still heavily influenced by the first selection. If the top-ranked match is incorrect or suboptimal, the subsequent options are less likely to provide a better alternative. This lack of diversity in the ranking process implies that the reliability of these methods is compromised if the top-ranked match is not ideal.
For further analysis, we employed two procedures to quantitatively evaluate the ranking results. To conduct a thorough evaluation, we began by plotting a histogram that represents the distribution of rankings for the top-1 match across all three methods. This histogram was compared to the annotation benchmark, which serves as the gold-standard manual annotation and is visually presented in Figure 4. The goal of this analysis is to assess how well the rankings of each method align with the reference benchmark across the various MST scale categories.
The top-1 ranking histogram shows the distribution of the highest-ranked match for each method. By comparing the histogram with the benchmark, we are able to assess how closely the rankings of each method resemble the distribution of samples observed in the manual annotations.
From the histograms in Figure 9, it is evident that the MST-AI stands out as the most accurate in closely matching the distribution of the gold-standard manual annotations across all MST scale bins. This indicates that the MST-AI can consistently estimate the skin tone distributions represented in the annotated dataset, whether the skin tones fall in the lighter or darker MST scale categories. The closeness of the histograms demonstrates that the KLD of PDFs is more in tune with the actual skin tone distributions, effectively capturing the variation in skin tones as represented by the annotated data. In contrast, the other two methods, although they may perform well in specific areas (e.g., lighter skin tones), show discrepancies when compared with manual annotations, particularly in terms of how they distribute the top-1 ranking. The ability of the MST-AI to closely align with the benchmark across all MST scale bins highlights its superior performance in reflecting the real-world diversity of skin tones in a more comprehensive manner.
Next, we investigated the comparison of individual sample rankings using ranking metrics. There are several well-known metrics to measure ranking performance, such as Precision at K, Recall at K, Kendall’s tau [39,40], Spearman’s Rank Correlation [41], and Position-based Accuracy. Although Precision and Recall at K show the performance of the estimator, they do not distinguish the order of ranks in the top-K results. Because the order of MST scale is important in our estimator, we use Kendall’s Tau, Spearman’s Rank Correlation, and Normalized Discounted Cumulative Gain (NDCG), which are all sensitive to rank orders.
Kendall’s tau is defined as:
τ = i = 1 k j = i + 1 k 1 , if   A i A j D i D j > 0 1 , otherwise k 2
where A i , D i , and k are the actual rank, detected rank, and the number of top results, respectively.
Spearman’s Rank Correlation is defined as
r s = 1 6 i = 1 k d i 2 k k 2 1 ; w h e r e   d i = A i D i
where, similar to Kendall’s Tau equation, A i , D i , and k are the actual rank, detected rank, and the number of top results, respectively. This equation is valid as long as there are no ties in the ranking, which is almost correct since the probability of ties is rare (if not zero) in our calculations, which can be ignored.
NDCG is defined as
N D C G = 1 i = 1 k 2 m d i 1 log 2 i + 1 i = 1 k 2 m a i 1 log 2 i + 1
where m a i and m d i are actual and detected membership scores for top-K results, respectively.
The Kendall’s Tau and Spearman’s Rank Correlation scores are between −1 and +1, where −1 indicates an extreme difference and +1 indicates complete similarity. NDCG score varies between 0 (dissimilar) and 1 (similar). The performance evaluation results are presented in Table 5.

4. Discussion

4.1. Statistical Analysis of the Results

Because Kendall’s Tau and Spearman’s correlation produce very similar results across individual models, we report only one for comparison. A two-tailed Student’s t-test was used to assess the statistical significance of the mean results. Both Kendall’s Tau and Spearman’s Correlation are treated as random uniform variables ranging between a = 1 and a = + 1 . Standard deviation of these samples would be as
σ = b a 2 12 0.5774
with similar calculations on NDCGs, which are between [0, +1], the standard deviation would be 0.2887.
The p-values for each pair of mean results were computed using a two-tailed Student’s t-test. Based on the conventional p-value of 5% in most medical applications, if the p-value is less than 0.05, then the difference between the two results is statistically significant, and we can assume that the method with the higher mean value is significantly better than the method with the lower mean value.
The p-values are presented in Table 6. The circles represent the methods listed in Table 5, the dashed line depicts the statistical difference, and the continuous line implies that the difference is not statistically significant.
For the top-1 results (first row in Table 6), correlation metrics cannot be computed due to their inherent definitions; however, NDCG scores are reported. Although the NDCG values in Table 5 appear close to each other, the differences between the methods are statistically significant (p < 0.01), likely due to the moderately large sample size (e.g., 1600 samples).
For the top-2 (second row in Table 6), the correlation scores between methods 1 and 2 are not significantly different. However, method 3 demonstrated a significant improvement over both methods 1 and 2, with a p-value less than 0.01. Additionally, when evaluating the NDCG scores at the top-2, all three methods showed significant differences in performance.
In the case of top-3 (third row in Table 6), the correlation scores between methods 1 and 2 are significantly different. However, the NDCG scores for the two methods do not show a significant difference. Regardless, method 3 consistently outperforms both methods 1 and 2 in both the correlation and NDCG metrics.
For top-4 (last row in Table 6), method 3 significantly outperforms both methods 1 and 2, but there is no significant difference between methods 1 and 2 in either metric (correlation or NDCG).
The statistical differences reported in Table 5 and Table 6 indicate that Kendall’s τ and Spearman’s correlation are more sensitive to changes in the ranking order of outputs than NDCG, which appears more tolerant to such variations. This suggests that rank-based correlation metrics may better capture subtle shifts in the prediction order, especially when evaluating consistency across methods.
Additionally, the results show that Skin K-Means outperforms Image K-Means in three of the seven evaluation metrics, emphasizing the effectiveness of frame and lesion removal in accurately identifying normal skin regions. This highlights the value of incorporating preprocessing steps tailored to address domain-specific challenges. Notably, MST-AI, our proposed method achieves a consistently superior performance, showing statistically significant improvements over both baseline methods, Image K-means and Skin K-means, across all three metrics. This demonstrates the robustness and overall advantage of our approach for this task.

4.2. Caveats and Limitations

We randomly sampled data from multiple years of the ISIC dataset. The test cohort comprised 2000 manually annotated samples, providing sufficient statistical power for reliable estimation with appropriate confidence levels. While this manually annotated dataset significantly contributes to our results, it also introduces several limitations, some inherent to the dataset and others that could be addressed in future research. Below, we outline these considerations.
First, the dataset lacks metadata on imaging conditions, including the specific instruments used and background lighting. As a result, there is no gold standard or reference available to evaluate how effectively any background illumination detection method (if present) compensates for the effects of instrumentation or lighting variations. This absence limits our ability to assess or correct for confounding variables.
Second, skin color can vary substantially across different regions of an individual’s body. Therefore, assigning a single skin color to a person based on a small, region-unknown sample may not be reliable.
Third, due to the cost of manual annotation, we relied on a single set of annotations. This process is inherently subjective, similar to most human annotations. Different annotators may produce different labels for the same image. Inter-rater agreement is a commonly used metric in medical image analysis for quantifying annotation subjectivity. Although our dataset is publicly available, we plan to expand it in the future to include annotations from multiple raters, enabling the calculation of inter-rater agreement scores.
In summary, while our method and annotated dataset aim to provide a robust unsupervised approach for skin color estimation using ISIC samples, they are constrained by the lack of information about the body region, environmental illumination, and imaging instrumentation. Without access to such real-world metadata, further improvements in reliability and validation are challenging.

5. Conclusions

Publicly available skin cancer datasets often lack skin color information, which is crucial for the diagnosis and prognosis of skin cancer. Our study explored various skin color scales and found that the MST scale offers a less biased alternative to the well-established Fitzpatrick scale, particularly in non-white populations, such as individuals with darker skin tones. Based on the MST scale, we proposed a novel skin color detection method that leverages large, publicly available skin cancer datasets, such as the ISIC archive.
In this research, we addressed key challenges, including the removal of image frames and lesions, by developing an unsupervised automatic frame detection and removal technique. We then used the cleaned data to construct a robust convolutional deep neural network for lesion segmentation. Once the frame and lesion were removed, the remaining normal skin was analyzed for its various colors, tones, and shades. We modeled the normal skin pixels as a PDF using the Variational Bayesian Gaussian Mixture Model (VB-GMM), which is effective in capturing the underlying data distribution and is less sensitive to the predefined number of clusters.
We compared the PDF of normal skin with those of the MST scale using the Kullback-Leibler Divergence (KLD) as a metric to quantify the similarity between the distributions. To address KLD’s sensitivity to computational under-/overflows, we developed an efficient approximation method to mitigate these issues.
Finally, we compared our proposed method with the K-means clustering approach applied to both whole images and skin regions. The results, validated against manual annotations of approximately 1600 images, showed that our method outperformed others in matching MST scale classifications to normal skin samples, as evidenced by Kendall’s Tau, Spearman Correlation, and NDCG metrics.
As a next step, we aim to extend this work to develop more equitable AI-based systems for skin cancer detection. By addressing skin tone imbalances present in large-scale dermatology datasets, our method provides a foundation for fairer and more accurate diagnostic tools. Future work will focus on integrating advanced deep learning models to enhance skin cancer detection across all skin tone scales, with the goal of improving diagnostic performance and reducing bias in real-world clinical applications.

Author Contributions

Conceptualization, H.L.; Methodology, V.K., A.B. and S.K.B.; Software, V.K.; Validation, V.K., J.N. and S.Z.-E.; Formal analysis, V.K.; Investigation, V.K., H.L. and C.R.; Resources, H.L. and E.P.M.; Data curation, J.N.; Writing—review & editing, H.L., S.Z.-E., C.R., A.A., A.B., E.P.M. and S.K.B.; Supervision, H.L. and S.K.B.; Project administration, S.K.B.; Funding acquisition, H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by new faculty funds from Fox Chase Cancer Center and the Cancer Center Support Grant P30CA006927 and Temple University.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The ISIC data is publicly available at https://challenge.isic-archive.com/ (accessed on 27 May 2025). Skin Color Dataset, which is annotated by our team, is also publicly available at https://www.kaggle.com/datasets/vahidkermani/skin-color-in-isic/data (accessed on 27 May 2025).

Acknowledgments

We acknowledge the use of https://methylome.ai (accessed on 27 May 2025), high-performance computing servers provided by Lee Lab, Cancer Epigenetics Institute, Nuclear Dynamics and Cancer Program, Fox Chase Cancer Center, which were essential for data processing and analysis in this study.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial Intelligence
ISICInternational Skin Imaging Collaboration
MSTMonk Skin Tone
GMMGaussian Mixture Model
VB-GMMVariational Bayesian Gaussian Mixture Model
KLDKullback-Leibler Divergence
DDIStanford Diverse Dermatology Images
FSTFitzpatrick Skin Types
ROIRegion of Interest
PDFProbability Distribution Function
RGBRed, Green, Blue
NDCGNormalized Discounted Cumulative Gain

Appendix A

Computational Considerations for Distance Measurements
Kullback-Leibler Divergence (KLD) on a Gaussian Mixture Model (GMM) can easily cause under/overflow. Here, we attempt to modify the equations to efficiently compute the KLD of the GMM without computational errors. Starting with the KLD definition:
D K L P Q = x χ P x log P x Q x
where P x and Q x are two PDFs.
Since color distributions are modeled with VB-GMM, their distribution is a Gaussian mixture and requires exponential computation.
P x , Q x ~ e x p o n e n t i a l   o f ( x μ ) 2
when x is close to one of the means of P and far from any means of Q, then P is exponentially larger than Q, and the fraction of P Q does not fit in the largest conventional floating-point storage, such as 64-bit and 128-bit floating-point. The use of unconventional floating-point containers also adds significant computational overhead and should be prohibited in computationally expensive applications.
P Q log P Q P Q log P Q + D K L P Q = x χ P x log P x Q x N a N
This computational error is an indirect impact of the exponential terms in the GMM. However, if we insert the exponential terms in the computation, we can solve this issue as follows.
Assume that
P x = 1 s p exp x p , Q x = 1 s q e x p ( x q )
where x p and x q are GMM logarithmic scores of P and Q at a single point of x, and s p and s q are the summations of samples of distributions P and Q, respectively. Based on these definitions, we can embed P and Q in KLD, expand and simplify as
D K L P Q = x χ P x log P x Q x = x χ 1 s p exp s p log 1 s p exp x p 1 s q exp x q = x χ 1 s p exp x p log 1 s p exp x p log 1 s q exp x q = x χ 1 s p exp x p log 1 s p + log exp x p log 1 s q log exp x q = x χ 1 s p exp x p x p log s p x q + log s q
We must notice that x p and x q are functions of x.
The final equation is unlikely to underflow or overflow as long as the input variables fall within the range of float32 or float64.

References

  1. Guy, G.P.; Thomas, C.C.; Thompson, T.; Watson, M.; Massetti, G.M.; Richardson, L.C.; Centers for Disease Control and Prevention (CDC). Vital Signs: Melanoma Incidence and Mortality Trends and Projections—United States, 1982–2030. MMWR Morb Mortal Wkly. Rep. 2015, 64, 591–596. [Google Scholar] [PubMed]
  2. Guy, G.P.; Machlin, S.R.; Ekwueme, D.U.; Yabroff, K.R. Prevalence and Costs of Skin Cancer Treatment in the U.S., 2002−2006 and 2007−2011. Am. J. Prev. Med. 2015, 48, 183–187. [Google Scholar] [CrossRef] [PubMed]
  3. Stern, R.S. Prevalence of a History of Skin Cancer in 2007: Results of an Incidence-Based Model. Arch Dermatol 2010, 146, 279–282. [Google Scholar] [CrossRef]
  4. American Cancer Society. Cancer Facts & Figures 2025; American Cancer Society: Atlanta, GA, USA, 2025. [Google Scholar]
  5. Osman, A.; Nigro, A.R.; Brauer, S.F.; Borda, L.J.; Roberts, A.A. Epidemiology and Primary Location of Melanoma in Asian Patients: A Surveillance, Epidemiology, and End Result–Based Study. JAAD Int. 2024, 16, 77–78. [Google Scholar] [CrossRef]
  6. Townsend, J.S.; Melkonian, S.C.; Jim, M.A.; Holman, D.M.; Buffalo, M.; Julian, A.K. Melanoma Incidence Rates Among Non-Hispanic American Indian/Alaska Native Individuals, 1999–2019. JAMA Dermatol. 2024, 160, 148. [Google Scholar] [CrossRef]
  7. Mahendraraj, K.; Sidhu, K.; Lau, C.S.M.; McRoy, G.J.; Chamberlain, R.S.; Smith, F.O. Malignant Melanoma in African–Americans: A Population-Based Clinical Outcomes Study Involving 1106 African–American Patients from the Surveillance, Epidemiology, and End Result (SEER) Database (1988–2011). Medicine 2017, 96, e6258. [Google Scholar] [CrossRef]
  8. Hu, S.; Soza-Vento, R.M.; Parker, D.F.; Kirsner, R.S. Comparison of Stage at Diagnosis of Melanoma Among Hispanic, Black, and White Patients in Miami-Dade County, Florida. Arch. Dermatol. 2006, 142, 704–708. [Google Scholar] [CrossRef]
  9. Qian, Y.; Johannet, P.; Sawyers, A.; Yu, J.; Osman, I.; Zhong, J. The Ongoing Racial Disparities in Melanoma: An Analysis of the Surveillance, Epidemiology, and End Results Database (1975–2016). J. Am. Acad. Dermatol. 2021, 84, 1585–1593. [Google Scholar] [CrossRef]
  10. Gregory, C. Mitro Skin Cancer Does Not Discriminate by Skin Tone. 2025. Available online: https://treatcancer.com/blog/skin-cancer-by-skin-tone/ (accessed on 27 May 2025).
  11. Gupta, A.K.; Bharadwaj, M.; Mehrotra, R. Skin Cancer Concerns in People of Color: Risk Factors and Prevention. Asian Pac. J. Cancer Prev. 2016, 17, 5257. [Google Scholar]
  12. Zheng, Y.J.; Ho, C.; Lazar, A.; Ortiz-Urda, S. Poor Melanoma Outcomes and Survival in Asian American and Pacific Islander Patients. J. Am. Acad. Dermatol. 2021, 84, 1725–1727. [Google Scholar] [CrossRef]
  13. Dawes, S.M.; Tsai, S.; Gittleman, H.; Barnholtz-Sloan, J.S.; Bordeaux, J.S. Racial Disparities in Melanoma Survival. J. Am. Acad. Dermatol. 2016, 75, 983–991. [Google Scholar] [CrossRef] [PubMed]
  14. Gloster, H.M.; Neal, K. Skin Cancer in Skin of Color. J. Am. Acad. Dermatol. 2006, 55, 741–760. [Google Scholar] [CrossRef] [PubMed]
  15. Huang, H.-Y.; Hsiao, Y.-P.; Mukundan, A.; Tsao, Y.-M.; Chang, W.-Y.; Wang, H.-C. Classification of Skin Cancer Using Novel Hyperspectral Imaging Engineering via YOLOv5. JCM 2023, 12, 1134. [Google Scholar] [CrossRef] [PubMed]
  16. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  17. Lin, T.-L.; Lu, C.-T.; Karmakar, R.; Nampalley, K.; Mukundan, A.; Hsiao, Y.-P.; Hsieh, S.-C.; Wang, H.-C. Assessing the Efficacy of the Spectrum-Aided Vision Enhancer (SAVE) to Detect Acral Lentiginous Melanoma, Melanoma In Situ, Nodular Melanoma, and Superficial Spreading Melanoma. Diagnostics 2024, 14, 1672. [Google Scholar] [CrossRef]
  18. Lin, T.-L.; Karmakar, R.; Mukundan, A.; Chaudhari, S.; Hsiao, Y.-P.; Hsieh, S.-C.; Wang, H.-C. Assessing the Efficacy of the Spectrum-Aided Vision Enhancer (SAVE) to Detect Acral Lentiginous Melanoma, Melanoma In Situ, Nodular Melanoma, and Superficial Spreading Melanoma: Part II. Diagnostics 2025, 15, 714. [Google Scholar] [CrossRef]
  19. Stephanie, W. Medical Student Melanoma Detection Rates in White and African American Skin Using Moulage and Standardized Patients. CRDOA 2015, 2, 1–4. [Google Scholar] [CrossRef]
  20. Daneshjou, R.; Smith, M.P.; Sun, M.D.; Rotemberg, V.; Zou, J. Lack of Transparency and Potential Bias in Artificial Intelligence Data Sets and Algorithms: A Scoping Review. JAMA Dermatol. 2021, 157, 1362. [Google Scholar] [CrossRef]
  21. Tschandl, P. Risk of Bias and Error From Data Sets Used for Dermatologic Artificial Intelligence. JAMA Dermatol. 2021, 157, 1271. [Google Scholar] [CrossRef]
  22. Daneshjou, R.; Vodrahalli, K.; Novoa, R.A.; Jenkins, M.; Liang, W.; Rotemberg, V.; Ko, J.; Swetter, S.M.; Bailey, E.E.; Gevaert, O.; et al. Disparities in Dermatology AI Performance on a Diverse, Curated Clinical Image Set. Sci. Adv. 2022, 8, eabq6147. [Google Scholar] [CrossRef]
  23. Fitzpatrick, T.B. The Validity and Practicality of Sun-Reactive Skin Types I Through VI. Arch. Dermatol. 1988, 124, 869. [Google Scholar] [CrossRef]
  24. Coleman, W.; Mariwalla, K.; Grimes, P. Updating the Fitzpatrick Classification: The Skin Color and Ethnicity Scale. Dermatol. Surg. 2023, 49, 725–731. [Google Scholar] [CrossRef]
  25. Hazirbas, C.; Bitton, J.; Dolhansky, B.; Pan, J.; Gordo, A.; Ferrer, C.C. Casual Conversations: A Dataset for Measuring Fairness in AI. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Nashville, TN, USA, 19–25 June 2021; pp. 2289–2293. [Google Scholar]
  26. Howard, J.J.; Sirotin, Y.B.; Tipton, J.L.; Vemury, A.R. Reliability and Validity of Image-Based and Self-Reported Skin Phenotype Metrics. IEEE Trans. Biom. Behav. Identity Sci. 2021, 3, 550–560. [Google Scholar] [CrossRef]
  27. Heldreth, C.M.; Monk, E.P.; Clark, A.T.; Schumann, C.; Eyee, X.; Ricco, S. Which Skin Tone Measures Are the Most Inclusive? An Investigation of Skin Tone Measures for Artificial Intelligence. ACM J. Responsib. Comput. 2024, 1, 1–21. [Google Scholar] [CrossRef]
  28. Bevan, P.J.; Atapour-Abarghouei, A. Detecting Melanoma Fairly: Skin Tone Detection and Debiasing for Skin Lesion Classification. In MICCAI Workshop on Domain Adaptation and Representation Transfer; Springer Nature: Cham, Switzerland, 2022. [Google Scholar]
  29. Caffery, L.J.; Clunie, D.; Curiel-Lewandrowski, C.; Malvehy, J.; Soyer, H.P.; Halpern, A.C. Transforming Dermatologic Imaging for the Digital Era: Metadata and Standards. J. Digit. Imaging 2018, 31, 568–577. [Google Scholar] [CrossRef]
  30. Codella, N.; Rotemberg, V.; Tschandl, P.; Celebi, M.E.; Dusza, S.; Gutman, D.; Helba, B.; Kalloo, A.; Liopyris, K.; Marchetti, M.; et al. Skin Lesion Analysis Toward Melanoma Detection 2018: A Challenge Hosted by the International Skin Imaging Collaboration (ISIC). arXiv 2019, arXiv:1902.03368. [Google Scholar]
  31. Gutman, D.; Codella, N.C.F.; Celebi, E.; Helba, B.; Marchetti, M.; Mishra, N.; Halpern, A. Skin Lesion Analysis toward Melanoma Detection: A Challenge at the International Symposium on Biomedical Imaging (ISBI) 2016, Hosted by the International Skin Imaging Collaboration (ISIC). arXiv 2016, arXiv:1605.01397. [Google Scholar]
  32. Codella, N.C.F.; Gutman, D.; Celebi, M.E.; Helba, B.; Marchetti, M.A.; Dusza, S.W.; Kalloo, A.; Liopyris, K.; Mishra, N.; Kittler, H.; et al. Skin Lesion Analysis toward Melanoma Detection: A Challenge at the 2017 International Symposium on Biomedical Imaging (ISBI), Hosted by the International Skin Imaging Collaboration (ISIC). In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; pp. 168–172. [Google Scholar]
  33. Monk, E. The Monk Skin Tone Scale. SocArXiv 2023. [Google Scholar] [CrossRef]
  34. Monk, E. Monk Skin Tone Scale. 2025. Available online: https://skintone.google (accessed on 27 May 2025).
  35. Barber, C.B.; Dobkin, D.P.; Huhdanpaa, H. The Quickhull Algorithm for Convex Hulls. ACM Trans. Math. Softw. 1996, 22, 469–483. [Google Scholar] [CrossRef]
  36. Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. arXiv 2014, arXiv:1411.4038. [Google Scholar] [CrossRef]
  37. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. arXiv 2019, arXiv:1912.01703. [Google Scholar]
  38. Schölkopf, B.; Smola, A.J. Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond; Adaptive computation and machine learning series; Reprint; MIT Press: Cambridge, MA, USA, 2002; ISBN 978-0-262-53657-8. [Google Scholar]
  39. Kendall, M.G. A New Measure of Rank Correlation. Biometrika 1938, 30, 81. [Google Scholar] [CrossRef]
  40. Kendall, M.G. THE TREATMENT OF TIES IN RANKING PROBLEMS. Biometrika 1945, 33, 239–251. [Google Scholar] [CrossRef]
  41. Kokoska, S.; Zwillinger, D. CRC Standard Probability and Statistics Tables and Formulae, Student Edition; CRC Press: Boca Raton, FL, USA, 2000; ISBN 978-0-429-18146-7. [Google Scholar]
Figure 1. Monk skin tone (MST) scale [34].
Figure 1. Monk skin tone (MST) scale [34].
Jimaging 11 00235 g001
Figure 2. Schematic of the skin color classifier model.
Figure 2. Schematic of the skin color classifier model.
Jimaging 11 00235 g002
Figure 3. (a) Flowchart of frame segmentation steps; (b) Example images for each step of the frame removal algorithm: steps 1–7 in Algorithm 1.
Figure 3. (a) Flowchart of frame segmentation steps; (b) Example images for each step of the frame removal algorithm: steps 1–7 in Algorithm 1.
Jimaging 11 00235 g003
Figure 4. Examples of Normal Skin Detection, from left to right: original image, detected lesion, detected frame, and remaining normal skin.
Figure 4. Examples of Normal Skin Detection, from left to right: original image, detected lesion, detected frame, and remaining normal skin.
Jimaging 11 00235 g004
Figure 5. MST scales in 3D RGB; each color corresponds to one of the MST scales shown in Figure 1.
Figure 5. MST scales in 3D RGB; each color corresponds to one of the MST scales shown in Figure 1.
Jimaging 11 00235 g005
Figure 6. The estimation of two MSTs with VB-GMM; (a) MST 1 samples, (b) VB-GMM estimation of MST 1, (c) MST 5 samples, (d) VB-GMM estimation of MST 5.
Figure 6. The estimation of two MSTs with VB-GMM; (a) MST 1 samples, (b) VB-GMM estimation of MST 1, (c) MST 5 samples, (d) VB-GMM estimation of MST 5.
Jimaging 11 00235 g006
Figure 7. Examples of MST scale pre-classification for the annotation process.
Figure 7. Examples of MST scale pre-classification for the annotation process.
Jimaging 11 00235 g007
Figure 8. Three examples of the included and excluded images. (a) Included image from left to right: original image, detected lesion, detected frame, normal skin; (b) Excluded due to insufficient skin area; (c) Excluded due to hair. The green rectangular box shows the manual annotation by the human annotator.
Figure 8. Three examples of the included and excluded images. (a) Included image from left to right: original image, detected lesion, detected frame, normal skin; (b) Excluded due to insufficient skin area; (c) Excluded due to hair. The green rectangular box shows the manual annotation by the human annotator.
Jimaging 11 00235 g008
Figure 9. The top-1 rankings.
Figure 9. The top-1 rankings.
Jimaging 11 00235 g009
Table 1. Classification task: number of samples and percentage of malignancy (vs benign) for different years in the ISIC datasets.
Table 1. Classification task: number of samples and percentage of malignancy (vs benign) for different years in the ISIC datasets.
YearTrainMalignant%ValidMalignant%TestMalignant%TotalMalignant%
201690019.2200.0037919.79127919.39
2017200018.7015020.0060019.50275018.95
201810,01519.5119322.80151220.3011,72019.67
201925,33136.8700.00N/AN/A25,33136.87
202033,1261.7600.00N/AN/A33,1261.76
Table 2. Segmentation task: number of samples and percentage of malignancy (vs. benign) for different years in the ISIC datasets (#segmented images/#total images).
Table 2. Segmentation task: number of samples and percentage of malignancy (vs. benign) for different years in the ISIC datasets (#segmented images/#total images).
YearTrainValidTestTotal
2016900/9000379/3791279/1279
20172000/2000150/150600/6002750/2750
20182594/10,015100/1931000/15123694/11,720
20190000
20200000
Table 3. Results of training and testing of FCN-50 on the ISIC 2016–2018 datasets.
Table 3. Results of training and testing of FCN-50 on the ISIC 2016–2018 datasets.
TrainValidationTest
Jaccard Loss0.00460.00710.0075
DICE94.07%89.50%89.39%
Table 4. Three examples of MST scale assignments for each method.
Table 4. Three examples of MST scale assignments for each method.
Original ImageJimaging 11 00235 i001Jimaging 11 00235 i002Jimaging 11 00235 i003
Annotated RegionJimaging 11 00235 i004Jimaging 11 00235 i005Jimaging 11 00235 i006
Image K-Means
MST scale
Memberships
Jimaging 11 00235 i007
5,4,6,7
0.151, 0.128, 0.113, 0.101
Jimaging 11 00235 i008
4,5,3,1
0.143, 0.137, 0.114, 0.108
Jimaging 11 00235 i009
6,7,5,8
0.152, 0.138, 0.128, 0.113
Skin K-Means
MST scale
Memberships
Jimaging 11 00235 i010
2,1,3,4
0.140, 0.138, 0.136, 0.120
Jimaging 11 00235 i011
3,2,1,4
0.140, 0.140, 0.136, 0.124
Jimaging 11 00235 i012
5,4,6,3
0.145, 0.128, 0.114, 0.102
MST-AI
MST scale
Memberships
Jimaging 11 00235 i013
1,2,3,4
0.133, 0.129, 0.118, 0.117
Jimaging 11 00235 i014
5,3,4,6
0.130, 0.127, 0.125, 0.121
Jimaging 11 00235 i015
8,1,2,7
0.149, 0.139, 0.120, 0.117
Table 5. Comparison of results for the three methods using three performance metrics (@k: top-k).
Table 5. Comparison of results for the three methods using three performance metrics (@k: top-k).
MetricImage K-MeansSkin K-MeansMST-AI
@1@2@3@4@1@2@3@4@1@2@3@4
Kendall’s Tau-0.120.12−0.02-0.120.17−0.05-0.780.730.74
Spearman’s Correlation-0.120.13−0.06-0.120.18−0.10-0.780.730.74
NDCG0.890.900.910.920.910.930.940.940.990.990.991.00
Table 6. p-values for two-tailed Student’s t-test between methods; dashed lines between methods indicate a statistically significant difference in their results, whereas solid lines represent differences that are not statistically significant.
Table 6. p-values for two-tailed Student’s t-test between methods; dashed lines between methods indicate a statistically significant difference in their results, whereas solid lines represent differences that are not statistically significant.
Top-kCorrelationsNDCG
top-1N/AJimaging 11 00235 i016
top-2Jimaging 11 00235 i017Jimaging 11 00235 i018
top-3Jimaging 11 00235 i019Jimaging 11 00235 i020
top-4Jimaging 11 00235 i021Jimaging 11 00235 i022
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Khalkhali, V.; Lee, H.; Nguyen, J.; Zamora-Erazo, S.; Ragin, C.; Aphale, A.; Bellacosa, A.; Monk, E.P.; Biswas, S.K. MST-AI: Skin Color Estimation in Skin Cancer Datasets. J. Imaging 2025, 11, 235. https://doi.org/10.3390/jimaging11070235

AMA Style

Khalkhali V, Lee H, Nguyen J, Zamora-Erazo S, Ragin C, Aphale A, Bellacosa A, Monk EP, Biswas SK. MST-AI: Skin Color Estimation in Skin Cancer Datasets. Journal of Imaging. 2025; 11(7):235. https://doi.org/10.3390/jimaging11070235

Chicago/Turabian Style

Khalkhali, Vahid, Hayan Lee, Joseph Nguyen, Sergio Zamora-Erazo, Camille Ragin, Abhishek Aphale, Alfonso Bellacosa, Ellis P. Monk, and Saroj K. Biswas. 2025. "MST-AI: Skin Color Estimation in Skin Cancer Datasets" Journal of Imaging 11, no. 7: 235. https://doi.org/10.3390/jimaging11070235

APA Style

Khalkhali, V., Lee, H., Nguyen, J., Zamora-Erazo, S., Ragin, C., Aphale, A., Bellacosa, A., Monk, E. P., & Biswas, S. K. (2025). MST-AI: Skin Color Estimation in Skin Cancer Datasets. Journal of Imaging, 11(7), 235. https://doi.org/10.3390/jimaging11070235

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop