Next Article in Journal
Implications of Podoplanin Overexpression in the Malignant Transformation of Oral Potentially Malignant Disorders: A Systematic Review and Meta-Analysis
Previous Article in Journal
Obesity and Metabolic Syndrome in Childhood Leukemia and in Long-Term Survivors: Causes and Personalized Treatments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Clinical Application of Vision Transformers for Melanoma Classification: A Multi-Dataset Evaluation Study

1
Department of Electrical and Computer Engineering, Worcester Polytechnic Institute, Worcester, MA 01609, USA
2
Facultad de Ingeniería Eléctrica, Universidad Tecnológica de Panamá, Avenida Universidad Tecnológica de Panamá, Panamá 0819-0728, Panama
3
College of Medicine, University of South Florida Morsani, Tampa, FL 33602, USA
4
Dermatology Service, James A. Haley Veterans Hospital, Tampa, FL 33637, USA
*
Author to whom correspondence should be addressed.
Cancers 2025, 17(21), 3447; https://doi.org/10.3390/cancers17213447
Submission received: 4 October 2025 / Revised: 21 October 2025 / Accepted: 25 October 2025 / Published: 28 October 2025
(This article belongs to the Topic Machine Learning and Deep Learning in Medical Imaging)

Simple Summary

Melanoma is a dangerous skin cancer that can be treated successfully when detected early, but it often looks similar to benign moles, which makes diagnosis difficult. This research uses Vision Transformers to help identify melanoma from skin images. The model was trained with real medical images and additional synthetic ones produced by a Deep Learning algorithm to improve learning. Its performance was tested against several Deep Learning classification models and a commercial diagnostic tool. The Vision Transformer achieved higher accuracy in separating cancerous and non-cancerous lesions. This approach may help doctors make faster and more confident assessments when examining skin images, supporting better detection of melanoma in clinical settings.

Abstract

Background: Melanoma is one of the most lethal skin cancers, with survival rates largely dependent on early detection, yet diagnosis remains difficult because of its visual similarity to benign nevi. Convolutional neural networks have achieved strong performance in dermoscopic analysis but often depend on fixed input sizes and local features, which can limit generalization. Vision Transformers, which capture global image relationships through self-attention, offer a promising alternative. Methods: A ViT-L/16 model was fine-tuned using the ISIC 2019 dataset containing more than 25,000 dermoscopic images. To expand the dataset and balance class representation, synthetic melanoma and nevus images were produced with StyleGAN2-ADA, retaining only high-confidence outputs. Model performance was evaluated on an external biopsy-confirmed dataset (MN187) and compared with CNN baselines (ResNet-152, DenseNet-201, EfficientNet-B7, ConvNeXt-XL, ViT-B/16) and the commercial MoleAnalyzer Pro system using ROC-AUC and DeLong’s test. Results: The ViT-L/16 model reached a baseline ROC-AUC of 0.902 on MN187, surpassing all CNN baselines and the MoleAnalyzer Pro system, though the difference was not statistically significant (p = 0.07). After adding 46,000 confidence-filtered GAN-generated images, the ROC-AUC increased to 0.915, giving a statistically significant improvement over the commercial MoleAnalyzer Pro system (p = 0.032). Conclusions: Vision Transformers show strong potential for melanoma classification, especially when combined with GAN-based augmentation, offering advantages in global feature representation and data expansion that support the development of dependable AI-driven clinical decision-support systems.

1. Introduction

Melanoma is a malignant tumor developing from melanocytes, the neural crest-derived cells responsible for melanin production in the epidermis. Unlike non-melanoma skin cancers, melanoma exhibits a high propensity for early invasion and metastasis, making it one of the most lethal forms of skin cancer despite its relatively low incidence rate [1].
The pathogenesis of melanoma is driven by a combination of environmental exposures and genetic alterations. Prominent driver mutations include activating alterations in the BRAF gene, with the V600E mutation detected in approximately 40–60% of cutaneous melanomas, along with NRAS (15–20%) and NF1 mutations (10–15%) [1,2,3]. These mutations activate the MAPK and PI3K/AKT pathways, promoting uncontrolled cell proliferation, survival, and migration. In addition, inactivating mutations in tumor suppressor genes like CDKN2A contribute to melanoma progression [4,5]. Intermittent intense sun exposure and other ultraviolet radiation sources induce DNA damage and mutagenesis, acting as a primary environmental initiator in susceptible individuals [6].
Globally, the incidence of melanoma has been rising, especially in fair-skinned populations. According to international epidemiological data, melanoma accounts for approximately 1–2% of all cancer diagnoses worldwide, with over 325,000 new cases and 57,000 deaths annually [7]. It ranks as one of the top five cancers in countries with high UV exposure, such as Australia, New Zealand, and parts of North America and Europe. The increasing incidence among younger individuals, combined with the rising cost of advanced immunotherapies and targeted therapies, contributes to the growing global burden of melanoma [8,9].
Melanoma often appears similar to a harmless mole (nevus), making early detection difficult. This challenge arises because certain benign lesions, such as dysplastic or Spitz nevi, share dermoscopic and structural features with early melanomas [10,11]. The clinical and dermoscopic variability among melanoma subtypes, including acral lentiginous, nodular, superficial spreading, and in situ forms, further complicates early diagnosis [12]. As a result, early-stage tumors are often overlooked or mistaken for benign growths, especially when they lack clear asymmetry or pigment irregularity.
Early-stage detection improves prognosis. Localized melanomas have a 5-year survival rate that exceeds 98%, while metastatic melanoma has a survival rate below 30% despite modern therapies [1]. Thus, improving early diagnostic accuracy remains a priority for reducing melanoma-related mortality. Advances in digital pathology, molecular biomarkers, and AI-assisted dermoscopy are being investigated to improve differentiation between benign and malignant lesions [4,13].
Among AI-assisted tools, convolutional neural networks (CNNs) have been reported to achieve good performance in classifying skin lesions with accuracy comparable to that of dermatologists [14,15,16,17,18,19].
CNN-based models have shown good performance in various clinical settings: Brinker et al. reported that a ResNet-50 CNN outperformed 136 of 157 dermatologists (86%) [15], while Haenssle et al. found that their Inception v4 CNN surpassed most of the 58 dermatologists in diagnostic accuracy [18]. Similarly, Tschandl et al. showed that CNN algorithms significantly outperformed 511 human readers, including 27 experts, in classifying pigmented skin lesions [20]. Only a small minority of dermatologists achieved performance comparable to or better than AI systems, typically fewer than 15–25% of participants. More recent architectures have maintained this trend: Yan et al. showed that their Vision Transformer-based model PanDerm outperformed clinicians by 10.2% in early melanoma detection and achieved diagnostic accuracy comparable to clinicians even without human assistance [21].
However, human–AI collaboration has emerged as the most effective approach, surpassing either modality alone. Hekler et al. showed that combining CNN predictions with dermatologist assessments using a fusion model achieved 89% sensitivity, outperforming both dermatologists alone (66%) and the CNN alone (86.1%) [22]. Tschandl et al. found that AI-based decision support improved human diagnostic accuracy by 13.3%, with the greatest benefits seen in less experienced clinicians [23]. Winkler et al. reported that dermatologists cooperating with a market-approved CNN achieved 100% sensitivity and 86.4% accuracy, significantly exceeding dermatologists alone (84.2% sensitivity, 74.1% accuracy) and CNN alone (81.6% sensitivity, 87.7% accuracy) [24]. In a randomized controlled trial, Han et al. showed that CNN assistance significantly improved nondermatology trainees’ accuracy from 29.7% to 54.7%, though dermatology residents did not show significant improvement [25].
CNNs process dermoscopic images to identify subtle malignancy indicators such as asymmetry, border irregularity, color variation, and diameter [26,27,28,29]. Transfer learning is frequently used in this domain, where models pre-trained on large datasets like ImageNet are fine-tuned on smaller, domain-specific datasets [30,31,32,33]. Emerging approaches using spectral enhancement and multimodal imaging have been explored to improve the representation of diagnostically relevant features [34].
On the other hand, CNNs have structural limitations when applied to diverse populations and imaging conditions. Their architecture relies on spatial locality through convolutional kernels, which restricts their ability to model long-range relationships between distant regions of a lesion and reduces adaptability to variations in imaging devices or illumination. The need for a fixed input size further limits flexibility and can result in the loss of important information during image resizing or cropping. These factors may lead to ignoring delicate dermoscopic textures, such as irregular pigment networks and subtle color gradients extending over multiple spatial scales. In addition, dermoscopic patterns such as regression areas and perifollicular pigmentation often appear in separate regions of an image, requiring analytical models capable of integrating information from distant areas rather than relying solely on local texture patterns [35,36]. Similar limitations have been noted in medical image segmentation studies, where convolutional layers struggle to capture long-range dependencies [37], leading to the development of hybrid CNN–Transformer architectures designed to address these constraints [38].
To address these architectural constraints, Vision Transformers (ViTs) have been introduced as an alternative for image analysis in medicine [39,40]. ViTs apply multi-head self-attention mechanisms that compute pairwise relations among image patches, allowing the model to capture broad contextual structure while maintaining fine detail. Their self-attention mechanism allows the model to capture global contextual dependencies that extend beyond the local neighborhood of pixels, which is valuable in dermoscopic images where diagnostically relevant cues such as pigment networks, vascular structures, and streaks are spatially scattered [36,41].
Unlike CNNs, which rely on local receptive fields, ViTs model global relationships between image patches, preserving fine-grained texture information that CNNs may lose during resizing or pooling of high-resolution images. In addition to modeling global structure, attention-based representations depend less on local pixel statistics that vary with acquisition settings and therefore adapt better when annotated data are limited or originate from heterogeneous clinical sources. Recent studies on ViTs show improved consistency among datasets on several medical imaging benchmarks, including dermatology datasets [35,36,42]. These properties make Vision Transformers appropriate for dermoscopic analysis, where diagnostic cues are subtle, spatially dispersed, and sensitive to imaging variability [43].
ViTs typically require large amounts of training data due to their high parameter count and the absence of inductive biases found in CNNs. This demand creates challenges in medical domains where annotated datasets are often limited. Although large datasets for skin lesions are publicly available, such as the ISIC archive [44], the number of images, which is on the order of tens of thousands, remains small compared with general purpose classification training datasets like ImageNet that contain millions of instances. To address this limitation, data augmentation techniques can be applied. A popular technique involves using Generative Adversarial Networks (GANs) to generate synthetic images or applying self-supervised learning to make use of unlabeled data [45,46,47]. In addition, transfer learning from large-scale pre-trained ViT models offers a practical strategy to reduce the impact of data scarcity. Motivated by these advances, this study applies a Vision Transformer architecture for melanoma classification and compares its performance with that of a widely used commercial system deployed in clinical practice.
In this study, a pretrained ViT model is evaluated to classify dermoscopic images into benign and malignant categories. The model is trained and fine-tuned on the ISIC 2019 dataset, which offers a large collection of high-quality dermoscopic images with expert-provided annotations. To improve the training process, a GAN-based data augmentation pipeline is implemented, generating synthetic images that strengthen the model’s generalization capacity.
The performance of the resulting model was assessed on a new dataset containing images from hospitals in the United States. The images in this dataset received initial classifications from the FotoFinder AI MoleAnalyzer Pro (version 230814_3; FotoFinder Systems GmbH, Bad Birnbach, Germany), a commercial diagnostic tool, and were later confirmed by biopsy. The FotoFinder system is a well-established tool in the domain of skin lesion image classification, as documented in several studies [24,48,49,50,51,52,53,54].
This article is structured as follows. Section 2 describes the experimental setup, including datasets, model architectures, training, augmentation, and evaluation. Section 3 presents the performance results, statistical tests, and the effect of synthetic data on the evaluated models. Section 4 discusses the architectural and clinical implications of the results. Finally, Section 5 summarizes the study, acknowledges its limitations, and outlines directions for future research.

2. Materials and Methods

2.1. Dataset Preparation

For training the Vision Transformer model in melanoma classification, the ISIC 2019 Challenge dataset was used. This dataset is a large and diverse collection of dermoscopic images compiled by the International Skin Imaging Collaboration [55]. It contains more than 25,000 high-resolution images of skin lesions annotated by expert dermatologists. The collection includes various skin conditions such as benign nevi, dysplastic nevi, and malignant melanomas, covering a broad range of clinical appearances.
The dataset was divided into training and validation, with an external test set. ISIC provides an official train and validation split, with the original ISIC-2019 test set repurposed as the validation set in this study. The training set contained 12,875 nevus images and 4522 melanoma images, while the validation set contained 2495 nevus images and 1327 melanoma images. To evaluate the model’s generalization to unseen clinical data, a new dataset, MN187, was used as the test set.
The MN187 dataset consisted of 187 de-identified dermoscopic images obtained from the VA dermatology clinic (College of Medicine, University of South Florida). It included 96 benign nevi and 91 malignant melanomas, all biopsy confirmed. Each image was initially analyzed by the FotoFinder AI MoleAnalyzer Pro, which produced preliminary malignancy probability scores used for comparative evaluation.
The FotoFinder MoleAnalyzer Pro is a commercially approved CNN-based software for skin lesion classification. It was the first AI application for dermatology to receive approval in the European market [54]. This system uses deep learning algorithms to analyze dermoscopic images and provide malignancy risk assessments with probability scores of 0.0 to 1.0 , where higher scores indicate a higher risk of malignancy [49]. The system can identify melanoma, basal cell carcinoma, and most squamous cell carcinomas, including actinic keratoses [50].
Evaluations of the MoleAnalyzer Pro in multiple clinical studies have produced variable outcomes, influenced by the study setting and population. In controlled clinical environments, the system has been reported to have high diagnostic accuracy, with sensitivities from 81.6 % to 97.1 % and specificities from 78.8 % to 94.0 % [53]. One prospective diagnostic accuracy study found a sensitivity of 88.1 % and specificity of 78.8 % , with performance similar to that of expert dermatologists [52]. Another study found that the AI’s diagnostic accuracy was similar to experienced dermatologists, achieving an area under the ROC curve of 0.702 [50].
In contrast, real-world clinical studies have reported more modest results. One study of a high-risk melanoma population found lower metrics, with a sensitivity of 70 % and specificity of 40 % compared to histopathology as ground truth (ROC-AUC of 0.68 ) [49]. Other research found even lower performance, with a sensitivity of 36.4 % and specificity of 42.1 % , and concluded that the system cannot take the place of dermatologist decision-making [48].
The images used in this study were resized and padded to a consistent size of 224 × 224 pixels, which is the input size required by most pre-trained CNN backbone architectures. Padding with black pixels was applied to make images square, which allows us to keep the original proportions of images within the padding and to eliminate distortions that could affect model performance. Although ViT models can handle variable input sizes, input dimensions were standardized in every model to enable fair comparisons.

2.2. Model Architecture

Modern deep learning classification models typically consist of a pretrained feature extraction backbone paired with a classification head composed of fully connected layers. The backbone, often a convolutional neural network or a transformer-based model, is trained on large datasets like ImageNet, while the classification head is fine-tuned for the specific task.
In this study, the Vision Transformer was chosen as the backbone architecture. ViTs represent a significant advancement in image classification by using self-attention mechanisms to capture global context. Introduced by Vaswani et al. (2017), attention mechanisms have revolutionized natural language processing and generative AI [56]. Although initially developed for text data, this architecture has been successfully adapted for image analysis [57].
Vision Transformers divide images into fixed-size patches, embed them linearly, and process them through transformer encoders. This allows ViTs to capture long-range dependencies and complex relationships between image regions. For this study, the ViT-L/16 model pretrained on the ImageNet-1k dataset was selected. This model is one of the largest publicly available pretrained Vision Transformers in the PyTorch library. It contains 24 transformer layers, each with 16 attention heads and a hidden size of 1024. Input images are divided into 16 × 16 pixel patches, producing a sequence of 196 patches for a standard 224 × 224 image. Each patch is embedded into a 1024-dimensional vector, with positional embeddings added to preserve spatial information. The transformer layers include multi-head self-attention mechanisms, feed-forward neural networks, layer normalization, and residual connections. With approximately 307 million parameters, the detailed architecture and implementation specifics, including the learning rate schedule, optimizer settings, dropout regularization, data augmentation, and training epochs, are summarized in Figure 1.
The classification head consists of four fully connected layers with ReLU activations: 256, 128, 64, and 1 neurons, respectively. The final layer uses a sigmoid activation function to generate a probability score between 0 and 1, indicating the probability of malignancy. Dropout layers with a 0.5 dropout rate were added after each fully connected layer to prevent overfitting.
For comparison, several CNN backbones were evaluated, including ResNet-152, DenseNet-201, EfficientNet-B7, and InceptionV4. These models, recognized as top performers in their respective architecture families, have achieved good results in image classification tasks, including skin lesion analysis [17,58,59,60]. Additionally, the ViT-B/16 model, a smaller variant of the ViT-L/16, was included. This model features 12 transformer layers, 12 attention heads, and a hidden size of 768, while maintaining a similar architecture to that shown in Figure 1. The classification head architecture was the same for all models to ensure a fair comparison.

2.3. Training Configuration and Optimization

All experiments were implemented in PyTorch (version 2.8.0) and executed on a pair of NVIDIA RTX A5000 GPUs with 48 GB total VRAM. For consistency among models, a unified training configuration was adopted. Images were normalized using the ImageNet mean and standard deviation ([0.485, 0.456, 0.406], [0.229, 0.224, 0.225]), in line with the pretrained weights. The training pipeline included random resized cropping to 224 pixels with a scale range of 0.7 to 1.0, random horizontal flipping, and RandAugment. The training set was loaded with a batch size of 128 and shuffled at each epoch, while the validation and test sets were processed with batch sizes of 64 images.
The Vision Transformer backbone was initialized with ImageNet-1k pretrained weights. The transformer backbone parameters were frozen, and only the classification head was trained.
Training optimization used binary cross-entropy with logits (BCEWithLogitsLoss) as the loss function. Parameters were updated with the AdamW optimizer configured with a learning rate of 2 × 10 4 and a weight decay of 0.05 . A cosine annealing learning rate scheduler with T max = 30 progressively reduced the learning rate through epochs.

2.4. Data Augmentation

To address class imbalance and broaden the training distribution, generative models were applied to the ISIC 2019 challenge dataset, which contains 12,875 nevus and 4522 melanoma images. Separate GAN models were trained for each class on images and resized to 256 × 256 pixels, allowing the networks to capture class-specific distributions.
Table 1 lists the principal parameters and settings used for the StyleGAN2-ADA models. The table summarizes hardware configuration, dataset preprocessing, optimizer setup, and transfer learning details applied during fine-tuning from a pretrained checkpoint.
The GAN models followed the StyleGAN2-ADA architecture described in [61] and used the official NVIDIA implementation [62], with modifications introduced for compatibility with newer PyTorch and CUDA versions. The StyleGAN2-ADA training procedure is based on ticks, where each tick processes a fixed number of images. Each model was trained for 1000 ticks, and snapshots were recorded every five ticks to monitor generator progress.
The FID50k_full metric, which computes the Fréchet Inception Distance (FID) over 50,000 generated images, served to monitor convergence. Each GAN required approximately two to three days of training under these conditions.
After training, each GAN produced 25,000 images for its class, resulting in a total of 50,000 synthetic samples. These generated images maintained consistency with real dermoscopic data and reflected intra-class variability. Generated outputs were filtered using the ViT-L/16 classifier to exclude samples with low confidence or unrealistic appearance before inclusion in the augmented dataset. This filtering step was applied only to synthetic images derived from the ISIC 2019 dataset. No data from the MN187 set were used, referenced, or accessed during GAN training or filtering, and MN187 remained fully reserved for final evaluation.
All synthetic images were resized and normalized to match the preprocessing of the real data ( 224 × 224 pixels, normalized with ImageNet statistics). The combined set of real and generated images formed a larger and more balanced training dataset that improved the performance of the classification model.

2.5. Model Performance Monitoring and Evaluation

The Receiver Operating Characteristic Area Under the Curve (ROC-AUC) serves as the primary evaluation metric in this study. This threshold-independent metric measures binary classification performance by quantifying the model’s capacity to separate positive and negative classes through all classification thresholds [63,64,65]. ROC-AUC is appropriate for imbalanced datasets and finds common application in medical imaging tasks such as skin lesion classification.
A high ROC-AUC score alone does not fully characterize model performance. For instance, a model might obtain a high score by correctly ranking a small number of straightforward cases, which could overstate its general discriminative capability [66].
To evaluate the statistical significance of performance differences, the ROC-AUC of each model was compared with a predefined baseline model using DeLong’s test [67]. This nonparametric method is a well-established approach for comparing the AUCs of two classifiers tested on the same dataset, accounting for the correlated nature of their predictions and the variability among paired samples.
The statistical hypotheses were defined as follows: the null hypothesis ( H 0 ) states that no difference in ROC-AUC exists between the baseline model and the evaluated model ( A U C baseline = A U C evaluated ); the alternative hypothesis ( H 1 ) states that a difference exists ( A U C baseline A U C evaluated ). A significance level of α = 0.05 was applied uniformly to all comparisons, and 95% confidence intervals were estimated to quantify the uncertainty associated with each AUC difference. A p-value from DeLong’s test below α ( p < 0.05 ) was interpreted as evidence of a statistically significant difference in model performance, whereas values of p 0.05 were considered consistent with random variation, implying no significant difference between models.

2.6. Model Interpretability

To improve the interpretability of the Vision Transformer–based skin lesion classifier, an attention-based visualization technique was applied to make explicit how the model distributes its focus across image regions during prediction. This approach provides a clear view of the spatial behavior of the network by presenting the self-attention responses computed within the transformer encoder.
In the ViT architecture, each input image I R H × W × 3 is divided into a sequence of N non-overlapping patches { x 1 , x 2 , , x N } , each projected into an D-dimensional embedding space [57]. Within each transformer layer l, the self-attention mechanism evaluates pairwise relationships among all patch tokens. For a given attention head h, the attention scores are computed as formulated in [56]:
A ( l , h ) = softmax Q ( l , h ) K ( l , h ) d k ,
where Q ( l , h ) = X ( l ) W Q ( l , h ) , K ( l , h ) = X ( l ) W K ( l , h ) , and X ( l ) denote the input patch embeddings to layer l. The resulting attention matrix A ( l , h ) R N × N represents the strength of interaction between image patches.
A single attention map per layer is obtained by averaging the attention matrices among all heads:
A ( l ) = 1 H h = 1 H A ( l , h ) ,
where H is the number of attention heads. The first row of A ( l ) , corresponding to the [CLS] token, is taken as the measure of global attention assigned to each image patch. This attention is reshaped into a spatial grid and displayed as a heatmap, with higher intensity values representing greater relevance to the model’s decision [68].
Through this process, the evolution of the model’s focus through transformer layers can be examined, from diffuse attention in the earlier stages to more spatially localized activation patterns in deeper layers. Examination of these attention distributions helps determine of whether the model concentrates on clinically relevant regions in the dermoscopic images, thus supporting interpretability in its diagnostic behavior [69].

3. Results

3.1. Model Performance

The proposed models were evaluated using the ROC-AUC metric. Table 2 summarizes the results for six deep learning classification models, each using distinct backbones for feature extraction while sharing the same classification head, as depicted in Figure 1. These models were trained on the ISIC-19 dataset, using standard image transformations (Resize, CenterCrop, and Normalization) integrated into the training pipeline. Validation was performed on the ISIC-19 dataset, and testing was performed on the MN187 dataset.
The outcomes of six runs for each model are detailed in Table 2. The ROC Curves and AUC scores for the top-performing model, ViT-L/16, on both validation and test datasets are presented in Figure 2.
Although the models were trained on the ISIC-19 dataset, they kept generalization capabilities when evaluated on the MN187 dataset, performing at a level similar to or exceeding the MoleAnalyzer Pro. The ViT-L/16 model achieved the highest ROC-AUC score of 0.902 on the MN187 dataset, exceeding all other models. The ConvNeXt-XL, a CNN-based model larger than ViT-L/16 in terms of parameters, achieved the second-highest ROC-AUC score of 0.885, comparable to the ViT-B/16 model. The other CNN models had lower performance, aligning more closely with the MoleAnalyzer Pro.
While most models achieved higher ROC-AUC scores than the MoleAnalyzer Pro, the statistical significance of these improvements was evaluated using DeLong’s test. Table 3 summarizes the results, comparing the ROC-AUC of each deep learning model to that of the MoleAnalyzer Pro.
The ViT-L/16 model achieved an ROC-AUC of 0.901, which exceeds the MoleAnalyzer Pro ROC-AUC of 0.856. However, the p-value of 0.070 means this difference is not statistically significant at the 0.05 significance level. The CNN models (EfficientNet-B7, DenseNet-201, ConvNeXt-XL) and the ViT-B/16 also had higher ROC-AUC scores than the MoleAnalyzer Pro, but their p-values suggest these differences are not statistically significant. None of the models had a performance improvement that could be attributed to factors other than chance, so the null hypothesis could not be rejected.

3.2. Results with Data Augmentation

While models like ViT-L/16 and ConvNeXt-XL achieved higher ROC-AUC scores and lower probabilities of the differences being due to chance, they did not reach a statistically significant improvement over the MoleAnalyzer Pro. Vision Transformers, due to their high parameter count and lack of inherent CNN inductive biases, often need large volumes of training data for optimal performance.
To address this, 50,000 synthetic images generated with GANs were added to the training dataset. As these synthetic images come from a trained GAN model, they might not match the true data distribution perfectly and could contain artifacts or biases. To manage this, the synthetic images were classified by the ViT-L/16 model, which had the highest ROC-AUC score. This model generated a new subset from the 50,000 synthetic images by choosing only those it classified with high confidence (probability > 0.6 for melanoma and <0.4 for nevus), as seen in Figure 3. This method added only high-quality synthetic images that were close to real images to the training set.
Figure 4 presents a sample of the synthetic images from the GANs. The left column has synthetic nevus images, and the right column has synthetic melanoma images. These images have a range of colors, textures, and patterns found in real dermoscopic images. The GANs produced realistic images that could be useful for training the classification model.
Before the addition of synthetic images, the ISIC-19 dataset had 12,875 nevus and 4522 melanoma images (class proportions of 74.0% nevus and 26.0% melanoma). After filtering with the ViT-L/16 model, 27,793 nevus and 18,438 melanoma synthetic images were chosen. This created a new training dataset with a total of 40,668 nevus and 22,960 melanoma images (adjusting the class proportions to 63.9% nevus and 36.1% melanoma), which increased the size and diversity of the training data.

3.3. Model Performance with Data Augmentation

Including high-confidence synthetic images in the training data produced performance gains for all model architectures when compared to their performance without synthetic data (presented in Table 2). Table 4 presents results from six independent training runs on both the validation (ISIC-2019) and external test (MN187) sets. The uniformity of these improvements over multiple runs supports the conclusion that the data augmentation strategy provided a general benefit.
The ViT-L/16 architecture remained the strongest performer, reaching a top ROC-AUC of 0.915 on the challenging MN187 test set, as presented in Figure 5. This result exceeds its previous best score of 0.902. Other models also showed classification performance improvements; the ConvNeXt-XL model advanced from a previous score of 0.885 to 0.894. Similar positive trends were observed for the ViT-B/16, ResNet-152, EfficientNet-B7, and DenseNet-201 models, confirming that the augmented dataset created a more effective training environment.
To evaluate the statistical significance of these improvements, the best run from each model was compared against the established benchmark, MoleAnalyzer Pro, using DeLong’s test. The results of this analysis are in Table 5. The standard ViT-L/16 model, with an AUC of 0.915, achieved a statistically significant advantage over the commercial software (p = 0.032).
All statistical tests were performed relative to the commercial MoleAnalyzer Pro system. No statistical comparison was conducted between the augmented and non-augmented ViT-L/16 models, as the focus was on demonstrating whether augmentation resulted in statistically validated performance relative to the clinical benchmark.
A refined version of the ViT-L/16 model was also developed for further improvement. This fine-tuned model, included in Figure 5 and Table 5, was created by initializing with the best ViT-L/16 run, freezing its classification head, and unfreezing the final six transformer layers. These layers were then trained for an additional 30 epochs with a reduced learning rate (1 × 10−4).
The model was trained on the same ISIC 2019 dataset used in the previous runs that optimized only the classification head, but this time, the focus was on updating the final six transformer layers of the encoder. Training followed the same structure as previous ViT runs, using identical data splits, augmentations, and optimization parameters. The AdamW optimizer with cosine annealing and a weight decay of 0.05 was applied, along with early stopping (patience = 15).
This stage began from the previously best-performing pretrained model, using the same learned weights as initialization. The earlier parameters were frozen to preserve their representations, and training was restarted with updates restricted to the last six of the 24 transformer layers. This design allowed the model to adjust its higher-level representations to the ISIC 2019 domain while maintaining stable lower-level features.
The outcome was a model that achieved an ROC-AUC of 0.926 on the MN187 test set. The statistical comparison with MoleAnalyzer Pro for this model produced a p-value of 0.006, providing evidence that its improved performance is statistically significant.

3.4. Attention Map Analysis

Attention maps from all 24 layers of the Vision Transformer were analyzed to observe how the model’s focus changed through the layers. In the initial layers, attention was distributed between both the lesion and the surrounding background, capturing structural features like overall shape, border contrast, and color patterns, similar to how a clinician begins a visual assessment.
In deeper layers, attention became more concentrated within smaller regions of the image, often centered on irregular pigment clusters or darker zones. This progression shows how the Vision Transformer gradually shifts from processing general spatial information to concentrating on detailed features that support accurate classification.
For visualization, the attention matrices from each layer were averaged among all heads and represented as heatmaps (Figure 6). This was generated by selecting one sample from the MN187 dataset and using the best-performing, fine-tuned Vision Transformer model (AUC 0.926). As shown, the attention maps evolve from broadly distributed patterns in the initial layers to more concentrated activations in the deeper layers.
An averaged attention map was produced by combining the responses from layers 3 to 5 (Figure 6). The overlaid heatmap representations were derived from the attention tensors of these layers, normalized across layers 3, 4, and 5, then averaged and rendered as a single heatmap. The choice of layers was guided by an entropy-based analysis performed on the validation subset of the ISIC dataset. Entropy was computed from the same attention tensors to measure the variability of activation among image regions, and the mean entropy per layer was obtained by averaging over all samples. Layers 3–5 showed the highest mean entropy values among the 24 layers, presenting greater diversity in spatial representations. The corresponding entropy values for all layers are presented in Table 6.
The visualization in Figure 7 shows how the model distributed its spatial attention. In several nevus examples, attention was spread across outer structures such as terminal hairs around the lesion, shown as darker blue areas in the heatmap. For instance, in the first nevus image, the model focused more on the thin terminal hairs extending from the lesion surface rather than the central pigmented region, indicating that these outer features were included in the model’s internal representation. This pattern of attention related to terminal hairs is consistent with dermoscopic observations, where hair growth within a lesion is often associated with benign nevi [70].
In contrast, melanoma samples often showed attention focused along the lesion borders or around uneven outer regions. This pattern matches known dermoscopic features such as irregular borders and uneven pigment distribution [71]. Some nevus heatmaps showed broader attention covering most of the lesion, while melanoma heatmaps concentrated on specific irregular areas or edges, suggesting that the model captured spatial patterns similar to those used in clinical dermoscopic assessment.

4. Discussion

The results of this study provide evidence for the potential of Vision Transformers in melanoma classification using dermoscopic images. The ViT-L/16 model achieved the highest ROC-AUC score among all evaluated models, which demonstrates its capacity to capture global contextual information in dermoscopic images. This performance supports the advantages of self-attention mechanisms when analyzing complex visual patterns found in skin lesions.
Despite promising results, the ROC-AUC differences compared with MoleAnalyzer Pro were not statistically significant in the initial experiments. This result shows the difficulty of achieving measurable performance gains in medical imaging, where high baselines and limited datasets restrict model development. Adding synthetic images produced by a GAN addressed data scarcity and produced a statistically significant improvement when compared with the commercial MoleAnalyzer Pro system (p = 0.032). This result implies that data augmentation can improve model generalization in domains with limited annotated data.
Using a GAN to generate synthetic images provided a practical approach to augment the training dataset. Filtering synthetic images based on model confidence reduced the chance of introducing artifacts or biases. This selection process limited the set to higher-quality synthetic images, which likely improved ViT-L/16 performance. However, reliance on synthetic data also raises concerns about possible limitations, including the risk of overfitting to the synthetic distribution.
When comparing ViT and CNN architectures, ViT generally outperformed the CNN-based models in this study. Although model size is an important factor, the superior performance of ViT-L/16 cannot be explained by parameter count alone. For example, the CNN-based ConvNeXt Large model contains more parameters than ViT-L/16 yet achieves lower performance. The results support the view that the Transformer architecture itself, with its self-attention mechanism, may be a better suited for the task. This result is consistent with recent reports showing stronger outcomes from ViT on a range of medical imaging applications [35,36,42]. However, ViT models require greater computational resources during both training and inference, which can limit their practicality in resource-constrained settings.
Another important point is the need for statistical validation when assessing model performance. Although ViT-L/16 produced higher ROC-AUC scores than MoleAnalyzer Pro, the absence of statistical significance in the initial comparison highlights the need for tests such as DeLong’s method to verify that the observed gains are not due to chance. In this study, all statistical comparisons were performed relative to the commercial MoleAnalyzer Pro system to determine whether the proposed models reached clinically validated performance levels. In clinical research, reporting metrics without such validation can be misleading, since a numerically higher value may not correspond to a real advancement. For applications that influence medical decisions, establishing statistical significance is an important step toward building confidence in a model’s reported results. This practice helps to ensure that performance claims are reliable and not simply the result of variability in the training process.
The fine-tuned ViT-L/16 variant, obtained by unfreezing only the last six transformer blocks and training them with a reduced learning rate, delivered an incremental improvement (ROC-AUC 0.926; p = 0.006 vs. MoleAnalyzer Pro) over the frozen-backbone configuration (best ROC-AUC 0.915). This suggests that limited, targeted adaptation of high-level self-attention layers helps align global contextual features with dermoscopic lesion morphology without incurring instability or overfitting. Expanding fine-tuning depth, applying layer-wise learning rate decay, or integrating lightweight adapters (such as LoRA) may further improve performance, but these strategies can lead to an increased risk of overfitting in limited-data and domain-shift settings [72,73,74].
Beyond performance, the Vision Transformer handles interpretability differently from convolution-based systems such as MoleAnalyzer Pro. Its self-attention layers produce maps showing which image regions contribute to the classification result, as described in the results. In dermoscopic images, these regions often align with visible characteristics such as uneven pigmentation, irregular borders, or streaks near the lesion edge. CNN-based approaches like Grad-CAM rely on gradient information to visualize relevant areas. The attention structure in Transformers provides an alternative representation, based on spatial relations within the model, which can be examined to see how attention is distributed during prediction and whether it corresponds to features used in dermatologic assessment.
Commercial skin analysis platforms such as MoleAnalyzer Pro, which currently rely on convolutional neural networks, could benefit from the integration of Vision Transformer models. These newer models generalize better, providing accurate diagnoses on data not seen during training. An important architectural advantage is their flexibility with input sizes, which removes the need for preprocessing steps such as padding and resizing. This combination of properties reduces manual image tuning and leads to more consistent performance in practice.

5. Conclusions

This study evaluated the clinical potential of Vision Transformers (ViTs) for melanoma classification using two dermoscopic datasets. Within a standardized experimental framework, the ViT-L/16 architecture consistently outperformed conventional CNN-based models and the commercial FotoFinder MoleAnalyzer Pro system. A statistically significant improvement in predictive performance was observed only when the GAN-augmented ViT-L/16 model was compared with MoleAnalyzer Pro (p = 0.032). This outcome supports the usefulness of synthetic data in addressing class imbalance and the limited availability of labeled clinical images.
Several conclusions can be drawn. First, ViTs effectively capture global contextual information within dermoscopic images, leading to more accurate separation of malignant and benign lesions. Second, when filtered carefully, GAN-based augmentation improves data diversity and generalization, reducing the constraints imposed by relatively small medical imaging datasets. Third, statistical evaluation with DeLong’s test is necessary to verify that observed performance gains are reproducible and clinically meaningful.
The study also has limitations. The external validation dataset, although biopsy-confirmed, was limited in size and diversity, which may affect general applicability. Furthermore, the analysis focused exclusively on binary classification (melanoma versus nevus), without considering other skin lesion categories. Expanding the approach to multi-class classification and incorporating additional clinical factors could improve practical relevance.
Future work should include larger and more diverse datasets, integration of clinical metadata and patient risk profiles to provide diagnostic context, and the adoption of interpretable AI approaches to strengthen clinician confidence. These steps would support the progression of Vision Transformer–based systems toward real-world dermatology applications and contribute to earlier melanoma detection and better patient outcomes.

Author Contributions

Conceptualization, A.G. and J.Z.; methodology, A.G.; software, A.G. and J.Z.; formal analysis, X.H.; investigation, A.G.; resources, G.P.-C. and T.B.; data curation, J.Z., G.P.-C. and T.B.; writing—original draft preparation, A.G.; writing—review and editing, X.H.; supervision, X.H., G.P.-C. and T.B.; project administration, X.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board of the University of South Florida (IRB #007837, approval date: 21 January 2025) and the James A. Haley Veterans Hospital—VA IRBNet Number: 1816729, approval date: 21 January 2025.

Informed Consent Statement

Patient consent was waived due to the retrospective nature of the study. All dermoscopic images were collected from a de-identified database, and it would not have been feasible to obtain informed consent from all patients, as some may have moved away or passed away. The IRB approved this waiver as described in the protocol: “We are requesting a waiver of consent to use the de-identified dermoscopy images of patients seen at the VA dermatology clinic, that are normally stored on the VA Sharepoint drive. This study cannot be conducted without biopsy proven dermoscopy pictures, and the recruitment timeline would be impracticably long to do this prospectively as patients come in. It would be impracticable to obtain consent from the patients whose images we intend to use because they are de-identified, so we would not be able to contact the patients, and some of them may have passed away or left the VA system.”

Data Availability Statement

The data supporting the findings of this study (including the MN187 clinical dermoscopic images and derived model outputs) are available from the corresponding author upon reasonable request. Due to patient privacy and institutional review constraints, the raw clinical images cannot be placed in a public repository. De-identified metadata, trained model weights, and synthetic GAN-generated images can be shared subject to a data use agreement.

Acknowledgments

We acknowledge the support provided to A. Garcia for his Ph.D. studies at Worcester Polytechnic Institute by the Secretaría Nacional de Ciencia, Tecnología e Innovación (SENACYT), and the Government of the Republic of Panama through the Fulbright-SENACYT fellowship.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Guo, W.; Wang, H.; Li, C. Signal pathways of melanoma and targeted therapy. Signal Transduct. Target. Ther. 2021, 6, 424. [Google Scholar] [CrossRef]
  2. Platz, A.; Egyhazi, S.; Ringborg, U.; Hansson, J. Human cutaneous melanoma; a review of NRAS and BRAF mutation frequencies in relation to histogenetic subclass and body site. Mol. Oncol. 2008, 1, 395–405. [Google Scholar] [CrossRef]
  3. Thomas, N.E.; Edmiston, S.N.; Alexander, A.; Groben, P.A.; Parrish, E.; Kricker, A.; Armstrong, B.K.; Anton-Culver, H.; Gruber, S.B.; From, L.; et al. Association Between NRAS and BRAF Mutational Status and Melanoma-Specific Survival Among Patients with Higher-Risk Primary Melanoma. JAMA Oncol. 2015, 1, 359. [Google Scholar] [CrossRef] [PubMed]
  4. Kim, H.J.; Kim, Y.H. Molecular Frontiers in Melanoma: Pathogenesis, Diagnosis, and Therapeutic Advances. Int. J. Mol. Sci. 2024, 25, 2984. [Google Scholar] [CrossRef] [PubMed]
  5. Palmieri, G.; Capone, M.; Ascierto, M.L.; Gentilcore, G.; Stroncek, D.F.; Casula, M.; Sini, M.C.; Palla, M.; Mozzillo, N.; Ascierto, P.A. Main roads to melanoma. J. Transl. Med. 2009, 7, 86. [Google Scholar] [CrossRef]
  6. Craig, S.; Earnshaw, C.H.; Virós, A. Ultraviolet light and melanoma. J. Pathol. 2018, 244, 578–585. [Google Scholar] [CrossRef] [PubMed]
  7. Arnold, M.; Singh, D.; Laversanne, M.; Vignat, J.; Vaccarella, S.; Meheus, F.; Cust, A.E.; De Vries, E.; Whiteman, D.C.; Bray, F. Global Burden of Cutaneous Melanoma in 2020 and Projections to 2040. JAMA Dermatol. 2022, 158, 495. [Google Scholar] [CrossRef]
  8. De Pinto, G.; Mignozzi, S.; La Vecchia, C.; Levi, F.; Negri, E.; Santucci, C. Global trends in cutaneous malignant melanoma incidence and mortality. Melanoma Res. 2024, 34, 265–275. [Google Scholar] [CrossRef]
  9. Gershenwald, J.E.; Guy, G.P. Stemming the Rising Incidence of Melanoma: Calling Prevention to Action. JNCI J. Natl. Cancer Inst. 2016, 108, djv381. [Google Scholar] [CrossRef]
  10. Elder, D.E. Precursors to melanoma and their mimics: Nevi of special sites. Mod. Pathol. 2006, 19, S4–S20. [Google Scholar] [CrossRef]
  11. Raluca Jitian (Mihulecea), C.; Frățilă, S.; Rotaru, M. Clinical-dermoscopic similarities between atypical nevi and early stage melanoma. Exp. Ther. Med. 2021, 22, 854. [Google Scholar] [CrossRef] [PubMed]
  12. Lu, C.T.; Lin, T.L.; Mukundan, A.; Karmakar, R.; Chandrasekar, A.; Chang, W.Y.; Wang, H.C. Skin Cancer: Epidemiology, Screening and Clinical Features of Acral Lentiginous Melanoma (ALM), Melanoma In Situ (MIS), Nodular Melanoma (NM) and Superficial Spreading Melanoma (SSM). J. Cancer 2025, 16, 3972–3990. [Google Scholar] [CrossRef]
  13. Liu, P.; Su, J.; Zheng, X.; Chen, M.; Chen, X.; Li, J.; Peng, C.; Kuang, Y.; Zhu, W. A Clinicopathological Analysis of Melanocytic Nevi: A Retrospective Series. Front. Med. 2021, 8, 681668. [Google Scholar] [CrossRef]
  14. Hekler, A.; Utikal, J.S.; Enk, A.H.; Solass, W.; Schmitt, M.; Klode, J.; Schadendorf, D.; Sondermann, W.; Franklin, C.; Bestvater, F.; et al. Deep learning outperformed 11 pathologists in the classification of histopathological melanoma images. Eur. J. Cancer 2019, 118, 91–96. [Google Scholar] [CrossRef] [PubMed]
  15. Brinker, T.J.; Hekler, A.; Enk, A.H.; Klode, J.; Hauschild, A.; Berking, C.; Schilling, B.; Haferkamp, S.; Schadendorf, D.; Fröhling, S.; et al. A convolutional neural network trained with dermoscopic images performed on par with 145 dermatologists in a clinical melanoma image classification task. Eur. J. Cancer 2019, 111, 148–154. [Google Scholar] [CrossRef]
  16. Brinker, T.J.; Hekler, A.; Enk, A.H.; Berking, C.; Haferkamp, S.; Hauschild, A.; Weichenthal, M.; Klode, J.; Schadendorf, D.; Holland-Letz, T.; et al. Deep neural networks are superior to dermatologists in melanoma image classification. Eur. J. Cancer 2019, 119, 11–17. [Google Scholar] [CrossRef]
  17. Jojoa Acosta, M.F.; Caballero Tovar, L.Y.; Garcia-Zapirain, M.B.; Percybrooks, W.S. Melanoma diagnosis using deep learning techniques on dermoscopic images. BMC Med. Imaging 2021, 21, 6. [Google Scholar] [CrossRef]
  18. Haenssle, H.; Fink, C.; Schneiderbauer, R.; Toberer, F.; Buhl, T.; Blum, A.; Kalloo, A.; Hassen, A.B.H.; Thomas, L.; Enk, A.; et al. Man against machine: Diagnostic performance of a deep learning convolutional neural network for dermoscopic melanoma recognition in comparison to 58 dermatologists. Ann. Oncol. 2018, 29, 1836–1842. [Google Scholar] [CrossRef]
  19. Nasr-Esfahani, E.; Samavi, S.; Karimi, N.; Soroushmehr, S.; Jafari, M.; Ward, K.; Najarian, K. Melanoma detection by analysis of clinical images using convolutional neural network. In Proceedings of the 2016 38th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Orlando, FL, USA, 16–20 August 2016; pp. 1373–1376. [Google Scholar] [CrossRef]
  20. Tschandl, P.; Codella, N.; Akay, B.N.; Argenziano, G.; Braun, R.P.; Cabo, H.; Gutman, D.; Halpern, A.; Helba, B.; Hofmann-Wellenhof, R.; et al. Comparison of the accuracy of human readers versus machine-learning algorithms for pigmented skin lesion classification: An open, web-based, international, diagnostic study. Lancet Oncol. 2019, 20, 938–947. [Google Scholar] [CrossRef]
  21. Yan, S.; Yu, Z.; Primiero, C.; Vico-Alonso, C.; Wang, Z.; Yang, L.; Tschandl, P.; Hu, M.; Ju, L.; Tan, G.; et al. A multimodal vision foundation model for clinical dermatology. Nat. Med. 2025, 31, 2691–2702. [Google Scholar] [CrossRef] [PubMed]
  22. Hekler, A.; Utikal, J.S.; Enk, A.H.; Hauschild, A.; Weichenthal, M.; Maron, R.C.; Berking, C.; Haferkamp, S.; Klode, J.; Schadendorf, D.; et al. Superior skin cancer classification by the combination of human and artificial intelligence. Eur. J. Cancer 2019, 120, 114–121. [Google Scholar] [CrossRef]
  23. Tschandl, P.; Rinner, C.; Apalla, Z.; Argenziano, G.; Codella, N.; Halpern, A.; Janda, M.; Lallas, A.; Longo, C.; Malvehy, J.; et al. Human–computer collaboration for skin cancer recognition. Nat. Med. 2020, 26, 1229–1234. [Google Scholar] [CrossRef]
  24. Winkler, J.K.; Blum, A.; Kommoss, K.; Enk, A.; Toberer, F.; Rosenberger, A.; Haenssle, H.A. Assessment of Diagnostic Performance of Dermatologists Cooperating with a Convolutional Neural Network in a Prospective Clinical Study: Human with Machine. JAMA Dermatol. 2023, 159, 621. [Google Scholar] [CrossRef]
  25. Han, S.S.; Kim, Y.J.; Moon, I.J.; Jung, J.M.; Lee, M.Y.; Lee, W.J.; Won, C.H.; Lee, M.W.; Kim, S.H.; Navarrete-Dechent, C.; et al. Evaluation of Artificial Intelligence–Assisted Diagnosis of Skin Neoplasms: A Single-Center, Paralleled, Unmasked, Randomized Controlled Trial. J. Investig. Dermatol. 2022, 142, 2353–2362.e2. [Google Scholar] [CrossRef]
  26. Nigar, N.; Umar, M.; Shahzad, M.K.; Islam, S.; Abalo, D. A Deep Learning Approach Based on Explainable Artificial Intelligence for Skin Lesion Classification. IEEE Access 2022, 10, 113715–113725. [Google Scholar] [CrossRef]
  27. Barata, C.; Celebi, M.E.; Marques, J.S. Explainable skin lesion diagnosis using taxonomies. Pattern Recognit. 2021, 110, 107413. [Google Scholar] [CrossRef]
  28. Shorfuzzaman, M. An explainable stacked ensemble of deep learning models for improved melanoma skin cancer detection. Multimed. Syst. 2021, 28, 1309–1323. [Google Scholar] [CrossRef]
  29. Mridha, K.; Uddin, M.M.; Shin, J.; Khadka, S.; Mridha, M.F. An Interpretable Skin Cancer Classification Using Optimized Convolutional Neural Network for a Smart Healthcare System. IEEE Access 2023, 11, 41003–41018. [Google Scholar] [CrossRef]
  30. Ali, M.S.; Miah, M.S.; Haque, J.; Rahman, M.M.; Islam, M.K. An enhanced technique of skin cancer classification using deep convolutional neural network with transfer learning models. Mach. Learn. Appl. 2021, 5, 100036. [Google Scholar] [CrossRef]
  31. Kassem, M.A.; Hosny, K.M.; Fouad, M.M. Skin Lesions Classification Into Eight Classes for ISIC 2019 Using Deep Convolutional Neural Network and Transfer Learning. IEEE Access 2020, 8, 114822–114832. [Google Scholar] [CrossRef]
  32. Alzubaidi, L.; Al-Amidie, M.; Al-Asadi, A.; Humaidi, A.J.; Al-Shamma, O.; Fadhel, M.A.; Zhang, J.; Santamaría, J.; Duan, Y. Novel Transfer Learning Approach for Medical Imaging with Limited Labeled Data. Cancers 2021, 13, 1590. [Google Scholar] [CrossRef]
  33. Hosny, K.M.; Kassem, M.A.; Foaud, M.M. Classification of skin lesions using transfer learning and augmentation with Alex-net. PLoS ONE 2019, 14, e0217293. [Google Scholar] [CrossRef] [PubMed]
  34. Lin, T.L.; Lu, C.T.; Karmakar, R.; Nampalley, K.; Mukundan, A.; Hsiao, Y.P.; Hsieh, S.C.; Wang, H.C. Assessing the Efficacy of the Spectrum-Aided Vision Enhancer (SAVE) to Detect Acral Lentiginous Melanoma, Melanoma In Situ, Nodular Melanoma, and Superficial Spreading Melanoma. Diagnostics 2024, 14, 1672. [Google Scholar] [CrossRef]
  35. Yang, G.; Luo, S.; Greer, P. A Novel Vision Transformer Model for Skin Cancer Classification. Neural Process. Lett. 2023, 55, 9335–9351. [Google Scholar] [CrossRef]
  36. Xin, C.; Liu, Z.; Zhao, K.; Miao, L.; Ma, Y.; Zhu, X.; Zhou, Q.; Wang, S.; Li, L.; Yang, F.; et al. An improved transformer network for skin cancer classification. Comput. Biol. Med. 2022, 149, 105939. [Google Scholar] [CrossRef]
  37. Khan, R.F.; Lee, B.D.; Lee, M.S. Transformers in medical image segmentation: A narrative review. Quant. Imaging Med. Surg. 2023, 13, 5048–5064. [Google Scholar] [CrossRef]
  38. Yuan, F.; Zhang, Z.; Fang, Z. An effective CNN and Transformer complementary network for medical image segmentation. Pattern Recognit. 2023, 139, 109520. [Google Scholar] [CrossRef]
  39. Al-hammuri, K.; Gebali, F.; Kanan, A.; Chelvan, I.T. Vision transformer architecture and applications in digital health: A tutorial and survey. Vis. Comput. Ind. Biomed. Art 2023, 6, 14. [Google Scholar] [CrossRef] [PubMed]
  40. Pu, Q.; Xi, Z.; Yin, S.; Zhao, Z.; Zhao, L. Advantages of transformer and its application for medical image segmentation: A survey. Biomed. Eng. Online 2024, 23, 14. [Google Scholar] [CrossRef]
  41. Khan, S.; Ali, H.; Shah, Z. Identifying the role of vision transformer for skin cancer—A scoping review. Front. Artif. Intell. 2023, 6, 1202990. [Google Scholar] [CrossRef]
  42. Arshed, M.A.; Mumtaz, S.; Ibrahim, M.; Ahmed, S.; Tahir, M.; Shafi, M. Multi-Class Skin Cancer Classification Using Vision Transformer Networks and Convolutional Neural Network-Based Pre-Trained Models. Information 2023, 14, 415. [Google Scholar] [CrossRef]
  43. Khan, S.; Khan, A. SkinViT: A transformer based method for Melanoma and Nonmelanoma classification. PLoS ONE 2023, 18, e0295151. [Google Scholar] [CrossRef]
  44. Codella, N.C.F.; Gutman, D.; Celebi, M.E.; Helba, B.; Marchetti, M.A.; Dusza, S.W.; Kalloo, A.; Liopyris, K.; Mishra, N.; Kittler, H.; et al. Skin lesion analysis toward melanoma detection: A challenge at the 2017 International symposium on biomedical imaging (ISBI), hosted by the international skin imaging collaboration (ISIC). In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; pp. 168–172. [Google Scholar] [CrossRef]
  45. Jeong, J.J.; Tariq, A.; Adejumo, T.; Trivedi, H.; Gichoya, J.W.; Banerjee, I. Systematic Review of Generative Adversarial Networks (GANs) for Medical Image Classification and Segmentation. J. Digit. Imaging 2022, 35, 137–152. [Google Scholar] [CrossRef]
  46. Goceri, E. GAN based augmentation using a hybrid loss function for dermoscopy images. Artif. Intell. Rev. 2024, 57, 234. [Google Scholar] [CrossRef]
  47. Behara, K.; Bhero, E.; Agee, J.T. Skin Lesion Synthesis and Classification Using an Improved DCGAN Classifier. Diagnostics 2023, 13, 2635. [Google Scholar] [CrossRef] [PubMed]
  48. Yazdanparast, T.; Shamsipour, M.; Ayatollahi, A.; Delavar, S.; Ahmadi, M.; Samadi, A.; Firooz, A. Comparison of the Diagnostic Accuracy of Teledermoscopy, Face-to-Face Examinations and Artificial Intelligence in the Diagnosis of Melanoma. Indian J. Dermatol. 2024, 69, 296–300. [Google Scholar] [CrossRef] [PubMed]
  49. Cerminara, S.E.; Cheng, P.; Kostner, L.; Huber, S.; Kunz, M.; Maul, J.T.; Böhm, J.S.; Dettwiler, C.F.; Geser, A.; Jakopović, C.; et al. Diagnostic performance of augmented intelligence with 2D and 3D total body photography and convolutional neural networks in a high-risk population for melanoma under real-world conditions: A new era of skin cancer screening? Eur. J. Cancer 2023, 190, 112954. [Google Scholar] [CrossRef] [PubMed]
  50. Crawford, M.E.; Kamali, K.; Dorey, R.A.; MacIntyre, O.C.; Cleminson, K.; MacGillivary, M.L.; Green, P.J.; Langley, R.G.; Purdy, K.S.; DeCoste, R.C.; et al. Using Artificial Intelligence as a Melanoma Screening Tool in Self-Referred Patients. J. Cutan. Med. Surg. 2024, 28, 37–43. [Google Scholar] [CrossRef]
  51. Hartman, R.I.; Trepanowski, N.; Chang, M.S.; Tepedino, K.; Gianacas, C.; McNiff, J.M.; Fung, M.; Braghiroli, N.F.; Grant-Kels, J.M. Multicenter prospective blinded melanoma detection study with a handheld elastic scattering spectroscopy device. JAAD Int. 2024, 15, 24–31. [Google Scholar] [CrossRef]
  52. MacLellan, A.N.; Price, E.L.; Publicover-Brouwer, P.; Matheson, K.; Ly, T.Y.; Pasternak, S.; Walsh, N.M.; Gallant, C.J.; Oakley, A.; Hull, P.R.; et al. The use of noninvasive imaging techniques in the diagnosis of melanoma: A prospective diagnostic accuracy study. J. Am. Acad. Dermatol. 2021, 85, 353–359. [Google Scholar] [CrossRef]
  53. Miller, I.; Rosic, N.; Stapelberg, M.; Hudson, J.; Coxon, P.; Furness, J.; Walsh, J.; Climstein, M. Performance of Commercial dermoscopic Systems That Incorporate Artificial Intelligence for the Identification of Melanoma in General Practice: A Systematic Review. Cancers 2024, 16, 1443. [Google Scholar] [CrossRef]
  54. Wei, M.L.; Tada, M.; So, A.; Torres, R. Artificial intelligence and skin cancer. Front. Med. 2024, 11, 1331895. [Google Scholar] [CrossRef]
  55. International Skin Imaging Collaboration (ISIC) Team. ISIC 2019 Challenge: Skin Lesion Classification. 2019. Available online: https://challenge.isic-archive.com/landing/2019/ (accessed on 15 October 2025).
  56. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is All you Need. In Advances in Neural Information Processing Systems; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
  57. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16 × 16 words: Transformers for image recognition at scale. In Proceedings of the International Conference on Learning Representations (ICLR), Online, 3–7 May 2021. [Google Scholar]
  58. Emara, T.; Afify, H.M.; Ismail, F.H.; Hassanien, A.E. A Modified Inception-v4 for Imbalanced Skin Cancer Classification Dataset. In Proceedings of the 2019 14th International Conference on Computer Engineering and Systems (ICCES), Cairo, Egypt, 17 December 2019; pp. 28–33. [Google Scholar] [CrossRef]
  59. Alzamel, M.; Iliopoulos, C.; Lim, Z. Deep learning approaches and data augmentation for melanoma detection. Neural Comput. Appl. 2024, 37, 10591–10604. [Google Scholar] [CrossRef]
  60. Safdar, K.; Akbar, S.; Shoukat, A. A Majority Voting based Ensemble Approach of Deep Learning Classifiers for Automated Melanoma Detection. In Proceedings of the 2021 International Conference on Innovative Computing (ICIC), Lahore, Pakistan, 9–10 November 2021; pp. 1–6. [Google Scholar] [CrossRef]
  61. Karras, T.; Aittala, M.; Hellsten, J.; Laine, S.; Lehtinen, J.; Aila, T. Training generative adversarial networks with limited data. In Proceedings of the Advances in Neural Information Processing Systems, Virtual, 6–12 December 2020; pp. 12104–12114. [Google Scholar]
  62. NVlabs. StyleGAN2-ADA PyTorch. 2020. Available online: https://github.com/NVlabs/stylegan2-ada-pytorch (accessed on 24 October 2025).
  63. Obuchowski, N.A.; Bullen, J.A. Receiver operating characteristic (ROC) curves: Review of methods with applications in diagnostic medicine. Phys. Med. Biol. 2018, 63, 07TR01. [Google Scholar] [CrossRef]
  64. Hajian-Tilaki, K. Receiver Operating Characteristic (ROC) Curve Analysis for Medical Diagnostic Test Evaluation. Casp. J. Intern. Med. 2013, 4, 627–635. [Google Scholar]
  65. Fawcett, T. An introduction to ROC analysis. Pattern Recognit. Lett. 2006, 27, 861–874. [Google Scholar] [CrossRef]
  66. He, Z.; Zhang, Q.; Song, M.; Tan, X.; Wang, W. Four overlooked errors in ROC analysis: How to prevent and avoid. BMJ Evid.-Based Med. 2025, 30, 208–211. [Google Scholar] [CrossRef] [PubMed]
  67. DeLong, E.R.; DeLong, D.M.; Clarke-Pearson, D.L. Comparing the Areas under Two or More Correlated Receiver Operating Characteristic Curves: A Nonparametric Approach. Biometrics 1988, 44, 837. [Google Scholar] [CrossRef]
  68. Chefer, H.; Gur, S.; Wolf, L. Transformer Interpretability Beyond Attention Visualization. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 782–791. [Google Scholar] [CrossRef]
  69. Wu, J.; Kang, W.; Tang, H.; Hong, Y.; Yan, Y. On the Faithfulness of Vision Transformer Explanations. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 17–21 June 2024; pp. 10936–10945. [Google Scholar] [CrossRef]
  70. Martín, J.; Bella-Navarro, R.; Jordá, E. Vascular Patterns in Dermoscopy. Actas-Dermo-Sifiliográficas (Engl. Ed.) 2012, 103, 357–375. [Google Scholar] [CrossRef]
  71. Longo, C.; Pampena, R.; Moscarella, E.; Chester, J.; Starace, M.; Cinotti, E.; Piraccini, B.M.; Argenziano, G.; Peris, K.; Pellacani, G. Dermoscopy of melanoma according to different body sites: Head and neck, trunk, limbs, nail, mucosal and acral. J. Eur. Acad. Dermatol. Venereol. 2023, 37, 1718–1730. [Google Scholar] [CrossRef]
  72. Hedegaard, L.; Alok, A.; Jose, J.; Iosifidis, A. Structured pruning adapters. Pattern Recognit. 2024, 156, 110724. [Google Scholar] [CrossRef]
  73. Na, G.S. Efficient learning rate adaptation based on hierarchical optimization approach. Neural Netw. 2022, 150, 326–335. [Google Scholar] [CrossRef]
  74. Razuvayevskaya, O.; Wu, B.; Leite, J.A.; Heppell, F.; Srba, I.; Scarton, C.; Bontcheva, K.; Song, X. Comparison between parameter-efficient techniques and full fine-tuning: A case study on multilingual news article classification. PLoS ONE 2024, 19, e0301738. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Overview of the Vision Transformer architecture. The input image is divided into patches, which are linearly embedded and passed through transformer layers. The output is a classification token used for final predictions. This diagram was adapted from the original Vision Transformer paper [57].
Figure 1. Overview of the Vision Transformer architecture. The input image is divided into patches, which are linearly embedded and passed through transformer layers. The output is a classification token used for final predictions. This diagram was adapted from the original Vision Transformer paper [57].
Cancers 17 03447 g001
Figure 2. ROC curves and AUC scores for the ViT-L/16 model on the validation (ISIC-19) and test (MN187) datasets.
Figure 2. ROC curves and AUC scores for the ViT-L/16 model on the validation (ISIC-19) and test (MN187) datasets.
Cancers 17 03447 g002
Figure 3. Distribution of predicted probabilities for the synthetic images generated by the GAN models. The green bars represent synthetic nevus images, while the red bars represent synthetic melanoma images. Gray bars represent uncertain classifications. The vertical dashed lines mark the thresholds used to select high-confidence synthetic images for training.
Figure 3. Distribution of predicted probabilities for the synthetic images generated by the GAN models. The green bars represent synthetic nevus images, while the red bars represent synthetic melanoma images. Gray bars represent uncertain classifications. The vertical dashed lines mark the thresholds used to select high-confidence synthetic images for training.
Cancers 17 03447 g003
Figure 4. Samples of synthetic images generated by the GAN models. The left column shows synthetic nevus images, while the right column shows synthetic melanoma images.
Figure 4. Samples of synthetic images generated by the GAN models. The left column shows synthetic nevus images, while the right column shows synthetic melanoma images.
Cancers 17 03447 g004
Figure 5. ROC curves and AUC scores for the multiple deep learning models trained with confidence-filtered GAN-augmented data on the validation (ISIC-2019) and external test (MN187) datasets.
Figure 5. ROC curves and AUC scores for the multiple deep learning models trained with confidence-filtered GAN-augmented data on the validation (ISIC-2019) and external test (MN187) datasets.
Cancers 17 03447 g005
Figure 6. Attention maps in each of the 24 layers of the Vision Transformer model. The sequence of maps shows a transition from broad and diffuse attention patterns in the early layers to more specific regions in the later stages of processing. In these visualizations, darker blue regions represent the highest attention scores in each layer, while colors shifting toward green and yellow indicate progressively lower attention scores. Areas without color correspond to regions where little or no attention is assigned.
Figure 6. Attention maps in each of the 24 layers of the Vision Transformer model. The sequence of maps shows a transition from broad and diffuse attention patterns in the early layers to more specific regions in the later stages of processing. In these visualizations, darker blue regions represent the highest attention scores in each layer, while colors shifting toward green and yellow indicate progressively lower attention scores. Areas without color correspond to regions where little or no attention is assigned.
Cancers 17 03447 g006
Figure 7. Examples of averaged attention maps obtained from layers 3–5 of the Vision Transformer model. Each pair of images shows the original dermoscopic input (left) and its corresponding attention-based visualization (right). Samples are organized by class to present the distribution of attention across different inputs. The color scale was adjusted according to the classification score, with images predicted as nevus displayed in blue surrounded by red, and those predicted as melanoma displayed in red surrounded by blue. All samples were selected from the MN187 dermoscopic dataset.
Figure 7. Examples of averaged attention maps obtained from layers 3–5 of the Vision Transformer model. Each pair of images shows the original dermoscopic input (left) and its corresponding attention-based visualization (right). Samples are organized by class to present the distribution of attention across different inputs. The color scale was adjusted according to the classification score, with images predicted as nevus displayed in blue surrounded by red, and those predicted as melanoma displayed in red surrounded by blue. All samples were selected from the MN187 dermoscopic dataset.
Cancers 17 03447 g007
Table 1. Summary of parameters and configurations used for training the GAN model (StyleGAN2-ADA). The –data file parameter shown in this table is dataset_nevus.zip, which contains the nevus samples used for training. For the melanoma model, a separate dataset_melanoma.zip file was used. Each GAN model required its corresponding training samples to be provided in a separate ZIP file, resulting in independently trained models that generated distinct samples.
Table 1. Summary of parameters and configurations used for training the GAN model (StyleGAN2-ADA). The –data file parameter shown in this table is dataset_nevus.zip, which contains the nevus samples used for training. For the melanoma model, a separate dataset_melanoma.zip file was used. Each GAN model required its corresponding training samples to be provided in a separate ZIP file, resulting in independently trained models that generated distinct samples.
CategoryParameterValue / SettingDescription / Notes
Hardware–gpus2Number of GPUs used (2× NVIDIA RTX A5000)
–workers3Number of DataLoader workers per process
Dataset–datadataset_nevus.zipPath to training dataset (datasets are provided to the model in zip compressed files. Both datasets, nevus and melanoma, are trained with the same parameters, only changing the dataset file)
–mirrorTrueEnables x-flip augmentation (data doubling)
–subsetNoneUses full dataset (no subset restriction)
Resolution256×256Input image resolution for generator and discriminator
Base Configuration–cfgautoAuto-selects model depth, learning rate, and γ
–kimg1000Total training duration (1000 k images = 1 M samples)
–batch16Global batch size
–gammaAuto (≈ 0.0002 × res 2 / batch )R1 regularization weight (≈ 13.1 at 256 × 256 )
–snap5Snapshot / FID computation interval (in ticks)
OptimizerAdamUsed for both Generator and Discriminator
Metrics[FID50k_full]Frechet Inception Distance computed every snapshot
Augmentation–augnoaugNo adaptive discriminator augmentation applied
–augpipeDefault (bgc)Pipeline defined but inactive due to noaug
–targetN/AADA target (0.6 default) unused in this configuration
OptimizationLearning rate0.002–0.0025Determined automatically by resolution and batch size
Betas(0, 0.99)Adam momentum parameters
ϵ 1 × 10−8Adam numerical stability constant
EMA ramp-upEnabled (10–20 kimg)Stabilizes generator weights early in training
Mixed precisionEnabled (FP16)Reduces GPU memory usage, improves throughput
TF32 / cuDNNEnabledAllows TensorFloat32 for convolutions; benchmarking on
Model ArchitectureLatent dimensions512StyleGAN2 latent and mapping vector size (z, w)
Mapping layers8Depth of style mapping network (MLP)
Channel base / max32,768/512Controls feature-map width scaling
–freezedNoneNo frozen discriminator layers
EMA decayAuto (0.999–0.9999)Exponential moving average rate for generator weights
EvaluationMetricFID50k_fullComputed on 50k generated images using Inception-V3
Stopping criterionFID < 50Early stopping once quality threshold achieved
Table 2. This table presents the results of running six different deep learning classification models, each using distinct backbones for feature extraction but the same classification head, as illustrated in Figure 1. The validation results are based on the ISIC-19 dataset, while the testing was conducted using the MN187 dataset.
Table 2. This table presents the results of running six different deep learning classification models, each using distinct backbones for feature extraction but the same classification head, as illustrated in Figure 1. The validation results are based on the ISIC-19 dataset, while the testing was conducted using the MN187 dataset.
ModelValidation (ISIC-2019)Testing (MN187)
Run 1Run 2Run 3Run 4Run 5Run 1Run 2Run 3Run 4Run 5
ViT-L/160.9210.9190.9190.9200.9190.9020.8960.8960.9020.902
ViT-B/160.9190.9160.9150.9140.9160.8870.8890.8770.8860.882
ResNet-1520.9090.9080.9070.9080.9080.8570.8610.8690.8600.861
EfficientNet-B70.8890.8870.8870.8890.8890.8470.8350.8380.8340.839
DenseNet-2010.9150.9160.9140.9140.9140.8420.8580.8550.8610.845
ConvNeXt-XL0.9260.9270.9290.9260.9280.8810.8770.8810.8840.885
Table 3. Summary of DeLong test results comparing model AUCs to the MoleAnalyzer Pro AUC. The p-values address whether the differences in AUCs are statistically significant.
Table 3. Summary of DeLong test results comparing model AUCs to the MoleAnalyzer Pro AUC. The p-values address whether the differences in AUCs are statistically significant.
ModelDeep Learning Model AUCMoleAnalyzer Pro AUCDeLong Test p-Value
ViT-L/160.9010.8560.070
ConvNeXt-XL0.8840.8560.226
ViT-B/160.8880.8560.240
ResNet-1520.8680.8560.654
EfficientNet-B70.8460.8560.709
DenseNet-2010.8590.8560.910
Table 4. ROC-AUC results (five runs each) for six models trained with the augmented dataset (real ISIC-2019 + confidence-filtered GAN synthetic images). Left block: validation (ISIC-2019). Right block: external test (MN187). Bold values indicate the best test run per model.
Table 4. ROC-AUC results (five runs each) for six models trained with the augmented dataset (real ISIC-2019 + confidence-filtered GAN synthetic images). Left block: validation (ISIC-2019). Right block: external test (MN187). Bold values indicate the best test run per model.
ModelValidation (ISIC-2019)Testing (MN187)
Run 1Run 2Run 3Run 4Run 5Run 1Run 2Run 3Run 4Run 5
ViT-L/160.9380.9360.9360.9370.9350.9140.9090.9080.9150.908
ViT-B/160.9240.9270.9270.9280.9270.8960.8920.8910.9010.899
ResNet-1520.9110.9090.9120.9120.9110.8750.8700.8730.8720.872
EfficientNet-B70.8950.8940.8930.8940.8930.8600.8660.8630.8660.869
DenseNet-2010.9210.9210.9200.9200.9210.8700.8680.8720.8750.861
ConvNeXt-XL0.9340.9360.9330.9330.9340.8910.8940.8850.8890.894
Table 5. Summary of DeLong test results comparing model AUCs to the MoleAnalyzer Pro AUC. The p-values address whether the differences in AUCs are statistically significant.
Table 5. Summary of DeLong test results comparing model AUCs to the MoleAnalyzer Pro AUC. The p-values address whether the differences in AUCs are statistically significant.
ModelDeep Learning Model AUCMoleAnalyzer Pro AUCDeLong Test p-Value
ViT-L/160.9200.8560.032
Fine tuned ViT-L/160.9260.8560.006
EfficientNet-B70.8700.8560.645
DenseNet-2010.8800.8560.466
ConvNeXt-XL0.8970.8560.154
ResNet-1520.8770.8560.498
ViT-B/160.9050.8560.098
Table 6. Average entropy ranking of attention layers in the Vision Transformer model. Higher entropy values represent greater diversity in spatial representations.
Table 6. Average entropy ranking of attention layers in the Vision Transformer model. Higher entropy values represent greater diversity in spatial representations.
RankLayerAverage Entropy
1Layer 30.8615
2Layer 40.8463
3Layer 50.8393
4Layer 10.8166
5Layer 60.8065
6Layer 20.7685
7Layer 70.7272
8Layer 240.7241
9Layer 230.7036
10Layer 80.6853
11Layer 90.6004
12Layer 210.5624
13Layer 180.5598
14Layer 170.5418
15Layer 110.5405
16Layer 160.5379
17Layer 130.5339
18Layer 120.5331
19Layer 150.5322
20Layer 100.5187
21Layer 140.5141
22Layer 220.4742
23Layer 200.4740
24Layer 190.4354
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Garcia, A.; Zhou, J.; Pinero-Crespo, G.; Beachkofsky, T.; Huang, X. Clinical Application of Vision Transformers for Melanoma Classification: A Multi-Dataset Evaluation Study. Cancers 2025, 17, 3447. https://doi.org/10.3390/cancers17213447

AMA Style

Garcia A, Zhou J, Pinero-Crespo G, Beachkofsky T, Huang X. Clinical Application of Vision Transformers for Melanoma Classification: A Multi-Dataset Evaluation Study. Cancers. 2025; 17(21):3447. https://doi.org/10.3390/cancers17213447

Chicago/Turabian Style

Garcia, Antony, Jixing Zhou, Gabriela Pinero-Crespo, Thomas Beachkofsky, and Xinming Huang. 2025. "Clinical Application of Vision Transformers for Melanoma Classification: A Multi-Dataset Evaluation Study" Cancers 17, no. 21: 3447. https://doi.org/10.3390/cancers17213447

APA Style

Garcia, A., Zhou, J., Pinero-Crespo, G., Beachkofsky, T., & Huang, X. (2025). Clinical Application of Vision Transformers for Melanoma Classification: A Multi-Dataset Evaluation Study. Cancers, 17(21), 3447. https://doi.org/10.3390/cancers17213447

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop