Next Article in Journal
Biomechanics and Motion Analysis: From Human Performance to Clinical Practice
Previous Article in Journal
Spatiotemporal Feature Learning for Daily-Life Cough Detection Using FMCW Radar
Previous Article in Special Issue
Automatic 3D Tracking of Liver Metastases: Follow-Up Assessment of Cancer Patients in Contrast-Enhanced MRI
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Generative Adversarial Networks in Dermatology: A Narrative Review of Current Applications, Challenges, and Future Perspectives

by
Rosa Maria Izu-Belloso
1,*,
Rafael Ibarrola-Altuna
2 and
Alex Rodriguez-Alonso
3
1
Medicine Faculty, Hospital Universitario Basurto, 48013 Bilbao, Spain
2
Hospital Universitario Galdakao, 48960 Bizkaia, Spain
3
Facultad de Medicina, Universidad del Pais Vasco/EHU, 48940 Leioa, Spain
*
Author to whom correspondence should be addressed.
Bioengineering 2025, 12(10), 1113; https://doi.org/10.3390/bioengineering12101113
Submission received: 21 August 2025 / Revised: 21 September 2025 / Accepted: 22 September 2025 / Published: 16 October 2025
(This article belongs to the Special Issue AI-Driven Imaging and Analysis for Biomedical Applications)

Abstract

Generative Adversarial Networks (GANs) have emerged as powerful tools in artificial intelligence (AI) with growing relevance in medical imaging. In dermatology, GANs are revolutionizing image analysis, enabling synthetic image generation, data augmentation, color standardization, and improved diagnostic model training. This narrative review explores the landscape of GAN applications in dermatology, systematically analyzing 27 key studies and identifying 11 main clinical use cases. These range from the synthesis of under-represented skin phenotypes to segmentation, denoising, and super-resolution imaging. The review also examines the commercial implementations of GAN-based solutions relevant to practicing dermatologists. We present a comparative summary of GAN architectures, including DCGAN, cGAN, StyleGAN, CycleGAN, and advanced hybrids. We analyze technical metrics used to evaluate performance—such as Fréchet Inception Distance (FID), SSIM, Inception Score, and Dice Coefficient—and discuss challenges like data imbalance, overfitting, and the lack of clinical validation. Additionally, we review ethical concerns and regulatory limitations. Our findings highlight the transformative potential of GANs in dermatology while emphasizing the need for standardized protocols and rigorous validation. While early results are promising, few models have yet reached real-world clinical integration. The democratization of AI tools and open-access datasets are pivotal to ensure equitable dermatologic care across diverse populations. This review serves as a comprehensive resource for dermatologists, researchers, and developers interested in applying GANs in dermatological practice and research. Future directions include multimodal integration, clinical trials, and explainable GANs to facilitate adoption in daily clinical workflows.

1. Introduction

Artificial Intelligence (AI) is revolutionizing modern medicine by offering advanced tools for the detection, diagnosis, and management of numerous diseases [1,2,3]. Dermatology, in particular, has greatly benefited from the Deep Learning (DL) techniques applied to skin image analysis, enabling automated classification, lesion segmentation, and the development of predictive models for a wide range of dermatoses [4,5].
Among DL methods, Generative Adversarial Networks (GANs) have shown remarkable potential in dermatology [6]. These networks are capable of generating high-fidelity synthetic skin images, addressing critical limitations in current datasets. For example, only 12% of existing dermatological image repositories include phototypes IV–VI, contributing to diagnostic biases in non-Caucasian populations [7]. GANs can augment these under-represented categories, improve image quality, simulate rare or atypical lesions, and support data anonymization—key aspects for both clinical research and AI development [8].
In addition to enhancing training data, GANs also have promising applications in medical education. A recent study showed that StyleGAN3 can generate realistic amelanotic melanoma images with 92% perceptual realism, offering a valuable tool for teaching complex diagnostic scenarios to dermatology trainees [5].
This review aims to provide a comprehensive overview of the current and emerging uses of GANs in dermatology, discussing their benefits, limitations, and future perspectives. Despite the existence of multiple reviews on artificial intelligence in dermatology, few have specifically focused on GANs and their unique clinical implications. Our contribution lies in offering a dermatologist’s perspective that bridges technical metrics with potential clinical applications, which we consider highly relevant given the current absence of clinical integration.

2. Materials and Methods

2.1. Objectives of the Work

The main objective of this work is to analyze the current state of the art regarding the use of GANs in dermatological image processing and to evaluate their impact on the diagnosis and management of skin diseases. To achieve this goal, the following specific objectives were defined:
  • To review fundamental concepts of AI in dermatology, with a focus on the role of GANs in this field.
  • To identify current applications of GANs for the generation, augmentation, and enhancement of dermatological images.
  • To analyze the technical, ethical, and regulatory challenges associated with the implementation of GANs in dermatological practice and research.
  • To explore the future perspectives of GANs and their potential for clinical integration.

2.2. Bibliographic Search Methodology

For this literature review, we consulted several scientific databases, including PubMed (National Library of Medicine, Bethesda, MD, USA), Web of Science (Clarivate Analytics, Philadelphia, PA, USA), Medline (U.S. National Library of Medicine, Bethesda, MD, USA), and Google Scholar (Google LLC, Mountain View, CA, USA), using the TOPA portal (Osakidetza Thesaurus of Words and Articles) of the Basque Health Service (Osakidetza, Vitoria-Gasteiz, Spain). The search strategy combined keywords such as: “GANs in dermatology,” “deep learning in dermatology,” “synthetic images in dermatology,” “medical image augmentation,” and “adversarial networks skin diseases.”
Inclusion criteria:
  • Articles published in indexed journals or academic works (e.g., Final Degree Project (FDP) or Master’s Thesis (MT)) focused on AI and dermatology, prioritizing review articles.
  • Publications from the last five years (2020–2025) to ensure relevance and currency, although some foundational studies published earlier were also included.
  • Full-text availability.
  • Articles in English or Spanish.
The PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) methodology was followed, encompassing four phases:
  • Identification: A preliminary search was conducted in selected databases.
  • Screening: Duplicates were removed, and titles/abstracts were reviewed for relevance.
  • Eligibility: Full texts were evaluated according to predefined inclusion and exclusion criteria.
  • Inclusion: Only studies that met all criteria were included, ensuring methodological rigor and thematic relevance.
This PRISMA diagram (Figure 1) summarizes the study selection process for the literature review on the use of GANs in dermatology.
For better organization and analysis of the articles included in the review, a classification was made according to the type of study (Table 1). This categorization makes it possible to distinguish the purpose, methodological approach, and level of applicability of each work, facilitating comparison between studies and the identification of trends, gaps, and needs in research on the use of GANs in dermatology.
The classification used was inspired by standard methodological criteria used in systematic reviews and adapted to the context of artificial intelligence in health.

2.3. Methodological Quality Assessment of Included Articles

Given the central focus of this review on GAN applications in dermatology, we not only compiled relevant studies but also evaluated their methodological rigor and scientific contribution. This qualitative analysis aimed to identify research strengths and limitations, minimize bias, and support evidence-based conclusions.
Each selected study specifically involving GANs was evaluated based on the following:
  • Type of study (experimental, comparative, methodological, review, etc.).
  • Objective and methodology (clarity of aim, GAN architecture, dataset characteristics, and validation strategy).
  • Level of evidence, adapted from AI-in-health classification systems:
    Level I–II: Studies with clinical validation or comparative trials with standard reference.
    Level III: Methodological studies using retrospective data or public datasets.
    Level IV–V: Foundational theoretical studies or narrative reviews without experimental validation.
  • Strengths (e.g., innovation, external validation, reproducibility, applicability).
  • Limitations (e.g., small sample size, lack of clinical testing, limited generalizability).
This structured evaluation enabled a robust comparison between studies, highlighting trends, unmet needs, and directions for future research—particularly regarding data diversity, clinical validation, and model generalizability.
The results of this assessment are summarized in Table 2.
Although the main objective of this review is to analyze the use of GANs in dermatology, it was decided to include some studies that do not focus exclusively on this medical specialty. This decision was based on the fact that certain articles, although applied to other areas of medicine or even of a theoretical–methodological nature, provide key knowledge for the understanding, development, or evaluation of NGS models subsequently used in dermatology.
The inclusion of these papers broadens the perspective of the review and allows us to enrich the analysis of the applications and limitations of GANs in dermatology, contextualizing them within the broader framework of medical AI.

3. Results

3.1. AI in Dermatology

3.1.1. Basic Concepts of AI in Medical Imaging

AI applied to medical image analysis has undergone a remarkable development in recent years, especially through advanced deep learning techniques [12,15]. There are two main types of neural networks used in AI: traditional neural networks (ANNs) and convolutional neural networks (CNNs), which have fundamental differences in their architecture, operation, and applications [2,5].
ANNs are fully connected, meaning each neuron in one layer is linked to every neuron in the next, resulting in a large number of parameters. In contrast, CNNs use local connectivity, where neurons are connected only to specific regions of the input, reducing the number of parameters and enabling efficient feature detection.
When it comes to data processing, ANNs handle input as a flat vector, which may ignore the spatial structure of data such as images. On the other hand, CNNs process structured data in two or three dimensions, allowing them to recognize spatial patterns like shapes and textures. This makes CNNs particularly effective at automatically extracting hierarchical features through layers of filters and kernels.
CNNs are more efficient in terms of parameter usage, as they reuse weights across the input space. ANNs, due to their dense connectivity, tend to generate many more parameters, which can increase the risk of overfitting and require more training data. Consequently, ANNs are often used in tasks such as classification and prediction of tabular or sequential data, while CNNs excel in visual recognition tasks like image classification, object detection, and medical image analysis.
Another key difference lies in spatial invariance. While ANNs do not inherently account for spatial relationships, CNNs are partially invariant to the position of objects in the input due to operations like pooling. Lastly, CNNs typically require more computational resources than ANNs, especially during training, as they involve multiple operations and deeper architectures. They often rely on GPUs to be trained efficiently.
Within medical imaging, CNNs have become the most widely used and effective method for automatic image interpretation due to their ability to recognize complex visual patterns. CNNs are computational models inspired by the functioning of the human visual system [2,12,40]. A CNN is a type of neural network specifically designed to process data with a mesh structure, such as images, videos, or audio signals. These networks work by successive layers performing mathematical operations called convolutions, applying specific filters on each pixel of an image to extract distinctive patterns such as edges, shapes, and textures. They are essential in computer vision tasks such as image classification, detection, and segmentation.
CNNs are mainly composed [2,12,40] of the following key elements (Figure 2):
-
Convolutional layers: This is the core of the CNN and where the main operations are performed. Each layer applies filters or kernels that apply mathematical filters on the original image to generate activation maps that highlight specific features (such as edges, textures, anomalies, or simple shapes) relevant for classification. The output of this operation generates feature maps representing the regions activated by the applied filter.
-
Pooling layers: They reduce the spatial dimensions of the processed images, keeping only the most relevant information and reducing the computational complexity. The most common techniques are max pooling (selection of the maximum value within a region) and average pooling (averaging of values).
-
Fully connected layers: After multiple convolutional and clustering layers, these layers receive the previously generated activation maps and use them to make final predictions or classifications. These layers work by integrating all previously extracted features to generate a specific diagnosis or classification.
The main advantages [2] of using CNNs are as follows:
-
Automation in feature extraction: CNNs eliminate the need to manually design filters to identify patterns in images.
-
Spatial invariance: They are able to recognize objects regardless of their position or orientation within an image.
-
Hierarchy in learning: They learn representations from simple features (edges) to complex (whole objects).
CCNs have been applied to issues such as face classification and recognition, semantic detection and segmentation in medical images [38], autonomous driving through visual analysis, and real-time object recognition. However, their use presents several challenges such as the intensive use of computational resources (i.e., training CNNs requires specialized hardware, such as GPUs) and the need for large, labeled datasets (i.e., for effective training, large volumes of representative data are required).
In dermatology, these networks have been successfully employed to automatically detect skin diseases from images obtained by digital photography or dermoscopy [12]. An emblematic example is the model developed by Google Health [20] that used CNNs trained with an extensive set of over 130,000 dermatological images to classify with high accuracy 26 different skin pathologies. This model achieved a diagnostic accuracy comparable to that of the obtained by certified dermatologists, thus demonstrating the real potential of these technological tools to support and complement clinical decision-making [12,40]. In the field of skin pathology, these models can identify pigmented and non-pigmented lesions with similar or even greater accuracy than that achieved by experienced dermatopathologists in certain specific tasks, such as differentiating squamous cell carcinomas from benign seborrheic keratoses or malignant melanomas from benign nevi [40]. In addition, due to their speed and low operating cost, CNNs facilitate rapid and reliable diagnoses, which is especially valuable in clinical settings with limited resources or high demand for care [2]. In short, the successful application of CNNs in dermatology clearly demonstrates their potential as complementary tools capable of optimizing clinical diagnosis through fast, accurate and accessible automated analyses. However, it is also essential to consider the challenges associated with their practical implementation, such as unrealistic expectations on the part of healthcare personnel or patients, ethical issues related to privacy and security of clinical data, and medico-legal aspects that must be carefully addressed before their widespread adoption [9].

3.1.2. AI Applications in Dermatology

International Skin Imaging Collaboration (ISIC—New York, NY, USA)
ISIC (https://www.isic-archive.com/ accessed on 12 May 2025) is a global initiative that brings together academic institutions and industry to improve skin cancer diagnosis through the use of digital skin imaging technologies. Its purpose is to reduce skin cancer mortality by promoting standards in dermatologic imaging and fostering collaboration between the dermatology and computer vision communities.
Its main activities and resources are as follows:
-
Public Image Archive: ISIC maintains an open-access public archive containing tens of thousands of dermatologic images. This resource is invaluable for teaching, research, and the development of artificial intelligence algorithms for skin disease diagnosis and is used in most AI projects in dermatological imaging.
-
Standards Development: The collaboration works on the creation and promotion of standards for imaging in dermatology, addressing aspects such as technology, technique, terminology, privacy, and the interoperability of dermatological imaging.
-
Machine Learning Challenges (ISIC Challenges): ISIC organizes annual competitions that invite artificial intelligence researchers to develop and evaluate algorithms for the analysis of skin lesion images. These challenges have contributed significantly to the advancement of computer-aided diagnostic techniques in dermatology, and some of them have included the use of GANs.
In addition, ISIC collaborates with platforms such as dermoscopedia.org, integrating images and educational resources to facilitate learning and dissemination of knowledge in dermoscopy (Figure 3).
AI in Dermatology
AI has been implemented in dermatology for a variety of functions, including the following:
-
AI-assisted diagnosis: Models such as the one developed by Han et al. [44] have demonstrated greater than 90% accuracy in detecting melanomas from dermoscopy images.
-
Detection of basal cell and squamous cell carcinomas: Here, again, the algorithm of Han et al. [44] showed an accuracy of 94.8% in the identification of malignant skin tumors.
-
Differential diagnosis in inflammatory diseases: Several models [3,5,45] allow for distinguishing between psoriasis, dermatitis, and rosacea from clinical photographs.
-
Segmentation of dermatological images: Some of the “ISIC Challenge” have driven accurate segmentation algorithms to delineate pigmented lesions and tumor margins in dermoscopic images and improve the quality of dermatological images [5].
-
Disease severity assessment: Applications such as SkinDeep-PASI use neural networks to automatically calculate the PASI index in patients with psoriasis [5]. In hidradenitis suppurativa, image-based AI has been explored to assess the Hurley stage and extent of involvement [3].
-
Follow-up of disease progression: Longitudinal follow-up tools, such as DermTrainer® (Silverchair Information Systems, Charlottesville, VA, USA), allow us to assess the evolution of skin lesions by comparing serial images, which is very useful, e.g., to evaluate the efficacy of treatments (https://www.silverchair.com/news/silverchair-launches-dermtrainer/ accessed on 12 May 2025)
-
AI-assisted teledermatology: AI-enabled mobile apps allow patients to obtain a pre-assessment of skin lesions without the need for an in-person consultation. The SkinVision® app (SkinVision BV, Amsterdam, The Netherlands) analyzes images of moles and suspicious lesions using AI and provides a risk recommendation to the patient. The First Derm® platform (iDoc24 AB, Gothenburg, Sweden) uses AI algorithms to assist in the triage of patient-referred dermatology cases (https://www.firstderm.com/ai-dermatology/ accessed on 12 May 2025)
-
Medical education and training: Simulators such as VisualDx’s Virtual Patient® (VisualDx, Rochester, NY, USA) use AI to generate clinical cases and evaluate students’ diagnostic reasoning.
-
Dermatological research: Analysis of large volumes of images from the HAM10000 project (Medical University of Vienna, Vienna, Austria) has advanced the study of patterns in common pigmented lesions [46].
-
Cosmetic applications and esthetic dermatology: Tools such as YouCam@ (Perfect Corp., Taipei, Taiwan) or Modiface@ (Modiface Inc., Toronto, ON, Canada) use AI to simulate cosmetic treatments, analyze skin aging, and predict esthetic results.

3.1.3. Current Challenges and Limitations

-
Biases in AI Models
One of the main problems in the implementation of AI in dermatology is bias in the quality and representativeness of training data: e.g., racial or phototype bias due to lack of diversity in databases or imbalances in the representation of rare diseases. Most datasets are predominantly composed of images of light-skinned patients, which reduces the effectiveness of the model in patients with darker skin [47]. This can lead to diagnostic errors and reduced equity of care.
-
Model Interpretability
Deep neural networks, including CNNs used in dermatology, are often considered “black boxes,” making it difficult to explain their decisions. Recent research has explored the use of heat maps and interpretability algorithms such as Grad-CAM to improve model transparency [48].
-
Regulatory and Legal Considerations
The use of AI in dermatology poses regulatory challenges, especially with regard to the approval of these systems by agencies such as the FDA and EMA. It is critical to ensure that models are rigorously evaluated clinically before implementation [49]. Another added problem is the slow approval by regulatory agencies.
-
Integration with Clinical Practice
In this respect, two factors are essential:
  • There is a need for interoperability with electronic medical record systems.
  • There is a reluctance on the part of professionals to rely on automated models. In addition, clinical validation is needed, which is difficult to achieve due to the lack of multicenter studies with prospective evaluation (i.e., it would be essential to have clinical trials that demonstrate real benefit in daily practice).
In conclusion, AI has proven to be a promising tool in dermatology, with applications ranging from automated diagnosis to clinical research. However, its widespread adoption requires overcoming significant technical, ethical, and regulatory challenges. The development of robust, explainable, clinically validated and data-driven models will be key to ensuring responsible and effective integration into the dermatologic care of the future. And this is where GANs can play a very important role.

3.2. Generative Adversarial Networks (GANs)

3.2.1. Principles and Functions of GANs

GANs were introduced by Goodfellow et al. [6] and have proven to be one of the most innovative techniques in the field of DL. These networks are based on the competition between two neural models: a generator and a discriminator.
-
Generator: transforms random noise vectors (z) into synthetic images. In dermatology, it is optimized to replicate skin textures (e.g., pores or vascularization) using deep convolutional layers [50].
-
Discriminator: evaluates whether an image is real or artificially generated, using adversarial loss functions.
During training, the generator improves its performance to fool the discriminator, while the discriminator is refined to detect false images. This process of adversarial feedback means that over time, the synthetic images become virtually indistinguishable from the real ones (Figure 4).

3.2.2. Relevant GAN Architectures in Medical Imaging

There are several GANs architectures used in medical imaging, including the following:
-
DCGAN (Deep Convolutional GANs): Introduces deep convolutions to improve the quality of the generated images [51]. For example, it standardizes stable training with batch normalization, generating 1024 × 1024-pixel images useful for analyzing melanoma patterns [8,39]. It is often used to synthesize dermoscopic patches that mimic lesion texture, helping to balance datasets and improve classifier training.
-
Pix2Pix: Used for image translation tasks, such as the conversion of grayscale images to color in dermatology [52]. Useful for paired image-to-image translation (e.g., clinical versus dermoscopic views when paired data exist), supporting reconstruction and education.
-
CycleGAN: Allows for the conversion of images between domains without the need for image pairs, useful in the synthesis of dermatoscopic dermatological images from clinical images [53]. Enables unpaired domain transfer (e.g., standardizing images from different cameras/centers), which can reduce device-related variability before analysis. The study by Han et al. [44] achieved an SSIM (Structural Similarity Index) of 0.89 in this task.
-
StyleGAN: Introduces fine control over the style and visual characteristics of the generated images [28,41]. Generates high-resolution, controllable images to synthesize rare phenotypes and darker phototypes for education and data balancing.

3.2.3. Evolution and Improvements in Models GANs

Recent advances in GANs have revolutionized their application in dermatology and medical diagnostics, with key technical improvements in five main areas:
  • Training stability
The introduction of Wasserstein GAN (WGAN) [26] with a gradient penalty solved problems of modal collapse (the model it stuck generating only one type of image although it should make many different ones) and instability (training a GAN is like a tug-of-war between the generator and the discriminator; if one becomes too strong, the other remains unlearned) in conventional GANs. This approach allows for the training of complex architectures while maintaining continuity in the loss surface—critical for medical applications where consistency is vital [54].
2.
Image quality and resolution
Models such as Progressive Growing GAN (PGGAN) generate high-resolution images (512 × 512 pixels), which have been used to preserve critical anatomical details in medical imaging [41]. StyleGAN2, on the other hand, has demonstrated a superior ability to produce very-high-resolution (1024 × 1024 pixels) dermatoscopic images, capturing subtle patterns of melanomas with clinical fidelity [55].
3.
Semantic control
Semantic control, facilitated by techniques such as attention maps and latent encoders, allows for the relevant features of medical images to be selectively modified. For example, these techniques can be used to adjust morphological parameters in dermatological lesions or to preserve critical anatomical structures during translation between modalities [56].
4.
Conditionality
cGANs have demonstrated efficacy in tasks such as multimodality translation (e.g., conversion of T1w to FLAIR images in MRI) and generation conditional on specific diagnoses. These capabilities enable the generation of synthetic images that simulate specific clinical conditions and facilitate the training of medical classification models [42,52].
5.
Multimodality
Hybrid systems [23] that combine different sources of information, such as clinical images and histopathological data, make it possible to generate composite or correlated representations. These techniques have been applied to combine multiple MRI sequences or to improve spatial resolution in angiographic images [57].
In dermatology, StyleGAN2 has enabled the generation of large synthetic datasets, such as more than 26,000 synthetic dermatoscopic images used to train AI classifiers. This has improved diagnostic sensitivity by 15% over conventional datasets and has facilitated the creation of balanced image banks that include rare pathological variants [46].

3.2.4. Limitations and Challenges in Implementing GANs

GANs have demonstrated enormous potential in the generation of synthetic medical images, but their implementation faces several technical, ethical, and operational challenges. The main limitations and challenges are detailed below:
-
Lack of explainability and interpretability: The “black box” of generative models makes their clinical validation difficult [3,9,46]. GANs present difficulties in explaining how they generate specific synthetic images. In the medical context, this lack of interpretability can be a major obstacle to their clinical adoption, as practitioners need to understand how the images are produced in order to rely on them as diagnostic tools.
-
Possibility of bias in the generated data: If the training dataset is limited or not representative, GANs can amplify existing biases [47]. Also in medical imaging, data can vary significantly depending on the quality of the equipment used, the experience of the technician generating the data, and the presence of artifacts. These inconsistencies can “foul” the data and compromise the accuracy of trained models [18,30].
-
Regulation and clinical approval: The lack of specific regulatory standards for generative models poses obstacles to their clinical implementation. Moreover, although these images do not correspond to real patients, their misuse could raise privacy and security issues. In addition, there is a risk of GANs being used to create false or misleading medical content [35].
-
Costs and operational complexity: The creation and maintenance of standardized databases to train GANs is a resource-intensive process. This includes technical harmonization between different sources, accurate annotation, and compliance with legal and ethical regulations. These difficulties limit the availability and representativeness of the data needed to develop reliable tools [9,41].

3.3. GANs in Dermatology

GANs have been used in dermatology for various applications [8,43], such as synthetic imaging of skin lesions, improved image segmentation, and color standardization in dermatological images.
The main algorithms and GANs investigated in Dermatology are as follows:
  • DermoCC-GAN [29]:
Application: Standardization of color in dermatological images.
Characteristics: Combines a customized heuristic algorithm with an antagonistic generative network to correct for variability in image illumination conditions. This generative network improves color constancy in dermatological images to standardize illumination by aiding lesion classification and segmentation and outperforming traditional color normalization algorithms (Figure 5).
2.
Efficient-GAN (EGAN) [24]
Application: Skin lesion segmentation.
Characteristics: It uses an adversarial learning architecture with a generator that incorporates a composite scaling path with excitation attention and asymmetric lateral connections. It includes a morphology-based loss function to generate smoother lesion contours. It is notable for its accuracy and a lightweight version called Mobile-GAN, suitable for devices with low computational power (Figure 6).
3.
DermGAN [22]
Application: Generation of synthetic clinical images of skin with specific pathologies.
Characteristics: Based on the Pix2Pix architecture (an image translation technique designed to transform images from one domain to another by supervised learning), this application allows for variance in the skin color, size, and location of the condition in the synthetic images, while maintaining the fidelity of the pathology represented (Figure 7).
4.
Skin Lesion Style-Based GANs [17]
Application: Skin lesion image synthesis for improved classification.
Characteristics: Modify the style control structure and noise input in the generator to produce high quality images, improving the performance of skin lesion classification models. The generated images were visually realistic and useful for augmenting the training data. The classifier trained with real and synthetic images obtained higher accuracy than the one trained only with real images (Figure 8).
5.
GLSFA-GAN (Generative Adversarial Network with Global and Local Semantic Feature Awareness) [34].
Application: Skin lesion segmentation.
Characteristics: It incorporates modules to fuse local features from multiple scales and extract the global context, improving accuracy in the segmentation of lesion areas (Figure 9).
6.
GAN with Hybrid Loss Function [23].
Application: Data enhancement for dermoscopic image classification.
Characteristics: They introduce a hybrid loss function to improve the generation of synthetic images, evaluating its effectiveness in the classification of seven types of skin lesions. The authors propose an unsupervised data augmentation solution that uses GANs and associated techniques in their latent space to generate controlled semantic variations in dermatoscopic images that are used to augment the training dataset. Recently, Veeramani et al. [33] proposed something similar—a new GAN called MELIIGAN—to improve the quality of dermoscopic images and facilitate fine feature extraction. This approach uses stacked residual blocks to handle large scaling factors and reconstruct intricate details. In addition, it incorporates a hybrid function of loss with total variation regularization and the Charbonnier function, which is more robust than the mean square error (Figure 10).
7.
“Conditional GANs for Dermatological Imaging [19].
Application: Generation of images of skin lesions conditioned by segmentation masks.
Characteristics: A cGAN is the same as a GAN, but with an additional condition. That is, you tell it what you want it to generate. It incorporates segmentation masks in the generation process to improve the structure of the lesions in the synthetic images (Figure 11).
There are many studies [10,11,13,14,21,32] using these techniques that have focused on the diagnosis of skin cancer, especially in the detection of melanomas.

3.4. Practical Applications in Clinical Dermatology

Once the latest advances in AI and GANs in dermatology are known, as clinical dermatologists, we ask ourselves “how can all this knowledge help us in our actual clinical practice?” All of the following come to mind:
-
Clinical support in consultation or automated triage: We could augment datasets and generate realistic synthetic images of skin lesions (melanoma, psoriasis, acne, etc.) to improve the training of AI algorithms when there is little data [16]. This will improve the accuracy of diagnostic models.
-
Simulation of disease progression [19]: We could show how a lesion would evolve over time if left untreated (melanoma, ulcers, psoriasis, etc.). This would be an extraordinary support in clinical decision-making, in visual explanations for patients, and in the selection of more aggressive treatments if unfavorable evolution is predicted.
-
Reconstruction of poor-quality images: We could improve blurred, pixelated, or poorly illuminated photos (e.g., sent by patients). This would allow us to make better remote diagnoses (teledermatology), even with suboptimal images.
-
Standardization of color and image conditions [30]: We could automatically correct for differences in light and skin tone or contrasts between images taken under different conditions. This would facilitate comparisons between visits, evolutionary follow-up, and improve the accuracy of AI or our own clinical assessment.
-
Lesion segmentation: We could automatically separate the lesion from the background (healthy skin) in an image. This could lead to a more objective measurement of the affected area (ulcer, psoriasis, eczema…), follow-ups on the response to treatment, and more accurate clinical documentation.
-
AI-assisted classification [25,27]: We could generate images to train systems that then classify lesions (melanoma vs. nevus [25], types of acne, etc.). This could be a source of diagnostic support in consultation, which is especially useful in patients with multiple lesions, or in the triaging of suspicious cases.
-
Medical education and training: We could create synthetic didactic images of rare cases or cases at different stages. These could serve as educational tool for students, residents, or even other specialists (e.g., primary care physicians).
-
Improve balance in clinical datasets: We could generate images of under-represented populations (e.g., dark skin tones [37]). This could contribute to more inclusive AI systems and avoid bias in automated diagnostics.
-
Creation of interactive tools for patients: We could show the patient how their skin will look with certain treatments or if left untreated (e.g., progression of rosacea, acne, vitiligo). This could improve adherence to treatment programs and understanding of the disease, especially in digital natives or young patients.
-
Simulation of esthetic or surgical results: We could generate simulated images of how the skin will look after an intervention (laser, cryotherapy, surgical resection). This could lead to clearer communication with the patient, the management of expectations, and support for informed consent.
-
Development of AI clinical applications: We could use generated images to create or enhance automatic assessment apps (e.g., for acne patients who send weekly photos). This could assist in the remote monitoring of chronic patients and the optimization of office resources and lead to more accurate referrals.
In addition to academic research, several commercial initiatives are already incorporating GAN-based solutions into dermatological tools. These developments often align with the use cases described in previous sections—such as data augmentation, rare lesion synthesis, or image standardization—and are designed to support clinicians in diagnostic, educational, and operational workflows.
Table 3 summarizes selected commercial applications identified through open-source and market searches (Google search/official websites), grouped according to the eleven use cases discussed. This overview illustrates the real-world translation of GAN technology into products aimed at clinical dermatologists and highlights the increasing relevance of generative models in industry-led innovation.

3.5. Reliability Assessment and Ethical Challenges

3.5.1. Validation of Images Generated by GANs

Synthetic images generated by GANs must undergo rigorous validation processes [8] before being used in clinical settings. There are multiple strategies to assess their quality (Table 4).
Comparison with Real Images Using Metrics:
In this review, some metrics help us judge how “real” synthetic images look compared with real ones, others reflect structural fidelity in paired tasks like denoising/super-resolution or estimate realism/diversity. These metrics are proxies: they should be interpreted alongside task-specific, clinically meaningful outcomes. The following are commonly used:
-
Fréchet Inception Distance (FID) [23,58]: FID measures the degree of similarity between images generated by a GAN and real images by comparing not only pixels but also deep features extracted by a trained neural network (usually Inception v3). The real images and the images generated by the Inception network are passed; this network extracts “features” (internal representations) from each set of images and compares the statistical distributions of those features (mean and covariance) using a mathematical formula called Fréchet distance. The lower the FID, the better; it indicates that the generated images are more similar to the real ones. An FID of 0 would be ideal (i.e., the fake images are indistinguishable from the real ones).
-
Structural Similarity Index (SSIM) [19,33]: The SSIM compares two images and evaluates their similarity structurally. It is very useful, for example, to see if a generated image is visually similar to an original image (e.g., in super-resolution or denoising tasks). It analyzes three things between the two images: luminance, contrast, and structure. A value between 0 and 1 is calculated, where 1 means identical images, and values close to 0 are completely different. A value greater than 0.9 is usually already considered to represent a high structural similarity.
-
Quality and realism metrics:
  • Inception Score (IS): Uses a pre-trained Inception network to calculate the entropy of the predicted labels on the generated images, but its limitation is that it does not compare directly with real images [26,58,59].
  • Peak Signal-to-Noise Ratio (PSNR) [19,28,33]: Measures the quality of the reconstruction (how much the generated image resembles the original, pixel by pixel), but it does not capture perceptual differences well; it only measures absolute intensity differences.
  • Mean Squared Error (MSE): Measures the average squared difference per pixel between real and generated images, but like PSNR, it does not reflect perceptual quality [19,28,33]
-
Perceptual metrics:
  • Learned Perceptual Image Patch Similarity (LPIPS) [26,34,48]: Measures perceptual similarity between images by calculating the distance between activations of deep neural networks (such as AlexNet, VGG) between two images. It has the advantage of a better correlation with human perception than SSIM or PSNR.
  • Perceptual Path Length (PPL) [34,50,55]: Measures the smoothness of the latent space of the GAN (how the image changes when interpolating between two latent vectors) and is used to evaluate the quality of the latent space in GANs such as StyleGAN.
-
Diversity Metrics:
  • Precision and Recall for Distributions [26,43,50]: Measures how realistic the generated images are and the diversity of the generated sample with respect to the real ones. It helps to evaluate the balance between fidelity and diversity, something that FID does not distinguish well.
  • Mode Score [26]: An extension of the Inception Score that penalizes the lack of class coverage and is used when you want to evaluate class diversity (e.g., digits 0 to 9 in MNIST).
-
Specific metrics in medical imaging:
The Dice Coefficient and Jaccard Index [22,24,34] measure the overlap between images (useful in segmented images or in controlled generation) and are used in GANs to compare synthetic images with real anatomical structures.
Evaluations by Clinical Experts
The use of synthetic images generated by GANs in dermatology requires rigorous validation by clinical experts. In this context, dermatologists review the generated images to determine whether they are indistinguishable from the real ones, which is a critical step to ensure the quality and diagnostic utility of these images. Tschandl et al. [46] conducted a study using the HAM10000 dataset, which is composed of dermoscopic images of various skin lesions. The experts evaluated both real and synthetic images generated by Convolutional Neural Networks (CNNs) and GANs, demonstrating that well-trained synthetic images can be almost indistinguishable from real ones. However, this process is highly dependent on the clinician’s experience and the quality of the generative algorithm used.
Testing on Models from Classification
Another key approach to evaluating the utility of synthetic images is their integration into automated classification and diagnostic models. Bissoto et al. [59] analyzed how data generated by GANs can improve the performance of classification algorithms in tasks such as melanoma detection. In their study, models were trained with datasets augmented by synthetic images, which allowed the authors to increase the diversity of the training set and to mitigate problems such as inter-class imbalance.
The results showed that models trained with augmented data achieved significant improvements in metrics such as diagnostic accuracy and sensitivity. However, limitations were also identified: the generated images may introduce biases if they do not adequately represent rare or complex clinical cases. In addition, the importance of proper preprocessing and the careful selection of metrics to assess the real impact of synthetic data on model performance was highlighted.

3.5.2. Biases and Problems in Generative Models

Despite its advantages, the implementation of GANs in the medical domain presents significant challenges related to biases in the training data and problems inherent to generative models, such as the following:
-
Limitation in the representation of skin phototypes: One of the most prominent problems in the use of GANs for medical imaging is the lack of representativeness in the training data, especially in relation to skin phototypes. Most of the datasets used to train these models are biased towards lighter skin tones, which may reduce the effectiveness of the images generated for patients with darker skin. According to Adamson and Smith [47], this bias limits diagnostic and therapeutic capability in under-represented populations, perpetuating inequities in access to advanced artificial intelligence-based tools. This problem is especially critical in dermatology, where the visual characteristics of skin lesions can vary significantly depending on the patient’s skin tone. The Fitzpatrick scale [60], used to classify phototypes according to the amount of melanin and sensitivity to the sun, could serve as a benchmark for developing more diverse and representative datasets. However, its implementation requires a coordinated effort to collect clinical images that include all phototypes, from I (very light skin) to VI (very dark skin).
-
Risk of overfitting to specific patterns which may affect the variability of the generated images. Overfitting is another critical challenge in GAN training. This phenomenon occurs when the model learns specific patterns from the training dataset and loses the ability to generalize to new data. In the medical context, this can result in synthetic images that reflect only features present in the original data, reducing the variability needed to adequately represent different clinical cases [61].
For example, if a GAN is trained exclusively with images from a single type of medical equipment or specific population, the images generated may not be useful for other clinical contexts. To mitigate this problem, it is essential to implement techniques such as dropout or regularization during model training to improve the model’s generalizability [61]. Dropout addresses this problem by randomly “turning off” certain neurons during the training process, following a predefined probability.
-
Difficulties in model calibration: Although GANs are capable of generating highly realistic images, they may contain incorrect or irrelevant clinical information. Yi et al. [8] point out that GANs may generate artifacts or anatomical inconsistencies that do not correspond to real features observed in patients. This represents a significant risk for clinical applications, as synthetic images could induce diagnostic errors if used without proper validation.
In addition, accurate model calibration is essential to ensure that the generated images are useful from a clinical point of view. This involves carefully tuning the generator and discriminator parameters to avoid problems such as mode collapse, where the model produces a limited variety of samples. The integration of medical experts during the validation process can help to identify these inconsistencies before the images are used in clinical applications.
Nevertheless, it is important to emphasize that these algorithmic metrics, while useful for benchmarking, do not by themselves guarantee clinical utility. To date, clinical validation of GAN-based models in dermatology remains scarce, and few studies [22,46,59] have evaluated patient-centered or clinically meaningful outcomes. Future work should therefore prioritize prospective validation trials and outcome measures that directly reflect diagnostic accuracy, safety, and usefulness in real-world dermatological practice.

3.5.3. Ethics and Regulations in the Use of Synthetic Imaging

Ethical Considerations
The integration of GANs in dermatology raises several ethical concerns [35,36]. These include the risk of bias amplification due to the under-representation of certain skin types in training datasets, the potential misuse of GANs for deep fakes [31] or fraudulent image manipulation, and the lack of transparency inherent to black-box models. Moreover, synthetic datasets used for model training or validation may not always ensure proper informed consent or data provenance, challenging the ethical principles of autonomy and non-maleficence. The deployment of GANs in clinical contexts also requires critical evaluation of fairness and explainability to safeguard trust and accountability in AI-assisted dermatology. In addition to these general concerns, several frameworks provide more concrete orientation. In the European Union [62,63], the Artificial Intelligence Act (AI Act, 2024) explicitly classifies medical AI as “high risk” and highlights the need for human oversight, transparency, and accountability. In the United States [64], the FDA regulates AI under its Software as a Medical Device (SaMD) framework, which emphasizes good machine learning practices and continuous oversight. Recent case discussions in dermatology also underline the importance of addressing bias in under-represented skin types and ensuring proper informed consent when synthetic images are created or used for training.
Regulatory Landscape
European Union—AI Act and MDR
In the European Union, the proposed Artificial Intelligence Act (AI Act) represents a pioneering attempt to regulate AI systems based on their potential risk to fundamental rights and safety. According to the AI Act, systems used in healthcare, including those for diagnostic support or image generation, are considered high-risk and must comply with strict requirements for transparency, human oversight, robustness, and data governance [62]. In parallel, medical AI tools must also comply with the Medical Device Regulation (MDR 2017/745), which mandates performance validation, clinical evaluation, and post-market surveillance. For GANs, which often fall into gray areas of regulation due to their generative nature, their classification as Software as a Medical Device (SaMD) remains under discussion [63].
United States—FDA Digital Health Initiatives
In the U.S., the Food and Drug Administration (FDA) regulates AI-based medical tools under the Digital Health Software Precertification Program and the Software as a Medical Device (SaMD) framework [64]. While no GAN-based tools have yet received specific approval, the FDA has issued guidance on “Good Machine Learning Practices” (GMLP), emphasizing transparency, reproducibility, and continuous learning oversight [65]. The GANs used for medical image generation would likely be scrutinized for their intended use, i.e., for diagnostic aid or training, and evaluated based on clinical validation, traceability, and bias mitigation strategies.
Summary and Perspectives
Current regulatory efforts recognize the transformative potential of AI in dermatology but are still evolving in response to novel architectures like GANs. While the EU adopts a more prescriptive and precautionary approach through the AI Act and MDR, the U.S. emphasizes adaptive oversight via guidance and pre-certification programs. For dermatologists and developers, early engagement with regulators and adherence to ethical AI design principles will be critical for the responsible adoption of GANs in clinical workflows.
The following table (Table 5) provides a comparative overview of the current regulatory landscape in the European Union (EU) and the United States (U.S.), highlighting key legislation, classification approaches, AI-specific strategies, and validation requirements. This comparison is essential to understand both the opportunities and the limitations imposed by each system, and it underscores the importance of harmonizing global regulatory efforts for emerging AI technologies in dermatology.

3.6. Future Perspectives and Clinical Applications

3.6.1. Integration into Clinical Practice

The integration of GANs in clinical dermatology represents not only a technological advance but also an organizational and ethical challenge [35,36]. Effective implementation of these tools requires overcoming several barriers:
  • Clinical and regulatory validation: Before being incorporated into routine care, GANs-based applications must undergo rigorous clinical trials that evaluate their safety, diagnostic efficacy and added value over conventional methods. In addition, they must comply with regulatory standards, such as the European Union’s Medical Device Regulation (MDR) or FDA approval in the USA.
  • Acceptance by healthcare personnel: For these technologies to be adopted, it is essential that dermatologists understand how they work, their limitations and their clinical applicability. Training in artificial intelligence (AI) and interpretation of results is key to fostering effective human–machine collaboration [46].
  • Technological infrastructure and digitization: Implementation of GANs models requires a robust computing, storage, and networking infrastructure. The hospitals must have interoperable and secure systems that allow for the seamless integration of these tools [40].
  • Interoperability with electronic health record systems (EHRs): GAN-based models can be integrated with EHRs to facilitate automated documentation, improve continuity of care, and provide more accurate assisted diagnoses [5]. This could also streamline screening for suspicious lesions through automatically generated clinical alerts.
Potential Impact on Assisted Diagnosis by AI
The use of GANs in dermatology has great potential to improve diagnostic accuracy and optimize healthcare resources. Some notable applications include the following:
-
Generation of extended and balanced datasets [59]: GANs can synthesize high-quality dermatological images with a variety of pathologies, which allows for the balancing of datasets with a scarcity of examples (e.g., images of rare diseases), reducing bias in deep learning models.
-
Enhanced teledermatology: The combination of GANs with telemedicine platforms allows for the simulation of diverse clinical cases, facilitating the training of models operating in remote contexts. This is especially useful for regions with limited access to dermatologists [20].
-
Support in differential diagnosis: GANs can generate images that represent different stages of a disease or atypical variations, helping dermatologists to recognize complex patterns and improve their clinical judgment. This could help reduce diagnostic errors in difficult cases [32].

3.6.2. Future Lines of Research

As GANs evolve, new research opportunities are opening up in the dermatological field. Some future lines include the following:
-
Improving explainability: One of the main criticisms of AI systems is their opacity. Therefore, GAN models are being developed [48] that integrate visual interpretation mechanisms, such as attention maps or Grad-CAM techniques, which allow for the identification of which regions of the image have influenced the model’s decision.
-
Equity and inclusion in datasets: To prevent algorithms from perpetuating inequalities, synthetic generation strategies focused on under-represented populations, including different ethnic groups and skin phototypes, are being promoted. Diversity in training data is essential to ensure truly equitable AI [47,60].
-
Personalized medicine [2]: GANs could be used to predict treatment response in chronic inflammatory diseases such as psoriasis or severe acne by simulating clinical evolution with and without treatment. This would open the door to individualized therapeutic decision models.
-
Collaboration between dermatologists and engineers will be key to ensure clinically relevant applications.
Finally, beyond GANs, other imaging modalities, such as Hyperspectral Imaging (HSI), are emerging as promising complementary tools for dermatologic diagnostics. Although recent papers on HSI were published after the cut-off date of our search (April 2025) and were therefore not included in this systematic selection, we acknowledge their relevance. Beyond the applications of GANs alone, hyperspectral imaging (HSI) represents a promising complementary modality in dermatologic AI. HSI captures spectral signatures across dozens or even hundreds of wavelengths, enabling the detection of subtle biochemical and structural differences in skin lesions that are invisible to conventional RGB imaging. When combined with GAN-based models, these high-dimensional spectral datasets could be exploited to generate synthetic hyperspectral images, augment under-represented spectral patterns and improve the robustness of classifiers trained on multimodal inputs. Such a synergistic approach may enhance early cancer detection, standardize image variability across acquisition devices, and open the door to novel diagnostic workflows where GANs support both spectral reconstruction and realistic data augmentation. Integrating GANs with HSI therefore holds potential to advance precision dermatology by uniting morphological and biochemical insights within a single AI framework.

4. Conclusions

4.1. Summary of Findings

The present work has explored the use of GANs in dermatology, highlighting their impact on the generation of synthetic images, the improvement of AI-assisted diagnosis and the anonymization of clinical data. Multiple clinical applications have been identified, ranging from data augmentation for training AI models to prediction of skin disease progression.
Results obtained in several studies suggest that GANs can improve the diagnostic accuracy of deep learning models and optimize the quality of medical images used in dermatology. In addition, these networks present great potential in preserving patient privacy by creating synthetic images that retain essential clinical features without compromising sensitive data.

4.2. Limitations of the Study

Despite following a structured methodology based on the PRISMA guidelines, this review has some limitations inherent to the selected approach.
First, the time criterion established (2020–2025) allowed us to focus the search on current studies, thus ensuring the relevance of the findings. However, this restriction could have excluded key papers published in previous years, especially between 2017 and 2019, when fundamental advances in the use of GANs in dermatology occurred. Although some earlier articles were included because of their relevance, their incorporation was not performed systematically. Second, the results were limited to articles published in English and Spanish. This decision, although it responds to the linguistic competence of the reviewer, may have excluded valuable scientific literature in other languages, such as German, Chinese, or Japanese, from countries with outstanding production in the field of medical artificial intelligence.
Furthermore, the review was restricted to articles with full access to the text, which may have introduced an availability bias, especially in databases such as Google Scholar, where many references are not fully accessible. This criterion may have affected the completeness of the review by leaving out potentially relevant studies.
Another limitation lies in the preferential selection of review articles and academic papers (FDP, MT), which may have reduced the representation of some recent original studies not yet included in reviews, or in the process of publication.
On the other hand, although broad keywords were defined and MeSH terms such as “GANs in dermatology,” “deep learning in dermatology,” “synthetic images in dermatology,” “medical image augmentation,” and “adversarial networks skin diseases” were combined, terminological variability in this emerging field may have resulted in the exclusion of publications that use alternative nomenclature to describe related techniques.

4.3. Recommendations for Future Research

Based on the analysis performed in this review, the following recommendations are proposed to guide future lines of research in the use of GANs applied to dermatology:
-
Formation of multidisciplinary teams: It is essential to foster collaboration between dermatologists, computer engineers, biomedical engineers, statisticians, and artificial intelligence experts. The correct implementation of GANs models in clinical settings requires not only sound technical development but also a thorough understanding of the dermatologic diagnostic processes, clinical needs, and ethical issues associated with the use of synthetic medical images. The presence of multidisciplinary teams ensures a more robust experimental design, a validation more in line with clinical practice, and a proper interpretation of the results.
-
Increasing the diversity and representativeness of the data: Many current studies use public databases with an under-representation of high phototypes or rare pathologies. It is crucial to encourage the collection and sharing of diverse and well-annotated dermatologic data to improve the generalizability of models and avoid algorithmic biases.
-
Clinical validation and comparison with specialists: It is recommended that the generated models be evaluated not only by quantitative metrics (FID, IS, IoU…) but also through comparative studies with dermatologic clinical practice. This includes the participation of specialists to assess the diagnostic utility of the synthetic images or of the segmentation or classification models trained with them.
-
Exploration of new clinical applications: beyond diagnosis, GANs could have applications in medical education, image quality improvement, dataset generation for rare diseases, or pre-surgical simulations. Future studies should explore these possibilities.
-
Development of accessible and explainable tools: It is recommended to prioritize the development of models that are not only accurate but also interpretable and understandable for clinicians. The algorithmic transparency and the possibility of auditing the model’s decisions will facilitate its acceptance and use in practice.
-
Ethical, legal, and regulatory considerations: Any development should incorporate from its early stages an assessment of the ethical, legal, and privacy risks, especially in relation to the generation of synthetic images of patients. Future studies should align with emerging regulatory frameworks on AI in healthcare.
In a broader context, it is worth noting that generative models have been extensively reviewed in other areas of healthcare, such as during the COVID-19 pandemic [66]. By contrast, our review was intentionally designed to focus on dermatology, aiming to translate these highly technical advances into clinically meaningful insights. In doing so, we hope to raise awareness among dermatologists and stimulate greater engagement with a rapidly evolving field that holds considerable promise for future integration into practice.
In conclusion, GANs represent an innovative tool with the potential to transform the field of dermatology. However, their successful implementation will depend on overcoming the technical, ethical, and regulatory challenges identified in this study.

Author Contributions

Conceptualization, R.M.I.-B.; methodology, R.M.I.-B.; investigation, R.M.I.-B., R.I.-A. and A.R.-A.; resources, R.M.I.-B. and A.R.-A.; writing—original draft preparation, R.M.I.-B., R.I.-A. and A.R.-A.; writing—review and editing, R.M.I.-B., R.I.-A. and A.R.-A.; visualization, R.M.I.-B., R.I.-A. and A.R.-A.; supervision, R.M.I.-B.; project administration, R.M.I.-B. and R.I.-A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AbbreviationMeaning
AIArtificial Intelligence
ANNArtificial Neural Network
CDRHCenter for Devices and Radiological Health
cGANConditional Generative Adversarial Network
CNNConvolutional Neural Network
DCGANDeep Convolutional GAN
DLDeep Learning
EGANEfficient-GAN
EMAEuropean Medicines Agency
EUEuropean Union
FDAFood and Drug Administration
FDPFinal Degree Project
FIDFréchet Inception Distance
FLAIRFluid Attenuated Inversion Recovery
GANGenerative Adversarial Network
GDPRGeneral Data Protection Regulation
GLSFAGlobal and Local Semantic Feature Awareness
GMLPGood Machine Learning Practice
HIPAAHealth Insurance Portability and Accountability Act
IoUIntersection over Union
ISInception Score
LPIPSLearned Perceptual Image Patch Similarity
MDRMedical Device Regulation
MELIIGANMelanoma Lesion Image Improvement GAN
MSEMean Squared Error
MTMaster’s Thesis
PMAPremarket Approval
PPLPerceptual Path Length
PRISMAPreferred Reporting Items for Systematic Reviews and Meta-Analyses
PSNRPeak Signal-to-Noise Ratio
ROCReceiver Operating Characteristic
SaMDSoftware as a Medical Device
SSIMStructural Similarity Index
VAEVariational Autoencoder
VGGVisual Geometry Group
WGANWasserstein GAN

References

  1. Annals RANM. Artificial Intelligence, Medical Imaging, and Precision Medicine: Advances and Perspectives. 2024. Available online: https://analesranm.es/revista/2024/141_02/14102_rev02 (accessed on 15 August 2025).
  2. Kourounis, G.; Elmahmudi, A.A.; Thomson, B.; Hunter, J.; Ugail, H.; Wilson, C. Computer image analysis with artificial intelligence: A practical introduction for medical professionals. Postgrad. Med. J. 2023, 99, 1287–1294. [Google Scholar] [CrossRef]
  3. Omiye, J.A.; Gui, H.; Daneshjou, R.; Cai, Z.R.; Muralidharan, V. Principles, applications, and future of artificial intelligence in dermatology. Front. Med. 2023, 10, 1278232. [Google Scholar] [CrossRef] [PubMed]
  4. Wongvibulsin, S.; Yan, M.J.; Pahalyants, V.; Murphy, W.; Daneshjou, R.; Rotemberg, V. Current State of Dermatology Mobile Applications With Artificial Intelligence Features. JAMA Dermatol. 2024, 160, 646–650. [Google Scholar] [CrossRef] [PubMed]
  5. Li, Z.; Koban, K.C.; Schenck, T.L.; Giunta, R.E.; Li, Q.; Sun, Y. Artificial Intelligence in Dermatology Image Analysis: Current Developments and Future Trends. J. Clin. Med. 2022, 11, 6826. [Google Scholar] [CrossRef] [PubMed]
  6. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. Adv. Neural Inf. Process. Syst. 2014, 27, 2672–2680. [Google Scholar] [CrossRef]
  7. Daneshjou, R.; Vodrahalli, K.; Novoa, R.A.; Jenkins, M.; Liang, W.; Rotemberg, V.; Ko, J.; Swetter, S.M.; Bailey, E.E.; Gevaert, O.; et al. Disparities in dermatology AI performance on a diverse, curated clinical image set. Sci. Adv. 2022, 8, eabq6147. [Google Scholar] [CrossRef]
  8. Yi, X.; Walia, E.; Babyn, P. Generative adversarial network in medical imaging: A review. Med. Image Anal. 2019, 58, 101552. [Google Scholar] [CrossRef]
  9. Behara, K.; Bhero, E.; Agee, J.T. AI in dermatology: A comprehensive review into skin cancer detection. PeerJ Comput. Sci. 2024, 10, e2530. [Google Scholar] [CrossRef]
  10. Furriel, B.C.R.S.; Oliveira, B.D.; Prôa, R.; Paiva, J.Q.; Loureiro, R.M.; Calixto, W.P.; Reis, M.R.C.; Giavina-Bianchi, M. Artificial intelligence for skin cancer detection and classification for clinical environment: A systematic review. Front. Med. 2024, 10, 1305954. [Google Scholar] [CrossRef]
  11. Hermosilla, P.; Soto, R.; Vega, E.; Suazo, C.; Ponce, J. Skin cancer detection and classification using neural network algorithms: A systematic review. Diagnostics 2024, 14, 454. [Google Scholar] [CrossRef]
  12. Iglesias-Puzas, A.; Boixeda, P. Deep learning and mathematical models in dermatology. Actas Dermosifiliogr. 2020, 111, 192–195. [Google Scholar] [CrossRef]
  13. Liu, Y.; Li, C.; Li, F.; Lin, R.; Zhang, D.; Lian, Y. Advances in deep learning-facilitated early detection of melanoma. Brief. Funct. Genom. 2025, 24, elaf002. [Google Scholar] [CrossRef]
  14. Nazari, S.; Garcia, R. Automatic skin cancer detection using clinical images: A comprehensive review. Life 2023, 13, 2123. [Google Scholar] [CrossRef] [PubMed]
  15. Vardasca, R.; Reddy, G.; Sanches, J.; Moreira, D. Skin cancer image classification using AI: A systematic review. J. Imaging 2024, 10, 265. [Google Scholar] [CrossRef]
  16. Alshardan, A.; Alahmari, S.; Alghamdi, M.; AL Sadig, M.; Mohamed, A.; Mohammed, G.P. GAN-based synthetic medical image augmentation for class imbalanced dermoscopic image analysis. Fractals 2025, 33, 2540039. [Google Scholar] [CrossRef]
  17. Behara, K.; Bhero, E.; Agee, J.T. Skin Lesion Synthesis and Classification Using an Improved DCGAN Classifier. Diagnostics 2023, 13, 2635. [Google Scholar] [CrossRef] [PubMed]
  18. Carrasco Limeros, S.; Camacho Vallejo, T.; Escudero, V.; Pérez, L.; Pomares, H.; Rojas, I. GAN-based generative modelling for dermatological applications: Comparative study. arXiv 2022. [Google Scholar] [CrossRef]
  19. Ding, S.; Zheng, J.; Liu, Z.; Zheng, Y.; Chen, Y.; Xu, X.; Lu, J.; Xie, J. High-resolution dermoscopy image synthesis with conditional generative adversarial networks. Biomed. Signal Process. Control 2021, 64, 102224. [Google Scholar] [CrossRef]
  20. Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017, 542, 115–118. [Google Scholar] [CrossRef]
  21. Efimenko, M.; Ignatev, A.; Koshechkin, K. Review of medical image recognition technologies to detect melanomas using neural networks. BMC Bioinform. 2020, 21 (Suppl. 11), 270. [Google Scholar] [CrossRef]
  22. Ghorbani, A.; Natarajan, V.; Coz, D.; Liu, Y. DermGAN: Synthetic Generation of Clinical Skin Images with Pathology. Proc. Mach. Learn. Health NeurIPS Workshop 2020, 116, 155–170. Available online: https://proceedings.mlr.press/v116/ghorbani20a.html (accessed on 12 May 2025).
  23. Goceri, E. GAN-based augmentation using a hybrid loss function for dermoscopy images. Artif. Intell. Rev. 2024, 57, 234. [Google Scholar] [CrossRef]
  24. Innani, S.; Dutande, P.; Baid, U.; Pokuri, V.; Bakas, S.; Talbar, S.; Baheti, B.; Guntuku, S.C. Generative adversarial networks based skin lesion segmentation. Sci. Rep. 2023, 13, 13467. [Google Scholar] [CrossRef] [PubMed]
  25. Heenaye-Mamode Khan, M.; Gooda Sahib-Kaudeer, N.; Dayalen, M.; Mahomedaly, F.; Sinha, G.R.; Nagwanshi, K.K.; Taylor, A. Multi-class skin problem classification using deep generative adversarial network (DGAN). Comput. Intell. Neurosci. 2022, 2022, 9480139. [Google Scholar] [CrossRef]
  26. Perez, E.; Ventura, S. Progressive growing of GANs for improving data augmentation and skin cancer diagnosis. Artif. Intell. Med. 2023, 141, 102556. [Google Scholar] [CrossRef]
  27. Qin, Z.; Song, K.; Leng, J.; Cheng, X.; Zheng, J. A GAN-based image synthesis method for skin lesion classification. Comput. Methods Programs Biomed. 2020, 195, 105568. [Google Scholar] [CrossRef]
  28. Ren, Z.; Liang, J.; Wan, W.; Wang, D.; Wang, Y.; Zhong, Y. Controllable medical image generation via GAN. J. Percept. Imaging 2022, 5, 5021. [Google Scholar] [CrossRef]
  29. Salvi, M.; Molinari, F.; Havaei, M.; Depeursinge, A. DermoCC-GAN: A new approach for standardizing dermatological images using generative adversarial networks. Comput. Methods Programs Biomed. 2022, 225, 107040. [Google Scholar] [CrossRef]
  30. Salvi, M.; Bassoli, S.; Molinari, F.; Havaei, M.; Depeursinge, A. Generative models for color normalization in pathology and dermatology. Expert. Syst. Appl. 2024, 245, 123105. [Google Scholar] [CrossRef]
  31. Sharafudeen, M.; Andrew, J.; Chandra, S.S.V. Leveraging Vision Attention Transformers for Detection of Artificially Synthesized Dermoscopic Lesion Deepfakes Using Derm-CGAN. Diagnostics 2023, 13, 825. [Google Scholar] [CrossRef]
  32. Jütte, L.; González-Villà, S.; Quintana, J.; Steven, M.; Garcia, R.; Roth, B. Integrating generative AI with ABCDE rule analysis for enhanced skin cancer diagnosis, dermatologist training and patient education. Front. Med. 2024, 11, 1445318. [Google Scholar] [CrossRef] [PubMed]
  33. Veeramani, N.; Jayaraman, P. A promising AI based super resolution image reconstruction technique for early diagnosis of skin cancer. Sci. Rep. 2025, 15, 5084. [Google Scholar] [CrossRef] [PubMed]
  34. Zou, R.; Zhang, J.; Wu, Y. Skin Lesion Segmentation through Generative Adversarial Networks with Global and Local Semantic Feature Awareness. Electronics 2024, 13, 3853. [Google Scholar] [CrossRef]
  35. Albisua, J.; Pacheco, P. Límites éticos en el uso de la inteligencia artificial (IA) en medicina. Open Respir. Arch. 2025, 7, 100383. [Google Scholar] [CrossRef]
  36. Bleher, H.; Braun, M. Reflections on putting AI ethics into practice: How three AI ethics approaches conceptualize theory and practice. Sci. Eng. Ethics 2023, 29, 21. [Google Scholar] [CrossRef]
  37. Rezk, E.; Eltorki, M.; El-Dakhakhni, W. Leveraging Artificial Intelligence to Improve the Diversity of Dermatological Skin Color Pathology: Protocol for an Algorithm Development and Validation Study. JMIR Res. Protoc. 2022, 11, e34896. [Google Scholar] [CrossRef]
  38. Higuera González, I.D. Evaluación de la Precisión y Automatización de Técnicas de Machine Learning en la Predicción de Enfermedades Mediante Imágenes Médicas. Monografía de Grado, Universidad Nacional Abierta y a Distancia, Bogota, Colombia, 2025. [Google Scholar]
  39. Redondo Hernández, S. Deep Learning para la Detección de Patologías de Cáncer de piel y Generación de Imágenes de Tejidos Humanos. Tesis de Máster, Universitat Politècnica de Catalunya, Barcelona, Spain, Universitat de Barcelona, Barcelona, Spain, 2020. Available online: https://upcommons.upc.edu/entities/publication/58aa1e22-041a-4456-bee0-f8fb040fad62 (accessed on 15 August 2025).
  40. Puri, P.; Shunatona, S.; Rosenbach, M.; Swerlick, R.; Anolik, R.; Chen, S.T.; Ko, C.J. Deep learning for dermatologists: Part II. Current applications. J. Am. Acad. Dermatol. 2022, 87, 1352–1360. [Google Scholar] [CrossRef]
  41. Karras, T.; Aila, T.; Laine, S.; Lehtinen, J. Progressive Growing of GANs for Improved Quality, Stability, and Variation. ICLR. 2018. Available online: https://github.com/tkarras/progressive_growing_of_gans (accessed on 12 May 2025).
  42. Mirza, M.; Osindero, S. Conditional generative adversarial nets. arXiv 2014. [Google Scholar] [CrossRef]
  43. Paladugu, P.S.; Ong, J.; Nelson, N.; Kamran, S.A.; Waisberg, E.; Zaman, N.; Kumar, R.; Dias, R.D.; Lee, A.G.; Tavakkoli, A. Generative Adversarial Networks in Medicine: Important Considerations for this Emerging Innovation in Artificial Intelligence. Ann. Biomed. Eng. 2023, 51, 2130–2142. [Google Scholar] [CrossRef]
  44. Han, S.S.S.; Kim, M.S.; Lim, W.; Park, G.H.; Park, I.; Chang, S.E. Classification of the clinical images for benign and malignant cutaneous tumors using a deep learning algorithm. J. Investig. Dermatol. 2018, 138, 1529–1538. [Google Scholar] [CrossRef]
  45. AlSuwaidan, L. Deep learning based classification of dermatological disorders. Biomed. Eng. Comput. Biol. 2023, 14, 11795972221138470. [Google Scholar] [CrossRef]
  46. Tschandl, P.; Rosendahl, C.; Kittler, H. The HAM10000 dataset, a large collection of multi-source dermatoscopic images of common pigmented skin lesions. Sci. Data 2018, 5, 180161. [Google Scholar] [CrossRef] [PubMed]
  47. Adamson, A.S.; Smith, A. Machine learning and health care disparities in dermatology. JAMA Dermatol. 2018, 154, 1247–1248. [Google Scholar] [CrossRef] [PubMed]
  48. Selvaraju, R.R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  49. Sengupta, D. Artificial Intelligence in Diagnostic Dermatology: Challenges and the Way Forward. Indian Dermatol. Online J. 2023, 14, 782–787. [Google Scholar] [CrossRef] [PubMed]
  50. Karras, T.; Laine, S.; Aittala, M.; Hellsten, J.; Lehtinen, J.; Aila, T. Analyzing and improving the image quality of StyleGAN. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8110–8119. [Google Scholar]
  51. Radford, A.; Metz, L.; Chintala, S. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv 2016, arXiv:1511.06434. [Google Scholar] [CrossRef]
  52. Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE Confrence Computer Vision Pattern Recognit, Honolulu, HI, USA, 21–26 July 2017; pp. 1125–1134. [Google Scholar]
  53. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using CycleGAN. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
  54. Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein GAN. arXiv 2017, arXiv:1701.07875. [Google Scholar]
  55. Karras, T.; Laine, S.; Aila, T. A style-based generator architecture for generative adversarial networks (StyleGAN). IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 4217–4228. [Google Scholar] [CrossRef]
  56. Schlegl, T.; Seeböck, P.; Waldstein, S.M.; Schmidt-Erfurth, U.; Langs, G. Unsupervised anomaly detection with generative adversarial networks to guide marker discovery. Inf. Process. Med. Imaging 2017, 10265, 146–157. [Google Scholar]
  57. Chartsias, A.; Joyce, T.; Dharmakumar, R.; Tsaftaris, S.A. Multimodal MR synthesis via modality-invariant latent representation. Med. Image Anal. 2018, 48, 118–131. [Google Scholar] [CrossRef]
  58. Heusel, M.; Ramsauer, H.; Unterthiner, T.; Nessler, B.; Hochreiter, S. GANs trained by a two time-scale update rule converge to a local Nash equilibrium. Adv. Neural Inf. Process. Syst. 2017, 30, 6626–6637. [Google Scholar]
  59. Bissoto, A.; Valle, E.; Avila, S. GAN-Based Data Augmentation and Anonymization for Skin-Lesion Analysis: A Critical Review. arXiv 2021, arXiv:2104.10603. [Google Scholar]
  60. Kim, Y.H.; Kobic, A.; Vidal, N.Y. Distribution of race and Fitzpatrick skin types in datasets for dermatology deep learning: A systematic review. J. Am. Acad. Dermatol. 2022, 87, 460–463. [Google Scholar] [CrossRef]
  61. EKEN, E. Determining overfitting and underfitting in generative adversarial networks using Fréchet distance. Turk. J. Electr. Eng. Comput. Sci. 2021, 29, 1524–1538. [Google Scholar] [CrossRef]
  62. European Commission. Proposal for a Regulation Laying Down Harmonised Rules on Artificial Intelligence (Artificial Intelligence Act). 2021. Available online: https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=CELEX:52021PC0206 (accessed on 5 June 2025).
  63. European Parliament and Council. Regulation (EU) 2017/745 on Medical Devices. 2017. Available online: https://eur-lex.europa.eu/eli/reg/2017/745/oj/eng (accessed on 5 June 2025).
  64. US Food & Drug Administration. Artificial Intelligence and Machine Learning (AI/ML)-Enabled Medical Devices. Available online: https://www.fda.gov/medical-devices/software-medical-device-samd/artificial-intelligence-and-machine-learning-aiml-enabled-medical-devices (accessed on 5 June 2025).
  65. US Food & Drug Administration. Good Machine Learning Practice for Medical Device Development: Guiding Principles. 2021. Available online: https://www.fda.gov/medical-devices/software-medical-device-samd/good-machine-learning-practice-medical-device-development-guiding-principles (accessed on 5 June 2025).
  66. Fei, Z.; Ryeznik, Y.; Sverdlov, O.; Tan, C.W.; Wong, W.K. An overview of healthcare data analytics with applications to the COVID-19 pandemic. IEEE Trans. Big Data 2022, 8, 1463–1480. [Google Scholar] [CrossRef]
Figure 1. PRISMA diagram with bibliography selection process.
Figure 1. PRISMA diagram with bibliography selection process.
Bioengineering 12 01113 g001
Figure 2. Illustration of the operational mode of a CNN [12].
Figure 2. Illustration of the operational mode of a CNN [12].
Bioengineering 12 01113 g002
Figure 3. “Screenshot” of the ISIC website.
Figure 3. “Screenshot” of the ISIC website.
Bioengineering 12 01113 g003
Figure 4. Architecture and Training Process of a Generative Adversarial Network.
Figure 4. Architecture and Training Process of a Generative Adversarial Network.
Bioengineering 12 01113 g004
Figure 5. Dermatological color constancy generative adversarial network (DermoCC-GAN) [29].
Figure 5. Dermatological color constancy generative adversarial network (DermoCC-GAN) [29].
Bioengineering 12 01113 g005
Figure 6. Segmentation models with the two proposed GANs: MGAN and EGAN [24].
Figure 6. Segmentation models with the two proposed GANs: MGAN and EGAN [24].
Bioengineering 12 01113 g006
Figure 7. Examples of real and synthetic images with the DermGan model [22].
Figure 7. Examples of real and synthetic images with the DermGan model [22].
Bioengineering 12 01113 g007
Figure 8. Flowchart proposed by the DCGAN architecture for skin lesion classification [17].
Figure 8. Flowchart proposed by the DCGAN architecture for skin lesion classification [17].
Bioengineering 12 01113 g008
Figure 9. Schematic of the GLSFA-GAN model [34].
Figure 9. Schematic of the GLSFA-GAN model [34].
Bioengineering 12 01113 g009
Figure 10. Comparison of images with different GANs and the one proposed by Goceri, E. [23].
Figure 10. Comparison of images with different GANs and the one proposed by Goceri, E. [23].
Bioengineering 12 01113 g010
Figure 11. Dermatoscopic images: Qualitative comparison of the Ding. S et al. model [19].
Figure 11. Dermatoscopic images: Qualitative comparison of the Ding. S et al. model [19].
Bioengineering 12 01113 g011
Table 1. Selected bibliography according to type of study.
Table 1. Selected bibliography according to type of study.
Type of StudyNumber of ArticlesAuthors and Year
Systematic and narrative reviews13Behara et al. (2024) [9], Daneshjou et al. (2022) [7], Furriel et al. (2024) [10], Hermosilla et al. (2024) [11], Iglesias-Puzas and Boixeda (2020) [12], Wongvibulsin et al. (2024) [4], Kourounis et al. (2023) [2], Liu et al. (2025) [13], Li et al. (2022) [5], Nazari and Garcia (2023) [14], Omiye et al. (2023) [3], Vardasca et al. (2024) [15], Yi et al. (2019) [8]
Methodological studies (development and validation of algorithms)20Alshardan et al. (2025) [16], Behara et al. (2023) [17], Carrasco Limeros et al. (2022) [18], Ding et al. (2021) [19], Esteva et al. (2017) [20], Efimenko et al. (2020) [21], Ghorbani et al. (2020) [22], Goceri (2024) [23], Innani et al. (2023) [24], Heenaye-Mamode Khan et al. (2022) [25], Perez and Ventura (2023) [26], Qin et al. (2020) [27], Ren et al. (2022) [28], Salvi et al. (2022) [29], Salvi et al. (2024) [30], Sharafudeen et al. (2023) [31], Jütte et al. (2024) [32], Veeramani and Jayaraman (2025) [33], Yi et al. (2019) [8], Zou et al. (2024) [34].
Ethical or regulatory studies3Albisua (2025) [35], Bleher and Braun (2023) [36], Rezk et al. (2022) [37]
Academic papers (FDP/MT)2Higuera González (2025) [38], Redondo Hernández (2020) [39]
Others (comments, proposals, hybrids not focused on GANs)4Puri et al. (2022) [40], Karras et al. (2018) [41], Mirza and Osindero (2014) [42], Paladugu et al. (2023) [43].
Table 2. Methodological quality assessment of studies focused on the use of GANs in dermatology.
Table 2. Methodological quality assessment of studies focused on the use of GANs in dermatology.
ArticleType of StudyObjectiveMethodLevel of EvidenceStrengthsLimitations
Alshardan et al. (2025) [16]Methodological studyImprove analysis of unbalanced dermoscopic imagesDCGAN to generate synthetic images; evaluation with ResNet50IIIGood performance with minority classesUse of only one dataset, limited external validation
Behara et al. (2023) [17]Methodological studySynthesis and classification of skin lesions using an improved DCGANDCGAN with modified architecture + CNN classifierIIIGood classification performance after augmentation, testing with multiple lesion typesLimited dataset, no external validation
Carrasco Limeros et al. (2022) [18]Comparative studyEvaluate different GAN models in dermatologyComparison of StyleGAN and CycleGAN architecturesIIIExtensive comparison, use of quantitative and qualitative metricsPreprint study, lacking peer review
Ding et al. (2021) [19]Methodological studyHigh-resolution dermoscopic image synthesiscGAN with improved architecture for dermoscopic imagingIIIHigh image quality generated, robust metricsEvaluation limited to synthetic quality, not clinical diagnosis
Ghorbani et al. (2020) [22]Methodological studySynthetic generation of clinical skin images with pathologyDermGAN (customized GAN with regularization to maintain pathologic features)IIIPreservation of relevant clinical features, morphological diversity, educational potentialNot clinically validated, dataset limited to Stanford images
Goceri (2024) [23]Methodological studyAugmentation of hybrid lossy imagesGAN with combined loss function (pixel + perception)IIIImproved class balance, direct applicability to classifier modelsLimited to a single GAN type and database
Innani et al. (2023) [24]Methodological studySegmentation of lesions with GANGAN personalized with encoder–decoder architectureIIIBetter delineation accuracy than U-NetNo comparison with other models
Heenaye-Mamode Khan et al. (2022) [25]Methodological studyMulticlass classification of skin diseasesDGAN (modified GAN) combined with CNNIIIGood multiclass performance, complete metricsExternal validation not performed
Karras et al. (2018) [41]Foundational studyImprove quality and stability of GANProgressive GAN proposal with progressive trainingIVHigh synthetic quality, significant methodological impactNot focused on dermatology
Mirza and Osindero (2014) [42]Foundational studyIntroduce the concept of conditional GANConditional GAN with class vectorIVCornerstone for cGANsNo direct application in health
Paladugu et al. (2023) [43]Methodological reviewReview of the use of cGANs in medicineNarrative review with classification of applicationsVBroad topic coverage, useful for developersDoes not evaluate empirical evidence
Perez and Ventura (2023) [26]Methodological studyData augmentation for skin cancer diagnosisProgressive HLG for dermoscopic image generationIIIImproved diagnostic accuracy using synthetic imagesSimulation-based assessment alone
Qin et al. (2020) [27]Methodological studySynthetic imaging classification of skin lesionsGAN for generation + CNN for classificationIIIImproved sensitivity in classes with less dataLimited data and no clinical validation
Ren et al. (2022) [28]Methodological studyControlled generation of medical imagesGAN with attribute controlIIIExplicit control of variables in synthetic imagingNot exclusively focused on dermatology
Salvi et al. (2022) [29]Methodological studyColor standardization in dermatological imagesDermoCC-GAN for visual standardizationIIIImproved comparability of images between devicesLack of direct clinical impact assessed
Salvi et al. (2024) [30]Methodological studyColor standardization in pathology and dermatologyGANs with style transfer networksIIICross-cutting application in digital medicineNon-dermatology specific
Sharafudeen et al. (2023) [31]Methodological studyDetection of false synthetic images (deepfakes)Vision Transformer on images generated with CGANIIIHigh precision, novel combination of technologiesLow sample size
Veeramani and Jayaraman (2025) [33]Methodological studyHigh-resolution melanoma image reconstructionMELIIGAN (enhanced GAN model with residual blocks)IIIIntricate details preserved, high objective metricsNot tested in clinical setting
Yi et al. (2019) [8]Narrative reviewReview of the use of GANs in medical imagingNarrative review with task analysis and architectureVDetailed overview, basis for future researchDoes not evaluate clinical efficacy directly
Zou et al. (2024) [34]Methodological studySegmentation of skin lesions using GANs with semantic awarenessGANs with global and local care modulesIIISignificant improvement in segmentation metrics, innovative architectureValidation on only two databases, no comparison to non-GAN methods
Table 3. Examples of commercial dermatology applications using GAN-based technology (date of the commercial tool search: April 2025).
Table 3. Examples of commercial dermatology applications using GAN-based technology (date of the commercial tool search: April 2025).
Clinical TaskHow a GAN HelpsClinical BenefitCommercial Apps® Examples
Data augmentation for AIGenerates realistic synthetic images of lesionsImproves performance of AI models even with little dataSkinVision, DermaSensor
Progression simulationVisually predicts lesion progressionAssists in treatment planning and patient educationFotoFinder, 3Derm
Image reconstructionRestores blurred or poorly illuminated imagesImproves diagnosis at a distance or with suboptimal imagesModMed EMA, Skinive
Image standardizationCorrects illumination and contrast differences contrastFacilitates tracking and comparison between consultationsDermoCC-GAN,
DermEngine
Lesion segmentationDistinguishes lesion from backgroundAllows for objective measurement of the affected area and clinical follow-
up
SkinIO, SLSNet
AI-assisted classificationImproves datasets to train classifiersAutomated diagnostic support or in
triage systems
SkinVision, DermaSensor
Medical educationCreates didactic images of rare casesEnhances training of students and residentsCrisalix, TouchMD
Skin tone inclusionGenerates images of diverse skin tonesReduces bias in AI, improves diagnostic fairness diagnosisFitzpatrick 17k, SkinIO
Patient toolsSimulates evolution/improvement with treatmentIncreases therapeutic adherence and patient understandingCrisalix, ModMed Patient Portal
Esthetic/surgical simulationVisualize effects of surgeries or treatmentsImproves communication and informed consentCrisalix, MirrorMe3D
Clinical AppsBasis for monitoring or self-monitoringChronic patient management and remote monitoringSkinive, DermaSensor
Table 4. Metrics to evaluate GANs in medical imaging.
Table 4. Metrics to evaluate GANs in medical imaging.
MetricWhat It MeasuresAdvantagesLimitationsApplication in Dermatology
FIDDistance between feature distributions (Inception)Sensitive to quality and diversity, correlates with human perceptionDepends on extraction model (Inception not trained on medical data)Widely used in studies of GANs in skin, although not specific
SSIMPerceptual structural similarityCaptures differences in luminance, contrast, and structureDoes not consider semantic content or general realismUseful in direct comparison (e.g., reconstruction or super-resolution)
ISDiversity and realism using classificationEasy to implementDoes not compare to real data, biased by base networkLess used in medicine, more in natural imaging
PSNRPixel-to-pixel error (based on noise)Simple and knownNon-perceptual, very sensitive to small differencesUsed in reconstruction or resolution enhancement tasks
MSEMean squared difference between imagesIntuitive and mathematically simpleVery limited for perceptual imagesSimilar to PSNR, useful for tasks where ground truth is known
LPIPSDistance between deep activations (perceptual)High correlation with human perception, robust to small deformationsMore computationally expensiveWell suited for comparing realism in synthetic lesions
PPLLatent space smoothnessEvaluates interpolations in latent space (important in GANs)More useful for model evaluation than for direct image comparisonIndirect, useful in GAN development (e.g., StyleGAN)
Precision/Recall (distributional)Fidelity vs. diversity of sample generatedAllows evaluation of
collapse and overfitting mode
More complex to calculateUseful when looking for variety of injury types
Mode ScoreRealism and class coverageEvaluates diversity when there are clear labelsRequires class annotationsApplicable in annotated datasets (e.g., lesion types)
Dice/JaccardSpatial overlap between masks or structuresWidely used in medical imaging with segmentationOnly applicable if segmentations are presentIdeal if images with defined anatomical
regions
Table 5. Regulatory landscape for GAN-based applications in dermatology: European Union vs. United States.
Table 5. Regulatory landscape for GAN-based applications in dermatology: European Union vs. United States.
DomainEuropean Union (EU)United States (U.S.)
Primary RegulationAI Act (2024, pending full implementation); Medical Device Regulation (MDR, 2017/745)FDA regulations under the Federal Food, Drug, and Cosmetic Act and Software as a Medical Device (SaMD) Framework
Regulatory AuthorityEuropean Commission; National Competent Authorities; European Medicines Agency (EMA)Food and Drug Administration (FDA); Center for Devices and Radiological Health (CDRH)
AI-specific FrameworkAI Act classifies AI systems by risk (unacceptable, high, limited, minimal); GANs used in diagnosis are high-riskFDA does not yet have a specific AI law but follows guidance on Good Machine Learning Practice (GMLP) and SaMD categories
Classification ApproachRisk-based: GANs for clinical decision-making would be “high-risk” under AI Act and Class IIa or higher under MDRFunction-based: based on intended use, risk level, and degree of autonomy; GANs used in diagnosis likely fall under Class II or III
Validation and SafetyStrict requirements under MDR for clinical evaluation, safety, and post-market surveillanceRequires premarket clearance or approval (510 (k), De Novo, PMA), and validation through clinical evidence
Ethical GuidelinesEuropean Ethics Guidelines for Trustworthy AI (2019); human oversight and transparency mandatoryAI-based tools must demonstrate safety, effectiveness, and transparency; bias mitigation is a growing focus (FDA AI/ML Action Plan)
Data Privacy and GovernanceGDPR compliance required, including data minimization and patient consentHIPAA compliance governs medical data use and patient privacy protection
Ongoing InitiativesEU AI Liability Directive; European Health Data Space; EUDAMED platformDigital Health Center of Excellence; AI/ML SaMD Action Plan; real-world performance monitoring frameworks
AI: Artificial Intelligence; GAN: Generative Adversarial Network; MDR: Medical Device Regulation; SaMD: Software as a Medical Device; FDA: Food and Drug Administration; CDRH: Center for Devices and Radiological Health; GMLP: Good Machine Learning Practice; GDPR: General Data Protection Regulation; HIPAA: Health Insurance Portability and Accountability Act; PMA: Premarket Approval; EUDAMED: European Database on Medical Devices.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Izu-Belloso, R.M.; Ibarrola-Altuna, R.; Rodriguez-Alonso, A. Generative Adversarial Networks in Dermatology: A Narrative Review of Current Applications, Challenges, and Future Perspectives. Bioengineering 2025, 12, 1113. https://doi.org/10.3390/bioengineering12101113

AMA Style

Izu-Belloso RM, Ibarrola-Altuna R, Rodriguez-Alonso A. Generative Adversarial Networks in Dermatology: A Narrative Review of Current Applications, Challenges, and Future Perspectives. Bioengineering. 2025; 12(10):1113. https://doi.org/10.3390/bioengineering12101113

Chicago/Turabian Style

Izu-Belloso, Rosa Maria, Rafael Ibarrola-Altuna, and Alex Rodriguez-Alonso. 2025. "Generative Adversarial Networks in Dermatology: A Narrative Review of Current Applications, Challenges, and Future Perspectives" Bioengineering 12, no. 10: 1113. https://doi.org/10.3390/bioengineering12101113

APA Style

Izu-Belloso, R. M., Ibarrola-Altuna, R., & Rodriguez-Alonso, A. (2025). Generative Adversarial Networks in Dermatology: A Narrative Review of Current Applications, Challenges, and Future Perspectives. Bioengineering, 12(10), 1113. https://doi.org/10.3390/bioengineering12101113

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop