Next Article in Journal
Integrating Spatial Omics and Deep Learning: Toward Predictive Models of Cardiomyocyte Differentiation Efficiency
Previous Article in Journal
Verification of the Effectiveness of a Token Economy Method Through Digital Intervention Content for Children with Attention-Deficit/Hyperactivity Disorder
Previous Article in Special Issue
Multimodal Integration Enhances Tissue Image Information Content: A Deep Feature Perspective
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Self-Attention Diffusion Models for Zero-Shot Biomedical Image Segmentation: Unlocking New Frontiers in Medical Imaging

by
Abderrachid Hamrani
* and
Anuradha Godavarty
*
Optical Imaging Laboratory, Department of Biomedical Engineering, Florida International University, 10555 West Flagler Street, EC 2675, Miami, FL 33174, USA
*
Authors to whom correspondence should be addressed.
Bioengineering 2025, 12(10), 1036; https://doi.org/10.3390/bioengineering12101036
Submission received: 14 July 2025 / Revised: 26 August 2025 / Accepted: 23 September 2025 / Published: 27 September 2025
(This article belongs to the Special Issue Medical Imaging Analysis: Current and Future Trends)

Abstract

Producing high-quality segmentation masks for medical images is a fundamental challenge in biomedical image analysis. Recent research has investigated the use of supervised learning with large volumes of labeled data to improve segmentation across medical imaging modalities and unsupervised learning with unlabeled data to segment without detailed annotations. However, a significant hurdle remains in constructing a model that can segment diverse medical images in a zero-shot manner without any annotations. In this work, we introduce the attention diffusion zero-shot unsupervised system (ADZUS), a new method that uses self-attention diffusion models to segment biomedical images without needing any prior labels. This method combines self-attention mechanisms to enable context-aware and detail-sensitive segmentations, with the strengths of the pre-trained diffusion model. The experimental results show that ADZUS outperformed state-of-the-art models on various medical imaging datasets, such as skin lesions, chest X-ray infections, and white blood cell segmentations. The model demonstrated significant improvements by achieving Dice scores ranging from 88.7% to 92.9% and IoU scores from 66.3% to 93.3%. The success of the ADZUS model in zero-shot settings could lower the costs of labeling data and help it adapt to new medical imaging tasks, improving the diagnostic capabilities of AI-based medical imaging technologies.

Graphical Abstract

1. Introduction

Biomedical image segmentation serves a crucial role in healthcare, facilitating diagnosis, treatment planning, and disease monitoring. Convolutional neural networks (CNNs), particularly the U-Net architecture, demonstrate effectiveness in this domain by utilizing a contracting path to capture context and an extensive path for precise localization [1,2]. These networks excel at biological image segmentation and cellular tracking across various microscopy modalities [3]. The success of deep convolutional networks stems from extensive labeled datasets like ImageNet [4,5]. However, biomedical image segmentation faces significant challenges due to the scarcity of annotated data [6,7]. While data augmentation techniques expand training datasets, they generate variations in existing samples without introducing true biological diversity. These synthetic modifications cannot capture the full spectrum of anatomical variations, pathological manifestations, and imaging conditions present in clinical settings [8].
Recent research explores unsupervised and zero-shot learning approaches to address these limitations [9,10,11,12]. Stable diffusion models exhibit the ability to learn inherent object concepts within their attention layers, enabling high-quality segmentation without extensive labeled datasets [13]. This capability proves particularly valuable in biomedical imaging, where manual annotation requires expert knowledge and significant time investment. The complexity of biomedical images varies significantly due to imaging modality, patient anatomy, and pathological variations, necessitating automated segmentation methods that operate with minimal supervision. Self-attention diffusion models present a promising solution [14,15]. These models excel at learning complex data distributions and capturing fine-grained details through their self-attention mechanisms. The ability to discern long-range dependencies and contextual information makes them particularly suited for biomedical image segmentation tasks [16,17].
This study introduces ADZUS (Attention Diffusion Zero-shot Unsupervised System), a deep learning model that leverages self-attention diffusion mechanisms for zero-shot biomedical image segmentation. The model generates precise segmentation masks through iterative merging of attention maps based on their Kullback–Leibler (KL) divergence [18,19], enabling coherent region identification while reducing computational redundancy. Our contributions include demonstrating the effectiveness of self-attention diffusion models for zero-shot biomedical image segmentation, achieving superior performance on benchmark datasets, and eliminating dependence on annotated data. This approach significantly reduces barriers to deploying advanced segmentation models in resource-constrained settings. Extensive experiments across various biomedical datasets validate the model’s versatility and robustness.

Motivation

Stable Diffusion models, initially designed for image generation, have demonstrated the capability to produce highly realistic and detailed images using only text prompts. Figure 1 illustrates this potential by comparing images generated by Stable Diffusion with real clinical cases across three different imaging modalities: white blood cells, skin lesions, and chest X-rays. The top row (a, b, c) presents synthetic images generated using descriptive text prompts in [20], while the bottom row (d, e, f) displays corresponding real medical cases.
To generate these images, Stable Diffusion was prompted with medical descriptions such as
  • “A highly detailed, realistic microscopic view of white blood cells, with clear distinction of nuclei and surrounding red blood cells, captured under high-resolution clinical lighting.”
  • “A close-up of a malignant skin lesion with irregular borders, dark pigmentation, and slight ulceration, presented in high-resolution medical imaging detail.”
  • “A high-resolution X-ray scan of a human chest, capturing clear details of the ribcage, lungs, and mediastinum, optimized for clinical radiographic examination.”
The synthetic images closely resemble real medical cases taken from databases in [21,22,23,24,25,26,27,28], accurately capturing critical features such as cellular morphology in microscopic imaging, pigmentation patterns in skin lesions, and structural details in chest radiographs. This high degree of realism suggests that the self-attention mechanisms within Stable Diffusion models inherently capture relevant medical imaging concepts, making them promising candidates for zero-shot segmentation tasks.
Given this capability, an important question arises: has Stable Diffusion been trained on medical images, particularly those used in this study? To address this, we conducted a data similarity verification analysis using the publicly available LAION-5B search tool, which forms the primary training dataset of Stable Diffusion. The images used in this study, including those in Figure 1d–f and those in the results section, were tested against the LAION-5B database to determine if they were present in Stable Diffusion’s pretraining corpus. Our analysis revealed no identical or closely matching images within the LAION-5B dataset. These findings confirm that our segmentation experiments were conducted without prior model exposure to the images used in this study, ensuring a fair evaluation of ADZUS for zero-shot medical image segmentation.

2. Methodology

A pre-trained stable diffusion model is leveraged by ADZUS, with utilization of its self-attention layers to generate high-quality segmentation masks. In Section 2.1, a concise overview of the stable diffusion model architecture will be provided, followed by a detailed introduction to ADZUS in Section 2.2.

2.1. Overview of the Stable Diffusion Model

The stable diffusion model [13], a well-known variant within the diffusion model family [29,30], is a generative model that operates through both forward and reverse passes. During the forward pass, Gaussian noise is incrementally added at each time step until the image becomes completely isotropic Gaussian noise. Conversely, in the reverse pass, the model is trained to progressively eliminate this Gaussian noise, thereby reconstructing the original clean image. The stable diffusion model [13] incorporates an encoder–decoder and U-Net architecture with attention layers [29].
Initially, an image x H × W × 3 is compressed into a latent space with reduced spatial dimensions z h × w × c using an encoder z = E x . This latent space can then be decompressed back into the image x ˜ = D z through a decoder. All diffusion processes occur within this latent space via the U-Net architecture, which is the primary focus of this paper’s investigation. The U-Net consists of modular blocks, including 16 specific blocks composed of ResNet layers and Transformer layers (Figure 2). The Transformer layer uses two attention mechanisms: self-attention to learn global attention across the image and cross-attention to learn attention between the image and optional text input.
The component of interest for our investigation is the self-attention layer in the Transformer layer. Specifically, there are 16 self-attention layers distributed across the 16 composite blocks, resulting in 16 self-attention tensors. Each attention tensor A k h k × w k × h k × w k is 4-dimensional. Inspired by DiffuMask [31], which demonstrates object grouping in the cross-attention layer, it is hypothesized that the unconditional self-attention also contains inherent object grouping information, which can be used to produce segmentation masks without text inputs.
For each spatial location (I, J) in the attention tensor, the corresponding 2D attention map A k I , J , : , : h k × w k captures the semantic correlation between all locations and the location (I, J). Each location (I, J) corresponds to a region in the original image pixel space, the size of which depends on the receptive field of the tensor.
Two important observations motivate the method proposed in the next section:
  • Intra-Attention Similarity: Within a 2D attention map A k I , J , : , : , locations tend to have strong responses if they correspond to the same object group as (I, J) in the original image space.
  • Inter-Attention Similarity: Between two 2D attention maps, e.g., A k I , J , : , : and A k I + 1 , J + 1 , : , : , they tend to share similar activations if (I, J) and (I + 1, J + 1) belong to the same object group in the original image space.
The resolution of the attention map dictates the size of its receptive field concerning the original image. Lower resolution maps (e.g., 8 × 8) provide better grouping of large objects, while higher resolution maps (e.g., 16 × 16) offer more fine-grained grouping of components within larger objects, potentially identifying smaller objects more effectively. The current stable diffusion model has attention maps in four resolutions: 8 × 8, 16 × 16, 32 × 32, and 64 × 64. Building on these observations, a simple heuristic is proposed to aggregate weights from different resolutions and an iterative method to merge all attention maps into valid segmentation masks. In our experiments, the stable diffusion pre-trained models from “Huggingface” are used [32]. Typically, these prompt-conditioned diffusion models run for 50 or more diffusion steps to generate new images. However, to efficiently extract attention maps for an existing clean image without conditional prompts, we use only the unconditioned latent and run the diffusion process once. The unconditional latent is calculated using an unconditioned text embedding. We set the time-step variable t to a large value (e.g., t = 300) so that real images are viewed as primarily denoised generated images from the diffusion model’s perspective.

2.2. ADZUS Model

Since the self-attention layers capture inherent object grouping information in spatial attention (probability) maps, we propose ADZUS, a simple post-processing method, to aggregate and merge attention tensors into a valid segmentation mask. As shown in Figure 3, the pipeline consists of three components: attention aggregation, iterative attention merging, and non-maximum suppression. ADZUS is built on pre-trained stable diffusion models. For our implementation, we use stable diffusion V1.4 [13].

2.2.1. Attention Aggregation

Given an input image passing through the encoder and U-Net, the stable diffusion model generates 16 attention tensors. Specifically, there are 5 tensors for each of the dimensions: (64 × 64 × 64 × 64), (32 × 32 × 32 × 32), (16 × 16 × 16 × 16), and (8 × 8 × 8 × 8). The goal is to aggregate attention tensors of different resolutions into the highest resolution tensor. To achieve this, the last 2 dimensions of all attention maps are up-sampled (bilinear interpolation) to 64 × 64, their highest resolution. Formally, for A k h k × w k × h k × w k :
A ˜ k = Bilinear - upsample A k h k × w k × 64 × 64
The first 2 dimensions indicate the locations to which attention maps are referenced. Therefore, we aggregate attention maps accordingly. For example, as shown in Figure 4, the attention map in the (0, 0) location in c is first upsampled and then repeatedly aggregated pixel-wise with the 4 attention maps (0, 0), (0, 1), (1, 0), (1, 1) in A z 16 × 16 . Formally, the final aggregated attention tensor A f 64 × 64 is:
A f I , J , : , : = k 1 , , 16 A ˜ k I / δ k , J / δ k , : , : R k
where δ k = 64 / w k and k R k = 1 . The aggregated attention map is normalized to ensure it is a valid distribution. The weights R are important hyper-parameters and are proportional to the resolution w k .

2.2.2. Iterative Attention Merging

In this step, the algorithm computes an attention tensor A f 64 × 64 . The goal is to merge the 64 × 64 attention maps in the tensor A f to a stack of object proposals where each proposal likely contains the activation of a single object or category. Instead of using a K-means algorithm, which requires specifying the number of clusters, we generate a sampling grid from which the algorithm can iteratively merge attention maps.
A set of M × M evenly spaced anchor points are generated. We then sample the corresponding attention maps from the tensor A f . This operation yields a list of M 2 2D attention maps as anchors:
L a = A f i m , j m , : , : 64 × 64 i m , j m M
To measure similarity between attention maps, we use KL divergence:
2 × D A f i , j , A f y , z = KL A f i , j A f y , z + KL A f y , z A f i , j
We start with N iterations of the merging process, where we compute the pair-wise distance between each element in the anchor list and all attention maps, averaging all attention maps with a distance smaller than a threshold τ . This process is repeated in subsequent iterations, reducing the number of proposals by merging maps with distances smaller than τ . A detailed sensitivity analysis of the KL divergence threshold (τ) is provided in Appendix A.
To consolidate the multiple attention maps into coherent object-level proposals, we employ an Iterative Attention Merging algorithm. This procedure progressively merges spatially and semantically similar attention maps, guided by KL-divergence similarity, until stable segmentation proposals are formed. The detailed steps are summarized in Algorithm 1.
Algorithm 1: Iterative Attention Merging
Required parameters:  L a , A f , N , τ
L a = 1 ν i , j ν A f i , j υ = 1 , , M 2
  where  v = i , j D L a υ , A f i , j < τ
for N-1 iterations do
    Initialize  L ˜ p = [ ]
   for  A in L p do
        V = 1 ν υ ν L p υ  # Merge attention maps
          where   v = υ D A , L p υ < τ
      Add  V to L ˜ p
      Remove  L p υ υ v from L p
    end for
        L p L ˜ p
end for

2.2.3. Non-Maximum Suppression

The iterative attention merging step yields a list L p N p × 64 × 64 of N p object proposals in the form of attention maps. To convert the list into a valid segmentation mask, we use non-maximum suppression (NMS). Each element is a probability distribution map, and the final segmentation mask S 512 × 512 is obtained by upsampling all elements in L p to the original resolution and taking the index of the largest probability at each spatial location across all maps. This methodology, combining attention aggregation, iterative merging, and non-maximum suppression, forms the core of the ADZUS approach for producing high-quality segmentation masks.
The iterative attention merging step yields a list of object proposals in the form of attention maps. To convert this list into a valid segmentation mask, we apply NMS, ensuring the selection of the most relevant segmented regions. Rather than directly identifying a specific anatomical structure, ADZUS generates a comprehensive segmentation mask that delineates multiple regions within the image, leveraging its self-attention mechanisms to outline the boundaries of all distinguishable structures. The model performs guided segmentation, generating a set of segmented regions without inherently classifying or labeling a specific area. Instead, it provides a structured segmentation output, allowing clinicians or users to interactively select the relevant region of interest based on the specific medical application.

3. Results

This section presents the evaluation of ADZUS across a range of medical image segmentation tasks to assess its zero-shot segmentation capabilities. The performance of ADZUS is benchmarked against state-of-the-art deep learning models in skin lesion segmentation, chest X-ray infection segmentation, and white blood cell segmentation. These tasks cover diverse imaging modalities, including dermoscopic images, wound photography, radiographic images, and microscopy, highlighting ADZUS’s versatility. Both quantitative metrics, such as dice similarity coefficient (DSC), intersection over union (IoU), precision, and recall, and qualitative visualizations are provided to demonstrate the model’s effectiveness in accurately delineating structures without the need for labeled training data. A detailed description of performance metrics used in this analysis is provided in Appendix A.

3.1. Skin Lesion Segmentation

In this section, the international skin imaging collaboration (ISIC) datasets from 2016 and 2017 have been utilized to provide a comprehensive and diverse benchmark for skin lesion segmentation. The ISIC 2016 dataset [21] included 900 dermoscopic lesion images for training and 379 images with ground truth annotations for testing, focusing on binary segmentation tasks to detect melanoma. The ISIC 2017 dataset [22] expanded the scope by providing 2000 dermoscopic lesion images for training and 600 images with ground truth masks for testing. For the evaluation of the ADZUS model, only the test data from these datasets are used, as ADZUS operates in a zero-shot manner and does not require training. The table below (Table 1) compares the performance of ADZUS with other segmentation models on the ISIC datasets 2016 and 2017.
The results in the table demonstrate that ADZUS achieved competitive performance compared to state-of-the-art segmentation models across the ISIC 2016 and 2017 datasets. In the ISIC 2016 dataset, ADZUS outperformed the top evaluation from the U-Net implementation, achieving a DSC score of 92.9% and an IoU score of 86.8%, compared to 91% and 84.3%, respectively. For the ISIC 2017 dataset, ADZUS achieved a DSC score of 88.7% and an IoU score of 80%, closely aligning with the performance of advanced models such as the EfficientNet-Based Model, which achieved a DSC score of 88% and an IoU score of 80.7%.
Figure 5 illustrates the qualitative performance of the ADZUS model on the ISIC 2016 and ISIC 2017 datasets. In both ISIC 2016 and 2017, the predicted boundaries (green) closely match the true boundaries (red), achieving high DSC scores (from 0.98 to 0.84). These results highlight ADZUS’s effectiveness as a zero-shot segmentation model for skin lesion boundaries.

3.2. Diabetic Foot Wound Segmentation

Building on the performance of ADZUS in skin lesion segmentation, this section transitions to evaluate its effectiveness in diabetic foot wound segmentation. To this end, ADZUS is compared against state-of-the-art segmentation models using the publicly available chronic wound dataset [37]. This dataset consists of 1010 labeled images of diabetic foot ulcers, with 810 images designated for training and 200 for inference. It provides a standardized platform for segmentation performance assessment within a supervised learning framework. The state-of-the-art models includes LinkNet-EffB1 + UNet-EffB2 [38,39], DeepLabV3Plus [40,41], Swin-Unet [39], DDRNet [39], SegFormer-b5 [42], and FUSegNet [39], which were used for the comparison.
As illustrated in Table 2, ADZUS demonstrates competitive performance (using only the test dataset of 200 images) when compared with state-of-the-art segmentation models. Notably, ADZUS outperforms the previously best-performing FUSegNet model by achieving the highest IoU of 86.68% compared to FUSegNet’s IoU of 86.40%. In terms of DSC score, ADZUS ranked slightly behind DeepLabV3Plus, LinkNet-EffB1 + UNet-EffB2 and FUSegNet (which achieved the highest DSC score of 92.70%). These findings collectively highlight the strength of ADZUS in delivering accurate and reliable segmentation results without the need for labeled training data, a significant advantage over supervised models such as FUSegNet and SegFormer-b5.
The qualitative results for the chronic wound dataset are illustrated in Figure 6. The segmentation outputs demonstrate the performance of ADZUS in accurately delineating wound regions, with original boundaries depicted in red and predicted boundaries in green. Additional qualitative results for the chronic wound dataset are presented in Figure 7, offering a comparative analysis of segmentation outputs from ADZUS and benchmark models. The figure highlights ADZUS’s ability to achieve precise segmentation, closely aligning with the original boundaries and minimizing deviations. With a DSC score of 93.56%, ADZUS demonstrates competitive performance, surpassing supervised models like FUSegNet and SegFormer, while performing slightly below DeepLabV3+.

3.3. Chest X-Rays Segmentation for COVID Dataset

In this section, we evaluate the ADZUS model for infection segmentation in chest X-ray images. This task focuses on identifying and segmenting infected regions in lung radiographs, which is crucial for diagnosing and monitoring respiratory diseases such as COVID-19 and pneumonia. The dataset used consists of 2913 chest X-ray images (583 images for test) [23,24,25,26].
Table 3 shows that ADZUS demonstrates superior performance in infection segmentation in chest X-ray images, outperforming state-of-the-art models across all evaluation metrics. ADZUS achieved the highest IoU (66.3%), precision (77.9%), and recall (83.1%), surpassing established deep learning architectures such as ResNet18, ResNet50, EfficientNet-b0, MobileNet_v2, and DenseNet121. Notably, ADZUS outperformed the best-performing DenseNet121 model in terms of IoU and precision scores, emphasizing its ability to accurately delineate infection regions with minimal false positives. The improved recall score suggests that ADZUS is particularly effective at capturing infected lung regions, reducing the likelihood of under-segmentation compared to traditional CNN-based approaches.
The qualitative results for infection segmentation in chest X-ray images are illustrated in Figure 8. The segmentation outputs demonstrate the performance of ADZUS in accurately delineating infection regions, with ground truth boundaries shown in red and predicted boundaries depicted in green. The figure emphasizes ADZUS’s ability to generate high-fidelity segmentation masks across different cases, achieving DSC of 90.39%, 84.94%, and 75.52%.

3.4. White Blood Cell Segmentation

The robustness of ADZUS is further demonstrated through its performance on the WbcMSBench dataset [26,27,28], a microscopy imaging dataset consisting of 80 test samples. Designed for multi-class segmentation across three distinct classes, WbcMSBench presents challenges due to the high variability in white blood cell morphology and overlapping structures. Table 4 shows that ADZUS achieves an IoU of 93.3%, precision of 96.5%, and recall of 96.7%, performing competitively with state-of-the-art models such as DenseNet121 (IoU: 93.8%) and EfficientNet-b0 (IoU: 93.7%). While ADZUS slightly trails these top-performing models in IoU, it matches them in precision and recall, highlighting its effectiveness in accurately delineating cellular boundaries and distinguishing between different cell classes. Figure 9 shows the qualitative results of ADZUS on the WbcMSBench dataset [26] for multi-class white blood cell segmentation. The predicted masks closely align with the ground truth, demonstrating ADZUS’s high accuracy across different cell types. All classes show strong performance, emphasizing the model’s robustness in handling complex cellular structures.

4. Discussion

The results presented in this study demonstrate the effectiveness of ADZUS as a zero-shot medical image segmentation model across multiple imaging modalities. ADZUS achieved competitive performance in skin lesion segmentation, chest X-ray infection segmentation, diabetic foot wound segmentation, and white blood cell segmentation, often surpassing or closely aligning with state-of-the-art supervised models.
In skin lesion segmentation, ADZUS outperformed the top U-Net implementation on the ISIC 2016 dataset, achieving a higher DSC score and IoU. On the ISIC 2017 dataset, its performance was slightly below the EfficientNet-based model in IoU, although ADZUS matched its DSC score. This outcome highlights ADZUS’s robustness in handling diverse lesion types, from melanomas to seborrheic keratosis, while also revealing its limitations when datasets exhibit high inter-class similarity. Such conditions can make fine distinctions between lesion types more challenging, which in turn impacts segmentation quality.
In the diabetic foot wound segmentation task, ADZUS exceeded the performance of the previously best-performing FUSegNet model in IoU and precision, despite operating in a completely unsupervised manner. This finding is particularly significant, as supervised models leverage extensive labeled data to optimize performance, whereas ADZUS delivers comparable accuracy without such data. However, its slightly lower DSC score compared to FUSegNet suggests that ADZUS may struggle with fine-grained boundaries in complex wound structures. We provide additional qualitative examples (Figure 6 and Figure 7) that illustrate these cases, showing both the strengths and the areas where segmentation completeness can be improved. ADZUS demonstrated superior performance in chest X-ray infection segmentation, outperforming architectures like DenseNet121 and EfficientNet-b0 across all key metrics. The model’s high recall indicates its strength in accurately capturing infected lung regions, minimizing the risk of under-segmentation, which is critical in clinical contexts such as COVID-19 diagnosis and monitoring. This result underscores ADZUS’s ability to handle macro-level infection patterns despite the variability and complexity inherent in radiographic images. In white blood cell segmentation, ADZUS performed competitively with state-of-the-art models such as DenseNet121 and EfficientNet-b0, achieving high IoU, precision, and recall. The qualitative results confirm that ADZUS effectively distinguishes between different cell types in multi-class segmentation tasks, handling challenges posed by overlapping structures and varying cell morphology. This performance further validates the model’s adaptability to microscopic imaging applications, extending its relevance beyond macroscopic medical imaging.
While ADZUS proves effective across a range of tasks, certain limitations warrant further exploration. Specifically: (1) Performance tends to be lower in datasets with high inter-class similarity, as observed in ISIC 2017. (2) ADZUS may also underperform when precise, fine-grained boundaries are required, as in some wound images. Addressing these challenges represents a promising avenue for future research. Semi-supervised fine-tuning or domain-specific adaptations could enhance performance in such scenarios. By leveraging the text-to-image generation capabilities inherent in stable diffusion models, ADZUS could also be extended to incorporate text-guided segmentation, where natural language descriptions of anatomical structures or pathological features guide the segmentation process. This enhancement would allow medical professionals to refine segmentation boundaries and identify specific regions of interest using descriptive text prompts.
In future work, ADZUS will benefit from expanding its evaluation to additional medical imaging modalities, such as retinal vessel segmentation or organ segmentation in CT scans, to further validate its generalizability. The text-guided approach could be particularly valuable in complex settings where precise anatomical descriptions can help distinguish between visually similar structures. Moreover, exploring the integration of self-supervised learning techniques, domain adaptation strategies, and runtime optimization will further improve its performance and feasibility in clinical environments. In addition, we report on computational feasibility. As noted previously, the use of a large stable diffusion backbone and iterative merging steps increases computational overhead, preventing the method from achieving strict real-time performance. Nevertheless, by requiring only a single forward pass through the diffusion model and a limited number of merging iterations, ADZUS remains practical for near real-time clinical use. Future work will focus on optimizing runtime, such as reducing anchor density, parallelizing attention merging, or leveraging lightweight diffusion backbones, to further improve efficiency for real-time deployment.

5. Conclusions

In this study, we introduced ADZUS, a zero-shot, unsupervised segmentation model leveraging self-attention diffusion mechanisms to address the challenges in biomedical image segmentation. The model’s ability to deliver high-quality segmentation results without the need for annotated datasets marks a significant advancement in the field, especially in scenarios where labeled data is scarce or difficult to obtain. ADZUS demonstrated competitive or superior performance across a variety of medical imaging tasks, including skin lesion segmentation, diabetic foot ulcer or any wound segmentation, chest X-ray infection segmentation, and white blood cell segmentation. These results underscore its versatility and robustness across diverse imaging modalities. The success of ADZUS highlights the potential of self-attention diffusion models to transform medical image analysis by minimizing reliance on extensive labeled datasets. This innovation enhances the accessibility of advanced segmentation tools in resource-constrained settings and paves the way for broader applications in clinical and diagnostic workflows. Future research will focus on expanding ADZUS’s capabilities to additional medical imaging domains, exploring integration with real-time diagnostic systems, and advancing toward a more autonomous segmentation framework. This next phase will enable ADZUS to recognize, label and segment specific zones of interest requested by clinicians, such as infections or abnormal regions, leveraging its learned attention patterns. The promising results of this study suggest that ADZUS could evolve into an intelligent, semi-supervised system capable of automatic region identification and preliminary labeling, thereby enhancing its adaptability for clinical applications while maintaining interpretability and expert oversight.

Author Contributions

Conceptualization, A.H.; methodology, A.H.; software, A.H.; validation, A.H.; formal analysis, A.H., A.G.; investigation, A.H.; resources, A.H.; data curation, A.H.; writing—original draft preparation, A.H.; writing—review and editing, A.H., A.G.; visualization, A.H.; supervision, A.H.; project administration, A.G.; funding acquisition, A.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by NIH-NIBIB (R01EB033413).

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Appendix A.1. Evaluation Metrics

In this study, the evaluation of segmentation performance was critical to understanding the effectiveness of the ADZUS model for medical image segmentation. Established metrics (expressed as percentages), including dice similarity coefficient (DSC), intersection-over-union (IoU), precision, and recall, were adopted.
D S C = 2 T P 2 T P + F P + F N × 100 ,
I o U = T P T P + F P + F N × 100 ,
p r e c i s i o n = T P T P + F P × 100 ,
r e c a l l = T P T P + F N × 100 ,
where TP, FP, and FN represent true positives, false positives, and false negatives, respectively.

Appendix A.2. Sensitivity Analysis of KL Divergence Threshold (τ)

The iterative attention merging step in ADZUS relies on the KL divergence threshold τ to decide whether two attention maps should be merged. As observed in the original DiffSeg framework, τ is the most sensitive hyperparameter in the merging process.
To evaluate its effect, we conducted a sensitivity analysis by varying τ across the range [0.7, 1.3] while keeping other parameters fixed (M2 = 256 anchors, N = 3 iterations, t = 300). The results are summarized as follows:
  • Low τ values (≤0.8): Segmentation tends to over-fragment, producing a large number of small regions and lower Dice/IoU scores.
  • Moderate τ values (0.9–1.1): Performance is stable, yielding the best balance between segmentation completeness and precision.
  • High τ values (≥1.2): Excessive merging occurs, collapsing multiple objects into a single region and reducing boundary accuracy.
Based on this study, we recommend using τ within [0.9–1.1] for general-purpose biomedical segmentation. For our experiments, τ = 1.1 was applied to skin lesions and wound segmentation tasks, while τ = 0.9 was used for chest X-ray datasets, which require finer boundary preservation.

References

  1. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. Unet++: A Nested u-Net Architecture for Medical Image Segmentation. In Lecture Notes in Computer Science; Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics; Springer: Berlin/Heidelberg, Germany, 2018; Volume 11045, pp. 3–11. [Google Scholar]
  2. Siddique, N.; Paheding, S.; Elkin, C.P.; Devabhaktuni, V. U-Net and Its Variants for Medical Image Segmentation: A Review of Theory and Applications. IEEE Access 2021, 9, 82031–82057. [Google Scholar] [CrossRef]
  3. Zunair, H.; Ben Hamza, A. Sharp U-Net: Depthwise Convolutional Network for Biomedical Image Segmentation. Comput. Biol. Med. 2021, 136, 104699. [Google Scholar] [CrossRef]
  4. Iglovikov, V.; Shvets, A. TernausNet: U-Net with VGG11 Encoder Pre-Trained on ImageNet for Image Segmentation. arXiv 2018, arXiv:1801.05746. [Google Scholar]
  5. Jiang, H.; Diao, Z.; Shi, T.; Zhou, Y.; Wang, F.; Hu, W.; Zhu, X.; Luo, S.; Tong, G.; Yao, Y.D. A Review of Deep Learning-Based Multiple-Lesion Recognition from Medical Images: Classification, Detection and Segmentation. Comput. Biol. Med. 2023, 157, 106726. [Google Scholar] [CrossRef] [PubMed]
  6. Wang, S.; Li, C.; Wang, R.; Liu, Z.; Wang, M.; Tan, H.; Wu, Y.; Liu, X.; Sun, H.; Yang, R.; et al. Annotation-Efficient Deep Learning for Automatic Medical Image Segmentation. Nat. Commun. 2021, 12, 1–13. [Google Scholar] [CrossRef] [PubMed]
  7. Yang, L.; Zhang, Y.; Chen, J.; Zhang, S.; Chen, D.Z. Suggestive Annotation: A Deep Active Learning Framework for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2017: 20th International Conference, Quebec, QC, Canada, 11–13 September 2017; Volume 10435, pp. 399–407. [Google Scholar] [CrossRef]
  8. Garcea, F.; Serra, A.; Lamberti, F.; Morra, L. Data Augmentation for Medical Imaging: A Systematic Literature Review. Comput. Biol. Med. 2023, 152, 106391. [Google Scholar] [CrossRef] [PubMed]
  9. Luo, G.; Xie, W.; Gao, R.; Zheng, T.; Chen, L.; Sun, H. Unsupervised Anomaly Detection in Brain MRI: Learning Abstract Distribution from Massive Healthy Brains. Comput. Biol. Med. 2023, 154, 106610. [Google Scholar] [CrossRef]
  10. Ghai, D.; Mishra, S.P.; Rani, S. Supervised and Unsupervised Techniques for Biomedical Image Classification. Min. Biomed. Text, Images Vis. Featur. Inf. Retr. 2025, 153–211. [Google Scholar] [CrossRef]
  11. Liu, J.; Zhao, J.; Xiao, J.; Zhao, G.; Xu, P.; Yang, Y.; Gong, S. Unsupervised Domain Adaptation Multi-Level Adversarial Learning-Based Crossing-Domain Retinal Vessel Segmentation. Comput. Biol. Med. 2024, 178, 108759. [Google Scholar] [CrossRef]
  12. Zhang, Y.; Shen, Z.; Jiao, R. Segment Anything Model for Medical Image Segmentation: Current Applications and Future Directions. Comput. Biol. Med. 2024, 171, 108238. [Google Scholar] [CrossRef]
  13. Rombach, R.; Blattmann, A.; Lorenz, D.; Esser, P.; Ommer, B. High-Resolution Image Synthesis with Latent Diffusion Models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–24 June 2022; pp. 10684–10695. [Google Scholar]
  14. Kang, S.; Song, J.; Kim, J. Advancing Medical Image Segmentation: Morphology-Driven Learning with Diffusion Transformer. arXiv 2024, arXiv:2408.00347. [Google Scholar] [CrossRef]
  15. Yan, P.; Li, M.; Zhang, J.; Li, G.; Jiang, Y.; Luo, H. Cold SegDiffusion: A Novel Diffusion Model for Medical Image Segmentation. Knowledge-Based Syst. 2024, 301, 112350. [Google Scholar] [CrossRef]
  16. Huang, Y.; Zhu, J.; Hassan, H.; Su, L.; Li, J.; Huang, B. Label-Efficient Multi-Organ Segmentation Method with Diffusion Model. arXiv 2024, arXiv:2402.15216. [Google Scholar]
  17. Shi, Z.; Zou, H.; Luo, F.; Huo, Z. MFSegDiff: A Multi-Frequency Diffusion Model for Medical Image Segmentation. In Proceedings of the 2024 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Lisbon, Portugal, 3–6 December 2024; pp. 2396–2401. [Google Scholar] [CrossRef]
  18. Zheng, Q.; Lu, Z.; Yang, W.; Zhang, M.; Feng, Q.; Chen, W. A Robust Medical Image Segmentation Method Using KL Distance and Local Neighborhood Information. Comput. Biol. Med. 2013, 43, 459–470. [Google Scholar] [CrossRef] [PubMed]
  19. Conforti, G.; Durmus, A.; Silveri, M.G. KL Convergence Guarantees for Score Diffusion Models under Minimal Data Assumptions. SIAM J. Math. Data Sci. 2025, 7, 86–109. [Google Scholar] [CrossRef]
  20. DreamStudio. Available online: https://dreamstudio.ai/generate (accessed on 20 February 2025).
  21. Gutman, D.; Codella, N.C.F.; Celebi, E.; Helba, B.; Marchetti, M.; Mishra, N.; Halpern, A. Skin Lesion Analysis toward Melanoma Detection: A Challenge at the International Symposium on Biomedical Imaging (ISBI) 2016, Hosted by the International Skin Imaging Collaboration (ISIC). arXiv 2016, arXiv:1605.01397. [Google Scholar] [CrossRef]
  22. Codella, N.C.F.; Gutman, D.; Celebi, M.E.; Helba, B.; Marchetti, M.A.; Dusza, S.W.; Kalloo, A.; Liopyris, K.; Mishra, N.; Kittler, H.; et al. Skin Lesion Analysis toward Melanoma Detection: A Challenge at the 2017 International Symposium on Biomedical Imaging (ISBI), Hosted by the International Skin Imaging Collaboration (ISIC). In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; pp. 168–172. [Google Scholar] [CrossRef]
  23. Chowdhury, M.E.H.; Rahman, T.; Khandakar, A.; Mazhar, R.; Kadir, M.A.; Mahbub, Z.B.; Islam, K.R.; Khan, M.S.; Iqbal, A.; Al Emadi, N.; et al. Can AI Help in Screening Viral and COVID-19 Pneumonia? IEEE Access 2020, 8, 132665–132676. [Google Scholar] [CrossRef]
  24. Rahman, T.; Khandakar, A.; Qiblawey, Y.; Tahir, A.; Kiranyaz, S.; Abul Kashem, S.B.; Islam, M.T.; Al Maadeed, S.; Zughaier, S.M.; Khan, M.S.; et al. Exploring the Effect of Image Enhancement Techniques on COVID-19 Detection Using Chest X-Ray Images. Comput. Biol. Med. 2021, 132, 104319. [Google Scholar] [CrossRef]
  25. Wang, X.; Peng, Y.; Lu, L.; Lu, Z.; Bagheri, M.; Summers, R.M. ChestX-Ray8: Hospital-Scale Chest X-Ray Database and Benchmarks on Weakly-Supervised Classification and Localization of Common Thorax Diseases. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 3462–3471. [Google Scholar] [CrossRef]
  26. Kuş, Z.; Aydin, M. MedSegBench: A Comprehensive Benchmark for Medical Image Segmentation in Diverse Data Modalities. Sci. Data 2024, 11, 1283. [Google Scholar] [CrossRef]
  27. Zheng, X.; Wang, Y.; Wang, G.; Liu, J. Fast and Robust Segmentation of White Blood Cell Images by Self-Supervised Learning. Micron 2018, 107, 55–71. [Google Scholar] [CrossRef]
  28. Acevedo, A.; Alférez, S.; Merino, A.; Puigví, L.; Rodellar, J. Recognition of Peripheral Blood Cell Images Using Convolutional Neural Networks. Comput. Methods Programs Biomed. 2019, 180, 105020. [Google Scholar] [CrossRef] [PubMed]
  29. Saharia, C.; Chan, W.; Saxena, S.; Li, L.; Whang, J.; Denton, E.L.; Ghasemipour, K.; Gontijo Lopes, R.; Karagol Ayan, B.; Salimans, T.; et al. Photorealistic Text-to-Image Diffusion Models with Deep Language Understanding. Adv. Neural Inf. Process. Syst. 2022, 35, 36479–36494. [Google Scholar]
  30. Ho, J.; Jain, A.; Abbeel, P. Denoising Diffusion Probabilistic Models. Adv. Neural Inf. Process. Syst. 2020, 33, 6840–6851. [Google Scholar]
  31. Wu, W.; Zhao, Y.; Shou, M.Z.; Zhou, H.; Shen, C. DiffuMask: Synthesizing Images with Pixel-Level Annotations for Semantic Segmentation Using Diffusion Models. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 1–6 October 2023; pp. 1206–1217. [Google Scholar] [CrossRef]
  32. Tian, J.; Aggarwal, L.; Colaco, A.; Kira, Z.; Gonzalez-Franco, M. Diffuse, Attend, and Segment: Unsupervised Zero-Shot Segmentation Using Stable Diffusion. arXiv 2023, arXiv:2308.12469. [Google Scholar]
  33. Yang, X.; Zeng, Z.; Yeo, S.Y.; Tan, C.; Tey, H.L.; Su, Y. A Novel Multi-Task Deep Learning Model for Skin Lesion Segmentation and Classification. arXiv 2017, arXiv:1703.01025. [Google Scholar] [CrossRef]
  34. Menegola, A.; Tavares, J.; Fornaciali, M.; Li, L.T.; Avila, S.; Valle, E. RECOD Titans at ISIC Challenge 2017. arXiv 2017, arXiv:1703.04819. [Google Scholar] [CrossRef]
  35. Goyal, M.; Yap, M.H.; Hassanpour, S. Multi-Class Semantic Segmentation of Skin Lesions via Fully Convolutional Networks. In Proceedings of the 13th International Joint Conference on Biomedical Engineering Systems and Technologies, Valletta, Malta, 24–26 February 2020; pp. 290–294. [Google Scholar] [CrossRef]
  36. Lama, N.; Hagerty, J.; Nambisan, A.; Stanley, R.J.; Van Stoecker, W. Skin Lesion Segmentation in Dermoscopic Images with Noisy Data. J. Digit. Imaging 2023, 36, 1712–1722. [Google Scholar] [CrossRef]
  37. Wang, C.; Anisuzzaman, D.M.; Williamson, V.; Dhar, M.K.; Rostami, B.; Niezgoda, J.; Gopalakrishnan, S.; Yu, Z. Fully Automatic Wound Segmentation with Deep Convolutional Neural Networks. Sci. Rep. 2020, 10, 21897. [Google Scholar] [CrossRef]
  38. Mahbod, A.; Schaefer, G.; Ecker, R.; Ellinger, I. Automatic Foot Ulcer Segmentation Using an Ensemble of Convolutional Neural Networks. In Proceedings of the 2022 26th International Conference on Pattern Recognition (ICPR), Montreal, QC, Canada, 21–25 August 2022; pp. 4358–4364. [Google Scholar] [CrossRef]
  39. Dhar, M.K.; Zhang, T.; Patel, Y.; Gopalakrishnan, S.; Yu, Z. FUSegNet: A Deep Convolutional Neural Network for Foot Ulcer Segmentation. Biomed. Signal Process. Control 2024, 92, 106057. [Google Scholar] [CrossRef]
  40. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 834–848. [Google Scholar] [CrossRef]
  41. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. In Proceedings of the Computer Vision—ECCV 2018, Munich, Germany, 8–14 September 2018; Volume 11211, pp. 833–851. [Google Scholar] [CrossRef]
  42. Xie, E.; Wang, W.; Yu, Z.; Anandkumar, A.; Alvarez, J.M.; Luo, P. SegFormer: Simple and Efficient Design for Semantic Segmentation with Transformers. Adv. Neural Inf. Process. Syst. 2021, 15, 12077–12090. [Google Scholar]
Figure 1. Comparison of Stable Diffusion-generated (ac) and real medical images (df) [21,22,23,24,25,26,27,28].
Figure 1. Comparison of Stable Diffusion-generated (ac) and real medical images (df) [21,22,23,24,25,26,27,28].
Bioengineering 12 01036 g001
Figure 2. Schematic of the stable diffusion configuration used in ADZUS model, consisting of 16 blocks, each containing transformer layers that produce a 4d self-attention tensor at various resolutions.
Figure 2. Schematic of the stable diffusion configuration used in ADZUS model, consisting of 16 blocks, each containing transformer layers that produce a 4d self-attention tensor at various resolutions.
Bioengineering 12 01036 g002
Figure 3. Flowchart of ADZUS model.
Figure 3. Flowchart of ADZUS model.
Bioengineering 12 01036 g003
Figure 4. Illustration of Attention Aggregation. A lower-resolution attention map is upsampled and duplicated to match the higher-resolution maps’ receptive field [32].
Figure 4. Illustration of Attention Aggregation. A lower-resolution attention map is upsampled and duplicated to match the higher-resolution maps’ receptive field [32].
Bioengineering 12 01036 g004
Figure 5. Sample of the qualitative evaluation of ADZUS model on ISIC 2016 and ISIC 2017 datasets: visual comparison of predicted boundaries (green) against true boundaries (red) with corresponding DSC score.
Figure 5. Sample of the qualitative evaluation of ADZUS model on ISIC 2016 and ISIC 2017 datasets: visual comparison of predicted boundaries (green) against true boundaries (red) with corresponding DSC score.
Bioengineering 12 01036 g005
Figure 6. Segmentation results on the chronic wound dataset [37]: original boundaries (red) and predicted boundaries (green) displayed on cropped images for enhanced visualization.
Figure 6. Segmentation results on the chronic wound dataset [37]: original boundaries (red) and predicted boundaries (green) displayed on cropped images for enhanced visualization.
Bioengineering 12 01036 g006
Figure 7. Sample comparison of wound segmentation results: ADZUS vs. benchmark models on chronic wound dataset [23], showing original boundaries (red) and predicted boundaries (green) on cropped images.
Figure 7. Sample comparison of wound segmentation results: ADZUS vs. benchmark models on chronic wound dataset [23], showing original boundaries (red) and predicted boundaries (green) on cropped images.
Bioengineering 12 01036 g007
Figure 8. Sample comparison of X-rays segmentation results: original boundaries (red) and predicted boundaries (green).
Figure 8. Sample comparison of X-rays segmentation results: original boundaries (red) and predicted boundaries (green).
Bioengineering 12 01036 g008
Figure 9. Sample qualitative results of ADZUS on WbcMSBench dataset [26] for multi-class white blood cell segmentation. The figure illustrates the input microscopic images (left), ADZUS-predicted segmentation masks (center), and ground truth annotations (right) for three white blood cell classes.
Figure 9. Sample qualitative results of ADZUS on WbcMSBench dataset [26] for multi-class white blood cell segmentation. The figure illustrates the input microscopic images (left), ADZUS-predicted segmentation masks (center), and ground truth annotations (right) for three white blood cell classes.
Bioengineering 12 01036 g009
Table 1. Performance comparison of ADZUS and state-of-the-art segmentation models on ISIC challenge datasets (2016–2017).
Table 1. Performance comparison of ADZUS and state-of-the-art segmentation models on ISIC challenge datasets (2016–2017).
YearModelDSC (%)IoU (%)
ISIC 2016 [21]U-Net implementation (Top evaluation) [21]91.084.3
ADZUS92.986.8
ISIC 2017 [22]Multi-task Deep Learning Model [33]N/A72.4
RECOD Titans Ensemble Approach [34]N/A79.3
Fully Convolutional Network (FCN-AlexNet) [35]81.9N/A
EfficientNet-Based Model [36]88.080.7
ADZUS88.780.0
Table 2. Performance comparison of ADZUS and state-of-the-art models in wound segmentation based on DSC and IoU metrics.
Table 2. Performance comparison of ADZUS and state-of-the-art models in wound segmentation based on DSC and IoU metrics.
ModelDSC (%)IoU (%)
DDRNet [39]73.1357.64
Swin-Unet [39]88.4679.30
SegFormer-b5 [42]91.0683.58
DeepLabV3Plus [40,41]92.0085.19
LinkNet-EffB1 + UNet-EffB2 [38,39]92.0785.51
FUSegNet [39]92.7086.40
ADZUS91.9886.68
Table 3. Performance comparison of ADZUS and state-of-the-art models in X-rays segmentation based on evaluation metrics.
Table 3. Performance comparison of ADZUS and state-of-the-art models in X-rays segmentation based on evaluation metrics.
Model [26]IoU (%)Precision (%)Recall (%)
Resnet18 62.774.182.4
Resnet5062.073.881.0
Efficientnet-b063.375.381.5
Mobilenet_v263.173.982.7
Densenet12164.77682.6
ADZUS66.377.983.1
Table 4. Performance comparison of ADZUS and state-of-the-art models in white blood cell segmentation based on evaluation metrics.
Table 4. Performance comparison of ADZUS and state-of-the-art models in white blood cell segmentation based on evaluation metrics.
Model [26]IoU (%)Precision (%)Recall (%)
Resnet18 9396.196.6
Resnet5093.196.296.6
Efficientnet-b093.796.596.8
Mobilenet_v292.695.996.3
Densenet12193.896.696.9
ADZUS93.396.596.7
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hamrani, A.; Godavarty, A. Self-Attention Diffusion Models for Zero-Shot Biomedical Image Segmentation: Unlocking New Frontiers in Medical Imaging. Bioengineering 2025, 12, 1036. https://doi.org/10.3390/bioengineering12101036

AMA Style

Hamrani A, Godavarty A. Self-Attention Diffusion Models for Zero-Shot Biomedical Image Segmentation: Unlocking New Frontiers in Medical Imaging. Bioengineering. 2025; 12(10):1036. https://doi.org/10.3390/bioengineering12101036

Chicago/Turabian Style

Hamrani, Abderrachid, and Anuradha Godavarty. 2025. "Self-Attention Diffusion Models for Zero-Shot Biomedical Image Segmentation: Unlocking New Frontiers in Medical Imaging" Bioengineering 12, no. 10: 1036. https://doi.org/10.3390/bioengineering12101036

APA Style

Hamrani, A., & Godavarty, A. (2025). Self-Attention Diffusion Models for Zero-Shot Biomedical Image Segmentation: Unlocking New Frontiers in Medical Imaging. Bioengineering, 12(10), 1036. https://doi.org/10.3390/bioengineering12101036

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop