Next Article in Journal
A Deep Convolutional Neural Network-Based Model for Aspect and Polarity Classification in Hausa Movie Reviews
Previous Article in Journal
Implementation of a Decision Support Mechanism on a Mobile Platform Using Clinical Evidence-Based Dynamic Insulin Dosage Adjustment for Artificial Intelligence-Enabled Diabetes Care (AIDCARE) System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Proceeding Paper

Resilience of UNet-Based Models Under Adversarial Conditions in Medical Image Segmentation †

1
School of Information Technology and Engineering, Kazakh-British Technical University, Almaty 050000, Kazakhstan
2
Department of Transportation System Engineering, Korea National University of Transportation, Uiwang-Si 27469, Republic of Korea
3
Department of Telecommunication, University of Ruse, 7004 Ruse, Bulgaria
*
Authors to whom correspondence should be addressed.
Presented at the International Conference on Electronics, Engineering Physics and Earth Science (EEPES 2025), Alexandroupolis, Greece, 18–20 June 2025.
Eng. Proc. 2025, 104(1), 3; https://doi.org/10.3390/engproc2025104003
Published: 21 August 2025

Abstract

Adversarial modifications of input data can degrade the stability of deep neural networks in medical image segmentation. This study evaluates the robustness of UNet and Att-UNet++ architectures using the NuInsSeg dataset with annotated nuclear regions from various tissue sources. Both models were trained and tested under eight perturbation types, including gradient-based, iterative, and stochastic methods, with identical parameter settings. In the absence of distortions, Att-UNet++ produced higher segmentation results with a Dice of 0.7160 and a mean IoU of 0.6190 compared to 0.6424 and 0.4732 for UNet. Under NI-FGSM and Gaussian noise, Att-UNet++ experienced a greater reduction in mean IoU, reaching 0.1215 and 0.0658, while UNet maintained 0.1968 and 0.2329. Loss landscape analysis showed smoother surfaces for Att-UNet++, yet revealed increased responsiveness to directional gradients. The findings suggest that improvements in segmentation accuracy through architectural modifications may be accompanied by increased vulnerability to input changes, highlighting the necessity of robustness evaluation in model development for medical image analysis.

1. Introduction

Medical image interpretation constitutes a central role in modern clinical workflows. It provides physicians with an efficient tool to diagnose and monitor the progression of various diseases [1]. Computational medical image processing includes steps such as image acquisition, enhancement, analysis, and segmentation [2]. With the introduction of deep learning (DL) approaches, these processes are becoming increasingly automated to extract complex features from different modalities [3]. Among these applications, medical image segmentation is extremely important to distinguish anatomical structures. However, the segmentation of medical images, particularly at the cellular and skin lesion level, remains challenging due to irregular structures and overlapping regions [4]. Another fundamental task is medical image classification, which aims to assign diagnostic labels to entire images and is widely used for disease detection [5]. In contrast to traditional image processing techniques, which rely on manual feature creation and explicit rule-based operations, deep learning (DL) methods derive hierarchical representations directly from the raw data through optimization [6]. This difference shifts the focus from manual feature extraction to data-driven model training, changing the approach to solving image analysis challenges. Convolutional neural networks (CNNs) have emerged as a foundational DL architecture in medical image analysis as it is capable of capturing patterns using convolutional operations [7]. Fully convolutional networks (FCNs), as well as models based on models that include U-Net and its variants, are often used to solve these problems due to their encoder and decoder design that preserves spatial resolution across the network [8,9]. Transfer learning [10] facilitates the adaptation of CNNs to novel medical imaging tasks by leveraging previously learned parameters from related domains, thereby reducing the reliance on large annotated datasets to a minimum [11]. Even slight changes in the input medical images could mislead the segmentation models, leading to confidently incorrect extraction of anatomical or pathological regions, such as missing tumor boundaries. Among the state-of-the-art (SOTA) approaches, a method using a cascaded autoencoder architecture with modified adaptive instance normalization stands out, which allows the modification of image feature statistics in order to create realistic images [12]. Another recent method, Mb-AdA [13], uses orthogonal image moments to remove information subtly without adding noise, thereby misleading classification and segmentation models while preserving the structural integrity of medical images. Furthermore, in another study of full and partial adversary attacks, such as FGSM and PGD, applied in different spatial domains, it was shown that even localized distortions might reduce the stability of the model by targeting salient areas [14]. Despite the overall influence of different attack strategies on DL-based medical image analysis, it remains understudied, especially in terms of their generalizability to different imaging.
This study establishes a systematic assessment paradigm for quantifying the resilience of UNet-based architectures by exposing them to various gradient, iterative, and stochastic perturbation algorithms under uniform experimental conditions, while utilizing loss landscape analysis as an auxiliary tool to determine model sensitivity.

2. Materials and Methods

The methodology incorporated the NuInsSeg dataset [15] for nuclear segmentation in histological specimens. The dataset comprises 665 image patches of 512 × 512 pixels, subsequently processed to 256 × 256 dimensions, containing around 30,700 annotated nuclei derived from 31 distinct human and murine organs. Data augmentation procedures expanded the training corpus to 1000 samples. Dataset partitioning followed an 80:10:10 ratio, yielding 800 training, 99 validation, and 66 test images.
This investigation evaluated two CNN architectures for nuclear segmentation on the NuInsSeg dataset UNet [16] and Att-UNet++ [17]. UNet implements a symmetric encoder–decoder structure with four stages comprising dual convolutional layers with ReLU activations and respective downsampling or upsampling operations. Att-UNet++ features a densely connected nested [18] architecture across five resolution scales, integrating attention gates at skip connections [19]. Both architectures underwent comparative analysis under adversarial changes.
The visual illustrations of the input sample, the applied perturbation and the resulting image after distortion are shown in Figure 1.
These distortions, when applied to the original image, create a subtly modified version designed to mislead the segmentation model. This research evaluated the resilience of segmentation models to eight types of attacks, including gradient, iterative, and noise exposures.
The fast gradient sign method (FGSM) [20] is a strategy for generating adversarial samples in which a distortion in the direction of the sign of the loss function gradient is added to the original image X distortion in the direction of the sign of the gradient:
X ˜ = X + ε s i g n ( X l ( X , y target ) ) ,
where X ˜ is an adversely perturbed input, X is the original image, and ε controls the perturbation magnitude, X l ( X , y ) [21].
Projected gradient descent (PGD) is a method [22] for generating adversarial examples based on the gradient lifting of the loss function with a projection of the result back into the allowable space.
Momentum iterative FGSM (MI-FGSM) is a gradient-based unfavorable perturbation method in which the perturbation vector is computed by recursively integrating the information over several iterations [23]. The algorithm differs from standard iterative approaches by including a momentum term that maintains the current average gradient directions:
g ( t + 1 ) = μ g ( t ) + x J ( x ( t ) , y ) x J ( x ( t ) , y ) 1 ,
where x(t) is the current perturbed image at iteration t, α is the update step size, μ is the momentum decay factor, J ( x , y ) is the loss function, and x j ( x , y ) is the gradient of the loss.
TI-FGSM (translation-invariant FGSM) is a method for generating bias-resistant attack examples [24]. Unlike classical methods, TI-FGSM considers not only the original image but also its shifts along the spatial axes, aggregating the gradients from these translations.
X ˜ = arg max z β ( x ) m , n w m n l ( τ m n ( z ) , c ) ,
where X ˜ is the final adversarial image, z is an intermediate perturbed image, x is the original clean input, τ m n is the translation operation shifting the image, w m n is the weight assigned to each translation, l is the loss function, and β ( x ) denotes the ball of permissible deviations with radius ϵ around x.
Diverse FGSM (DIFGSM) is an extension of FGSM that also improves the tolerance of unfavorable examples by applying random transformations to the input data before computing the gradients. DI-FGSM introduces stochastic transformations of the input data at each iteration of the gradient computation to generate more diverse external changes.
x k + 1 = Pr δ z k + η s i g n ( u l ( R ( z k ) , c ) ) ,
where xk+1 is the adversarial example at iteration k, η is the update step size, R is the random transformation applied to the input, and Pr δ is the projection onto the l ball of radius δ .
Scale-invariant and diverse input FGSM (SI-DI-FGSM) is an advanced attack strategy that increases the tolerance of changes by combining two mechanisms: scale-invariant gradient [25] computation and diverse input transformation. The scale-invariant component generates unfavorable gradients by averaging the loss gradients over multiple modified versions of the input image, which reduces sensitivity to the scale of the input.
Nesterov iterative FGSM (NI-FGSM) is a gradient-based adversarial strategy that complements the traditional iterative FGSM system by integrating Nesterov’s momentum mechanism to improve transferability by computing gradients at look-ahead point updates [26].
Gaussian noise (GN) perturbation represents a non-adversarial robustness assessment methodology characterized by the superimposition of stochastic additive disturbances sampled from a normal distribution with zero mean and predetermined variance onto the input signal.
x ˜ = x + ,     N ( 0 , σ 2 ) ,
where x denotes the clean input, x ˜ represents the noise-perturbed image, is the additive Gaussian noise, N ( 0 , σ 2 ) defines a normal distribution with mean 0 and variance σ 2 , and σ controls the magnitude of the noise.
All adversarial attacks were conducted at the same settings to ensure a valid comparison. The epsilon perturbation value was fixed at 0.03, the alpha step size was fixed at 0.005, and the number of iterations was fixed at ten for all iterative methods. For the moment-based attacks, a fading factor of 1.0 or 1.5 was used. The DI-FGSM attack used stochastic transformations with a probability of 0.7. The GN method introduced additive variations with a standard deviation of 0.1.
In this research, the behavior of segmentation models was assessed through quantitative analysis using established performance indicators. These include the Dice coefficient, mean intersection over union (mean IoU), accuracy, and Dice loss [27].
The Dice coefficient reflects the degree of spatial correspondence between the predicted (P) segmentation domain and the real annotation (G), which serves as a measure of topological correspondence.
Dice   coefficient = 2 P G P + G
Mean IoU estimates the average overlap across all classes, offering a balanced metric.
M e a n   I O U = 1 N i = 1 N P i G i P i G i
Accuracy defines the fraction of identified pixels with respect to the full set of image pixels [28]. Dice loss [29] is determined by the Dice coefficient and serves the purpose of optimization by penalizing for poor agreement between predicted and actual segmentations during model training.
The experimental setup used was an NVIDIA A100 GPU. The software environment included Python 3.10 and TensorFlow 2.13.0.

3. Results

In this study, a dataset comprising 532 annotated medical images was employed. Two segmentation models, UNet and Attention-UNet++, were trained using a mini-batch size of 8 for 60 epochs. A comparative assessment of the segmentation efficiency of the models under consideration is summarized in Table 1.
As summarized in Table 1, the performance of the Att-Unet++ model exceeds that of UNet in all metrics considered. The Dice metric for Att-Unet++ is 0.7160 compared to 0.6424 for UNet, indicating an increase of about 11.46%. Average IOU improves from 0.4732 to 0.6190, representing a relative increase of 30.77 percent. Accuracy improves from 0.9009 to 0.9292, corresponding to an improvement of 3.14 percent. The quantitative evaluation of model efficiency was conducted using the Dice similarity metric calculated on a subset of scores with retention, and detailed comparative results are presented in Table 2.
On baseline, undisturbed test images, Att-Unet++ shows a 6.34% reduction in Dice performance compared to UNet, indicating a slightly less optimal baseline generalization. Under gradient attacks such as FGSM, PGD. and TI-FGSM, Att-Unet++ demonstrates a further reduction from 13% to 20%, confirming a higher sensitivity to directional changes. Notably, Att-Unet++ achieves marginal improvements with MI-FGSM and DI-FGSM, with respective relative increases of 1.00% and 7.08%, indicating limited robustness in the presence of pulse and input diversity. However, the performance with NI-FGSM and GN deteriorates considerably, with Dice scores decreasing by 42.66% and 55.45%, respectively. The comparative performance of segmentation models based on mean IOU under different adverse distortions is systematically presented in Table 3.
In the baseline analysis, Att-Unet++ exhibits a slightly lower average IOU compared to UNet, with a relative decrease of 4.31 percent. When exposed to hostile perturbations, the degree of performance degradation differs by attack type. Moderate degradation is observed for FGSM and TI-FGSM attacks, resulting in 15.21 and 15.62 percent performance degradation, respectively. A more severe degradation is observed for iterative attacks such as PGD and NI-FGSM, where the average IOU drops by 24.58 and 38.27 percent, respectively. The introduction of GN leads to the most pronounced decrease, with the average IOU for Att-Unet++ dropping to 0.0658, which corresponds to a 71.73 percent decrease compared to UNet. Interestingly, under DI-FGSM conditions, Att-Unet++ shows a slight improvement over UNet by 6.82%, which may reflect the advantage of the attention mechanism in processing spatially transformed input data. The loss values obtained under adversarial and noise-based test-time conditions for both segmentation models are shown in Table 4.
Under baseline conditions, the loss value for Att-Unet++ is 37.6 percent lower compared to UNet. Across all attack types, Att-Unet++ consistently demonstrates lower loss values. The most substantial reductions are observed under PGD and NI-FGSM attacks, with decreases of 35.76 percent and 33.99 percent, respectively. Moderate improvements are noted under MI-FGSM and TI-FGSM, with reductions of 41.70 percent and 41.09 percent. The smallest difference is found under Gaussian noise, where the loss decreases by 24.21 percent. These findings highlight the enhanced robustness of Att-Unet++ over UNet, particularly under gradient-based adversarial conditions.
Following the quantitative results presented in the previous tables, Figure 2 illustrates the trends in model validation performance during training (Figure 2).
The validation dynamics presented in Figure 3 reveal that Att-Unet++ consistently outperforms UNet in both Dice and average IOU across all training epochs. Att-Unet++ reaches a peak Dice around epoch 50 and maintains stable performance thereafter, while UNet reaches a lower plateau earlier. Similarly, the average IOU for Att-Unet++ increases dramatically and stabilizes around epoch 30, outperforming UNet, which exhibits slower convergence.
A comparative graphical representation of the model behavior under attack and noise conditions is presented in Figure 3 and Figure 4. Figure 3 represents the Dice coefficient performance responses, while Figure 4 depicts the corresponding loss values.
Figure 3 illustrates the Dice coefficient performance of UNet and Att-Unet++ models under a range of adversarial and noise-based attack conditions on the test set. In the clean setting, UNet outperformed Att-Unet++ by 6.8%. Both models exhibited substantial performance degradation when subjected to gradient-based deviations, with PGD and NI-FGSM attacks causing the most significant drops. UNet’s Dice coefficient decreased by 60.8% and 52.5% under PGD and NI-FGSM, respectively, whereas Att-Unet++ declined by 66.4% and 70.9% under the same conditions. Exposure to GN resulted in a 59.5% reduction for UNet and 80.8% for Att-Unet++. Across all distortions, UNet consistently maintained higher Dice scores.
Figure 4 presents the loss values of UNet and Att-Unet++ under various attack conditions. UNet demonstrates higher sensitivity, with loss increasing by 991.1% under PGD and 1079.6% under NI-FGSM. In contrast, Att-Unet++ shows comparatively lower increases of 1024.2% and 1147.2% under the same attacks. Across all scenarios, Att-Unet++ consistently maintains lower loss.
Under PGD, the Dice coefficient of UNet decreases by 60.77 percent and that of Att-Unet++ by 66.43 percent. NI-FGSM results in a Dice reduction of 52.45 percent for UNet and 70.91 percent for Att-Unet++. Corresponding loss increases are observed: for PGD, the rise is 993.86 percent in UNet and 1024.89 percent in Att-Unet++, while for NI-FGSM, the increases reach 1079.98 percent and 1148.24 percent, respectively.
The local behavior of the loss function near adversarial noise was investigated using 3D surface visualization. Figure 5 depicts the loss landscapes for Att-UNet++ and UNet under NIFGSM distortions along the main adversarial and orthogonal directions.
Figure 5 illustrates the loss landscapes of the segmentation models Att-UNet++ (a) and UNet (b) under deviations generated by the NIFGSM method. The x-axis corresponds to the perturbation along the adversarial direction ε1, while the y-axis denotes the orthogonal direction ε2. The loss values are encoded using a perceptually uniform colormap. The surface in subplot (a) reveals a smooth and monotonic increase in loss for Att-UNet++, with the maximum value reaching approximately 2.25 at ε1 = 0.02 and ε2 = 0.02. In contrast, the surface for UNet in subplot (b) is characterized by irregular elevations and sharper gradients. The peak loss reaches around 2.0 but occurs across a more fluctuating region.
The Att-UNet++ loss surface exhibits higher regularity and reduced gradient dispersion in ε-space, implying increased robustness to local perturbation biases. The UNet topology has sharper discontinuities and increased curvature, which potentially indicates greater susceptibility to hostile gradient fluctuations. This divergence in loss landscape geometry suggests that Att-UNet++ maintains a more stable optimization trajectory under adverse conditions, whereas irregularities in the UNet surface may contribute to gradient bias. Figure 6 displays the effect of the different attacks.
Among PGD attacks, MI-FGSM, MI-FGSM, TI-FGSM, SI-DI-FGSM, and NI-FGSM cause substantial degradations, leading to widespread segmentation errors, as evidenced by the prominent red and blue areas. DI-FGSM also exhibits notable sensitivity, especially when over-segmenting non-nuclear regions. In contrast, FGSM and GN have relatively minor effects on the model. The segmentation discrepancies resulting from different perturbations are examined in subfigure (A), where the pixel-level difference maps generated by the UNet model give an indication of the prediction shifts before and after the exposure of the attack. In comparison with (A), (B) illustrates identical changes applied to Att-UNet++, and the UNet architecture demonstrates increased vulnerability to adversarial noise. Difference maps reveal more extensive chromatic variations, indicating reduced segmentation stability under perturbation conditions. Iterative and momentum-based methods produce significant alterations to predicted segmentation masks. Under FGSM perturbation, UNet exhibits considerable deviation from baseline predictions, while Att-UNet++ maintains greater consistency.

4. Conclusions and Future Work

This study investigated the vulnerability of CNN architectures to unfavorable factors used for nuclei segmentation in histopathological images. The analysis was performed on the NuInsSeg dataset containing hematoxylin and eosin-stained nuclei from different tissues. Two models, UNet and Att-UNet++, were evaluated under eight types of distortions to assess robustness. Although Att-UNet++ showed better segmentation performance on undisturbed data, with a Dice coefficient of 0.7160 and IoU of 0.6190, it showed greater performance degradation under certain adversary attacks. Specifically, under NI-FGSM and Gaussian noise, the model’s average IoU decreased to 0.1215 and 0.0658, respectively. In comparison, UNet reached 0.1968 and 0.2329 under the same conditions. The loss landscape analysis also showed that Att-UNet++ retains a smoother topological profile but remains more sensitive to gradient perturbations. These results suggest that while architectural complexity can improve baseline performance, it can also create different vulnerability profiles. Nevertheless, the current scope of the study is limited to binary segmentation performed on a single bivariate dataset. Validation on different datasets was not performed, which limits the generalizability of the results obtained. Furthermore, the analysis was limited to two CNN architectures, excluding SOTA lightweight or transformer-based models. Prospective research will include comparative evaluations of different datasets and modalities. Furthermore, subsequent studies may investigate innovative defense strategies [30], including attention-driven noise regularization.

Author Contributions

Conceptualization, D.K. J.W.K., T.I. and A.B.; methodology, A.M., D.K., J.W.K. and A.B.; software, D.K. and A.M.; validation, D.K., A.M. and J.W.K., A.B.; formal analysis, A.M. and D.K.; investigation, J.W.K., A.M., D.K. and A.B.; resources, J.W.K.; data curation, D.K. and A.M.; writing—original draft preparation, D.K. and T.I.; writing—review and editing, D.K. and A.M.; visualization, D.K. and A.M.; supervision, T.I.; project administration, T.I., J.W.K. and A.B. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Korea Institute for Advancement of Technology (KIAT) grant funded by the Korea Government (MOTIE) (RS-2022-KI002562, HRD Program for Industrial Innovation) and this research was supported by the Ministry of Trade, Industry and Energy (MOTIE) and the Korea Institute for Advancement of Technology through the “Support for Middle Market Enterprises and Regional innovation Alliances (R&D-2025 year)” program.

Institutional Review Board Statement

Ethical approval was waived due to the use of publicly available datasets.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Li, X.; Zhang, L.; Yang, J.; Teng, F. Role of artificial intelligence in medical image analysis: A review of current trends and future directions. J. Med. Biol. Eng. 2024, 44, 231–243. [Google Scholar] [CrossRef]
  2. Xu, Y.; Quan, R.; Xu, W.; Huang, Y.; Chen, X.; Liu, F. Advances in Medical Image Segmentation: A Comprehensive Review of Traditional, Deep Learning and Hybrid Approaches. Bioengineering 2024, 11, 1034. [Google Scholar] [CrossRef]
  3. Archana, R.; Jeevaraj, P.E. Deep learning models for digital image processing: A review. Artif. Intell. Rev. 2024, 57, 11. [Google Scholar] [CrossRef]
  4. Gabdullin, M.T.; Mukasheva, A.; Koishiyeva, D.; Umarov, T.; Bissembayev, A.; Kim, K.S.; Kang, J.W. Automatic cancer nuclei segmentation on histological images: Comparison study of deep learning methods. Biotechnol. Bioprocess Eng. 2024, 29, 1034–1047. [Google Scholar] [CrossRef]
  5. Jiang, H.; Diao, Z.; Shi, T.; Zhou, Y.; Wang, F.; Hu, W.; Yao, Y.D. A review of deep learning-based multiple-lesion recognition from medical images: Classification, detection and segmentation. Comput. Biol. Med. 2023, 157, 106726. [Google Scholar] [CrossRef]
  6. Trigka, M.; Dritsas, E. A Comprehensive Survey of Deep Learning Approaches in Image Processing. Sensors 2025, 25, 531. [Google Scholar] [CrossRef] [PubMed]
  7. Kshatri, S.S.; Singh, D. Convolutional neural network in medical image analysis: A review. Arch. Comput. Methods Eng. 2023, 30, 2793–2810. [Google Scholar] [CrossRef]
  8. Azad, R.; Aghdam, E.K.; Rauland, A.; Jia, Y.; Avval, A.H.; Bozorgpour, A.; Merhof, D. Medical image segmentation review: The success of u-net. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 10076–10095. [Google Scholar] [CrossRef] [PubMed]
  9. Mukasheva, A.; Koishiyeva, D.; Sergazin, G.; Sydybayeva, M.; Mukhammejanova, D.; Seidazimov, S. Modification of U-Net with Pre-Trained ResNet-50 and Atrous Block for Polyp Segmentation: Model TASPP-UNet. Eng. Proc. 2024, 70, 16. [Google Scholar] [CrossRef]
  10. Zoetmulder, R.; Gavves, E.; Caan, M.; Marquering, H. Domain-and task-specific transfer learning for medical segmentation tasks. Comput. Methods Programs Biomed. 2022, 214, 106539. [Google Scholar] [CrossRef]
  11. Ma, X.; Niu, Y.; Gu, L.; Wang, Y.; Zhao, Y.; Bailey, J.; Lu, F. Understanding adversarial attacks on deep learning based medical image analysis systems. Pattern Recognit. 2021, 110, 107332. [Google Scholar] [CrossRef]
  12. Lee, W.; Ju, M.; Sim, Y.; Jung, Y.K.; Kim, T.H.; Kim, Y. Adversarial Attacks on Medical Segmentation Model via Transformation of Feature Statistics. Appl. Sci. 2024, 14, 2576. [Google Scholar] [CrossRef]
  13. Maliamanis, T.V.; Apostolidis, K.D.; Papakostas, G.A. How resilient are deep learning models in medical image analysis? The case of the moment-based adversarial attack (Mb-AdA). Biomedicines 2022, 10, 2545. [Google Scholar] [CrossRef]
  14. Pal, S.; Rahman, S.; Beheshti, M.; Habib, A.; Jadidi, Z.; Karmakar, C. The Impact of Simultaneous Adversarial Attacks on Robustness of Medical Image Analysis. IEEE Access 2024, 12, 66478–66494. [Google Scholar] [CrossRef]
  15. Mahbod, A.; Polak, C.; Feldmann, K.; Khan, R.; Gelles, K.; Dorffner, G.; Ellinger, I. Nuinsseg: A fully annotated dataset for nuclei instance segmentation in H&E-stained histological images. Sci. Data 2024, 11, 295. [Google Scholar] [CrossRef]
  16. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, Proceedings of the MICCAI 2015, Munich, Germany, 5–9 October 2015; Navab, N., Hornegger, J., Wells, W., Frangi, A., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Berlin/Heidelberg, Germany, 2015; Volume 9351. [Google Scholar]
  17. Li, Z.; Zhang, H.; Li, Z.; Ren, Z. Residual-Attention UNet++: A Nested Residual-Attention U-Net for Medical Image Segmentation. Appl. Sci. 2022, 12, 7149. [Google Scholar] [CrossRef]
  18. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Proceedings of the 4th International Workshop, DLMIA 2018, and 8th International Workshop, ML-CDS 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, 20 September 2018. [Google Scholar] [CrossRef]
  19. Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Rueckert, D. Attention u-net: Learning where to look for the pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar] [CrossRef]
  20. Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.; Fergus, R. Intriguing properties of neural networks. arXiv 2013, arXiv:1312.6199. [Google Scholar] [CrossRef]
  21. Kurakin, A.; Goodfellow, I.; Bengio, S. Adversarial machine learning at scale. arXiv 2016, arXiv:1611.01236. [Google Scholar] [CrossRef]
  22. Madry, A.; Makelov, A.; Schmidt, L.; Tsipras, D.; Vladu, A. Towards deep learning models resistant to adversarial attacks. arXiv 2017, arXiv:1706.06083. [Google Scholar] [CrossRef]
  23. Dong, Y.; Liao, F.; Pang, T.; Su, H.; Zhu, J.; Hu, X.; Li, J. Boosting adversarial attacks with momentum. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 9185–9193. [Google Scholar] [CrossRef]
  24. Xie, C.; Zhang, Z.; Zhou, Y.; Bai, S.; Wang, J.; Ren, Z.; Yuille, A.L. Improving transferability of adversarial examples with input diversity. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 2725–2734. [Google Scholar] [CrossRef]
  25. Lin, Z.; Peng, A.; Wei, R.; Yu, W.; Zeng, H. An enhanced transferable adversarial attack of scale-invariant methods. In Proceedings of the 2022 IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 16–19 October 2022; pp. 3788–3792. [Google Scholar] [CrossRef]
  26. Lian, Y.; Tang, Y.; Zhou, S. Research on three-step accelerated gradient algorithm in deep learning. Stat. Theory Relat. Fields 2022, 6, 40–57. [Google Scholar] [CrossRef]
  27. Koishiyeva, D.; Bissembayev, A.; Iliev, T.; Kang, J.W.; Mukasheva, A. Classification of Skin Lesions using PyQt5 and Deep Learning Methods. In Proceedings of the 2024 5th International Conference on Communications, Information, Electronic and Energy Systems (CIEES), Veliko Tarnovo, Bulgaria, 20–22 November 2024; pp. 1–7. [Google Scholar] [CrossRef]
  28. Ming, Q.; Xiao, X. Towards accurate medical image segmentation with gradient-optimized dice loss. IEEE Signal Process. Lett. 2023, 31, 191–195. [Google Scholar] [CrossRef]
  29. Dietrich, N.; Gong, B.; Patlas, M.N. Adversarial artificial intelligence in radiology: Attacks, defenses, and future considerations. Diagn. Interv. Imaging, 2025; in press. [Google Scholar] [CrossRef]
  30. Tolkynbekova, A.; Koishiyeva, D.; Bissembayev, A.; Mukhammejanova, D.; Mukasheva, A.; Kang, J.W. Comparative Analysis of the Predictive Risk Assessment Modeling Technique Using Artificial Intelligence. J. Electr. Eng. Technol. 2025; in press. [Google Scholar] [CrossRef]
Figure 1. Visual comparison of histologic images after attack: (a) original image; (b) perturbation after attack, where colors represent the intensity and spatial distribution of adversarial noise, with yellow indicating high and blue indicating low perturbation; (c) model output after attack.
Figure 1. Visual comparison of histologic images after attack: (a) original image; (b) perturbation after attack, where colors represent the intensity and spatial distribution of adversarial noise, with yellow indicating high and blue indicating low perturbation; (c) model output after attack.
Engproc 104 00003 g001
Figure 2. Validation performance curves of segmentation models: (a) dice coefficient over epochs; (b) mean IoU over epochs.
Figure 2. Validation performance curves of segmentation models: (a) dice coefficient over epochs; (b) mean IoU over epochs.
Engproc 104 00003 g002
Figure 3. Dice coefficient on the test set under different attack scenarios.
Figure 3. Dice coefficient on the test set under different attack scenarios.
Engproc 104 00003 g003
Figure 4. Loss values on the test set under different attack scenarios.
Figure 4. Loss values on the test set under different attack scenarios.
Engproc 104 00003 g004
Figure 5. The 3D loss surface for NIFGSM distortions along the principal direction (ε1) and orthogonal direction (ε2) for Att-UNet++ (a) and UNet (b).
Figure 5. The 3D loss surface for NIFGSM distortions along the principal direction (ε1) and orthogonal direction (ε2) for Att-UNet++ (a) and UNet (b).
Engproc 104 00003 g005
Figure 6. Visualization of segmentation deviations under different aggressive distortions using the UNet (A) and Att-UNet++ (B): difference maps highlight pixel-by-pixel differences between pre- and post-attack predictions, where red tones indicate increased activation after a perturbation, blue tones indicate decreased activation, and neutral tones (white) represent minimal or no change.
Figure 6. Visualization of segmentation deviations under different aggressive distortions using the UNet (A) and Att-UNet++ (B): difference maps highlight pixel-by-pixel differences between pre- and post-attack predictions, where red tones indicate increased activation after a perturbation, blue tones indicate decreased activation, and neutral tones (white) represent minimal or no change.
Engproc 104 00003 g006
Table 1. Comparative analysis of segmentation performance based on validation metrics.
Table 1. Comparative analysis of segmentation performance based on validation metrics.
Metrics UNetAtt-Unet++
Dice0.64240.7160
Mean IOU0.47320.6190
Accuracy0.90090.9292
Table 2. Dice coefficient of segmentation models under adversarial and noise-based attacks on the test set.
Table 2. Dice coefficient of segmentation models under adversarial and noise-based attacks on the test set.
AttackUNetAtt-Unet++
Original image0.81180.7603
FGSM0.66070.5717
PGD0.31850.2553
MI-FGSM0.31850.3217
TI-FGSM0.33500.2882
DI-FGSM0.28840.3088
SI-DI-FGSM0.36670.3208
NI-FGSM0.38590.2212
GN0.32840.1463
Table 3. Mean IOU of segmentation models under adversarial and noise-based attacks on the test set.
Table 3. Mean IOU of segmentation models under adversarial and noise-based attacks on the test set.
AttackUNetAtt-Unet++
Original0.68450.6550
FGSM0.48990.4153
PGD0.18930.1428
MI-FGSM0.24410.1896
TI-FGSM0.20030.1690
DI-FGSM0.16740.1788
SI-DI-FGSM0.22460.1899
NI-FGSM0.19680.1215
GN0.23290.0658
Table 4. Comparative loss values for segmentation under synthetic distortions.
Table 4. Comparative loss values for segmentation under synthetic distortions.
AttackUNetAtt-Unet++
Original0.34080.2127
FGSM1.02010.6375
PGD3.72752.3940
MI-FGSM2.51361.4659
TI-FGSM3.41772.0134
DI-FGSM3.43651.9829
SI-DI-FGSM3.17381.9152
NI-FGSM4.02432.6552
GN2.01011.5234
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Koishiyeva, D.; Kang, J.W.; Iliev, T.; Bissembayev, A.; Mukasheva, A. Resilience of UNet-Based Models Under Adversarial Conditions in Medical Image Segmentation. Eng. Proc. 2025, 104, 3. https://doi.org/10.3390/engproc2025104003

AMA Style

Koishiyeva D, Kang JW, Iliev T, Bissembayev A, Mukasheva A. Resilience of UNet-Based Models Under Adversarial Conditions in Medical Image Segmentation. Engineering Proceedings. 2025; 104(1):3. https://doi.org/10.3390/engproc2025104003

Chicago/Turabian Style

Koishiyeva, Dina, Jeong Won Kang, Teodor Iliev, Alibek Bissembayev, and Assel Mukasheva. 2025. "Resilience of UNet-Based Models Under Adversarial Conditions in Medical Image Segmentation" Engineering Proceedings 104, no. 1: 3. https://doi.org/10.3390/engproc2025104003

APA Style

Koishiyeva, D., Kang, J. W., Iliev, T., Bissembayev, A., & Mukasheva, A. (2025). Resilience of UNet-Based Models Under Adversarial Conditions in Medical Image Segmentation. Engineering Proceedings, 104(1), 3. https://doi.org/10.3390/engproc2025104003

Article Metrics

Back to TopTop