Next Article in Journal
Association of SLC22A1, SLC22A2, SLC47A1, and SLC47A2 Polymorphisms with Metformin Efficacy in Type 2 Diabetic Patients
Next Article in Special Issue
Artificial Intelligence Driven Biomedical Image Classification for Robust Rheumatoid Arthritis Classification
Previous Article in Journal
Effect of MNCQQ Cells on Migration of Human Dermal Fibroblast in Diabetic Condition
Previous Article in Special Issue
Deep Learning for Bone Mineral Density and T-Score Prediction from Chest X-rays: A Multicenter Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

How Resilient Are Deep Learning Models in Medical Image Analysis? The Case of the Moment-Based Adversarial Attack (Mb-AdA)

by
Theodore V. Maliamanis
,
Kyriakos D. Apostolidis
and
George A. Papakostas
*
MLV Research Group, Department of Computer Science, International Hellenic University, 65404 Kavala, Greece
*
Author to whom correspondence should be addressed.
Biomedicines 2022, 10(10), 2545; https://doi.org/10.3390/biomedicines10102545
Submission received: 20 September 2022 / Revised: 8 October 2022 / Accepted: 10 October 2022 / Published: 12 October 2022
(This article belongs to the Special Issue Artificial Intelligence in Biological and Biomedical Imaging 2.0)

Abstract

:
In the past years, deep neural networks (DNNs) have become popular in many disciplines such as computer vision (CV). One of the most important challenges in the CV area is Medical Image Analysis (MIA). However, adversarial attacks (AdAs) have proven to be an important threat to vision systems by significantly reducing the performance of the models. This paper proposes a new black-box adversarial attack, which is based οn orthogonal image moments named Mb-AdA. Additionally, a corresponding defensive method of adversarial training using Mb-AdA adversarial examples is also investigated, with encouraging results. The proposed attack was applied in classification and segmentation tasks with six state-of-the-art Deep Learning (DL) models in X-ray, histopathology and nuclei cell images. The main advantage of Mb-AdA is that it does not destroy the structure of images like other attacks, as instead of adding noise it removes specific image information, which is critical for medical models’ decisions. The proposed attack is more effective than compared ones and achieved degradation up to 65% and 18% in terms of accuracy and IoU for classification and segmentation tasks, respectively, by also presenting relatively high SSIM. At the same time, it was proved that Mb-AdA adversarial examples can enhance the robustness of the model.

1. Introduction

Computer Vision has evolved dramatically with the introduction of Deep Learning (DL) models [1] as the accuracy and the effectiveness of deep convolutional neural networks [2] became impressive. DL has been applied with great success to nearly all CV tasks e.g., classification [2], semantic segmentation [3], object detection [4], pose estimation [5], quality assessment [6] and depth prediction [7]. The application of CV in Medical Image Analysis [8] is mostly referred to segmentation and classification problems. The application of CV to MIA compared to other CV applications, seems to have an advantage due to the fact that medical images are mostly ideally captured, lacking occlusion and miss-orientation problems and minimizing distortion, deformation and mis-illumination problems that other images generally present. Lately, the pandemic of COVID-19 showed that doctors and personnel can never be enough and that everything that can help them is valuable. This underlines the need for the implementation of computer vision-based automated medical image analysis in a safe way.
When Adversarial Attacks appeared on DL models [9], CV became seemingly unreliable. A great research field of Adversarial Computer Vision (AdCV) was born in order to create greater security in the models [10,11,12]. Driven by the success of adversarial attacks on natural images, researchers implement adversarial examples on medical images to investigate the MIA models’ robustness [13]. The majority of studies have been done on classification and segmentation tasks. Dermoscopy images were tested in [14,15] for classification tasks. MRIs have been tested on classification tasks [16,17] and segmentation tasks [18,19]. For X-ray images, researchers in [20,21] proved that general purpose attacks are able to importantly decrease the models’ accuracy. Moreover, attacks were implemented in fundoscopy images in [22,23] presenting high attack accuracy. The aforementioned studies used several attacks such as Fast Gradient Sign Method (FGSM), Projected Gradient Descent (PGD), DeepFool, Jacobian-based Saliency Map Attack (JSMA), Universal Adversarial Perturbations (UAPs), Basic Iterative Method (BIM) and Carlini & Wagner (C&W), which significantly reduced the MIA models’ accuracy.
Additionally, several custom medical imaging attacks were developed. In [24] created an attack for Ultrasound (US) images for classification tasks. This attack was tested on the InceptionResNetV2 model. The authors in [25] created an attack for segmentation on fundoscopy and dermoscopy images, using the U-Net model. In addition, [26] developed an attack for fundoscopy images, which can be implemented on segmentation and classification tasks. Kugler et al. [27] tried to lead five models into misclassification on dermoscopy images, (ResNet, InceptionV3, InceptionResNetV2, MobileNet, Xception). Vatian et al. [28] created an attack, which is based on the “natural” noise of medical imaging systems. Experiments were carried out on CT scans and brain MR images. Chen et al. [29] generate adversarial examples for medical image segmentation, which are tested on CT scans. Another interesting attack was created by Tian et al. [30] and it is based on a biased field phenomenon.
DL is more efficient than traditional techniques because it provides powerful models that can achieve very high accuracy in difficult medical problems just by retraining some pretrained models. That is why DL models should be resilient to adversarial attacks, as it is a critical research task that will help CV to become reliable. Trustworthiness together with the explainability of DL models, which means the ability to explain how and what features are modeled inside the DL model to extract a decision, are two qualities of the utmost importance to achieve the goal of integrating this technology into medical practice. The contribution of this work is to:
(1)
Highlight the vulnerability of medical images to attacks and try to explain why some tasks are more vulnerable to attacks than others.
(2)
Propose a generalized attack that significantly affects the operation of DL models.
(3)
Investigate adversarial training using the proposed attack method as a universal defense method.
The proposed attack has the following features:
(1)
Can be characterized as a black box attack since there is no need for any knowledge of the MIA task, the DL model structure, or the dataset used for its training.
(2)
Has several degrees of freedom and can be adapted to any DL model and any image resolution.
(3)
Its effect is adjustable.
(4)
The way it affects DL models is fully explainable, adding clues to how we can get closer to the DL interpretability or explainability goal.

2. Materials and Methods

Image moments are image projections on a basis produced by monomials or polynomials [31]. The first set of moments introduced were “geometrical moments” that can be produced using the following equation:
M p q = + + x p y q f ( x , y )   d x d y
where f(x,y) is the density distribution function of a single image channel and p, q are integer numbers. For color images, the moment set consists of three sets one for each image channel produced using the same equation.
The uniqueness theorem presented by Hu [32] proves that a piecewise continuous and fractured function f ( x , y ) , that doesn’t have infinite non-zero values, can be uniquely determined by a sequence of moments { M p q } and conversely. Digital images are represented by a density distribution function f ( x , y ) , which obviously satisfies Hu theorem’s conditions, except continuation. Such a function is discrete and therefore not continuous, that is the reason why we need discrete polynomials to produce the basis. In the above means the Equation (1) for discrete polynomials can be written as:
M p q = N F ( p , q ) i = 1 N j = 1 M K e r n e l p q ( x i , y j ) f ( x i , y j )
where K e r n e l p q is the product of specific polynomials [33] and is the appropriate normalization factor for the polynomial family selected.
It can easily be concluded that the maximum order of the moments is equal to the maximum of the values of the image dimensions N, M, so the produced moment set has the same dimension as the image ( N × M ).
Any discrete polynomial set that can produce a basis can be used to produce moments, but orthogonal polynomial families are mostly used. The orthogonality property simply means that the moments production procedure can be easily inverted using the same products of polynomials used for moment production.
When we collect the sequence of moments { M p q } , we can use them to reconstruct the image in a simple way, due to the orthogonality property of the polynomials used in K e r n e l p q (2), using the equation:
f ^ ( x , y ) = p = 0 K 1 q = 0 L 1 K e r n e l p q ( x , y ) M p q
Noting that in (3) according to the uniqueness theorem, if the limits K and L are equal to image dimensions N and M, respectively, the estimated density distribution function f ^ ( x , y ) is theoretically identical to the initial function f ( x , y ) . Moment sets can fully describe and reconstruct the image, which explains why they were broadly used as image descriptors.
The explanation of moments in physics and mathematics shows that lower order geometrical moments carry the greatest amount of information and higher order moments carry details. In image moments, lower order geometric moments represent some well-known image properties such as the total mass, the center of mass, the orientation and many others. In addition, moment functions which are invariant to scaling, translation (positioning), rotation and reflection, are called “moment invariants” [34], and they are generated using lower order moments. Lower order moment features describe the image, while on the other hand higher order moments contain information that is relatively useless for the task of image classification [35]. The way that the image information is distributed among image moments depends on the polynomial family that is used for moment production and on the family of expanding image moments. Specifically in image tasks, for the reasons mentioned earlier, discrete orthogonal image moments [31] are used. Most representative families are the Tchebichef, Krawtchouk, dual Hahn and Racah [32]. It is worth mentioning that moment set’s production is a computationally “heavy” task that can have large approximation errors, especially when many orders are needed. Numerous computational strategies [33] have been proposed, to minimize approximation errors and ease computational load.
The exclusion of some moments M p q of specific orders used for reconstruction by excluding some values of p or q from the sums of the reconstruction Equation (3), or practically substituting the specific moment values to be excluded with zeros, concludes in the production of an approximation of the estimated density distribution function f ^ ( x , y ) . Obviously, a good approximation is obtained using the optimal moments, those that carry most of the image’s information.
The proposed adversarial attack that we call Moment based- Adversarial Attack (Mb-AdA) produces the adversarial examples as follows:
  • The image is transformed in a moment set { M p q } using a discrete orthogonal image moments family.
  • Excluding some moments of specific orders, the image is reconstructed producing an approximation.
  • The product can be used as an adversarial example to attack every DL model.
Τhe motivation behind the proposed attack is based on the fact that by eliminating some moment orders (possibly non-robust features) participating in the reconstructed image (attacked) the model will fail to detect this information and consequently will not compute the desired features within its deep structure.
Mb-AdA can be applied to any image form or resolution. DL models have a standard input resolution and that is not obvious to an image analysis system (IAS) user, because usually the system internally converts the image to the model’s standard color space and resolution before feeding it to the model’s input. In that case, every Mb-AdA can be used, even if it is constructed for the attacked model’s input resolution or not. The effect of Mb-AdA remains after any conversion made in the image processing part of an IAS.
The above, together with the fact that the attack is based only on image features, makes the proposed attack to be applicable to any DL model, without the need for any knowledge of its structure or the dataset that was used for training. Moreover, the goal of the proposed attack is not to misclassify an image in a specific category. That is why Mb-AdA can be clearly characterized as an untargeted black box attack.
One other characteristic of Mb-AdA is that its effect is adjustable. As previously explained, the lower the moment’s order is, the more significant information is carried, which is why the exclusion of the last ordered moments is preferred. Using a smaller subset of moments during image reconstruction can increase the attack’s effect but with the cost of decreased image quality. On the other hand, acquiring better image quality of the adversarial examples excluding a smaller set of the last ordered moments, leads to a less efficient attack.
Adversarial Attack methods usually add an inexplainable noise to images, that is optimized in a different way in every attack method. Mb-AdA is the only that excludes specific known features of the image that makes its effect on DL models fully explainable. Moreover, studying the effect of several ordered Mb-AdAs we can explain in which moment features, is the DL model trained to base its decisions.
Obviously, for preparing only a specific Mb-AdA there is no need actually to compute entire the moment sets of the images in the first step but only the chosen subset of the moments that will be used for the reconstruction, lowering the computational cost of the procedure.
The experiments following the proposed adversarial attack uses the Tchebichef image moments family, that are constructed using (2) and the image reconstruction can be done using (3). The used K e r n e l p q in both formulas is the product of pth and qth order of the scaled Tchebichef polynomials given in (4) below. The used normalization factor N F ( p , q ) , is the inverse of the product of the squared norms ρ ( p , N ) given in (5).
t ˜ n ( x ) = ( 1 / b ( n , N ) ) i = 0 n [ ( N 1 i n i ) ( n + i n ) ( x i ) ]
ρ ( p , N ) = x = 0 N 1 [ t ˜ p ( x ) ] 2
where n is the order of the polynomial, N − 1 is the maximum order and equal to the corresponding dimension value of the image and b ( n , N ) is a scaling constant, usually set equal to Nn.
For constructing K-th order MB-AdA for the following experiments we used the following algorithm:
  • Read image dimensions N, M.
  • Select attack order K min { N , M }
  • Compute the polynomials t ˜ n ( x ) using (4) and add them to a matrix.
  • Compute normalization factor N F ( p , q ) using (5) and add them to another matrix.
  • Compute moment set { M p q } for every image channel using Equation (2) with the above matrices. Compute the K e r n e l p q ( x i , y j ) , in every iteration of the sum from 1 to the selected order K, using the appropriate values from the polynomial matrix and saving them in a new Kernel matrix.
  • Reconstruct the image using the above moment set   { M p q } and the saved Kernel matrix using (3).
It should be mentioned that the reason we do not compute the entire moment set in the above algorithm is that, as explained earlier, we exclude only the last ordered moments from the reconstruction procedure. So, their computation would be not only useless but also computationally expensive.
For measuring the quality of the reconstructed (attacked) image, mean structural similarity index measure (SSIM) [36,37] is used. SSIM is one of the most representative ways to measure the difference between two images as it is correlated with the quality and perception of the human visual system [38] and it is given by:
SSIM ( x , y ) = ( l ( x , y ) ) α ( c ( x , y ) ) β ( s ( x , y ) ) γ
where α   ,   β   a n d   γ   are weights (usually set to 1) of the terms luminance (l), contrast (c) and structure (s), that are given below as functions of mean values μ A   a n d   μ B , variance σ A   a n d   σ B and covariance σ A B respectively:
l ( x , y ) = 2 μ A μ B + C 1 μ A 2 + μ B 2 + C 1  
c ( x , y ) = 2 σ A σ B + C 2 σ A 2 + σ B 2 + C 2
s ( x , y ) = σ A B + C 3 σ A σ B + C 3
where C 1 ,   C 2   a n d   C 3   are small constants used to stabilize the metric for the case where the means and variances become very small.

3. Results

3.1. Attack Results

In order to measure the effectiveness of the proposed Mb-AdA to MIA tasks, we applied the attack, with several orders, in classification and segmentation tasks. The classification task was tested on X-ray [39] and histopathological [40] datasets. The first consists of three categories of lung X-rays (299 × 299 pixels size) while the second, which is more difficult, consists of four categories of cancer (2048 × 1536 pixels size). All datasets were divided into training, validation and test set. For the X-ray dataset, we used 3160 images for training 360 for validation and 365 for testing while in histopathological dataset we used 280/60/60 for training, validation and testing, respectively. These datasets, were used to train five DL models namely DenseNet 201 [41], Inception V3 [42], MobileNet V2 [43], DenseNet 169 [41] and Inception ResNet [44]. The segmentation task was tested on nuclei dataset [45] using the U-Net model [46]. For this task, we used 495 images (256 × 256 pixels size) for training and from 120 for training and validation. In Figure 1 examples of each dataset of the tasks are shown. The proposed attack was compared to FGSM [47], PGD [48] and Square Attack [49]. FGSM and PGD were trained in an irrelevant dataset in order for them to behave as black box attacks. All these attacks were created with Adversarial Robustness Toolbox (ART) [50]. All experiments were performed in Python with the Keras library. Additionally, we have experimented with several moment order values but for space saving reasons we present the most significant. As metrics we used Accuracy for classification task and Intersection over Union (IoU) for segmentation task.
A c c u r a c y = T r u e   P o s i t i v e + T r u e   N e g a t i v e T r u e   P o s i t i v e + T r u e   N e g a t i v e + F a l s e   P o s i t i v e + F a l s e   N e g a t i v e
I o U = A r e a   o f   I n t e r s e c t i o n   o f   t w o   m a s k s   ( p r e d i c t e d   a n d   g r o u n d   t r u t h ) A r e a   o f   U n i o n   o f   t w o   m a s k s   ( p r e d i c t e d   a n d   g r o u n d   t r u t h )
The FGSM attack extracts the adversarial gradient and decreases or increases the value of pixels so that the loss function increases. It perturbs a clean sample for a one-step update along the direction of gradient descend. Projected Gradient Descent (PGD) attack like an Iterative FGSM. Perturbations are constrained by projecting adversarial samples from each iteration into L or L2 neighbor of the clean image. Square Attack (SA) is based on a randomized search scheme which selects localized square shaped updates at random positions so that at each iteration the perturbation is situated approximately at the boundary of the feasible set.
In Table 1, nuclei segmentation results are presented with the SSIM (Structure Similarity Index Measure) and IoU (Intersection over Union) percentage that measures the accuracy of the segmentation for every attack. The indicating order number corresponds to the maximum order up to which is used to reconstruct the image (e.g., “Order 200” means that moments from 0 order up to 200 order, 201 × 201 moments in total are participating in the image reconstruction).
In Table 2 and Table 3, the results accuracy percentage for X-rays and histopathology images are presented, respectively.

3.2. Defence Results

The impact of Adversarial Training with Mb-AdA examples, augmenting the training dataset of the previously used ML models, as a defensive technique against adversarial attacks, is a reasonable question. Training DL models with images that lack non-robust features, as Mb-AdA examples are, should theoretically make them highly resistible to adversarial attacks. Numerous experiments were performed with several combinations of orders of images but for space saving, we will show indicatively the influence of adversarial training on MobileNetV2 with the histopathological dataset, which was the most vulnerable model. Table 4 shows the MobileNetV2 results with adversarial training using the initial training set and several moment orders.

4. Discussion

The main parameter of the proposed attack is the order of moments. Mb-AdA is an attack family that does not add some specific or optimized noise to an image as usually other adversarial attacks do, but it rather removes a part of the image. The removed part is not abstract but it consists of the last ordered moment features. It is worth noting that the first moment features carry the main image information, and the last ordered ones less. Figure 2 and Figure 3 are presented some examples of images under the proposed attack.
According to the results on X-ray dataset, Mb-AdA gets stronger as the order limit drops, because more and more of the higher ordered moment features are excluded, but as it is rational the image quality drops also. DenseNet 201, Inception V3 and Inception ResNet seem to be more resistant to higher order Mb-AdA than MobileNet V2 and DenseNet 169.
Additionally, compared to the other attacks, Mb-AdA seems to keep higher the image quality when it starts to significantly drop the accuracy rates. For example, Mb-AdA order 80 drops the DenseNet 201’s accuracy score to 82.74 keeping SSIM to 91.35 when FGSM ϵ = 0.07 drops accuracy to the slightly worse 84.03 with SSIM to 76.35. PGD ϵ = 0.007 drops accuracy to the slightly worse 85 and also drops the SSIM to 77.87. The same happens with Square Attack ϵ = 0.07, which drops accuracy to 85.56 dropping the SSIM to 82.59, which is better than FGSM and PGD, nevertheless it can only be compared to Mb-AdA with order 40 and SSIM equals to 82 that drops accuracy score to 53.97. The proposed attack performs globally better classification than the other attacks in X-ray. On the other hand, the classification of histopathological images behaves differently. While SSIM drops, PGD, FGSM and Square Attack perform better than Mb-AdA. For example, Mb-AdA with order 100 and SSIM 66.88, achieved 71.67% accuracy while PGD ϵ = 0.3 with SSIM 65.32 and Square Attack ϵ = 0.2 with SSIM 65 achieved 53.33% and 61.67% accuracy, respectively. However, in high SSIM, the proposed attack outperforms the others attacks, which is our main goal. Τhe segmentation problem, strongly highlights the superiority of the proposed attack, as it globally shows a significantly greater reduction in terms of IoU compared to the other attacks, presenting a much higher image quality at the same time. The other attacks, achieved 3.5% IoU degradation even with very low SSIM, while our proposed attack achieved nearly 20% IoU degradation with significantly higher SSIM.
Mb-AdA is an attack that does not add some random or optimized noise to an image as usually other adversarial attacks do [13], but it rather removes a part of the image. The removed part is not abstract but it consists of the last ordered moment features. Removing from an image the last ordered non-robust features should have a small or no effect on DL models, since the image quality remains high. However, this effect is in most cases stronger than the other attacks. The reason can now be explained as the DL model during training, learning these non-robust features that are removed by the Mb-AdA, together with robust ones. The other attacks affect the non-robust and the robust features with the same weight factor, and that is why they importantly drop image quality in order to achieve the same scores with Mb-AdA.
Moreover, medical image analysis tasks are quite difficult because models learn to make decisions according to some small details in the pictures. Mb-AdA has the ability to remove features that describe these details without destroying the structure of the image, and that is why it achieves high model’s performance degradation with high SSIM. The fact that the proposed attack does not destroy the structure of the image is also justified by the GradCAM algorithm [51] applied in Figure 4, which shows that models look at the almost same coordinates. This means that the attack preserves the structure of the images and does not disorient the model but removing useful information leads to misdiagnosis.
In addition to that, the data augmentation for adversarial training proves that it not only improves performance under Mb-AdA but improves performance even under other attacks. The MobileNetV2 model, which was the most vulnerable, has become quite robust after training with augmented data from the proposed attack in histopathological classification. The results are presented detailed in Table 4 but some representative examples are the following. The model trained with the Original Training Set (OTS) achieved 63.33% accuracy under FGSM ϵ = 0.03, while the same model trained with Augmented Training Set (ATS) achieved 90% accuracy. Additionally, the model with OTS under PGD ϵ = 0.1 achieved 60% accuracy while with ATS achieved 81.67%. These results come from the classification of histopathological images which was also the most difficult problem. It is clear that augmenting the data with examples of Mb-AdA significantly enhances the robustness of the model.

5. Conclusions

In this study, a black-box adversarial attack for medical images based on a moment exclusion strategy is proposed. The introduced attack was applied in classification and segmentation problems, achieving significant degradation in models’ performance while maintaining good image quality. Compared to other black-box attacks, the Mb-AdA shows better attacking results with additional improved imperceptibility. Additionally, adversarial learning was applied as a defense, proving that it can significantly enhance the robustness of models even under other attacks. The proposed attack can be useful for developing more robust deep learning models towards enforcing the integration of Artificial Intelligence tools in critical applications such as medical image analysis.
This work is sparking the future investigation into several directions, such as the optimization of the excluded moment orders per case, the incorporation of other moment families (e.g., Krawtchouk, dual Hahn, Fractional Moments, etc.) in the same attacking scheme and the examination of other medical image modalities.

Author Contributions

Conceptualization, G.A.P.; methodology, T.V.M. and G.A.P.; software, T.V.M. and K.D.A.; validation, T.V.M. and K.D.A.; investigation, T.V.M. and K.D.A.; resources, K.D.A.; data curation, K.D.A.; writing—original draft preparation, T.V.M.; writing—review and editing, G.A.P. and T.V.M.; visualization, K.D.A.; supervision, G.A.P.; project administration, G.A.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

This work was supported by the MPhil program “Advanced Technologies in Informatics and Computers”, hosted by the Department of Computer Science, International Hellenic University, Greece, Kavala.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  2. Rawat, W.; Wang, Z. Deep Convolutional Neural Networks for Image Classification: A Comprehensive Review. Neural Comput. 2017, 29, 2352–2449. [Google Scholar] [CrossRef] [PubMed]
  3. Wang, R.; Lei, T.; Cui, R.; Zhang, B.; Meng, H.; Nandi, A.K. Medical Image Segmentation Using Deep Learning: A Survey. IET Image Process. 2022, 16, 1243–1267. [Google Scholar] [CrossRef]
  4. Zhao, Z.-Q.; Zheng, P.; Xu, S.; Wu, X. Object Detection with Deep Learning: A Review. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3212–3232. [Google Scholar] [CrossRef] [Green Version]
  5. Zheng, C.; Wu, W.; Chen, C.; Yang, T.; Zhu, S.; Shen, J.; Kehtarnavaz, N.; Shah, M. Deep Learning-Based Human Pose Estimation. A Survey. arXiv 2022, arXiv:2012.13392v4. [Google Scholar]
  6. Ryu, J. A Visual Saliency-Based Neural Network Architecture for No-Reference Image Quality Assessment. Appl. Sci. 2022, 12, 9567. [Google Scholar] [CrossRef]
  7. Li, Q.; Zhu, J.; Liu, J.; Cao, R.; Li, Q.; Jia, S.; Qiu, G. Deep Learning based Monocular Depth Prediction: Datasets, Methods and Applications. arXiv 2020, arXiv:2011.04123. [Google Scholar]
  8. Rahim, T.; Usman, M.A.; Shin, S.Y. A survey on contemporary computer-aided tumor, polyp, and ulcer detection methods in wireless capsule endoscopy imaging. Comput. Med. Imaging Graph. 2020, 85, 101767. [Google Scholar] [CrossRef]
  9. Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.; Fergus, R. Intriguing properties of neural networks. arXiv 2013, arXiv:1312.6199. [Google Scholar]
  10. Maliamanis, T.; Papakostas, G. Adversarial computer vision: A current snapshot. In Proceedings of the Twelfth International Conference on Machine Vision (ICMV 2019), Amsterdam, Netherlands, 16–18 November 2019; Osten, W., Nikolaev, D.P., Eds.; SPIE: Amsterdam, The Netherlands, 2020; p. 121. [Google Scholar]
  11. Schneider, J.; Apruzzese, G. Concept-based Adversarial Attacks: Tricking Humans and Classifiers Alike. arXiv 2022, arXiv:220310166. [Google Scholar]
  12. Yao, Z.; Gholami, A.; Xu, P.; Keutzer, K.; Mahoney, M.W. Trust region based adversarial attack on neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2019, Long Beach, CA, USA, 15–20 June 2019; pp. 11350–11359. [Google Scholar]
  13. Apostolidis, K.D.; Papakostas, G.A. A Survey on Adversarial Deep Learning Robustness in Medical Image Analysis. Electronics 2021, 10, 2132. [Google Scholar] [CrossRef]
  14. Huq, A.; Pervin, M.T. Analysis of Adversarial Attacks on Skin Cancer Recognition. In Proceedings of the 2020 International Conference on Data Science and Its Applications (ICoDSA), Bandung, Indonesia, 5–6 August 2020; IEEE: Bandung, Indonesia, 2020; pp. 1–4. [Google Scholar]
  15. Paschali, M.; Conjeti, S.; Navarro, F.; Navab, N. Generalizability vs. Robustness: Investigating Medical Imaging Networks Using Adversarial Examples. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2018; Frangi, A.F., Schnabel, J.A., Davatzikos, C., Alberola-López, C., Fichtinger, G., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2018; Volume 11070, pp. 493–501. ISBN 978-3-030-00927-4. [Google Scholar]
  16. Li, Y.; Zhang, H.; Bermudez, C.; Chen, Y.; Landman, B.A.; Vorobeychik, Y. Anatomical context protects deep learning from adversarial perturbations in medical imaging. Neurocomputing 2020, 379, 370–378. [Google Scholar] [CrossRef] [PubMed]
  17. Risk Susceptibility of Brain Tumor Classification to Adversarial Attacks | SpringerLink. Available online: https://link.springer.com/chapter/10.1007/978-3-030-31964-9_17 (accessed on 4 June 2021).
  18. Cheng, G.; Ji, H. Adversarial Perturbation on MRI Modalities in Brain Tumor Segmentation. IEEE Access 2020, 8, 206009–206015. [Google Scholar] [CrossRef]
  19. Anand, D.; Tank, D.; Tibrewal, H.; Sethi, A. Self-Supervision vs. Transfer Learning: Robust Biomedical Image Analysis Against Adversarial Attacks. In Proceedings of the 2020 IEEE 17th International Symposium on Biomedical Imaging (ISBI), Iowa City, IA, USA, 3–7 April 2020; IEEE: Iowa City, IA, USA, 2020; pp. 1159–1163. [Google Scholar]
  20. Hirano, H.; Minagi, A.; Takemoto, K. Universal adversarial attacks on deep neural networks for medical image classification. BMC Med. Imaging 2021, 21, 9. [Google Scholar] [CrossRef]
  21. Ma, X.; Niu, Y.; Gu, L.; Wang, Y.; Zhao, Y.; Bailey, J.; Lu, F. Understanding Adversarial Attacks on Deep Learning Based Medical Image Analysis Systems. Pattern Recognit. 2021, 110, 107332. [Google Scholar] [CrossRef]
  22. Finlayson, S.G.; Chung, H.W.; Kohane, I.S.; Beam, A.L. Adversarial Attacks Against Medical Deep Learning Systems. arXiv 2018, arXiv:1804.05296. [Google Scholar]
  23. Shah, A.; Lynch, S.; Niemeijer, M.; Amelon, R.; Clarida, W.; Folk, J.; Russell, S.; Wu, X.; Abramoff, M.D. Susceptibility to misdiagnosis of adversarial images by deep learning based retinal image analysis algorithms. In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; IEEE: Washington, DC, USA, 2018; pp. 1454–1457. [Google Scholar]
  24. Byra, M.; Styczynski, G.; Szmigielski, C.; Kalinowski, P.; Michalowski, L.; Paluszkiewicz, R.; Ziarkiewicz-Wroblewska, B.; Zieniewicz, K.; Nowicki, A. Adversarial attacks on deep learning models for fatty liver disease classification by modification of ultrasound image reconstruction method. In Proceedings of the 2020 IEEE International Ultrasonics Symposium (IUS), Las Vegas, NV, USA, 7–11 September 2020. [Google Scholar]
  25. Ozbulak, U.; Van Messem, A.; De Neve, W. Impact of Adversarial Examples on Deep Learning Models for Biomedical Image Segmentation. arXiv 2019, arXiv:1907.13124. [Google Scholar]
  26. Yao, Q.; He, Z.; Lin, Y.; Ma, K.; Zheng, Y.; Zhou, S.K. A Hierarchical Feature Constraint to Camouflage Medical Adversarial Attacks. arXiv 2021, arXiv:2012.09501. [Google Scholar]
  27. Kugler, D. Physical Attacks in Dermoscopy: An Evaluation of Robustness for clinical Deep-Learning. In Proceedings of the Medical Imaging with Deep Learning (MIDL), London, UK, 8–10 July 2019; pp. 1–11. [Google Scholar]
  28. Vatian, A.; Gusarova, N.; Dobrenko, N.; Dudorov, S.; Nigmatullin, N.; Shalyto, A.; Lobantsev, A. Impact of Adversarial Examples on the Efficiency of Interpretation and Use of Information from High-Tech Medical Images. In Proceedings of the 2019 24th Conference of Open Innovations Association (FRUCT), Moscow, Russia, 8–12 April 2019; IEEE: Moscow, Russia, 2019; pp. 472–478. [Google Scholar]
  29. Chen, L.; Bentley, P.; Mori, K.; Misawa, K.; Fujiwara, M.; Rueckert, D. Intelligent image synthesis to attack a segmentation CNN using adversarial learning. arXiv 2019, arXiv:1909.11167. [Google Scholar]
  30. Tian, B.; Guo, Q.; Juefei-Xu, F.; Chan, W.L.; Cheng, Y.; Li, X.; Xie, X.; Qin, S. Bias Field Poses a Threat to DNN-based X-ray Recognition. arXiv 2021, arXiv:2009.09247. [Google Scholar]
  31. Papakostas, G.A. Over 50 years of image moments and moment invariants. Moments Moment Invariants-Theory Appl. 2014, 1, 3–32. [Google Scholar]
  32. Hu, M.-K. Visual pattern recognition by moment invariants. IEEE Trans. Inf. Theory 1962, 8, 179–187. [Google Scholar]
  33. Flusser, J.; Zitova, B.; Suk, T. Moments and Moment Invariants in Pattern Recognition; John Wiley & Sons: Hoboken, NJ, USA, 2009. [Google Scholar]
  34. Shu, H.; Luo, L.; Coatrieux, J.L. Derivation of Moments Invariants; Science Gate Publishing: Xanthi, Greece, 2014. [Google Scholar]
  35. Papakostas, G.A. Improving the recognition performance of moment features by selection. In Feature Selection for Data and Pattern Recognition; Springer: Berlin/Heidelberg, Germany, 2015; pp. 305–327. [Google Scholar]
  36. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Wang, Z.; Simoncelli, E.P.; Bovik, A.C. Multiscale structural similarity for image quality assessment. In Proceedings of the The Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, Pacific Grove, CA, USA, 9–12 November 2003; Volume 2, pp. 1398–1402. [Google Scholar]
  38. Setiadi, D.R.I.M. PSNR vs. SSIM: Imperceptibility quality assessment for image steganography. Multimed. Tools Appl. 2021, 80, 8423–8444. [Google Scholar] [CrossRef]
  39. Chowdhury, M.E.H.; Rahman, T.; Khandakar, A.; Mazhar, R.; Kadir, M.A.; Mahbub, Z.B.; Islam, K.R.; Khan, M.S.; Iqbal, A.; Emadi, N.A.; et al. Can AI Help in Screening Viral and COVID-19 Pneumonia? IEEE Access 2020, 8, 132665–132676. [Google Scholar] [CrossRef]
  40. Aresta, G.; Araújo, T.; Kwok, S.; Chennamsetty, S.S.; Safwan, M.; Alex, V.; Marami, B.; Prastawa, M.; Chan, M.; Donovan, M.; et al. BACH: Grand challenge on breast cancer histology images. Med. Image Anal. 2019, 56, 122–139. [Google Scholar] [CrossRef] [PubMed]
  41. Huang, G.; Liu, Z.; van der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. arXiv 2018, arXiv:1608.06993. [Google Scholar]
  42. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR); IEEE: Las Vegas, NV, USA, 2016; pp. 2818–2826. [Google Scholar]
  43. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. arXiv 2019, arXiv:180104381. [Google Scholar]
  44. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning. arXiv 2016, arXiv:160207261. [Google Scholar] [CrossRef]
  45. 2018 Data Science Bowl | Kaggle. Available online: https://www.kaggle.com/c/data-science-bowl-2018/data (accessed on 18 March 2022).
  46. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. arXiv 2015, arXiv:1505.04597. [Google Scholar]
  47. Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and Harnessing Adversarial Examples. arXiv 2015, arXiv:14126572. [Google Scholar]
  48. Madry, A.; Makelov, A.; Schmidt, L.; Tsipras, D.; Vladu, A. Towards Deep Learning Models Resistant to Adversarial Attacks. arXiv 2019, arXiv:170606083. [Google Scholar]
  49. Andriushchenko, M.; Croce, F.; Flammarion, N.; Hein, M. Square Attack: A Query-Efficient Black-Box Adversarial Attack via Random Search. In Proceedings of the Computer Vision—ECCV 2020, Glasgow, UK, 23–28 August 2020; Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 484–501. [Google Scholar]
  50. Nicolae, M.-I.; Sinn, M.; Tran, M.N.; Buesser, B.; Rawat, A.; Wistuba, M.; Zantedeschi, V.; Baracaldo, N.; Chen, B.; Ludwig, H.; et al. Adversarial Robustness Toolbox v1.0.0. arXiv 2019, arXiv:180701069. [Google Scholar]
  51. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
Figure 1. Examples from datasets were used. (a) Histopathology (scale bar—0.42 μm), (b) chest X-rays, (c) nuclei cells.
Figure 1. Examples from datasets were used. (a) Histopathology (scale bar—0.42 μm), (b) chest X-rays, (c) nuclei cells.
Biomedicines 10 02545 g001
Figure 2. Images from the histopathological dataset (scale bar—0.42 μm). (a) Original image. Mb-Ada with order, (b) 200, (c) 120, (d) 50.
Figure 2. Images from the histopathological dataset (scale bar—0.42 μm). (a) Original image. Mb-Ada with order, (b) 200, (c) 120, (d) 50.
Biomedicines 10 02545 g002
Figure 3. Images from the nuclei cell dataset. (a) Original image. Mb-Ada with order, (b) 200, (c) 120, (d) 80.
Figure 3. Images from the nuclei cell dataset. (a) Original image. Mb-Ada with order, (b) 200, (c) 120, (d) 80.
Biomedicines 10 02545 g003
Figure 4. (a) Original image, (b) the corresponding GradCAM, (c) image under Mb-AdA with order 60, and (d) the corresponding GradCAM.
Figure 4. (a) Original image, (b) the corresponding GradCAM, (c) image under Mb-AdA with order 60, and (d) the corresponding GradCAM.
Biomedicines 10 02545 g004
Table 1. Nuclei segmentation results under all attacks in terms of SSIM and IoU (in %).
Table 1. Nuclei segmentation results under all attacks in terms of SSIM and IoU (in %).
ATTACKSSIMIoU
Attack Free10075.22
Mb-AdA Order 20092.1774.37
Mb-AdA Order 14091.8373.80
Mb-AdA Order 8088.9669.70
Mb-AdA Order 6086.2565.50
Mb-AdA Order 5081.7361.47
Mb-AdA Order 4074.2357.57
FGSM ϵ = 0.0189.8474.10
FGSM ϵ = 0.0767.3174.00
FGSM ϵ = 0.0954.6574.00
FGSM ϵ = 0.1241.5771.90
PGD ϵ = 0.0189.7374.12
PGD ϵ = 0.0579.4074.16
PGD ϵ = 0.0770.4574.15
PGD ϵ = 0.0960.8874.12
Square Attack ϵ = 0.0190.1474.00
Square Attack ϵ = 0.0869.3872.95
Square Attack ϵ = 0.165.0072.28
Table 2. Image quality and accuracy in histopathological images under attacks (in %).
Table 2. Image quality and accuracy in histopathological images under attacks (in %).
ATTACKSSIMDenseNet 201Inception V3MobileNet V2DenseNet 169Inception ResNet
Original10081.6771.6785.0071.6774.00
Mb-AdA Order 20095.2270.0066.6776.6770.0073.33
Mb-AdA Order 18092.1868.3363.3371.6770.0071.67
Mb-AdA Order 16088.1165.0063.3375.0071.6766.67
Mb-AdA Order 14082.8668.3358.3365.0073.3370.00
Mb-AdA Order 10066.8871.6763.3356.6761.6770.00
Mb-AdA Order 8056.0056.6758.3353.3346.6756.67
FGSM ϵ = 0.0388.2673.3368.3363.3373.3368.33
FGSM ϵ = 0.177.2965.0065.0050.0060.0068.00
FGSM ϵ = 0.342.1745.0053.3335.0046.6754.00
PGD ϵ = 0.0189.4675.0066.6776.6775.0076.67
PGD ϵ = 0.183.6071.6775.0060.0066.6773.33
PGD ϵ = 0.265.3263.3358.3340.0055.0066.67
PGD ϵ = 0.356.5553.3355.0030.0036.6751.67
Square Attack ϵ = 0.0189.7076.6770.0081.6773.3373.33
Square Attack ϵ = 0.0587.3078.3365.0075.0068.3373.33
Square Attack ϵ = 0.180.9066.6756.6760.0065.0075.00
Square Attack ϵ = 0.265.0061.6743.3351.6761.6761.67
Table 3. Image quality and accuracy in chest X-ray images under attacks (in %).
Table 3. Image quality and accuracy in chest X-ray images under attacks (in %).
ATTACKSSIMDenseNet 201Inception V3MobileNet V2DenseNet 169Inception ResNet
Original10099.1893.9797.8197.2697.26
Mb-AdA Order 20099.0398.9093.4291.5193.4296.99
Mb-AdA Order 16098.1398.9092.0587.6794.5296.44
Mb-AdA Order 12095.2696.9089.0480.5589.5994.25
Mb-AdA Order 8091.3582.7481.1056.1669.3286.58
Mb-AdA Order 5085.4360.8259.7343.8436.7061.37
Mb-AdA Order 4082.0053.9740.5542.4733.1560.00
Mb-AdA Order 3077.8940.5534.5240.2732.8854.25
FGSM ϵ = 0.0198.7198.3691.2391.7893.4295.34
FGSM ϵ = 0.0394.6096.4084.9380.2788.4992.60
FGSM ϵ = 0.0584.1090.1473.1553.1570.4184.11
FGSM ϵ = 0.0776.3584.0366.1142.3068.3576.75
FGSM ϵ = 0.0964.2373.1157.1439.7864.7169.75
PGD ϵ = 0.0198.1598.9091.6790.0094.2095.90
PGD ϵ = 0.0393.8496.4083.3075.5682.7087.50
PGD ϵ = 0.0586.7991.6776.1155.8375.2877.78
PGD ϵ = 0.0777.8785.0065.5649.4468.6170.56
PGD ϵ = 0.0968.2176.9459.7246.3963.9860.83
Square Attack ϵ = 0.0199.2998.6191.9491.9493.6196.67
Square Attack ϵ = 0.0588.0090.2872.2261.3983.8992.50
Square Attack ϵ = 0.0782.5985.5658.0655.8377.5091.39
Square Attack ϵ = 0.0976.3979.7246.6758.0670.2885.00
Table 4. Accuracy after adversarial training in histopathological images with MobileNetV2 (in %).
Table 4. Accuracy after adversarial training in histopathological images with MobileNetV2 (in %).
ATTACKNormal Training SetAugmented Orders
20–200
Augmented Orders
100–200
Original8588.3381.6
Mb-AdA Order 20076.679073.33
Mb-AdA Order 18071.679071.67
Mb-AdA Order 1607586.6775
Mb-AdA Order 140658573.33
Mb-AdA Order 1206588.3376.67
Mb-AdA Order 110658580
Mb-AdA Order 10056.6786.6775
Mb-AdA Order 8053.3381.6776.67
Mb-AdA Order 60457063.33
Mb-AdA Order 504568.3368.33
FGSM ϵ = 0.0363.339076.67
FGSM ϵ = 0.056088.3358.33
FGSM ϵ = 0.15071.6761.67
PGD ϵ = 0.0176.6788.3376.67
PGD ϵ = 0.16081.6770
PGD ϵ = 0.33058.3335
Square Attack ϵ = 0.028086.6785
Square Attack ϵ = 0.05758570
Square Attack ϵ = 0.16068.3375
Square Attack ϵ = 0.251.676036.33
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Maliamanis, T.V.; Apostolidis, K.D.; Papakostas, G.A. How Resilient Are Deep Learning Models in Medical Image Analysis? The Case of the Moment-Based Adversarial Attack (Mb-AdA). Biomedicines 2022, 10, 2545. https://doi.org/10.3390/biomedicines10102545

AMA Style

Maliamanis TV, Apostolidis KD, Papakostas GA. How Resilient Are Deep Learning Models in Medical Image Analysis? The Case of the Moment-Based Adversarial Attack (Mb-AdA). Biomedicines. 2022; 10(10):2545. https://doi.org/10.3390/biomedicines10102545

Chicago/Turabian Style

Maliamanis, Theodore V., Kyriakos D. Apostolidis, and George A. Papakostas. 2022. "How Resilient Are Deep Learning Models in Medical Image Analysis? The Case of the Moment-Based Adversarial Attack (Mb-AdA)" Biomedicines 10, no. 10: 2545. https://doi.org/10.3390/biomedicines10102545

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop