The Role of Feature Vector Scale in the Adversarial Vulnerability of Convolutional Neural Networks
Abstract
1. Introduction
- We propose a novel concept, “unfair learning of data”, which attributes adversarial vulnerability through feature vector scale rather than loss magnitude or sample imbalance.
- We provide empirical evidence on CIFAR-10 and VGG16, showing that samples with small-scale feature vectors exhibit significantly higher adversarial vulnerability.
- We highlight the potential of scale-aware learning as a new research direction for the improvement of adversarial robustness.
2. Background
2.1. Vanilla Learning Based on Softmax
2.2. Angle-Based Learning
2.3. Adversarial Examples
3. Scale of Feature Vector
3.1. Feature Vectors in CNN Models
3.2. Geometric Interpretation of Feature Vector Training
3.3. Hypothesis for Scale Sensitivity
4. Experiments
4.1. Baseline Model and Dataset
4.2. Scale Section
4.3. Similarity Comparison
4.4. Comparison of Error Rates
4.5. Analysis of Feature Vector Scale Under Adversarial Examples
4.6. Analysis of Feature Vector Scale Under Gabor Noise
4.7. Discussion on Generality
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
CNN | Convolutional Neural Network |
FC | Fully Connected |
References
- Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and harnessing adversarial examples. arXiv 2014, arXiv:1412.6572. [Google Scholar]
- Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.; Fergus, R. Intriguing properties of neural networks. arXiv 2013, arXiv:1312.6199. [Google Scholar]
- Papernot, N.; McDaniel, P.; Jha, S.; Fredrikson, M.; Celik, Z.B.; Swami, A. The limitations of deep learning in adversarial settings. In Proceedings of the 2016 IEEE European Symposium on Security and Privacy (EuroS&P), Saarbruecken, Germany, 21–24 March 2016; pp. 372–387. [Google Scholar]
- Moosavi-Dezfooli, S.M.; Fawzi, A.; Frossard, P. Deepfool: A simple and accurate method to fool deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2574–2582. [Google Scholar]
- Liu, Y.; Chen, X.; Liu, C.; Song, D. Delving into Transferable Adversarial Examples and Black-box Attacks. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
- Tramèr, F.; Papernot, N.; Goodfellow, I.; Boneh, D.; McDaniel, P. The space of transferable adversarial examples. arXiv 2017, arXiv:1704.03453. [Google Scholar]
- Tanay, T.; Griffin, L. A boundary tilting persepective on the phenomenon of adversarial examples. arXiv 2016, arXiv:1608.07690. [Google Scholar]
- Ilyas, A.; Santurkar, S.; Tsipras, D.; Engstrom, L.; Tran, B.; Madry, A. Adversarial examples are not bugs, they are features. In Proceedings of the NIPS’19: 33rd International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; Volume 32, pp. 125–136. [Google Scholar] [CrossRef]
- Gilmer, J.; Metz, L.; Faghri, F.; Schoenholz, S.S.; Raghu, M.; Wattenberg, M.; Goodfellow, I. Adversarial spheres. arXiv 2018, arXiv:1801.02774. [Google Scholar]
- Schmidt, L.; Santurkar, S.; Tsipras, D.; Talwar, K.; Madry, A. Adversarially robust generalization requires more data. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018; Volume 31, pp. 5019–5031. [Google Scholar] [CrossRef]
- Koh, P.W.; Liang, P. Understanding black-box predictions via influence functions. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 1885–1894. [Google Scholar] [CrossRef]
- Bengio, Y.; Louradour, J.; Collobert, R.; Weston, J. Curriculum learning. In Proceedings of the 26th Annual International Conference on Machine Learning, Montreal, QC, Canada, 14–19 June 2009; pp. 41–48. [Google Scholar] [CrossRef]
- Wang, H.; Wang, Y.; Zhou, Z.; Ji, X.; Gong, D.; Zhou, J.; Li, Z.; Liu, W. Cosface: Large margin cosine loss for deep face recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 5265–5274. [Google Scholar] [CrossRef]
- Deng, J.; Guo, J.; Xue, N.; Zafeiriou, S. Arcface: Additive angular margin loss for deep face recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 5962–5979. [Google Scholar] [CrossRef] [PubMed]
- Park, J.; Chai, J.C.L.; Yoon, J.; Teoh, A.B.J. Understanding the Feature Norm for Out-of-Distribution Detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 1–6 October 2023; pp. 1557–1567. [Google Scholar] [CrossRef]
- Wen, Y.; Liu, W.; Weller, A.; Raj, B.; Singh, R. SphereFace2: Binary Classification is All You Need for Deep Face Recognition. In Proceedings of the International Conference on Learning Representations (ICLR), Virtual Event, 25–29 April 2022. [Google Scholar]
- Huang, Y.; Wang, Y.; Tai, Y.; Liu, X.; Shen, P.; Li, S.; Li, J.; Huang, F. Curricularface: Adaptive curriculum learning loss for deep face recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 5900–5909. [Google Scholar] [CrossRef]
- Meng, Q.; Zhao, S.; Huang, Z.; Zhou, F. MagFace: A Universal Representation for Face Recognition and Quality Assessment. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 14220–14229. [Google Scholar] [CrossRef]
- Kim, M.; Jain, A.K.; Liu, X. AdaFace: Quality Adaptive Margin for Face Recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 18729–18738. [Google Scholar] [CrossRef]
- Muñoz-González, L.; Lupu, E.C. Sensitivity of deep convolutional networks to Gabor noise. In Proceedings of the ICML 2019 Workshop on Identifying and Understanding Deep Learning Phenomena, Long Beach, CA, USA, 15 June 2019. [Google Scholar]
- Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
- Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar] [CrossRef]
- Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269. [Google Scholar] [CrossRef]
- Liu, W.; Wen, Y.; Yu, Z.; Yang, M. Large-margin softmax loss for convolutional neural networks. In Proceedings of the ICML, New York, NY, USA, 19–24 June 2016; Volume 2, pp. 507–516. [Google Scholar] [CrossRef]
- Liu, W.; Wen, Y.; Yu, Z.; Li, M.; Raj, B.; Song, L. Sphereface: Deep hypersphere embedding for face recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6738–6746. [Google Scholar] [CrossRef]
- Madry, A.; Makelov, A.; Schmidt, L.; Tsipras, D.; Vladu, A. Towards deep learning models resistant to adversarial attacks. arXiv 2017, arXiv:1706.06083. [Google Scholar]
- Tabacof, P.; Valle, E. Exploring the space of adversarial images. In Proceedings of the 2016 International Joint Conference on Neural Networks (IJCNN), Vancouver, BC, Canada, 24–29 July 2016; pp. 426–433. [Google Scholar] [CrossRef]
- Shafahi, A.; Huang, W.R.; Studer, C.; Feizi, S.; Goldstein, T. Are adversarial examples inevitable? arXiv 2018, arXiv:1809.02104. [Google Scholar]
- Fawzi, A.; Fawzi, H.; Fawzi, O. Adversarial vulnerability for any classifier. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018; Volume 31, pp. 1186–1195. [Google Scholar] [CrossRef]
- Mahloujifar, S.; Diochnos, D.I.; Mahmoody, M. The curse of concentration in robust learning: Evasion and poisoning attacks from concentration of measure. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 4536–4543. [Google Scholar] [CrossRef]
- Wu, D.; Xia, S.T.; Wang, Y. Adversarial weight perturbation helps robust generalization. In Proceedings of the NIPS’20: 34th International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 6–12 December 2018; Volume 33, pp. 2958–2969. [Google Scholar]
- Gowal, S.; Qin, C.; Uesato, J.; Mann, T.; Kohli, P. Uncovering the limits of adversarial training against norm-bounded adversarial examples. arXiv 2020, arXiv:2010.03593. [Google Scholar]
- Benz, P.; Zhang, C.; Kweon, I.S. Batch normalization increases adversarial vulnerability and decreases adversarial transferability: A non-robust feature perspective. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 7798–7807. [Google Scholar] [CrossRef]
- Zhao, M.; Zhang, L.; Ye, J.; Lu, H.; Yin, B.; Wang, X. Adversarial training: A survey. arXiv 2024, arXiv:2410.15042. [Google Scholar]
- Hu, J.; Ye, J.; Feng, Z.; Yang, J.; Liu, S.; Yu, X.; Jia, L.; Song, M. Improving adversarial robustness via feature pattern consistency constraint. In Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence, Jeju, Republic of Korea, 3–9 August 2024; pp. 848–856. [Google Scholar] [CrossRef]
- Liu, D.; Chen, T.; Peng, C.; Wang, N.; Hu, R.; Gao, X. Improving Adversarial Robustness via Decoupled Visual Representation Masking. IEEE Trans. Inf. Forensics Secur. 2025, 20, 5678–5689. [Google Scholar] [CrossRef]
- Zhou, N.; Zhou, D.; Liu, D.; Wang, N.; Gao, X. Mitigating feature gap for adversarial robustness by feature disentanglement. In Proceedings of the AAAI Conference on Artificial Intelligence, Philadelphia, PA, USA, 25 February–4 March 2025; No. 1203. pp. 10825–10833. [Google Scholar] [CrossRef]
Class | Section Size | Scale Section | |||||||||
---|---|---|---|---|---|---|---|---|---|---|---|
1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | ||
Bird | 21.0 | 100 | 100 | 98.5 | 98.0 | 94.5 | 82.9 | 55.7 | 27.4 | 10.8 | 0 |
Airplane | 19.8 | 100 | 100 | 98.0 | 95.3 | 90.7 | 75.8 | 49.5 | 20.7 | 4.3 | 0 |
Dog | 20.8 | n/a | 100 | 100 | 100 | 99.8 | 98.2 | 89.9 | 62.6 | 21.3 | 0 |
Cat | 5.1 | n/a | 0 | 17.6 | 16.9 | 13.4 | 9.7 | 0 | 0 | 0 | 0 |
Ship | 29.9 | 100 | 100 | 99.3 | 94.4 | 80.3 | 51.9 | 19.6 | 6.3 | 0 | 0 |
Frog | 20.0 | n/a | 100 | 99.5 | 98.8 | 94.3 | 81.2 | 57.4 | 29.7 | 0 | 0 |
Automobile | 36.3 | n/a | 100 | 100 | 99.0 | 88.4 | 64.6 | 42.2 | 22.5 | 11.4 | 0 |
Truck | 35.0 | 100 | 100 | 100 | 97.9 | 93.6 | 72.6 | 38.2 | 14.1 | 5 | 0 |
Deer | 33.3 | n/a | n/a | 100 | 100 | 99.9 | 97.3 | 87.9 | 59.4 | 24.4 | 0 |
Horse | 45.7 | n/a | n/a | 100 | 100 | 98.6 | 85.8 | 56.3 | 38.1 | 22.3 | 26.6 |
Class | Section Size | Scale Section | |||||||||
---|---|---|---|---|---|---|---|---|---|---|---|
1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | ||
Bird | 21.0 | 100 | 100 | 44.9 | 26.3 | 12.6 | 4.4 | 2.1 | 0.3 | 1.3 | 0 |
Airplane | 19.8 | 0 | 33.3 | 25.2 | 15.7 | 8.8 | 4.5 | 1.4 | 0.3 | 0 | 0 |
Dog | 20.8 | n/a | 80 | 65.5 | 54.4 | 33.0 | 22.8 | 12.0 | 4.8 | 3.3 | 0 |
Cat | 5.1 | n/a | 0 | 6.5 | 9.5 | 12.8 | 14.4 | 33.3 | 33.3 | 0 | 0 |
Ship | 29.9 | 100 | 45.4 | 35.4 | 19.3 | 11.7 | 6.1 | 3.8 | 0.3 | 0 | 0 |
Frog | 20.0 | n/a | 56.0 | 37.6 | 16.8 | 5.5 | 1.5 | 0.3 | 0 | 0 | 0 |
Automobile | 36.3 | n/a | 0 | 50 | 24.7 | 9.2 | 3.5 | 1.0 | 0.3 | 0.3 | 0 |
Truck | 35.0 | 100 | 0 | 43.4 | 17.9 | 10.9 | 4.7 | 0.9 | 0.3 | 0 | 0 |
Deer | 33.3 | n/a | n/a | 100 | 62.5 | 41.2 | 22.9 | 10.7 | 4.5 | 1.4 | 0 |
Horse | 45.7 | n/a | n/a | 100 | 51.0 | 22.5 | 7.7 | 1.8 | 1.4 | 0 | 0 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Park, H.-C.; Lee, S.-W. The Role of Feature Vector Scale in the Adversarial Vulnerability of Convolutional Neural Networks. Mathematics 2025, 13, 3026. https://doi.org/10.3390/math13183026
Park H-C, Lee S-W. The Role of Feature Vector Scale in the Adversarial Vulnerability of Convolutional Neural Networks. Mathematics. 2025; 13(18):3026. https://doi.org/10.3390/math13183026
Chicago/Turabian StylePark, Hyun-Cheol, and Sang-Woong Lee. 2025. "The Role of Feature Vector Scale in the Adversarial Vulnerability of Convolutional Neural Networks" Mathematics 13, no. 18: 3026. https://doi.org/10.3390/math13183026
APA StylePark, H.-C., & Lee, S.-W. (2025). The Role of Feature Vector Scale in the Adversarial Vulnerability of Convolutional Neural Networks. Mathematics, 13(18), 3026. https://doi.org/10.3390/math13183026