Explainable AI-Based Intrusion Detection Systems for Industry 5.0 and Adversarial XAI: A Systematic Review
Abstract
1. Introduction
1.1. Industry 5.0: Characteristics and AI Integration Context
1.2. Scope of the Survey
1.3. Related Surveys
1.4. Contributions
- We provide a clear and comprehensive taxonomy of XAI systems with categorization of ante hoc and post hoc methods, analyzing their applicability and limitations in cybersecurity contexts.
- We provide a detailed overview of current state-of-the-art IDSs, their limitations, and the deployment of XAI approaches in IDSs, systematically analyzing 135 empirical studies to identify implementation patterns, commonly used datasets, and explainability technique preferences.
- We systematically discuss the exploitation of XAI methods for launching more advanced adversarial attacks on IDSs, mapping specific XAI techniques to documented attack vectors and vulnerability levels.
- We analyze Industry 5.0-specific cybersecurity challenges and identify research directions for adversarially robust, human-centered explainable security systems, including federated learning architectures and SOC workflow integration.
2. Methodology
2.1. Research Questions
- What are the key cybersecurity challenges in Industry 5.0, and why are explainable AI-based intrusion detection systems (X-IDSs) essential to addressing these threats?
- What techniques and methods enhance transparency and interpretability in X-IDS implementations?
- What are the primary challenges and limitations of X-IDS in cybersecurity applications?
- What are the security implications of adversaries exploiting X-IDSs decision mechanisms, and how can these systems be protected against such attacks?
- What are the emerging trends and future research directions for X-IDSs in Industry 5.0 contexts?
2.2. Search Strategy
2.3. Study Selection
2.4. Inclusion and Exclusion Criteria
- Investigated XAI-based IDSs or Adv-XIDSs within the cybersecurity context of Industry 5.0.
- Employed deep learning architectures (e.g., transformers and LSTMs) or shallow computational models for intrusion detection.
- Comprised peer-reviewed articles or high-quality gray literature containing empirical findings or theoretical frameworks with substantive insights.
- Non-English studies that would introduce language-based analytical barriers.
- Articles with peripheral relevance to XAI-based IDSs or cybersecurity paradigms.
- Publications with insufficient methodological detail or inadequate empirical support (e.g., conference abstracts and letters to editors).
2.5. Data Extraction and Synthesis
3. Explainable AI (XAI) Taxonomies
3.1. Ante Hoc Explainability
3.2. Post Hoc Explainability
Global and Local Explanations
4. Industry 5.0 and Associated Cybersecurity Challenges
5. Intrusion Detection Systems for Cybersecurity in Industry 5.0
5.1. Traditional Intrusion Detection System (IDS)
5.2. Explainable IDS (X-IDS)
5.2.1. Self-Model Explainability
5.2.2. Pre-Modeling Explainability
5.2.3. Post-Model Explainability
6. Adversarial XAI and IDSs
6.1. Adversarial Attacks Without Utilizing Explainability
6.2. Adversarial Attacks Utilizing Explainability
6.3. Synthesis and Implications
7. XAI-Based IDSs: Lessons Learned, Challenges, and Future Research Directions
7.1. Synthesis of Key Findings
7.2. Industry 5.0-Specific Insights and Implications
7.3. Critical Research Gaps and Challenges
7.4. Future Research Priorities for Industry 5.0
- Adversarially Robust Explainability Architectures: Developing explanation mechanisms that maintain transparency while resisting adversarial exploitation through techniques such as explanation obfuscation, multi-level explanation hierarchies, and dynamic explanation strategies adapted to threat contexts.
- Cybersecurity-Native Explanation Methods: Creating explanation techniques designed specifically for cybersecurity’s temporal, contextual, and multi-modal data characteristics, moving beyond adaptations of general-purpose XAI methods to domain-optimized approaches.
- Human-Centric Explanation Design: Developing explanation systems optimized for human cognitive factors, stress conditions, and decision-making requirements specific to cybersecurity incident response in Industry 5.0’s collaborative environments.
- Federated Explainable Learning Frameworks: Establishing explanation methods for distributed learning environments that preserve privacy while enabling effective cross-organizational threat intelligence sharing essential for Industry 5.0’s interconnected ecosystems. Federated explainable learning frameworks must address concrete Industry 5.0 deployment scenarios: distributed manufacturing networks where multiple facilities collaboratively train IDS models while maintaining proprietary data privacy, supply chain ecosystems requiring shared threat intelligence without exposing sensitive operational details, and edge–cloud hybrid architectures balancing local real-time detection with centralized model updates. Technical challenges include compact explanation representation for bandwidth-constrained industrial networks, privacy-preserving explanation aggregation techniques, and development of standardized explanation formats that ensure interoperability across heterogeneous industrial systems. These frameworks must demonstrate compliance with both cybersecurity requirements and industrial safety protocols to achieve practical deployment in human-centric industrial environments.
- Real-Time Explanation Generation: Advancing efficient algorithms capable of generating meaningful explanations within Industry 5.0’s real-time operational constraints without compromising detection accuracy or explanation quality.
- Standardized Evaluation Methodologies: Developing comprehensive evaluation frameworks specifically designed for cybersecurity applications, incorporating metrics for explanation accuracy, actionability, temporal consistency, and adversarial robustness.
8. Conclusions
Supplementary Materials
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Speith, T. A Review of Taxonomies of Explainable Artificial Intelligence (XAI) Methods. In Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency, Seoul, Republic of Korea, 21–24 June 2022; pp. 2239–2250. [Google Scholar]
- Alsamhi, S.H.; Shvetsov, A.V.; Hawbani, A.; Shvetsova, S.V.; Kumar, S.; Zhao, L. Survey on Federated Learning enabling indoor navigation for industry 4.0 in B5G. Future Gener. Comput. Syst. 2023, 148, 250–265. [Google Scholar] [CrossRef]
- Rane, N.L.; Kaya, Ö.; Rane, J. Artificial Intelligence, Machine Learning, and Deep Learning for Sustainable Industry 5.0; Deep Science Publishing: San Francisco, CA, USA, 2024. [Google Scholar]
- Holzinger, A.; Saranti, A.; Molnar, C.; Biecek, P.; Samek, W. Explainable AI methods—A brief overview. In Proceedings of the International Workshop on Extending Explainable AI Beyond Deep Models and Classifiers, Vienna, Austria, 18 July 2022; Springer: Cham, Switzerland, 2022; pp. 13–38. [Google Scholar]
- Gunning, D.; Aha, D. DARPA’s explainable artificial intelligence (XAI) program. AI Mag. 2019, 40, 44–58. [Google Scholar]
- Alexandrov, N. Explainable AI decisions for human-autonomy interactions. In Proceedings of the 17th AIAA Aviation Technology, Integration, and Operations Conference, Denver, Colorado, 5–9 June 2017; p. 3991. [Google Scholar]
- Capuano, N.; Fenza, G.; Loia, V.; Stanzione, C. Explainable Artificial Intelligence in CyberSecurity: A Survey. IEEE Access 2022, 10, 93575–93600. [Google Scholar] [CrossRef]
- Yayla, A.; Haghnegahdar, L.; Dincelli, E. Explainable artificial intelligence for smart grid intrusion detection systems. IT Prof. 2022, 24, 18–24. [Google Scholar] [CrossRef]
- Scalas, M.; Rieck, K.; Giacinto, G. Improving malware detection with explainable machine learning. In Explainable Deep Learning AI; Elsevier: Amsterdam, The Netherlands, 2023; pp. 217–238. [Google Scholar]
- Marino, D.L.; Wickramasinghe, C.S.; Manic, M. An Adversarial Approach for Explainable AI in Intrusion Detection Systems. In Proceedings of the IECON 2018-44th Annual Conference of the IEEE Industrial Electronics Society, Washington, DC, USA, 21–23 October 2018; pp. 3237–3243. [Google Scholar]
- Baniecki, H.; Biecek, P. Adversarial Attacks and Defenses in Explainable Artificial Intelligence: A Survey. arXiv 2023, arXiv:2306.06123. [Google Scholar] [CrossRef]
- Sharma, D.K.; Mishra, J.; Singh, A.; Govil, R.; Srivastava, G.; Lin, J.C.W. Explainable Artificial Intelligence for Cybersecurity. Comput. Electr. Eng. 2022, 103, 108356. [Google Scholar] [CrossRef]
- Yu, J.; Shvetsov, A.V.; Alsamhi, S.H. Leveraging machine learning for cybersecurity resilience in industry 4.0: Challenges and future directions. IEEE Access 2024, 12, 159579–159596. [Google Scholar] [CrossRef]
- Kiran, A.; Prakash, S.W.; Kumar, B.A.; Sameeratmaja, T.; Charan, U.S.S.R. Intrusion Detection System Using Machine Learning. In Proceedings of the 2023 International Conference on Computer Communication and Informatics (ICCCI), Coimbatore, India, 23–25 January 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–4. [Google Scholar]
- Salam, A.; Ullah, F.; Amin, F.; Abrar, M. Deep learning techniques for web-based attack detection in industry 5.0: A novel approach. Technologies 2023, 11, 107. [Google Scholar] [CrossRef]
- Gadekallu, T.R.; Maddikunta, P.K.R.; Boopathy, P.; Deepa, N.; Chengoden, R.; Victor, N.; Wang, W.; Wang, W.; Zhu, Y.; Dev, K. Xai for industry 5.0-concepts, opportunities, challenges and future directions. IEEE Open J. Commun. Soc. 2024, 6, 2706–2729. [Google Scholar] [CrossRef]
- Hussain, F.; Hussain, R.; Hassan, S.A.; Hossain, E. Machine learning in IoT security: Current solutions and future challenges. IEEE Commun. Surv. Tutor. 2020, 22, 1686–1721. [Google Scholar] [CrossRef]
- Chou, D.; Jiang, M. A survey on data-driven network intrusion detection. ACM Comput. Surv. (CSUR) 2021, 54, 1–36. [Google Scholar] [CrossRef]
- Alazab, M.; KP, S.; Srinivasan, S.; Venkatraman, S.; Pham, Q.V.; Simran. Deep Learning for Cyber Security Applications: A Comprehensive Survey. 2021. Available online: https://d197for5662m48.cloudfront.net/documents/publicationstatus/162402/preprint_pdf/1db580dedba693379c64ee6ebfbf4792.pdf (accessed on 30 August 2025).
- Markevych, M.; Dawson, M. A review of enhancing intrusion detection systems for cybersecurity using artificial intelligence (ai). In Proceedings of the International Conference Knowledge-Based Organization; Paradigm: Boston, UK, 2023; Volume 29, pp. 30–37. [Google Scholar]
- Sowmya, T.; Anita, E.M. A comprehensive review of AI based intrusion detection system. Meas. Sens. 2023, 28, 100827. [Google Scholar] [CrossRef]
- Sauka, K.; Shin, G.Y.; Kim, D.W.; Han, M.M. Adversarial robust and explainable network intrusion detection systems based on deep learning. Appl. Sci. 2022, 12, 6451. [Google Scholar] [CrossRef]
- Maddikunta, P.K.R.; Pham, Q.V.; Prabadevi, B.; Deepa, N.; Dev, K.; Gadekallu, T.R.; Ruby, R.; Liyanage, M. Industry 5.0: A survey on enabling technologies and potential applications. J. Ind. Inf. Integr. 2022, 26, 100257. [Google Scholar] [CrossRef]
- Czeczot, G.; Rojek, I.; Mikołajewski, D.; Sangho, B. AI in IIoT Management of Cybersecurity for Industry 4.0 and Industry 5.0 Purposes. Electronics 2023, 12, 3800. [Google Scholar] [CrossRef]
- Taj, I.; Zaman, N. Towards industrial revolution 5.0 and explainable artificial intelligence: Challenges and opportunities. Int. J. Comput. Digit. Syst. 2022, 12, 295–320. [Google Scholar] [CrossRef]
- Bobek, S.; Nowaczyk, S.; Gama, J.; Pashami, S.; Ribeiro, R.P.; Taghiyarrenani, Z.; Veloso, B.; Rajaoarisoa, L.H.; Szelazek, M.; Nalepa, G.J. Why Industry 5.0 Needs XAI 2.0? In Proceedings of the xAI (Late-Breaking Work, Demos, Doctoral Consortium), Lisbon, Portugal, 26–28 July 2023; pp. 1–6. [Google Scholar]
- Rane, N. ChatGPT and similar Generative Artificial Intelligence (AI) for building and construction industry: Contribution, Opportunities and Challenges of large language Models for Industry 4.0, Industry 5.0, and Society 5.0. Oppor. Chall. Large Lang. Model. Ind. 2023, 4. [Google Scholar] [CrossRef]
- Moosavi, S.; Farajzadeh-Zanjani, M.; Razavi-Far, R.; Palade, V.; Saif, M. Explainable AI in Manufacturing and Industrial Cyber–Physical Systems: A Survey. Electronics 2024, 13, 3497. [Google Scholar] [CrossRef]
- Beg, O.A.; Khan, A.A.; Rehman, W.U.; Hassan, A. A Review of AI-Based Cyber-Attack Detection and Mitigation in Microgrids. Energies 2023, 16, 7644. [Google Scholar] [CrossRef]
- Habib, G.; Qureshi, S. XAI and Machine Learning for Cyber Security: A Systematic Review. In Medical Data Analysis and Processing Using Explainable Artificial Intelligence; CRC Press: Boca Raton, FL, USA, 2023; pp. 91–104. [Google Scholar]
- Bac, T.P.; Ha, D.T.; Tran, K.D.; Tran, K.P. Explainable Articial Intelligence for Cybersecurity in Smart Manufacturing. In Artificial Intelligence for Smart Manufacturing: Methods, Applications, and Challenges; Springer: Cham, Switzerland, 2023; pp. 199–223. [Google Scholar]
- Ahmad, I.; Rodriguez, F.; Kumar, T.; Suomalainen, J.; Jagatheesaperumal, S.K.; Walter, S.; Asghar, M.Z.; Li, G.; Papakonstantinou, N.; Ylianttila, M.; et al. Communications security in Industry X: A survey. IEEE Open J. Commun. Soc. 2024, 5, 982–1025. [Google Scholar] [CrossRef]
- Bhattacharya, P.; Obaidat, M.S.; Sanghavi, S.; Sakariya, V.; Tanwar, S.; Hsiao, K.F. Internet-of-explainable-digital-twins: A case study of versatile corn production ecosystem. In Proceedings of the 2022 International Conference on Communications, Computing, Cybersecurity, and Informatics (CCCI), Dalian, China, 17–19 October 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–5. [Google Scholar]
- Minh, D.; Wang, H.X.; Li, Y.F.; Nguyen, T.N. Explainable artificial intelligence: A comprehensive review. Artif. Intell. Rev. 2022, 55, 3503–3568. [Google Scholar] [CrossRef]
- Buijsman, S. Defining explanation and explanatory depth in XAI. Minds Mach. 2022, 32, 563–584. [Google Scholar] [CrossRef]
- Dwivedi, R.; Dave, D.; Naik, H.; Singhal, S.; Omer, R.; Patel, P.; Qian, B.; Wen, Z.; Shah, T.; Morgan, G.; et al. Explainable AI (XAI): Core ideas, techniques, and solutions. ACM Comput. Surv. 2023, 55, 1–33. [Google Scholar] [CrossRef]
- Yang, W.; Wei, Y.; Wei, H.; Chen, Y.; Huang, G.; Li, X.; Li, R.; Yao, N.; Wang, X.; Gu, X.; et al. Survey on Explainable AI: From Approaches, Limitations and Applications Aspects. Hum.-Centric Intell. Syst. 2023, 3, 161–188. [Google Scholar] [CrossRef]
- Chandre, P.R.; Vanarote, V.; Patil, R.; Mahalle, P.N.; Shinde, G.R.; Nimbalkar, M.; Barot, J. Explainable AI for Intrusion Prevention: A Review of Techniques and Applications. In Proceedings of the International Conference on Information and Communication Technology for Intelligent Systems; Springer: Singapore, 2023; pp. 339–350. [Google Scholar]
- Moustafa, N.; Koroniotis, N.; Keshk, M.; Zomaya, A.Y.; Tari, Z. Explainable Intrusion Detection for Cyber Defences in the Internet of Things: Opportunities and Solutions. IEEE Commun. Surv. Tutor. 2023, 25, 1775–1807. [Google Scholar] [CrossRef]
- Ahmad, K.; Maabreh, M.; Ghaly, M.; Khan, K.; Qadir, J.; Al-Fuqaha, A. Developing future human-centered smart cities: Critical analysis of smart city security, Data management, and Ethical challenges. Comput. Sci. Rev. 2022, 43, 100452. [Google Scholar] [CrossRef]
- Nwakanma, C.I.; Ahakonye, L.A.C.; Njoku, J.N.; Odirichukwu, J.C.; Okolie, S.A.; Uzondu, C.; Ndubuisi Nweke, C.C.; Kim, D.S. Explainable artificial intelligence (xai) for intrusion detection and mitigation in intelligent connected vehicles: A review. Appl. Sci. 2023, 13, 1252. [Google Scholar] [CrossRef]
- Castro, O.E.L.; Deng, X.; Park, J.H. Comprehensive survey on AI-based technologies for enhancing IoT privacy and security: Trends, challenges, and solutions. Hum.-Centric Comput. Inf. Sci. 2023, 13, 39. [Google Scholar]
- Alsamhi, S.H.; Hawbani, A.; Sahal, R.; Srivastava, S.; Kumar, S.; Zhao, L.; Al-qaness, M.A.; Hassan, J.; Guizani, M.; Curry, E. Towards sustainable industry 4.0: A survey on greening IoE in 6G networks. Ad Hoc Netw. 2024, 165, 103610. [Google Scholar] [CrossRef]
- Sinclair, C.; Pierce, L.; Matzner, S. An application of machine learning to network intrusion detection. In Proceedings of the Proceedings 15th Annual Computer Security Applications Conference (ACSAC’99), Phoenix, AZ, USA, 6–10 December 1999; IEEE: Piscataway, NJ, USA, 1999; pp. 371–377. [Google Scholar]
- Mahbooba, B.; Timilsina, M.; Sahal, R.; Serrano, M. Explainable artificial intelligence (XAI) to enhance trust management in intrusion detection systems using decision tree model. Complexity 2021, 2021, 6634811. [Google Scholar] [CrossRef]
- Manoj, V.; Wenda, S.; Sihan, N.; Rouff, C.; Watkins, L.; Rubin, A. Explainable Autonomic Cybersecurity For Industrial Control Systems. In Proceedings of the 2023 IEEE 13th Annual Computing and Communication Workshop and Conference (CCWC), Vegas, NV, USA, 8–11 March 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 0900–9006. [Google Scholar]
- Fazzolari, M.; Ducange, P.; Marcelloni, F. An Explainable Intrusion Detection System for IoT Networks. In Proceedings of the 2023 IEEE International Conference on Fuzzy Systems (FUZZ), Songdo Incheon, Republic of Korea, 13–17 August 2023; pp. 1–6. [Google Scholar] [CrossRef]
- Yang, B.; Hoffman, M.; Brown, N.J. Bayesian Networks for Interpretable Cyberattack Detection. In Proceedings of the 56th Hawaii International Conference on System Sciences (HICSS-56), Online, 3–7 January 2023; p. 3. [Google Scholar]
- Zolanvari, M.; Yang, Z.; Khan, K.; Jain, R.; Meskin, N. TRUST XAI: Model-Agnostic Explanations for AI with a Case Study on IIoT Security. IEEE Internet Things J. 2021, 10, 2967–2978. [Google Scholar] [CrossRef]
- Le, T.T.H.; Kim, H.; Kang, H.; Kim, H. Classification and explanation for intrusion detection system based on ensemble trees and SHAP method. Sensors 2022, 22, 1154. [Google Scholar] [CrossRef] [PubMed]
- Alani, M.M.; Mashatan, A.; Miri, A. XMal: A lightweight memory-based explainable obfuscated-malware detector. Comput. Secur. 2023, 133, 103409. [Google Scholar] [CrossRef]
- Alani, M.M.; Miri, A. Towards an explainable universal feature set for IoT intrusion detection. Sensors 2022, 22, 5690. [Google Scholar] [CrossRef]
- Alani, M.M. An explainable efficient flow-based Industrial IoT intrusion detection system. Comput. Electr. Eng. 2023, 108, 108732. [Google Scholar] [CrossRef]
- Gürbüz, E.; Turgut, Ö.; Kök, İ. Explainable AI-Based Malicious Traffic Detection and Monitoring System in Next-Gen IoT Healthcare. In Proceedings of the 2023 International Conference on Smart Applications, Communications and Networking (SmartNets), Istanbul, Turkey, 25–27 July 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–6. [Google Scholar]
- Patil, S.; Varadarajan, V.; Mazhar, S.M.; Sahibzada, A.; Ahmed, N.; Sinha, O.; Kumar, S.; Shaw, K.; Kotecha, K. Explainable Artificial Intelligence for Intrusion Detection System. Electronics 2022, 11, 3079. [Google Scholar] [CrossRef]
- Zebin, T.; Rezvy, S.; Luo, Y. An explainable AI-based intrusion detection system for DNS over HTTPS (DoH) Attacks. IEEE Trans. Inf. Forensics Secur. 2022, 17, 2339–2349. [Google Scholar] [CrossRef]
- Sivamohan, S.; Sridhar, S. An optimized model for network intrusion detection systems in industry 4.0 using XAI based Bi-LSTM framework. Neural Comput. Appl. 2023, 35, 11459–11475. [Google Scholar] [CrossRef]
- Wang, Y.; Xu, L.; Liu, W.; Li, R.; Gu, J. Network intrusion detection based on explainable artificial intelligence. Wirel. Pers. Commun. 2023, 131, 1115–1130. [Google Scholar] [CrossRef]
- Tanuwidjaja, H.C.; Takahashi, T.; Lin, T.N.; Lee, B.; Ban, T. Hybrid Explainable Intrusion Detection System: Global vs. Local Approach. In Proceedings of the 2023 Workshop on Recent Advances in Resilient and Trustworthy ML Systems in Autonomous Networks, Copenhagen, Denmark, 30 November 2023; pp. 37–42. [Google Scholar]
- Mills, A.; Spyridopoulos, T.; Legg, P. Efficient and interpretable real-time malware detection using random-forest. In Proceedings of the 2019 International Conference on Cyber Situational Awareness, Data Analytics and Assessment (Cyber SA), Oxford, UK, 3-4 June 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–8. [Google Scholar]
- Ables, J.; Kirby, T.; Anderson, W.; Mittal, S.; Rahimi, S.; Banicescu, I.; Seale, M. Creating an Explainable Intrusion Detection System Using Self Organizing Maps. In Proceedings of the 2022 IEEE Symposium Series on Computational Intelligence (SSCI), Singapore, 4–7 December 2022; pp. 404–412. [Google Scholar] [CrossRef]
- Ables, J.; Kirby, T.; Mittal, S.; Banicescu, I.; Rahimi, S.; Anderson, W.; Seale, M. Explainable Intrusion Detection Systems Using Competitive Learning Techniques. arXiv 2023, arXiv:2303.17387. [Google Scholar] [CrossRef]
- Lundberg, H.; Mowla, N.I.; Abedin, S.F.; Thar, K.; Mahmood, A.; Gidlund, M.; Raza, S. Experimental Analysis of Trustworthy In-Vehicle Intrusion Detection System Using eXplainable Artificial Intelligence (XAI). IEEE Access 2022, 10, 102831–102841. [Google Scholar] [CrossRef]
- AL-Essa, M.; Andresini, G.; Appice, A.; Malerba, D. Xai to explore robustness of features in adversarial training for cybersecurity. In Proceedings of the International Symposium on Methodologies for Intelligent Systems, Cosenza, Italy, 3–5 October 2022; Springer: Cham, Switzerland, 2022; pp. 117–126. [Google Scholar]
- Al-Essa, M.; Andresini, G.; Appice, A.; Malerba, D. An XAI-based adversarial training approach for cyber-threat detection. In Proceedings of the 2022 IEEE International Conference on Dependable, Autonomic and Secure Computing (DASC), Pervasive Intelligence and Computing (PICom), Cloud & Big Data Computing (CBDCom) & Cyber Science & Technology Congress (CyberSciTech), Falerna, Calabria, Italy, 12–15 September 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–8. [Google Scholar]
- Lanfer, E.; Sylvester, S.; Aschenbruck, N.; Atzmueller, M. Leveraging Explainable AI Methods Towards Identifying Classification Issues on IDS Datasets. In Proceedings of the 2023 IEEE 48th Conference on Local Computer Networks (LCN), Daytona Beach, Fl, USA, 2–5 October 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–4. [Google Scholar]
- Sharma, B.; Sharma, L.; Lal, C.; Roy, S. Explainable artificial intelligence for intrusion detection in IoT networks: A deep learning based approach. Expert Syst. Appl. 2024, 238, 121751. [Google Scholar] [CrossRef]
- Lu, Z.; Thing, V.L.L. “How Does It Detect a Malicious App?” Explaining the Predictions of AI-Based Malware Detector. In Proceedings of the 2022 IEEE 8th International Conference on Big Data Security on Cloud (BigDataSecurity), High Performance and Smart Computing (HPSC), and Intelligent Data and Security (IDS), Jinan, China, 6–8 May 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 194–199. [Google Scholar]
- Sarhan, M.; Layeghy, S.; Portmann, M. Evaluating Standard Feature Sets Towards Increased Generalisability and Explainability of ML-Based Network Intrusion Detection. Big Data Res. 2022, 30, 100359. [Google Scholar] [CrossRef]
- Oseni, A.; Moustafa, N.; Creech, G.; Sohrabi, N.; Strelzoff, A.; Tari, Z.; Linkov, I. An Explainable Deep Learning Framework for Resilient Intrusion Detection in IoT-Enabled Transportation Networks. IEEE Trans. Intell. Transp. Syst. 2022, 24, 1000–1014. [Google Scholar] [CrossRef]
- Alani, M.M.; Awad, A.I.; Barka, E. ARP-PROBE: An ARP spoofing detector for Internet of Things networks using explainable deep learning. Internet Things 2023, 23, 100861. [Google Scholar] [CrossRef]
- Kalutharage, C.S.; Liu, X.; Chrysoulas, C.; Pitropakis, N.; Papadopoulos, P. Explainable AI-based DDOS attack identification method for IoT networks. Computers 2023, 12, 32. [Google Scholar] [CrossRef]
- Muna, R.K.; Hossain, M.I.; Alam, M.G.R.; Hassan, M.M.; Ianni, M.; Fortino, G. Demystifying machine learning models of massive IoT attack detection with Explainable AI for sustainable and secure future smart cities. Internet Things 2023, 24, 100919. [Google Scholar] [CrossRef]
- Abou El Houda, Z.; Brik, B.; Senouci, S.M. A novel iot-based explainable deep learning framework for intrusion detection systems. IEEE Internet Things Mag. 2022, 5, 20–23. [Google Scholar] [CrossRef]
- da Silveira Lopes, R.; Duarte, J.C.; Goldschmidt, R.R. False Positive Identification in Intrusion Detection Using XAI. IEEE Lat. Am. Trans. 2023, 21, 745–751. [Google Scholar] [CrossRef]
- Szczepański, M.; Choraś, M.; Pawlicki, M.; Kozik, R. Achieving explainability of intrusion detection system by hybrid oracle-explainer approach. In Proceedings of the 2020 International Joint Conference on neural networks (IJCNN), Glasgow, UK, 19–24 July 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–8. [Google Scholar]
- Wang, M.; Zheng, K.; Yang, Y.; Wang, X. An explainable machine learning framework for intrusion detection systems. IEEE Access 2020, 8, 73127–73141. [Google Scholar] [CrossRef]
- Nguyen, T.L.; Nguyen, X.H.; Le, K.H. Enhancing Explainability of Machine Learning-based Intrusion Detection Systems. In Proceedings of the 2022 RIVF International Conference on Computing and Communication Technologies (RIVF), Ho Chi Minh, Vietnam, 20–22 December 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 606–611. [Google Scholar]
- Siganos, M.; Radoglou-Grammatikis, P.; Kotsiuba, I.; Markakis, E.; Moscholios, I.; Goudos, S.; Sarigiannidis, P. Explainable AI-based Intrusion Detection in the Internet of Things. In Proceedings of the 18th International Conference on Availability, Reliability and Security, Benevento, Italy, 29 August–1 September 2023; ACM: New York, NY, USA, 2023. ARES ’23. [Google Scholar] [CrossRef]
- Roy, S.; Li, J.; Pandey, V.; Bai, Y. An Explainable Deep Neural Framework for Trustworthy Network Intrusion Detection. In Proceedings of the 2022 10th IEEE International Conference on Mobile Cloud Computing, Services, and Engineering (MobileCloud), San Francisco, CA, USA, 15–18 August 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 25–30. [Google Scholar]
- Das, S.; Shiva, S. Machine Learning application lifecycle augmented with explanation and security. In Proceedings of the 2021 IEEE 12th Annual Ubiquitous Computing, Electronics & Mobile Communication Conference (UEMCON), New York, NY, USA, 1–4 December 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 0171–0177. [Google Scholar]
- Mane, S.; Rao, D. Explaining network intrusion detection system using explainable AI framework. arXiv 2021, arXiv:2103.07110. [Google Scholar] [CrossRef]
- Khan, I.A.; Moustafa, N.; Pi, D.; Sallam, K.M.; Zomaya, A.Y.; Li, B. A New Explainable Deep Learning Framework for Cyber Threat Discovery in Industrial IoT Networks. IEEE Internet Things J. 2021, 9, 11604–11613. [Google Scholar] [CrossRef]
- Ha, D.T.; Hoang, N.X.; Hoang, N.V.; Du, N.H.; Huong, T.T.; Tran, K.P. Explainable anomaly detection for industrial control system cybersecurity. IFAC-PapersOnLine 2022, 55, 1183–1188. [Google Scholar] [CrossRef]
- Nguyen, Q.P.; Lim, K.W.; Divakaran, D.M.; Low, K.H.; Chan, M.C. GEE: A gradient-based explainable variational autoencoder for network anomaly detection. In Proceedings of the 2019 IEEE Conference on Communications and Network Security (CNS), Washington, DC, USA, 10–12 June 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 91–99. [Google Scholar]
- Antwarg, L.; Miller, R.M.; Shapira, B.; Rokach, L. Explaining anomalies detected by autoencoders using Shapley Additive Explanations. Expert Syst. Appl. 2021, 186, 115736. [Google Scholar] [CrossRef]
- Aguilar, D.L.; Medina-Perez, M.A.; Loyola-Gonzalez, O.; Choo, K.K.R.; Bucheli-Susarrey, E. Towards an interpretable autoencoder: A decision-tree-based autoencoder and its application in anomaly detection. IEEE Trans. Dependable Secur. Comput. 2022, 20, 1048–1059. [Google Scholar] [CrossRef]
- Lanvin, M.; Gimenez, P.F.; Han, Y.; Majorczyk, F.; Mé, L.; Totel, E. Towards Understanding Alerts raised by Unsupervised Network Intrusion Detection Systems. In Proceedings of the The 26th International Symposium on Research in Attacks, Intrusions and Defenses (RAID 2023), Hong Kong, China, 16–18 October 2023. [Google Scholar]
- Javeed, D.; Gao, T.; Kumar, P.; Jolfaei, A. An explainable and resilient intrusion detection system for industry 5.0. IEEE Trans. Consum. Electron. 2023, 70, 1342–1350. [Google Scholar] [CrossRef]
- Shoukat, S.; Gao, T.; Javeed, D.; Saeed, M.S.; Adil, M. Trust my IDS: An explainable AI integrated deep learning-based transparent threat detection system for industrial networks. Comput. Secur. 2025, 149, 104191. [Google Scholar] [CrossRef]
- Lin, Y.; Chang, X. Towards interpretable ensemble learning for image-based malware detection. arXiv 2021, arXiv:2101.04889. [Google Scholar] [CrossRef]
- Iadarola, G.; Martinelli, F.; Mercaldo, F.; Santone, A. Towards an interpretable deep learning model for mobile malware detection and family identification. Comput. Secur. 2021, 105, 102198. [Google Scholar] [CrossRef]
- Andresini, G.; Appice, A.; Caforio, F.P.; Malerba, D.; Vessio, G. ROULETTE: A neural attention multi-output model for explainable network intrusion detection. Expert Syst. Appl. 2022, 201, 117144. [Google Scholar] [CrossRef]
- Piplai, A.; Chukkapalli, S.S.L.; Joshi, A. NAttack! Adversarial Attacks to bypass a GAN based classifier trained to detect Network intrusion. In Proceedings of the 2020 IEEE 6th Intl Conference on Big Data Security on Cloud (BigDataSecurity), IEEE Intl Conference on High Performance and Smart Computing, (HPSC) and IEEE Intl Conference on Intelligent Data and Security (IDS), Baltimore, MD, USA, 25–27 May 2020; pp. 49–54. [Google Scholar] [CrossRef]
- Ayub, M.A.; Johnson, W.A.; Talbert, D.A.; Siraj, A. Model evasion attack on intrusion detection systems using adversarial machine learning. In Proceedings of the 2020 54th Annual Conference on Information Sciences and Systems (CISS), Princeton, NJ, USA, 18–20 March 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–6. [Google Scholar]
- Alshahrani, E.; Alghazzawi, D.; Alotaibi, R.; Rabie, O. Adversarial attacks against supervised machine learning based network intrusion detection systems. PLoS ONE 2022, 17, e0275971. [Google Scholar] [CrossRef]
- Duy, P.T.; Khoa, N.H.; Hien, D.T.T.; Do Hoang, H.; Pham, V.-H. Investigating on the robustness of flow-based intrusion detection system against adversarial samples using Generative Adversarial Networks. J. Inf. Secur. Appl. 2023, 74, 103472. [Google Scholar] [CrossRef]
- Zhang, S.; Xie, X.; Xu, Y. A brute-force black-box method to attack machine learning-based systems in cybersecurity. IEEE Access 2020, 8, 128250–128263. [Google Scholar] [CrossRef]
- Lan, T.; Demetrio, L.; Nait-Abdesselam, F.; Han, Y.; Aonzo, S. Trust Under Siege: Label Spoofing Attacks against Machine Learning for Android Malware Detection. arXiv 2025, arXiv:2503.11841. [Google Scholar]
- Qiu, H.; Dong, T.; Zhang, T.; Lu, J.; Memmi, G.; Qiu, M. Adversarial attacks against network intrusion detection in IoT systems. IEEE Internet Things J. 2020, 8, 10327–10335. [Google Scholar] [CrossRef]
- Chen, J.; Wu, D.; Zhao, Y.; Sharma, N.; Blumenstein, M.; Yu, S. Fooling intrusion detection systems using adversarially autoencoder. Digit. Commun. Netw. 2021, 7, 453–460. [Google Scholar] [CrossRef]
- Jiang, H.; Lin, J.; Kang, H. FGMD: A robust detector against adversarial attacks in the IoT network. Future Gener. Comput. Syst. 2022, 132, 194–210. [Google Scholar] [CrossRef]
- Ravikrishnan, B.; Sriram, I.; Mahadevan, S. ARDL-IDS: Adversarial Resilience in Deep Learning-based Intrusion Detection Systems. In Proceedings of the 2023 International Conference on Wireless Communications Signal Processing and Networking (WiSPNET), Chennai, India, 29–31 March 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–5. [Google Scholar]
- Li, S.; Wang, J.; Wang, Y.; Zhou, G.; Zhao, Y. EIFDAA: Evaluation of an IDS with function-discarding adversarial attacks in the IIoT. Heliyon 2023, 9, e13520. [Google Scholar] [CrossRef]
- Debicha, I.; Cochez, B.; Kenaza, T.; Debatty, T.; Dricot, J.M.; Mees, W. Adv-Bot: Realistic adversarial botnet attacks against network intrusion detection systems. Comput. Secur. 2023, 129, 103176. [Google Scholar] [CrossRef]
- Li, P.; Liu, Q.; Zhao, W.; Wang, D.; Wang, S. Bebp: An poisoning method against machine learning based idss. arXiv 2018, arXiv:1803.03965. [Google Scholar] [CrossRef]
- Xu, J.; Wen, Y.; Yang, C.; Meng, D. An approach for poisoning attacks against rnn-based cyber anomaly detection. In Proceedings of the 2020 IEEE 19th International Conference on Trust, Security and Privacy in Computing and Communications (TrustCom), Guangzhou, China, 29 December 2020–1 January 2021; IEEE: Piscataway, NJ, USA, 2020; pp. 1680–1687. [Google Scholar]
- Nguyen, T.D.; Rieger, P.; Miettinen, M.; Sadeghi, A.R. Poisoning Attacks on Federated Learning-Based IoT Intrusion Detection System. In Proceedings of the Workshop on Decentralized IoT Systems and Security (DISS), San Diego, CA, USA, 23–26 February 2020; pp. 1–7. Available online: https://www.ndss-symposium.org/wp-content/uploads/2020/04/diss2020-23003-paper.pdf (accessed on 30 August 2025).
- Rosenberg, I.; Meir, S.; Berrebi, J.; Gordon, I.; Sicard, G.; David, E.O. Generating end-to-end adversarial examples for malware classifiers using explainability. In Proceedings of the 2020 International Joint Conference on Neural Networks (IJCNN), Glasgow, UK, 19–24 July 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–10. [Google Scholar]
- Zhang, H.; Han, D.; Liu, Y.; Wang, Z.; Sun, J.; Zhuang, S.; Liu, J.; Dong, J. Explainable and Transferable Adversarial Attack for ML-Based Network Intrusion Detectors. arXiv 2024, arXiv:2401.10691. [Google Scholar] [CrossRef]
- Kuppa, A.; Le-Khac, N.A. Black box attacks on explainable artificial intelligence (XAI) methods in cyber security. In Proceedings of the 2020 International Joint Conference on Neural Networks (IJCNN), Glasgow, UK, 19–24 July 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–8. [Google Scholar]
- Kuppa, A.; Le-Khac, N.A. Adversarial XAI Methods in Cybersecurity. IEEE Trans. Inf. Forensics Secur. 2021, 16, 4924–4938. [Google Scholar] [CrossRef]
- Severi, G.; Meyer, J.; Coull, S.; Oprea, A. {Explanation-Guided} backdoor poisoning attacks against malware classifiers. In Proceedings of the 30th USENIX Security Symposium (USENIX Security 21), Online, 11–13 August 2021; pp. 1487–1504. [Google Scholar]
- Okada, S.; Jmila, H.; Akashi, K.; Mitsunaga, T.; Sekiya, Y.; Takase, H.; Blanc, G.; Nakamura, H. XAI-driven adversarial attacks on network intrusion detectors. In Proceedings of the 2024 European Interdisciplinary Cybersecurity Conference, Xanthi, Greece, 5–6 June 2024; pp. 65–73. [Google Scholar]
- Sun, R.; Guo, S.; Guo, J.; Xing, C.; Yang, L.; Guo, X.; Pan, Z. Instance Attack: An Explanation-based Vulnerability Analysis Framework Against DNNs for Malware Detection. arXiv 2022, arXiv:2209.02453. [Google Scholar] [CrossRef]
- Hoang, N.V.; Trung, N.D.; Trung, D.M.; Duy, P.T.; Pham, V.H. ADV-Sword: A Framework of Explainable AI-Guided Adversarial Samples Generation for Benchmarking ML-Based Intrusion Detection Systems. In Proceedings of the 2024 International Conference on Advanced Technologies for Communications (ATC), Ho Chi Minh City, Vietnam, 17–19 October 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 885–890. [Google Scholar]
- Alani, M.M.; Mashatan, A.; Miri, A. Adversarial Explainability: Utilizing Explainable Machine Learning in Bypassing IoT Botnet Detection Systems. arXiv 2023, arXiv:2310.00070. [Google Scholar] [CrossRef]
- Shu, Z.; Yan, G. EAGLE: Evasion attacks guided by local explanations against Android malware classification. IEEE Trans. Dependable Secur. Comput. 2023, 21, 3165–3182. [Google Scholar] [CrossRef]
- Page, M.J.; McKenzie, J.E.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. The PRISMA 2020 statement: An updated guideline for reporting systematic reviews. BMJ 2021, 372, n71. [Google Scholar] [CrossRef]
- Sarkar, A.; Vijaykeerthy, D.; Sarkar, A.; Balasubramanian, V.N. A Framework for Learning Ante-hoc Explainable Models via Concepts. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 10286–10295. [Google Scholar]
- Arrieta, A.B.; Díaz-Rodríguez, N.; Del Ser, J.; Bennetot, A.; Tabik, S.; Barbado, A.; García, S.; Gil-López, S.; Molina, D.; Benjamins, R.; et al. Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Inf. Fusion 2020, 58, 82–115. [Google Scholar] [CrossRef]
- Islam, S.R.; Eberle, W.; Ghafoor, S.K.; Ahmed, M. Explainable artificial intelligence approaches: A survey. arXiv 2021, arXiv:2101.09429. [Google Scholar] [CrossRef]
- Hanif, A.; Zhang, X.; Wood, S. A Survey on Explainable Artificial Intelligence Techniques and Challenges. In Proceedings of the 2021 IEEE 25th International Enterprise Distributed Object Computing Workshop (EDOCW), Gold Coast, Australia, 25–29 October 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 81–89. [Google Scholar]
- Došilović, F.K.; Brčić, M.; Hlupić, N. Explainable artificial intelligence: A survey. In Proceedings of the 2018 41st International Convention on Information and Communication Technology, Electronics and Microelectronics (MIPRO), Opatija, Croatia, 21–25 May 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 0210–0215. [Google Scholar]
- Liu, Q.; Hagenmeyer, V.; Keller, H.B. A review of rule learning-based intrusion detection systems and their prospects in smart grids. IEEE Access 2021, 9, 57542–57564. [Google Scholar] [CrossRef]
- van der Velden, B.H. Explainable AI: Current status and future potential. Eur. Radiol. 2023, 34, 1187–1189. [Google Scholar] [CrossRef] [PubMed]
- Kalasampath, K.; Spoorthi, K.; Sajeev, S.; Kuppa, S.S.; Ajay, K.; Angulakshmi, M. A Literature review on applications of explainable artificial intelligence (XAI). IEEE Access 2025, 13, 41111–41140. [Google Scholar] [CrossRef]
- Bobek, S.; Nalepa, G.J. Local universal rule-based explainer (lux). SoftwareX 2025, 30, 102102. [Google Scholar] [CrossRef]
- Vale, D.; El-Sharif, A.; Ali, M. Explainable artificial intelligence (XAI) post-hoc explainability methods: Risks and limitations in non-discrimination law. AI Ethics 2022, 2, 815–826. [Google Scholar] [CrossRef]
- Covert, I.; Lundberg, S.; Lee, S.I. Explaining by removing: A unified framework for model explanation. J. Mach. Learn. Res. 2021, 22, 1–90. [Google Scholar]
- Confalonieri, R.; Coba, L.; Wagner, B.; Besold, T.R. A historical perspective of explainable Artificial Intelligence. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2021, 11, e1391. [Google Scholar] [CrossRef]
- Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why Should I Trust You?”: Explaining the Predictions of Any Classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; ACM: New York, NY, USA, 2016. KDD ’16. pp. 1135–1144. [Google Scholar] [CrossRef]
- Lundberg, S.M.; Lee, S.I. A Unified Approach to Interpreting Model Predictions. In Proceedings of the Advances in Neural Information Processing Systems; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
- Alenezi, R.; Ludwig, S.A. Explainability of cybersecurity threats data using SHAP. In Proceedings of the 2021 IEEE Symposium Series on Computational Intelligence (SSCI), Virtual, 5–7 December 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–10. [Google Scholar]
- Belle, V.; Papantonis, I. Principles and practice of explainable machine learning. Front. Big Data 2021, 4, 688969. [Google Scholar] [CrossRef] [PubMed]
- Kim, S.; Jeong, M.; Ko, B.C. Lightweight surrogate random forest support for model simplification and feature relevance. Appl. Intell. 2022, 52, 471–481. [Google Scholar] [CrossRef]
- Tan, S.; Caruana, R.; Hooker, G.; Lou, Y. Distill-and-compare: Auditing black-box models using transparent model distillation. In Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society, New Orleans, LA, USA, 2–3 February 2018; pp. 303–310. [Google Scholar]
- Liu, S. Improved model search based on distillation framework. In Proceedings of the 2nd International Conference on Computer Vision, Image, and Deep Learning, Liuzhou, China, 25–27 June 2021; SPIE: Bellingham, WA, USA, 2021; Volume 11911, pp. 399–406. [Google Scholar]
- Johansson, U.; Konig, R.; Niklasson, L. Inconsistency-friend or foe. In Proceedings of the 2007 International Joint Conference on Neural Networks, Orlando, FL, USA, 12–17 August 2007; IEEE: Piscataway, NJ, USA, 2007; pp. 1383–1388. [Google Scholar]
- Konig, R.; Johansson, U.; Niklasson, L. G-REX: A versatile framework for evolutionary data mining. In Proceedings of the 2008 IEEE International Conference on Data Mining Workshops, Pisa, Italy, 15–19 December; IEEE: Piscataway, NJ, USA, 2008; pp. 971–974. [Google Scholar]
- Hassan, M.A.; Zardari, S.; Farooq, M.U.; Alansari, M.M.; Nagro, S.A. Systematic Analysis of Risks in Industry 5.0 Architecture. Appl. Sci. 2024, 14, 1466. [Google Scholar] [CrossRef]
- Rajabion, L. Industry 5.0 and Cyber Crime Security Threats. In Advanced Research and Real-World Applications of Industry 5.0; IGI Global: Hershey, PA, USA, 2023; pp. 66–76. [Google Scholar]
- Adel, A. Future of industry 5.0 in society: Human-centric solutions, challenges and prospective research areas. J. Cloud Comput. 2022, 11, 40. [Google Scholar] [CrossRef]
- Rjoub, G.; Bentahar, J.; Wahab, O.A.; Mizouni, R.; Song, A.; Cohen, R.; Otrok, H.; Mourad, A. A Survey on Explainable Artificial Intelligence for Cybersecurity. IEEE Trans. Netw. Serv. Manag. 2023, 20, 5115–5140. [Google Scholar] [CrossRef]
- Kiruthika, M.; Moorthi, K.; Devi, M.A.; Roseline, S.A. Role of XAI in building a super smart society 5.0. In XAI Based Intelligent Systems for Society 5.0; Elsevier: Amsterdam, The Netherlands, 2024; pp. 295–326. [Google Scholar]
- Khan, A.; Jhanjhi, N.Z.; Haji, D.H.T.B.A.; bin Haji Omar, H.A.H. The Need for Explainable AI in Industry 5.0. In Advances in Explainable AI Applications for Smart Cities; IGI Global: Hershey, PA, USA, 2024; pp. 1–30. [Google Scholar]
- Alsamhi, S.H.; Curry, E.; Hawbani, A.; Kumar, S.; Hassan, U.U.; Rajput, N.S. DataSpace in the Sky: A Novel Decentralized Framework to Secure Drones Data Sharing in B5G for Industry 4.0 toward Industry 5.0. Preprints 2023. [Google Scholar] [CrossRef]
- Alnajjar, I.A.; Almazaydeh, L.; Odeh, A.A.; Salameh, A.A.; Alqarni, K.; Ban Atta, A.A. Anomaly Detection Based on Hierarchical Federated Learning with Edge-Enabled Object Detection for Surveillance Systems in Industry 4.0 Scenario. Int. J. Intell. Eng. Syst. 2024, 17, 649–665. [Google Scholar] [CrossRef]
- Nascita, A.; Aceto, G.; Ciuonzo, D.; Montieri, A.; Persico, V.; Pescapé, A. A survey on explainable artificial intelligence for internet traffic classification and prediction, and intrusion detection. IEEE Commun. Surv. Tutor. 2024, 27, 3165–3198. [Google Scholar] [CrossRef]
- Eltomy, R.; Lalouani, W. Explainable Intrusion Detection in Industrial Control Systems. In Proceedings of the 2024 IEEE 7th International Conference on Industrial Cyber-Physical Systems (ICPS), St. Louis, MO, USA, 12–15 May 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1–8. [Google Scholar]
- Arisdakessian, S.; Wahab, O.A.; Mourad, A.; Otrok, H.; Guizani, M. A Survey on IoT Intrusion Detection: Federated Learning, Game Theory, Social Psychology, and Explainable AI as Future Directions. IEEE Internet Things J. 2023, 10, 4059–4092. [Google Scholar] [CrossRef]
- Liao, H.J.; Lin, C.H.R.; Lin, Y.C.; Tung, K.Y. Intrusion detection system: A comprehensive review. J. Netw. Comput. Appl. 2013, 36, 16–24. [Google Scholar] [CrossRef]
- Liu, H.; Lang, B. Machine learning and deep learning methods for intrusion detection systems: A survey. Appl. Sci. 2019, 9, 4396. [Google Scholar] [CrossRef]
- Ahmad, Z.; Shahid Khan, A.; Wai Shiang, C.; Abdullah, J.; Ahmad, F. Network intrusion detection system: A systematic study of machine learning and deep learning approaches. Trans. Emerg. Telecommun. Technol. 2021, 32, e4150. [Google Scholar] [CrossRef]
- Zhang, Z.; Hamadi, H.A.; Damiani, E.; Yeun, C.Y.; Taher, F. Explainable Artificial Intelligence Applications in Cyber Security: State-of-the-Art in Research. IEEE Access 2022, 10, 93104–93139. [Google Scholar] [CrossRef]
- Kaur, R.; Gabrijelčič, D.; Klobučar, T. Artificial intelligence for cybersecurity: Literature review and future research directions. Inf. Fusion 2023, 97, 101804. [Google Scholar] [CrossRef]
- Divekar, A.; Parekh, M.; Savla, V.; Mishra, R.; Shirole, M. Benchmarking datasets for Anomaly-based Network Intrusion Detection: KDD CUP 99 alternatives. In Proceedings of the 2018 IEEE 3rd International Conference on Computing, Communication and Security (ICCCS), Kathmandu, Nepal, 25–27 October 2018; pp. 1–8. [Google Scholar] [CrossRef]
- Lippmann, R.; Haines, J.W.; Fried, D.J.; Korba, J.; Das, K. The 1999 DARPA off-line intrusion detection evaluation. Comput. Netw. 2000, 34, 579–595. [Google Scholar] [CrossRef]
- Tavallaee, M.; Bagheri, E.; Lu, W.; Ghorbani, A.A. A detailed analysis of the KDD CUP 99 data set. In Proceedings of the 2009 IEEE Symposium on Computational Intelligence for Security and Defense Applications, Ottawa, ON, Canada, 8–10 July 2009; pp. 1–6. [Google Scholar] [CrossRef]
- Moustafa, N.; Slay, J. UNSW-NB15: A comprehensive data set for network intrusion detection systems (UNSW-NB15 network data set). In Proceedings of the 2015 Military Communications and Information Systems Conference (MilCIS), Canberra, Australia, 10–12 November 2015; pp. 1–6. [Google Scholar] [CrossRef]
- Faraj, O.; Megías, D.; Ahmad, A.M.; Garcia-Alfaro, J. Taxonomy and challenges in machine learning-based approaches to detect attacks in the internet of things. In Proceedings of the 15th International Conference on Availability, Reliability and Security, Virtual, 25–28 August 2020; pp. 1–10. [Google Scholar]
- Vigano, L.; Magazzeni, D. Explainable security. In Proceedings of the 2020 IEEE European Symposium on Security and Privacy Workshops (EuroS&PW), Genoa, Italy, 7–11 September 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 293–300. [Google Scholar]
- Islam, S.R.; Eberle, W. Domain Knowledge-Aided Explainable Artificial Intelligence. In Explainable Artificial Intelligence for Cyber Security: Next Generation Artificial Intelligence; Springer: Cham, Switzerland, 2022; pp. 73–92. [Google Scholar]
- Neupane, S.; Ables, J.; Anderson, W.; Mittal, S.; Rahimi, S.; Banicescu, I.; Seale, M. Explainable intrusion detection systems (x-ids): A survey of current methods, challenges, and opportunities. IEEE Access 2022, 10, 112392–112415. [Google Scholar] [CrossRef]
- Lipton, Z.C. The mythos of model interpretability: In machine learning, the concept of interpretability is both important and slippery. Queue 2018, 16, 31–57. [Google Scholar] [CrossRef]
- Hoenig, A.; Roy, K.; Acquaah, Y.; Yi, S.; Desai, S. Explainable AI for cyber-physical systems: Issues and challenges. IEEE Access 2024, 12, 73113–73140. [Google Scholar] [CrossRef]
- Pourret, O.; Na, P.; Marcot, B. Bayesian Networks: A Practical Guide to Applications; John Wiley & Sons: Hoboken, NJ, USA, 2008. [Google Scholar]
- Kohonen, T. Self-organized formation of topologically correct feature maps. Biol. Cybern. 1982, 43, 59–69. [Google Scholar] [CrossRef]
- Tcydenova, E.; Kim, T.W.; Lee, C.; Park, J.H. Detection of adversarial attacks in AI-based intrusion detection systems using explainable AI. Hum.-Centric Comput. Inf. Sci. 2021, 11, 35. [Google Scholar]
- Gaitan-Cardenas, M.C.; Abdelsalam, M.; Roy, K. Explainable AI-Based Intrusion Detection Systems for Cloud and IoT. In Proceedings of the 2023 32nd International Conference on Computer Communications and Networks (ICCCN), Honolulu, HI, USA, 24–27 July 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–7. [Google Scholar]
- Charte, D.; Charte, F.; del Jesus, M.J.; Herrera, F. An analysis on the use of autoencoders for representation learning: Fundamentals, learning task case studies, explainability and challenges. Neurocomputing 2020, 404, 93–107. [Google Scholar] [CrossRef]
- Yoon, M.K.; Mohan, S.; Choi, J.; Sha, L. Memory heat map: Anomaly detection in real-time embedded systems using memory behavior. In Proceedings of the 52nd Annual Design Automation Conference, San Francisco, CA, USA, 7–11 June 2015; pp. 1–6. [Google Scholar]
- Hariharan, S.; Velicheti, A.; Anagha, A.; Thomas, C.; Balakrishnan, N. Explainable artificial intelligence in cybersecurity: A brief review. In Proceedings of the 2021 4th International Conference on Security and Privacy (ISEA-ISAP), Dhanbad, India, 27–30 October 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–12. [Google Scholar]
- Charmet, F.; Tanuwidjaja, H.C.; Ayoubi, S.; Gimenez, P.F.; Han, Y.; Jmila, H.; Blanc, G.; Takahashi, T.; Zhang, Z. Explainable artificial intelligence for cybersecurity: A literature survey. Ann. Telecommun. 2022, 77, 789–812. [Google Scholar] [CrossRef]
- Duddu, V. A survey of adversarial machine learning in cyber warfare. Def. Sci. J. 2018, 68, 356. [Google Scholar] [CrossRef]
- Ling, X.; Wu, L.; Zhang, J.; Qu, Z.; Deng, W.; Chen, X.; Qian, Y.; Wu, C.; Ji, S.; Luo, T.; et al. Adversarial attacks against Windows PE malware detection: A survey of the state-of-the-art. Comput. Secur. 2023, 128, 103134. [Google Scholar] [CrossRef]
- Shokri, R.; Stronati, M.; Song, C.; Shmatikov, V. Membership inference attacks against machine learning models. In Proceedings of the 2017 IEEE Symposium on Security and Privacy (SP), San Jose, CA, USA, 22–26 May 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 3–18. [Google Scholar]
- Hall, M.; Harborne, D.; Tomsett, R.; Galetic, V.; Quintana-Amate, S.; Nottle, A.; Preece, A. A systematic method to understand requirements for explainable AI (XAI) systems. In Proceedings of the IJCAI Workshop on eXplainable Artificial Intelligence (XAI 2019), Macau, China, 11 August 2019; Volume 11. [Google Scholar]
- Mohseni, S.; Zarei, N.; Ragan, E.D. A multidisciplinary survey and framework for design and evaluation of explainable AI systems. ACM Trans. Interact. Intell. Syst. (TiiS) 2021, 11, 1–45. [Google Scholar] [CrossRef]
- Srivastava, G.; Jhaveri, R.H.; Bhattacharya, S.; Pandya, S.; Maddikunta, P.K.R.; Yenduri, G.; Hall, J.G.; Alazab, M.; Gadekallu, T.R. XAI for cybersecurity: State of the art, challenges, open issues and future directions. arXiv 2022, arXiv:2206.03585. [Google Scholar] [CrossRef]
- Chakraborty, A.; Alam, M.; Dey, V.; Chattopadhyay, A.; Mukhopadhyay, D. A survey on adversarial attacks and defences. CAAI Trans. Intell. Technol. 2021, 6, 25–45. [Google Scholar] [CrossRef]
- Sarker, I.H. Multi-aspects AI-based modeling and adversarial learning for cybersecurity intelligence and robustness: A comprehensive overview. Secur. Priv. 2023, 6, e295. [Google Scholar] [CrossRef]
- Apruzzese, G.; Colajanni, M.; Ferretti, L.; Marchetti, M. Addressing adversarial attacks against security systems based on machine learning. In Proceedings of the 2019 11th International Conference on Cyber Conflict (CyCon), Tallinn, Estonia, 28–31 May 2019; IEEE: Piscataway, NJ, USA, 2019; Volume 900, pp. 1–18. [Google Scholar]
- Alatwi, H.A.; Aldweesh, A. Adversarial black-box attacks against network intrusion detection systems: A survey. In Proceedings of the 2021 IEEE World AI IoT Congress (AIIoT), Virtual, 10–13 May 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 0034–0040. [Google Scholar]
- Apruzzese, G.; Andreolini, M.; Ferretti, L.; Marchetti, M.; Colajanni, M. Modeling realistic adversarial attacks against network intrusion detection systems. Digit. Threat. Res. Pract. (DTRAP) 2022, 3, 1–19. [Google Scholar] [CrossRef]
- Merzouk, M.A.; Cuppens, F.; Boulahia-Cuppens, N.; Yaich, R. A deeper analysis of adversarial examples in intrusion detection. In Proceedings of the Risks and Security of Internet and Systems: 15th International Conference, CRiSIS 2020, Paris, France, 4–6 November 2020; Revised Selected Papers 15. Springer: Cham, Switzerland, 2021; pp. 67–84. [Google Scholar]
- Dunmore, A.; Jang-Jaccard, J.; Sabrina, F.; Kwak, J. A Comprehensive Survey of Generative Adversarial Networks (GANs) in Cybersecurity Intrusion Detection. IEEE Access 2023, 11, 76071–76094. [Google Scholar] [CrossRef]
- Alkadi, S.; Al-Ahmadi, S.; Ismail, M.M.B. Better Safe Than Never: A Survey on Adversarial Machine Learning Applications towards IoT Environment. Appl. Sci. 2023, 13, 6001. [Google Scholar] [CrossRef]
- Pujari, M.; Cherukuri, B.P.; Javaid, A.Y.; Sun, W. An approach to improve the robustness of machine learning based intrusion detection system models against the carlini-wagner attack. In Proceedings of the 2022 IEEE International Conference on Cyber Security and Resilience (CSR), Virtual, 27–29 July 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 62–67. [Google Scholar]
- Merzouk, M.A.; Delas, J.; Neal, C.; Cuppens, F.; Boulahia-Cuppens, N.; Yaich, R. Evading Deep Reinforcement Learning-based Network Intrusion Detection with Adversarial Attacks. In Proceedings of the 17th International Conference on Availability, Reliability and Security, Vienna, Austria, 23–26 August 2022; pp. 1–6. [Google Scholar]
- He, K.; Kim, D.D.; Asghar, M.R. Adversarial machine learning for network intrusion detection systems: A comprehensive survey. IEEE Commun. Surv. Tutor. 2023, 25, 538–566. [Google Scholar] [CrossRef]
- McCarthy, A.; Ghadafi, E.; Andriotis, P.; Legg, P. Functionality-preserving adversarial machine learning for robust classification in cybersecurity and intrusion detection domains: A survey. J. Cybersecur. Priv. 2022, 2, 154–190. [Google Scholar] [CrossRef]
- Rehman, A.; Farrakh, A.; Khan, S. Explainable AI in Intrusion Detection Systems: Enhancing Transparency and Interpretability. Int. J. Adv. Sci. Comput. 2023, 2, 7–20. [Google Scholar]
- Nadeem, A.; Vos, D.; Cao, C.; Pajola, L.; Dieck, S.; Baumgartner, R.; Verwer, S. Sok: Explainable machine learning for computer security applications. In Proceedings of the 2023 IEEE 8th European Symposium on Security and Privacy (EuroS&P), Delft, The Netherlands, 3–7 July 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 221–240. [Google Scholar]







| Ref. | XAI | Cyber | Adv-XAI | I5.0 | Focus and Key Limitation |
|---|---|---|---|---|---|
| [23,25,26] | ✓ | ✓ | ✗ | ● | XAI role in Industry 5.0 digital transformation (e.g., AI, IoT, robotics, and cybersecurity). Limitation: General overview without adversarial XAI analysis or human-centric security workflows. |
| [37] | ✓ | ✓ | ✗ | ✗ | XAI applications across multiple domains (e.g., healthcare, finance, cybersecurity, and education). Limitation: Broad scope without depth in adversarial robustness or Industry 5.0 context. |
| [34,35,36] | ✓ | ✗ | ✗ | ✗ | XAI taxonomies, counterfactual explanations, and technique surveys. Limitation: General XAI methods without cybersecurity or IDS application. |
| [35] | ✓ | ✗ | ✗ | ✗ | Theoretical framework for XAI explanation design using counterfactuals. Limitation: No empirical application to cybersecurity or industrial contexts. |
| Ours | ✓ | ✓ | ✓ | ✓ | First systematic review integrating (1) XAI-based IDS taxonomy, (2) Industry 5.0 human-centric cybersecurity threats, (3) adversarial XAI exploitation (Adv-XIDS), (4) legacy vs. modern dataset analysis, and (5) federated learning and SOC workflow considerations for collaborative industrial environments. |
| Ref. | Threat Addressed | IDS Data Type | Dataset | Detection Model | XAI Algorithm | Explanation |
|---|---|---|---|---|---|---|
| [44] | Intrusion detection | Network-based IDS (NIDS) | Analyst-designed training sets from archived network events | Genetic algorithm, ID3 | Rule-based explainability | Generates rules for distinguishing normal network connections from an anomalous one, based on expert-domain knowledge. |
| [45] | DOS, R2L, U2R, and PROBING | Network-based IDS (NIDS) | KDD-99 | ID3 algorithm | Rule-based explainability | Generates rules by Rattle package in R and visualizing in exploratory plots. |
| [46,47] | DOS, R2L, U2R, and PROBING | Network-based IDS (NIDS) | SCADA VM, and N-BaIoT | DT, RF, GNB, SVM, and K-NN | Rule-based explainability | Generates rules by Tree nodes to visualize the decision-making process as an exploratory plot. |
| [48] | DOS, R2L, U2R, and PROBING | Host-based IDS (NIDS) | Tracer FIRE 9 (TF9) and Tracer FIRE 10 (TF10) | Bayesian network (BNs) | Rule-based explainability | Visualizes network graph of the Bayes’ Rule, where the relation of a single feature to the target variable is found via conditional probability tables (CPTs). |
| [49] | Industrial IoT Security | IIoT-based IDS | WUSTL-IIoT, NSL-KDD, and UNSW-NB15 | Artificial Neural Network (ANN) | Transparency Relying Upon Statistical Theory (TRUST) system | Employs mutual information for ranking variables and selects the most impactful ones on the ANN’s outputs, naming them as representatives of the classes. |
| [50] | Industrial IoT Security | IoT-based IDS | IoTID20, NF-BoT-IoT-v2, and NF-ToN-IoT-v2 | Ensemble Trees (DT and RF) | SHapley Additive exPlanations (SHAP) | The ensemble model’s outputs are plotted in the form of heatmaps and decision plots using SHAP explanation techniques. |
| [51] | Android devices Malware detection | Android application- based IDS | MalMem-2022, Drebin-215, Malgenome-215, and CICMalDroid2020 | RF, LR, DT, GNB, and XGB | SHAP | SHAP values are employed to interpret model predictions by highlighting the most influential features contributing to malware detection decisions. |
| [52,53] | Industrial IoT Security | IoT-based IDS | Aposemat IoT-23, IoTID20, NF-BoT-IoT-v2, NF-ToN-IoT-v2, and WUSTL-IIOT-2021 | RF, LR, DT, GNB, XGB, and SVM | SHAP | The study uses SHAP to provide a global interpretation of model behavior, identifying critical IoT traffic features influencing intrusion detection outcomes. |
| [54] | Malicious Traffic Detection in IoT Healthcare Networks | IoT-based IDS) | Intensive Care Unit (ICU) dataset | RF and DT | SHAP, LIME, ELI5, and Integrated Gradients (IGs) | Visualizes the contribution of each feature in the model’s decision using Shapash Monitor explanation interface. |
| [55] | Man-In-The-Middle (MITM), DoS, Mirai botnet, Port/OS scanning, and Host scanning | IoT-based IDS | CICIDS-2017 | Voting Classifier | Local Interpretable Model-agnostic Explanation (LIME) | Plots the contribution of each feature in the model’s decision using LIME. |
| [56] | DNS over HTTPS (DoH) attacks | Network-based IDS | CIRA-CIC-DoHBrw-2020 | Random Forest (RF) | SHAP | Highlights the features which are contributing to the underlying decision of the model using SHAP values. |
| [57] | Glastopt, Dionaea, Cowrie, Canarytokens, DoS, R2L, U2R, and Probe attacks | Network-based IDS (NIDS) | Honeypot and NSL-KDD datasets | Bidirectional Long Short-Term Memory (BiLSTM) | LIME and SHAP | Focuses on generating global and local faithful explanations by approximating the behavior of the BiLSTM model around a specific instance of interest. |
| [58] | DOS, R2L, U2R, and PROBING | Network-based IDS (NIDS) | KDD-99 | CNN-LSTM | LIME and SHAP | LIME mechanism enables the model to interpret each individual factor and their impact on output. A Decision Tree is generated from the top-most influential features, which are then visualized using SHAP interpretation. |
| [59] | DOS, R2L, U2R, and PROBING | Network-based IDS (NIDS) | Ton-IOT Windows | RF | LIME and SHAP | Employed three primary techniques—variable importance plot, individual value plot, and partial dependence plot—to explain the decision-making process of the RF model. |
| [60] | Malware detection | File Content Analysis | VX Vault- and Virus Share-based generated dataset | Random Forest classifier | Visualizing Decision Trees | Presents the Trees that had classified a process as malware or benign and the relevant decision nodes. |
| [61,62] | Web based, Brute Force, DoS, DDoS, Infiltration, Heartbleed, and Bot and Scan | Host-based and network-based IDSs | NSL-KDD and CIC-IDS-2017 | Population-based Self-Organizing Maps (POPSOM) implementation | Self-Organizing Map (SOM)-based X-IDS | Produces robust, explanatory visualizations of the SOM model and create accurate IDS predictions |
| [63] | DoS and ID fabrication | In-vehicle IDS (IV-IDS) | Survival Analysis Dataset for automobile IDS | Deep Neural Network (DNN) | Visualization-based Explanation, (VisExp) | A dual swarm plot is created to display normal Controller Area Network (CAN) traffic at the top and intruder’s traffic at the bottom based on SHAP- value distribution. |
| [64] | Adware, Banking malware, SMS malware, Riskware, Brute Force FTP, and DoS | File Content Analysis and Network-based IDS | MalDroid20, CIC-IDS2017 | Deep Neural Network (DNN) | DALEX framework | DALEX employs a permutation-based algorithm to find the significance of individual variables, enhancing DNN prediction performance. |
| [65] | Adware, Banking malware, SMS malware, Riskware, Brute Force FTP, DoS | File Content Analysis and Network-based IDS | MalDroid20 and CIC-IDS2017 | Deep Neural Network (DNN) | SHAP | Fine-tunes DNN prediction performance through adversarial training and XAI combination. |
| [66] | Denial-of-Service (DoS) and Probe attack types | Network-based IDS (NIDS) | NSL-KDD | Random Forest (RF) | SHAP | Utilizes SHAP beeswarm plots to visualize explanations of the target class individually. |
| [67] | IoT Network Security | IoT-based IDS | NSL-KDD and UNSW-NB15 | DNN and CNN | LIME and SHAP | A deep learning-based IDS employing DNN and CNN models for attack classification, with feature selection using a filter-based approach. Model explanations are generated using LIME for local interpretability and SHAP for global feature importance. |
| [68] | Android malware detection | File Content Analysis | DREBIN | SVM and BERT | Feature importance | Inspired by MPT, minimizes variance in prediction score changes and attribution values for impactful feature attribution. |
| [69] | Brute Force, Bot, DoS, DDoS, Infiltration, and Web attacks | Network-based IDS (NIDS) | CSE-CIC-IDS2018, ToN-IoT, and Bot-IoT | Multi-Layer Perceptron (MLP) and Random Forest (RF) | SHAP | Calculates Shapley values to assess feature contributions and identify key influencers in the dataset. |
| [70] | DoS and DDoS in IoT/IoV networks | Network-based IDS (NIDS) | ToN_IoT dataset | Deep Neural Network | Deep SHAP technique | Combines SHAP values from neural network parts via DeepLIFT’s multipliers for full network interpretation. |
| [71] | Command injection, DoS, Reconnaissance, and backdoors | Industrial IoT-based IDS (IIoT-IDS) | WUSTL-IIOT-2021 | Deep Neural Network (DNN) | SHAP | Uses DeepExplainer to provide insights into DeepIIoT’s decision making via SHAP values. |
| [72] | DDoS attacks on IoT and traditional networks | IoT-based IDS (IoT-IDS) | USB-IDS dataset | Fully connected autoencoder with RELU | Kernel SHAP | Identifies top-R features contributing to reconstruction errors using SHAP values. |
| [73] | Industrial IoT Security | IoT-based IDS (IoT-IDS) | IoTID20 dataset | XG-Boost | LIME, TreeSHAP, and ELI5 | LIME explains contributions, SHAP combines importance and effects, and ELI5 reveals weights. |
| [74] | IoT-network security | IoT-based IDS (IoT-IDS) | UNSW-NB15 | DNN | RuleFit and SHAP | Calculates feature importance values for the decision model. |
| [10,75] | DOS | Network-based IDS (NIDS) | NSL-KDD99, LYCOS-IDS2017 | Linear Model (LM) and Multi-Layer Perceptron (MLP) | SHAP and Adversarial ML | Generates visual explanations for misclassifications, identifying responsible features. |
| [76] | DDoS, XSS, and SQL Injection attacks | Anomaly-based IDS (AIDS) | CICIDS2017 | ANN with PCA | Decision Trees with microaggregation | Uses dtreeviz to plot tree structure and highlight key features in predictions. |
| [77] | DOS, R2L, and U2R | Network-based IDS (NIDS) | NSL-KDD | One-versus-all classifier and multiclass classifier | SHAP | Combines local and global explanations to enhance IDS interpretation. |
| [78] | DOS, R2L, and U2R | Network-based IDS (NIDS) | KDD99 and CICIDS2017 | CNN and DT | SHAP | Combines local and global explanations to improve IDS interpretation. |
| [79] | IoT-network security | Network-based IDS (NIDS) | CIC-IoT-Dataset-2022 | DNN | SHAP | Combines local and global explanations to improve IDS interpretation. |
| [80,81] | DDoS, XSS, and SQL Injection attacks | Network-based IDS (NIDS) | KDD99 and CICIDS2017 | Deep Neural Network (DNN) and ensemble models | SHAP and LIME | Generates model-centric and subject-centric explanations from DNN predictions. |
| [82] | Anomaly detection | Network-based IDS (NIDS) | NSL-KDD | Deep Neural Network (DNN) | SHAP, BRCG, LIME, ProtoDash, and CEM | Plots SHAP values, extracts rules with BRCG, generates local explanations with LIME, summarizes data with ProtoDash, and calculates minimal perturbations with CEM. |
| [83] | Data injection and poisoning in Industrial IoT | Anomaly-based IDS (AIDS) | Real-world GSP time-series data | Conv-LSTM-based autoencoder | LIME | Illustrates relevant attributes and weights for interpretation. |
| [84] | Industrial control system anomaly detection | Anomaly-based IDS (AIDS) | SCADA dataset | LSTM-based autoencoder | SHAP | Visualizes feature influence on model output globally. |
| [85] | Low-rate DoS, Port scanning, Botnet, Spam, and Blacklist | Cyclostationarity-based network IDS (NIDS) | UGR’16 dataset | Variational autoencoder (VAE) framework | Gradient-based explanation | The interpretability of variational autoencoders is generated by utilizing gradients for clustering anomalies and deriving attack-related fingerprints. |
| [86] | Anomaly detection | Anomaly-based IDS (AIDS) | Warranty claims, KDD Cup 1999, Credit Card Fraud Detection, and artificial dataset | Autoencoder framework | Kernel SHAP | Computes SHAP values for reconstructed features and links them to true anomalous input values to explain prediction errors. |
| [87] | Anomaly detection | Anomaly-based IDS (AIDS) | UCI Machine Learning Repository | Decision Tree-based autoencoder | Rule-based explainability | The correlation values among different categorical attributes provide explanations behind the Decision Tree. |
| [88] | DDoS, XSS and SQL Injection attacks | Network-based IDS (NIDS) | CICIDS2017 | Sec2Graph technique | Explanation based on AE-pvalues | Explanation about the anomaly alert is produced by using the p-value of the empirical distribution of the dimension-wise reconstruction error to flag abnormal feature values. |
| [89] | DDoS | Network-based IDS (NIDS) | CICIDDoS2019 | BiLSTM + BiGRU + CNN | SHAP | Uses SHAP decision graphs including decision plots, Waterfall Plots, and Summary Plots to demonstrate the important features that contributed the most to detection. |
| [55] | Network Intrusion Detection | ML-based IDS | CICIDS-2017 | DT, RF, SVM, and Voting Classifier | LIME | An ensemble-based IDS combining Decision Tree, Random Forest, and SVM models with a Voting Classifier to enhance detection accuracy and reduce false positives. LIME is applied to interpret model predictions and enhance trust in the black-box ensemble system. |
| [90] | Mirai/Gafgyt botnets, DoS/DDoS, SQL Injection, and backdoors | Network-based IDS (NIDS) | N-BaIoT, Edge-IIoTset, and CIC-IDS2017 | LSTM-AutoEncoder for encoding; Attention-based GRU with softmax for multiclass classification | SHAP | Feature attribution scores for each prediction; highlights key traffic features responsible for malicious activity classification, improving SOC trust and traceability. |
| [91] | Malware detection | File Content Analysis | Malimg dataset | Selective Deep Ensemble Learning-based (SDEL) detector | Ensemble Deep Taylor Decomposition (EDTD) | EDTD converts the SDEL prediction into a heatmap, where brighter pixels indicate the most suspicious parts in the malware binary image. |
| [92] | Mobile malware detection | File Content Analysis | Android Malware Dataset (Argus Lab) | Convolutional Neural Network (CNN) | Grad-CAM | Generates heatmap for visualizing the predictions made by image-based CNN mode. |
| [93] | DoS, Probe, R2L, U2R, Fuzzers, Analysis, backdoors, Exploits, Generic, Reconnaissance, Shellcode, and worms | Network-based IDS (NIDS) | NSL-KDD and UNSW-NB15 | Convolutional Neural Network (CNN) | Attention mechanism of ROULETTE | Explainability involves utilizing the attention weights generated by the neural model to provide insights into the classification decisions made by the model for network traffic data. |
| Ref. | Data Type | Dataset | Attack Type | Detection Model | XAI Targeted/No XAI |
|---|---|---|---|---|---|
| [94] | Network event logs | IEEE BigData 2019 Cup: Suspicious Network Event Recognition | Perturbation | GAN | ✗ |
| [95] | Network-based | CICIDS 2017 and TRAbID 2017 | Perturbation | MLP | ✗ |
| [96] | Network-based | CICIDS2017 | Evasion | DT and LR | ✗ |
| [97] | Network-based | CICIDS2018 and InSDN | Evasion | DT, LR, CNN, MLP, and LSTM | ✗ |
| [98] | Host-based, network-based, and application-based | ADFA-LD, NSL-KDD, and DREBIN | Perturbation | DT, LR, MLP, NB, and RF | ✗ |
| [99] | Android APKs | 40K samples (20,769 benign from Google Play) | Label Spoofing | LSVM, GBT, NN, and RF | ✗ |
| [100] | IoT network-based | Mirai, Falsifying Video streaming application | Perturbation | DNN-based autoencoder | ✗ |
| [101] | Network-based | NSL-KDD, UNSW-NB15, and CICIDS2017 | Evasion | Autoencoder | ✗ |
| [102] | IoT Network-based | MedBIoT and IoTID | Perturbation | LSTM and RNN | ✗ |
| [103] | Network-based | KDDCup’99 | Evasion | DNN | ✗ |
| [104] | IoT network-based | X-IIoTID | Evasion | SVM, DT, RF, KNN, CNN, GRU, and HyDL-IDS | ✗ |
| [105] | Network-based | CTU-13 and CSE-CIC-IS2018D | Evasion | MLP, RF, and KNN | ✗ |
| [106] | Host-based | KDDCUP99, NSL-KDD, and Kyoto 2006+ | Poisoning | NB-Gaussian, LR, and SVM-sigmoid | ✗ |
| [107] | Network-based | LANL network security dataset | Poisoning | LSTM, B-LSTM, and T-LSTM | ✗ |
| [108] | Network-based | D¨IoT-Benign, UNSW-Benign, and D¨IoT-Attack | Poisoning | Federated Learning-based DNN | ✗ |
| [109] | Portable Executable (PE Files) | EMBER | Adversarial examples | GBDT | Integrated Gradients, DeepLIFT, and Layer-wise Relevance Propagation (LRP) |
| [110] | Network-based IDS | CIC-IDS2017 and Kitsune | Adversarial examples | MLP, AlertNet, IDSNet, DeepNet, RF, Xgboost, Multi-attribute Markov Probability Fingerprints (MaMPF), Flow Sequence Network (FS-Net), KitNET, and Diff-RF | SAGE (Shapley Additive Global Explanation) |
| [111] | Network-based and PE Files | Malicious/benign PDF files, Android apps, and UGR16 | Perturbation | MLP and adversarial autoencoder | Gradient-based XAI |
| [112] | Network-based and PE Files | Leaked Password, CICIDS17, and VirusShare | Evasion, oracle, and poisoning | Autoencoder and Gradient Boosting Model (GBM), Neural Network (NN) | Latent counterfactual, permute attack, and diverse counterfactual |
| [109] | PE Files | Ember (1 M samples) | Feature modification | GBDT | Integrated Gradients, DeepLIFT, -LRP, and SHAP |
| [113] | PE Files | EMBER, Contagio (PDFs), and Drebin (Android executables) | Evasion, oracle, and poisoning | Autoencoder, Gradient Boosting Model (GBM), and Neural Network (NN) | Latent counterfactual, permute attack, and diverse counterfactual |
| [114] | Network-based | CIC-IDS2017 and TON_IoT | Perturbation | DNN (Feedforward NN) | Integrated Gradients and KernelSHAP |
| [115] | PE Files | Microsoft Malware classification Challenge | Evasion | Deep Neural Network (DNN) | Superpixels |
| [116] | Network-based | InSDN | SHAP-guided evasion with AMM | LightGBM, RF, and CNN | SHAP |
| [117] | Network-based | IoT network intrusion dataset | Evasion | Extreme Gradient Boosting (XGB) | SHAP |
| [118] | PE Files | Drebin (Android executables) | Evasion | Random Forest (RF) and Multi-Layer Perceptron (MLP) | LIME |
| XAI Technique | Type | Model | Primary Applications | Vuln. | Documented Exploitations | Protection Mechanisms |
|---|---|---|---|---|---|---|
| Rule-based Explanations (ID3, DT, and BN) | Global | Inherently interpretable models | Policy compliance, expert knowledge integration, and audit trails | Low | Rule extraction and logic manipulation | Rule obfuscation, ensemble rules, and expert validation |
| SHAP (all variants) | Local + global | Model-agnostic | Feature importance, attack attribution, and pattern analysis | Very high | Feature manipulation, transferability attacks, and SAGE exploitation | Explanation randomization, input validation, and ensemble explanations |
| LIME | Local | Model-agnostic | Individual alert validation, incident forensics, and false-positive analysis | High | Local perturbation, feature evasion, and Android malware evasion | Instance validation, surrogate diversity, and perturbation bounds |
| Integrated Gradients | Local | Deep learning | Deep model interpretation, malware analysis, and healthcare IoT | Very high | Gradient manipulation, adversarial examples, and malware evasion | Gradient masking, defensive distillation, and input smoothing |
| Gradient-based methods | Local | Deep learning | Anomaly clustering, attack fingerprinting, and VAE interpretation | High | Gradient-based XAI exploitation and Manifold Manipulation | Gradient noise injection and multi-path gradients |
| Feature importance (Gini and permutation) | Global | Tree-based and ensemble | Model debugging, system optimization, and strategic threat analysis | Medium | Importance ranking manipulation and feature masking | Permutation testing, cross-validation, and feature redundancy |
| Attention mechanisms | Local + global | Neural Networks | Sequence analysis, traffic classification, and multiclass detection | Medium | Attention weight manipulation focus redirection | Attention regularization, multi-head validation, and attention dropout |
| Visualization techniques (Grad-CAM and Saliency) | Local | CNN and image-based | Malware visualization, binary analysis, and spatial feature mapping | Medium | Visual manipulation, Superpixels exploitation, and heatmap distortion | Multi-view validation, statistical verification, and ensemble visualization |
| Autoencoder-based (AE-pvalues and reconstruction) | Local + global | Autoencoders | Anomaly detection, reconstruction error analysis, and Industrial IoT | High | Latent space manipulation, reconstruction evasion, and counterfactual attacks | Latent space regularization, ensemble autoencoders, and adversarial training |
| Hybrid techniques (TRUST, DALEX, and RuleFit) | Global | Model-specific | Statistical analysis, Industrial IoT Security, and multi-modal explanation | Low–medium | Statistical manipulation and component-wise attacks | Statistical robustness testing, component isolation, and hybrid validation |
| Advanced methods (CEM, ProtoDash, and BRCG) | Local | Model-agnostic | Contrastive analysis, data summarization, and rule generation | Medium | Contrastive manipulation and summary poisoning | Robustness constraints, multi-method validation, and outlier detection |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Khan, N.; Ahmad, K.; Al Tamimi, A.; Alani, M.M.; Bermak, A.; Khalil, I. Explainable AI-Based Intrusion Detection Systems for Industry 5.0 and Adversarial XAI: A Systematic Review. Information 2025, 16, 1036. https://doi.org/10.3390/info16121036
Khan N, Ahmad K, Al Tamimi A, Alani MM, Bermak A, Khalil I. Explainable AI-Based Intrusion Detection Systems for Industry 5.0 and Adversarial XAI: A Systematic Review. Information. 2025; 16(12):1036. https://doi.org/10.3390/info16121036
Chicago/Turabian StyleKhan, Naseem, Kashif Ahmad, Aref Al Tamimi, Mohammed M. Alani, Amine Bermak, and Issa Khalil. 2025. "Explainable AI-Based Intrusion Detection Systems for Industry 5.0 and Adversarial XAI: A Systematic Review" Information 16, no. 12: 1036. https://doi.org/10.3390/info16121036
APA StyleKhan, N., Ahmad, K., Al Tamimi, A., Alani, M. M., Bermak, A., & Khalil, I. (2025). Explainable AI-Based Intrusion Detection Systems for Industry 5.0 and Adversarial XAI: A Systematic Review. Information, 16(12), 1036. https://doi.org/10.3390/info16121036

