Next Article in Journal
Recent Advancement in Fluorescent Probes for Peroxynitrite (ONOO)
Previous Article in Journal
A Preliminary Study for Ultraviolet Optical Methyl Salicylate Monitoring in Agriculture
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

Vision-Language Model-Based Local Interpretable Model-Agnostic Explanations Analysis for Explainable In-Vehicle Controller Area Network Intrusion Detection

1
School of Electrical Engineering, Korea University, Seoul 02841, Republic of Korea
2
Department of Data Science, Duksung Women’s University, Seoul 01370, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(10), 3020; https://doi.org/10.3390/s25103020 (registering DOI)
Submission received: 7 April 2025 / Revised: 8 May 2025 / Accepted: 9 May 2025 / Published: 10 May 2025
(This article belongs to the Special Issue AI-Based Intrusion Detection Techniques for Vehicle Networks)

Abstract

The Controller Area Network (CAN) facilitates efficient communication among vehicle components. While it ensures fast and reliable data transmission, its lightweight design makes it susceptible to data manipulation in the absence of security layers. To address these vulnerabilities, machine learning (ML)-based intrusion detection systems (IDS) have been developed and shown to be effective in identifying anomalous CAN traffic. However, these models often function as black boxes, offering limited transparency into their decision-making processes, which hinders trust in safety-critical environments. To overcome these limitations, this paper proposes a novel method that combines Local Interpretable Model-agnostic Explanations (LIME) with a vision-language model (VLM) to generate detailed textual interpretations of an ML-based CAN IDS. This integration mitigates the challenges of visual-only explanations in traditional XAI and enhances the intuitiveness of IDS outputs. By leveraging the multimodal reasoning capabilities of VLMs, the proposed method bridges the gap between visual and textual interpretability. The method supports both global and local explanations by analyzing feature importance with LIME and translating results into human-readable narratives via VLM. Experiments using a publicly available CAN intrusion detection dataset demonstrate that the proposed method provides coherent, text-based explanations, thereby improving interpretability and end-user trust.
Keywords: vision-language model; controller area network; anomaly detection; vehicle intrusion detection; explainable artificial intelligence; local interpretable model-agnostic explanation vision-language model; controller area network; anomaly detection; vehicle intrusion detection; explainable artificial intelligence; local interpretable model-agnostic explanation

Share and Cite

MDPI and ACS Style

Lee, J.; Rew, J. Vision-Language Model-Based Local Interpretable Model-Agnostic Explanations Analysis for Explainable In-Vehicle Controller Area Network Intrusion Detection. Sensors 2025, 25, 3020. https://doi.org/10.3390/s25103020

AMA Style

Lee J, Rew J. Vision-Language Model-Based Local Interpretable Model-Agnostic Explanations Analysis for Explainable In-Vehicle Controller Area Network Intrusion Detection. Sensors. 2025; 25(10):3020. https://doi.org/10.3390/s25103020

Chicago/Turabian Style

Lee, Jaeseung, and Jehyeok Rew. 2025. "Vision-Language Model-Based Local Interpretable Model-Agnostic Explanations Analysis for Explainable In-Vehicle Controller Area Network Intrusion Detection" Sensors 25, no. 10: 3020. https://doi.org/10.3390/s25103020

APA Style

Lee, J., & Rew, J. (2025). Vision-Language Model-Based Local Interpretable Model-Agnostic Explanations Analysis for Explainable In-Vehicle Controller Area Network Intrusion Detection. Sensors, 25(10), 3020. https://doi.org/10.3390/s25103020

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop