This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Open AccessArticle
Vision-Language Model-Based Local Interpretable Model-Agnostic Explanations Analysis for Explainable In-Vehicle Controller Area Network Intrusion Detection
by
Jaeseung Lee
Jaeseung Lee 1
and
Jehyeok Rew
Jehyeok Rew 2,*
1
School of Electrical Engineering, Korea University, Seoul 02841, Republic of Korea
2
Department of Data Science, Duksung Women’s University, Seoul 01370, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(10), 3020; https://doi.org/10.3390/s25103020 (registering DOI)
Submission received: 7 April 2025
/
Revised: 8 May 2025
/
Accepted: 9 May 2025
/
Published: 10 May 2025
Abstract
The Controller Area Network (CAN) facilitates efficient communication among vehicle components. While it ensures fast and reliable data transmission, its lightweight design makes it susceptible to data manipulation in the absence of security layers. To address these vulnerabilities, machine learning (ML)-based intrusion detection systems (IDS) have been developed and shown to be effective in identifying anomalous CAN traffic. However, these models often function as black boxes, offering limited transparency into their decision-making processes, which hinders trust in safety-critical environments. To overcome these limitations, this paper proposes a novel method that combines Local Interpretable Model-agnostic Explanations (LIME) with a vision-language model (VLM) to generate detailed textual interpretations of an ML-based CAN IDS. This integration mitigates the challenges of visual-only explanations in traditional XAI and enhances the intuitiveness of IDS outputs. By leveraging the multimodal reasoning capabilities of VLMs, the proposed method bridges the gap between visual and textual interpretability. The method supports both global and local explanations by analyzing feature importance with LIME and translating results into human-readable narratives via VLM. Experiments using a publicly available CAN intrusion detection dataset demonstrate that the proposed method provides coherent, text-based explanations, thereby improving interpretability and end-user trust.
Share and Cite
MDPI and ACS Style
Lee, J.; Rew, J.
Vision-Language Model-Based Local Interpretable Model-Agnostic Explanations Analysis for Explainable In-Vehicle Controller Area Network Intrusion Detection. Sensors 2025, 25, 3020.
https://doi.org/10.3390/s25103020
AMA Style
Lee J, Rew J.
Vision-Language Model-Based Local Interpretable Model-Agnostic Explanations Analysis for Explainable In-Vehicle Controller Area Network Intrusion Detection. Sensors. 2025; 25(10):3020.
https://doi.org/10.3390/s25103020
Chicago/Turabian Style
Lee, Jaeseung, and Jehyeok Rew.
2025. "Vision-Language Model-Based Local Interpretable Model-Agnostic Explanations Analysis for Explainable In-Vehicle Controller Area Network Intrusion Detection" Sensors 25, no. 10: 3020.
https://doi.org/10.3390/s25103020
APA Style
Lee, J., & Rew, J.
(2025). Vision-Language Model-Based Local Interpretable Model-Agnostic Explanations Analysis for Explainable In-Vehicle Controller Area Network Intrusion Detection. Sensors, 25(10), 3020.
https://doi.org/10.3390/s25103020
Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details
here.
Article Metrics
Article Access Statistics
For more information on the journal statistics, click
here.
Multiple requests from the same IP address are counted as one view.