Next Article in Journal
LW-YOLO11: A Lightweight Arbitrary-Oriented Ship Detection Method Based on Improved YOLO11
Previous Article in Journal
Research on Intelligent Monitoring and Protection Equipment of Vital Signs of Underground Personnel in Coal Mines: Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cockpit-Llama: Driver Intent Prediction in Intelligent Cockpit via Large Language Model

1
College of Automotive Engineering, Jilin University, Changchun 130025, China
2
Graduate School of Information and Science Technology, The University of Tokyo, Tokyo 113-8654, Japan
3
National Key Laboratory of Automotive Chassis Integration and Bionics, Jilin University, Changchun 130025, China
*
Authors to whom correspondence should be addressed.
Sensors 2025, 25(1), 64; https://doi.org/10.3390/s25010064
Submission received: 24 October 2024 / Revised: 29 November 2024 / Accepted: 24 December 2024 / Published: 25 December 2024
(This article belongs to the Section Vehicular Sensing)

Abstract

:
The cockpit is evolving from passive, reactive interaction toward proactive, cognitive interaction, making precise predictions of driver intent a key factor in enhancing proactive interaction experiences. This paper introduces Cockpit-Llama, a novel language model specifically designed for predicting driver behavior intent. Cockpit-Llama predicts driver intent based on the relationship between current driver actions, historical interactions, and the states of the driver and cockpit environment, thereby supporting further proactive interaction decisions. To improve the accuracy and rationality of Cockpit-Llama’s predictions, we construct a new multi-attribute cockpit dataset that includes extensive historical interactions and multi-attribute states, such as driver emotional states, driving activity scenarios, vehicle motion states, body states and external environment, to support the fine-tuning of Cockpit-Llama. During fine-tuning, we adopt the Low-Rank Adaptation (LoRA) method to efficiently optimize the parameters of the Llama3-8b-Instruct model, significantly reducing training costs. Extensive experiments on the multi-attribute cockpit dataset demonstrate that Cockpit-Llama’s prediction performance surpasses other advanced methods, achieving BLEU-4, ROUGE-1, ROUGE-2, and ROUGE-L scores of 71.32, 80.01, 76.89, and 81.42, respectively, with relative improvements of 92.34%, 183.61%, 95.54%, and 201.27% compared to ChatGPT-4. This significantly enhances the reasoning and interpretative capabilities of intelligent cockpits.

1. Introduction

Automotive intelligent cockpits are equipped with advanced hardware and software systems, providing human–machine interaction, network services, and scenario expansion capabilities that integrate humans, machines, and the environment. These cockpits offer a comprehensive experience for drivers and passengers, including safety, intelligence, efficiency, and comfort. Based on the type of functionality provided, the interactive functions of intelligent cockpits can be divided into three dimensions: driving control, riding comfort, and infotainment. The driving control domain focuses on improving intelligent vehicle driving quality, the riding comfort domain aims to enhance the passenger experience, and the infotainment domain provides value-added information services and entertainment. According to the level of intelligence, the evolution of intelligent cockpits from “traditional function realization” to “comprehensive intelligent cognition” is divided into five levels (L0 to L4): Functional Cockpit, Perception Intelligent Cockpit, Partially Cognitive Intelligent Cockpit, Highly Cognitive Intelligent Cockpit, and Fully Cognitive Intelligent Cockpit [1]. Among these, proactive interaction capabilities are the key to advancing cockpit intelligence. Its working mechanism generally involves: processing and analyzing data by the intelligent system, predicting the functions the user might need in the current scenario, determining the composition and importance of these functions, and then interacting with the user through appropriate means to provide services [2]. In this entire interaction process, intent prediction serves as the system’s core “brain”, and accurate prediction of user functional needs forms the basis of proactive cockpit interaction.
Human–cockpit interaction describes how humans interact with the cockpit, which can be divided into explicit and implicit interactions [3]. Explicit interaction involves users consciously and clearly expressing their intentions and needs, characterized by certainty and directness. Examples include voice-based interaction [4,5,6,7], haptic-based interaction [8,9,10,11], and gesture-based interaction [12,13,14], etc. Implicit interaction, on the other hand, focuses on non-direct user interactions, detecting and interpreting non-specific behaviors such as body movements, physiological signals, facial expressions, and environmental factors to intelligently understand user states and needs and respond autonomously [15]. Achieving higher-level intelligence relies on accurately understanding implicit interaction intentions. However, the aforementioned implicit intention inference methods face two limitations. First, traditional rule-based or data-driven lightweight model approaches lack scalability and comprehension capabilities, only applicable to independent, single-task intent prediction. In the cockpit environment, driving intentions, comfort, and information interactions are closely interwoven. Driver operations involve not only basic driving tasks but also controlling systems such as air conditioning, music, and navigation to meet comfort and information needs. Cockpits encompass dozens of different types of intentions, such as adjusting seat positions, changing air conditioning temperatures, making calls, playing music, or interacting with the navigation system during driving. These actions require coordination between multiple intelligent systems, making it difficult for single-task approaches to support high-level intelligence. Second, the reasoning and decision-making process of the entire system lacks transparency, resembling a “black box” operation, and lacks human social context knowledge, making it challenging to fully analyze human behavior and difficult for humans to understand.
The rapid development of large language models (LLMs) brings hope for addressing these challenges. LLMs possess strong generalization and common-sense reasoning abilities, allowing them to infer unseen information. Since they are pre-trained on extensive interdisciplinary data, including social interactions and human behavior, LLMs have prior common knowledge about the human world, potentially bridging the social intelligence gap in behavior prediction systems. LLMs have been used for predictive tasks in various fields, such as predicting future vehicle behaviors in autonomous driving, including lane changes, turns, and braking [2,16,17,18]. Compared to autonomous driving, cockpit intentions are more complex, encompassing three domains—driving control, riding comfort, and infotainment—with dozens or even nearly a hundred specific intentions. Moreover, most intention prediction works are still independent single tasks, while driver intentions are interrelated and holistic. Single-task predictions lead to inaccurate intention forecasts, a lack of coordination between multiple intentions, and potential conflicts.
This paper further proposes a comprehensive multi-task intention prediction method for intelligent cockpits based on LLMs, named Cockpit-Llama. Its prediction scope covers dozens of intentions in driving control, riding comfort, and infotainment. As LLMs are pre-trained on broad interdisciplinary data, including social interactions and human behavior, they have prior common knowledge about the human world. Utilizing this general knowledge, Cockpit-Llama achieves accurate multi-functional intention prediction within the cockpit. To ensure the rationality and accuracy of Cockpit-Llama’s predictions, we construct a dedicated multi-attribute cockpit dataset for cockpit intention prediction to fine-tune the model. During fine-tuning, adapter methods were used to optimize model parameters and reduce training costs. The main contributions of this work can be summarized as follows:
(1) Dataset Construction: A novel dataset is constructed containing approximately 60,000 sequences of multi-attribute states (including driver emotional states, driving activity scenes, vehicle motion states, body states, multimedia states and external environment) and driver historical interaction information, comprehensively reflecting the relationship between the driver’s future intentions, historical interaction sequences, and driving context.
(2) Novel Prediction Model: A novel prediction model named Cockpit-Llama is proposed, comprising three main components: modal data encoder, LLM backbone, and behavior intent decoder, specifically designed for the intention prediction task in intelligent cockpits. The model is fine-tuned on the constructed dataset. During fine-tuning, the LoRA adapter method is used to efficiently optimize model parameters, significantly reducing training costs. Cockpit-Llama demonstrates excellent performance in intent prediction across various scenarios (such as suburb drive, commute work, etc.), showcasing its strong generalization and reasoning capabilities.
(3) Performance Evaluation: Extensive experimental evaluation demonstrates that Cockpit-Llama outperforms large language models such as ChatGPT-4 and ChatGLM-3 in driver behavior prediction, achieving significant performance improvements. Compared to ChatGPT-4, BLEU-4 improved by approximately 92.34%, ROUGE-1 improved by 183.61%, ROUGE-2 by 95.54%, and ROUGE-L by 201.27%.

2. Related Works

2.1. Intelligent Cockpit Intent Prediction

Currently, intelligent cockpit intent prediction technology has been applied to varying degrees in driving control, riding comfort, and infotainment. In driving control, interaction research mainly focuses on predicting driver safety intentions derived from cognitive states. By monitoring driver facial expressions, eye movement, heart rate, and driving behavior changes, models can infer whether the driver is distracted or fatigued, allowing for timely safety interventions. Zhang et al. [19] used binary logistic regression and Fisher discriminant analysis to predict normal and distracted driving for safety warnings. Tran et al. [20] employed convolutional neural networks (CNN) to predict the driver’s distraction and fatigue states, effectively adapting to changes in lighting and shadows. Fasanmade et al. [21] developed a dynamic Bayesian (DDB) distraction severity classification model based on fuzzy logic, which predicts the severity of driver distraction by inputting the driver’s facial direction, activities, hand movements, and previous distraction history, enabling timely preventive measures. Riaz et al. [22] proposed a fuzzy logic-based driver distraction assessment system that calculates the driver’s level of distraction using fuzzy set algorithms. Additionally, they developed an emotion-enabled cognitive driving assistance model (EECDAM) to detect the driver’s emotional intensity and support the safety of driving decisions. Teyeb et al. [23] designed a multi-variable system based on fuzzy logic to infer driver vigilance levels through eye blinking analysis and head posture estimation, Xiao et al. [24] proposed a transfer learning-based method that utilizes driver facial images to recognize emotions in on-road environments.
Research in the riding comfort domain mainly focuses on emotion-driven proactive soothing interactions and intention prediction for vehicle component adjustments. In emotion-driven proactive soothing interactions, Li et al. [25], proposed using convolutional techniques to take driver facial expressions and cognitive process features as inputs for predicting driver emotions, thereby achieving affective human–machine interaction. Oehl et al. [26] proposed using self-assessment models to predict driver emotions based on the force exerted by the driver on the steering wheel. In vehicle component adjustment intention prediction, Hu et al. [27] proposed a dual-model coupling method, combining a habitual temperature prediction model and a time-series temperature prediction model, to predict air conditioning settings in real time.
In the infotainment domain, current research includes situational awareness, navigation intent prediction and music recommendation. Ferenc et al. [28] explored the effectiveness of fuzzy logic in context awareness prediction, using a set of 14 key prediction factors, including decision time, significance, eye-related indicators, and driver experience, to achieve accurate context awareness assessment. Li et al. [29] constructed a fuzzy inference system (FIS), using point-of-interest (POI) data from navigation maps as semantic cues to effectively infer various driving environments around the vehicle, such as shopping areas, tourist spots, public stations, service areas, and safety zones. Ye et al. [30] used power-law distribution (PD) to model user check-in behavior and proposed a collaborative point of interest (POI) algorithm based on geographical influence for prediction and recommendation using naive Bayes (NB). Sayak et al. [31] proposed a music recommendation method using content-based filtering and K-means clustering. However, these methods have inherent limitations. On one hand, the entire reasoning process resembles a “black box”, lacking intuitive interpretability and transparency. Additionally, they lack human social background knowledge, making it challenging to comprehensively analyze human behavior effectively, and difficult for humans to understand.

2.2. Large Language Models

Currently, LLMs have made rapid advancements, demonstrating significant potential in simulating human intelligence. These models are trained on vast amounts of internet data to understand and generate human-like text, showcasing outstanding performance in natural language processing. One of the most remarkable features of LLMs is their emerging capabilities, such as in-context learning (ICL), following instructions, and chain-of-thought (CoT) reasoning. Generative Pre-trained Transformers (GPT) [32] represents a pioneering effort, proposing the use of GPT to solve text understanding and generation tasks. Subsequent versions, GPT-3.5 and GPT-4 [33], have also demonstrated impressive conversational and reasoning abilities. Recently released LLMs, such as PaLM [34], Vicuna [35], LLaMA and LLaMA 2 [36,37], generate corresponding textual feedback based on human-followed instructions, enhancing the instruction-following capability of LLMs.
LLMs have been applied in multiple fields, including predicting the future behavior of autonomous vehicles, such as lane changes, turns, and braking. Zhao et al. [16] proposed an LLM model for autonomous driving decision-making, capable of predicting vehicle paths and performing motion planning accordingly. Xu et al. [17] introduced a multi-modal LLM-based end-to-end autonomous driving system, which can predict turning behaviors in an interpretable manner. Compared to autonomous driving, intent prediction in the cockpit is more complex, covering three domains: driving control, ride comfort, and infotainment, with dozens or even nearly a hundred specific intentions. In the cockpit, LLMs are currently mainly applied to single tasks, such as the HiVeGPT [38], which supports dialogue and facilitates interaction between human drivers and driving assistance systems. However, these intent predictions remain independent single tasks, whereas driver intent is interrelated and holistic. This may lead to inaccurate intent predictions, lack of coordination between multiple intents, and potential conflicts.

3. Methodology

This section proposes an LLM-based comprehensive multi-task intent prediction method for intelligent cockpits, Cockpit-Llama, which takes vehicle-centered human–vehicle interaction behaviors and multi-attribute states as inputs, and outputs the driver’s future intent. Figure 1 shows an overview of Cockpit-Llama, which mainly consists of three parts: multi-attribute cockpit dataset construction, Cockpit-Llama model architecture design and model fine-tuning strategy.

3.1. Multi-Attribute Cockpit Dataset Construction

Due to the scarcity of cockpit data, publicly available proprietary datasets for driver behavior prediction are almost nonexistent. Therefore, we specifically construct a novel multi-attribute cockpit dataset for driver behavior intent prediction. The participants include five laboratory members and fifteen recruited individuals. Data collection is conducted over the course of one month with five experimental vehicles, and each participant conducts at least one week of natural driving experiments. During the driving process, we collect a large amount of human–vehicle interaction data through embedded points in event tracking technology and vehicle status data via the Controller Area Network (CAN) bus. Additionally, after each trip, we infer the driver’s emotional state based on feedback from the driver and emotional records provided by the co-driver. By integrating and processing these multi-source data, we ultimately construct a large-scale novel cockpit dataset. This dataset includes human-cockpit historical interactions and multi-attribute states (including driver emotional states, driving activity scenarios, vehicle motion states, body states, multimedia states and external environment), which are used for fine-tuning Cockpit-Llama. In this study, we design a toolkit to pre-process, augment, pair the data, and package it into a text format for model training.

3.1.1. Sensor Data Pre-Processing

Our goal is to process raw sensor data into the desired large-scale cockpit dataset. We deploy various sensors on five vehicles, including temperature sensors, humidity sensors, weather sensors, etc., to collect a large amount of raw sensor data. We also design a toolkit containing a set of functional functions to extract and integrate information about the cockpit, environment, and driver interactions from the raw sensor data, transforming them into model-learnable types. Specifically, we use the state extraction function F s ( · ) from the toolkit to extract sensor data about the cockpit and the environment. By processing this data, we obtain information about the status of various in-cabin devices and the surrounding environment of the vehicle. Additionally, we use the driver behavior detection function F a ( · ) from the toolkit to extract the corresponding driver interaction data.

3.1.2. Data Augmentation

Due to limitations in technology and other conditions, the collected state data and behavior data often fail to comprehensively reflect the complexity of the driving environment. In other words, the existing behavioral attributes and state attributes are insufficient to fully characterize cockpit behaviors and their contexts. Therefore, in the research of cockpit data, the transformation between state and behavior is crucial to enhance data attributes and scale. In this work, we further processed the pre-processed data to perform data augmentation.
  • State Attribute to Behavior Attribute Transformation. The existing behavior attributes mainly reflect the driver’s usage of cockpit infotainment systems, such as radio, music, calls, navigation, and functional settings, lacking the capture of behaviors like air conditioning adjustment, door operations, window adjustment, and seat adjustment. To capture these behaviors, we construct a state-to-behavior attribute transformation algorithm, F s to a ( · ) . The specific process of transforming seat status into behavior is detailed in Algorithm 1. First, the algorithm initializes the sensor data δ t o b s at the current time t o b s and the sensor data δ t o b s 1 at the previous time t o b s 1 (line 1). Then it traverses all state data and uses the state-to-behavior function F s to a ( · ) from the toolkit to extract the corresponding seat heating states s b o d , t o b s s e a t and s b o d , t o b s 1 s e a t from the sensor data δ t o b s at present time t o b s and the sensor data δ t o b s 1 at previous time t o b s 1 (lines 2 to 3). The value of s b o d , t o b s s e a t can be 0, 1, 2, or 3, representing seat heating off, level one, level two, and level three heating, respectively. Then, further judge s b o d , t o b s s e a t and s b o d , t o b s 1 s e a t , when s b o d , t o b s s e a t and s b o d , t o b s 1 s e a t are different, return the corresponding behavior value h s e a t , t o b s (lines 4 to 20). By analyzing the transformation of various state information within the cockpit through this algorithm, driver behaviors can be inferred, enriching the behavior data.
  • Behavior Attribute to State Attribute Transformation. Similar to the transformation from state attributes to behavior attributes, the existing state attributes are mainly limited to vehicle driving states and vehicle body states, lacking descriptions of the in-cabin device states, which can be inferred from behavior attributes. For instance, behaviors such as starting music playback indicate that the music state is on, while pausing music playback or ending the driving journey represents the music state being off. Therefore, when capturing the driver’s behavior of starting music playback, it can be transformed into the action of turning on the music player. Similarly, when capturing the driver’s behavior of stopping music playback, other common state attributes that can be inferred include radio, phone, and navigation driving scenes. The specific construction method is detailed in Algorithm 2. First, initialize the detected driver behavior H (line 1), then traverse all the behavior data H to extract multimedia-related behaviors (lines 2 to 3). Next, judge the corresponding value of h m u l , t o b s and return the state s m u l , t o b s corresponding to h m u l , t o b s (lines 4 to 16). Algorithm 2 can convert cockpit behavior attributes into state attributes. By analyzing driver behavior data, the usage status of these devices can be inferred, enriching the state attributes.
Algorithm 1 Seat heating state to action algorithm
Input: present sensor data δ t o b s , previous sensor data δ t o b s 1
Output: The Action h t o b s + 1
1:
Initialize δ t o b s , δ t o b s 1
2:
for  t 0 in N do
3:
     δ Extract s b o d , t o b s s e a t δ t o b s , s b o d , t o b s 1 s e a t δ t o b s 1
4:
    if  s b o d , t o b s s e a t = 0  then
5:
        if  s b o d , t o b s 1 s e a t 0  then
6:
            h b o d , t o b s : = T u r n o f f t h e s e a t h e a t i n g
7:
        end if
8:
    else if  s b o d , t o b s s e a t = 1  then
9:
        if  s b o d , t o b s 1 s e a t 1  then
10:
            h b o d , t o b s : = A d j u s t s e a t h e a t i n g l e v e l t o 1
11:
        end if
12:
    else if  s b o d , t o b s s e a t = 2  then
13:
        if  s b o d , t o b s 1 s e a t 2  then
14:
            h b o d , t o b s : = A d j u s t s e a t h e a t i n g l e v e l t o 2
15:
        end if
16:
    else if  s b o d , t o b s s e a t = 3  then
17:
        if  s b o d , t o b s 1 s e a t 3  then
18:
            h b o d , t o b s : = A d j u s t s e a t h e a t i n g l e v e l t o 3
19:
        end if
20:
    end if
21:
end for
Algorithm 2 Action to state algorithm
Input: Action data H
Output: State s m u l , t o b s
1:
Initialize H
2:
for  t 0 in N do
3:
     Δ Extract h m u l , t o b s H
4:
    if  h m u l , t o b s = P l a y t h e m u s i c " or C o n t i n u e t h e m u s i c "  then
5:
         s m u l , t o b s : = M u s i c O n
6:
    else if  h m u l , t o b s = S t o p t h e m u s i c "  then
7:
         s m u l , t o b s : = n M u s i c O f f "
8:
    else if  h m u l , t o b s = P l a y t h e r a d i o then
9:
         s m u l , t o b s : = R a d i o O n
10:
    else if  h m u l , t o b s = S t o p t h e r a d i o then
11:
         s m u l , t o b s : = R a d i o O f f
12:
    else if  h m u l , t o b s = E n t e r t h e N a v i g a t i o n then
13:
         s m u l , t o b s : = N a v i g a t i o n O n
14:
    else
15:
         s m u l , t o b s : = N a v i g a t i o n O f f
16:
    end if
17:
end for

3.1.3. Dataset Text Generation

In this work, the pre-processed data and augmented data are integrated to form a global feature set X T , which consists of the multi-attribute states set s t o b s and historical behavior sequences h ( t o b s j ) : t o b s of the driver at time j:
X T = { s t o b s , h ( t o b s j ) : t o b s }
where behavior sequence h ( t obs j ) : t obs = { h t obs j , h t o b s j + 1 , , h t o b s } records the specific contents of driver’s current interaction and the previous j historical interactions. Multi-attribute states set s t obs includes the driver emotional states s emo , driving activity scenarios s sce , vehicle motion states s dri , body states s bod , multimedia states s mul and external environment s env :
s t obs = { s emo , s sce , s dri , s bod , s mul , s env }
Driver emotional states s emo include seven types: anger, disgust, happiness, fear, sadness, surprise, and neutral:
s e m o = { s emo anger , s emo disgust , s emo happiness , s emo fear , s emo sadness , s emo surprise , s emo neutral }
Driving activity scenarios s sce sre divided into six types: daily scenario, weekend travel scenario, far driving scenario, suburb driving scenario, commute work scenario, and commute home scenario.
s s c e = { s sce daily , s sce week , s sce suberb , s sce far , s sce work , s sce home }
Vehicle motion states s dri include vehicle speed, steering angle, braking, etc:
s dri = { s dri speed , s dri yaw , s dri brake }
Vehicle body states s bod include air conditioning status, window openness, sunroof openness, door status, ambient light status, seat status, cabin temperature, etc:
s b o d = { s bod air , s bod window , s bod sunroof , s bod door , s bod light , s bod seat , s bod temp }
where air conditioning status values s bod air include air conditioning mode, fan speed, and temperature. Seat status values s bod seat include seat adjustment angle and seat heating level.
  • Multimedia states s mul include music playback status, radio playback status, and navigation status:
s mul = { s mul music , s mul radio , s mul navigation }
External environment s env includes external temperature, ambient light, and rainfall status:
s env = { s env temp , s env s unlight , s env rainfall }
To enable the model to produce higher-quality answers to prediction tasks, we also carefully design text instructions X I constitutes the input for constructing the cockpit instruction-following dataset. Finally, we combine the textual instructions X I and the global dataset X T content to form the model’s input dataset, which contains approximately 60k entries. As shown in Figure 2, we present an example sample from a multi-attribute scenario dataset. Moreover, we provide statistics on some representative behaviors from the dataset, as shown in Figure 3.

3.2. Cockpit-Llama Model Architecture Design

As shown in Figure 4, the Cockpit-Llama model architecture mainly consists of three major components: modal text data encoding, lightweight fine-tuning for the LLM backbone, and driver behavior intent decoding. First, a text tokenizer is used to convert the text instructions into a format that the LLM can process, which is then input into the LLM for processing. Next, to ensure that Cockpit-Llama can precisely adapt to the current task, an adapter method is used to fine-tune the model, with adapters added to the LLM backbone. This adapter fine-tuning method enhances the LLM’s ability to understand and process text data, enabling it to generate effective prediction labels. It also optimizes the parameter size of the LLM, significantly reducing training costs while maintaining high performance and efficiency. Finally, the prediction labels generated by the LLM are decoded using a text detokenizer. To improve the efficiency of extracting inference results, driver behavior intents are decoded hierarchically, and during the decoding process, the predicted driver behavior intents are output in text format.

3.2.1. Modal Text Data Encoding

The input to Cockpit-Llama consists of the global feature set X T and a piece of instruction text X l used to guide the large language model. The text-tokenizer F token ( · ) is used to convert all the contents of our constructed multi-modal dataset into corresponding text token R L :
R L = F t o k e n ( X T X I )
where R L belongs to the text embedding space, which can be understood and processed by the LLM backbone.

3.2.2. Lightweight Fine-Tuning of LLM Backbone Using Adapters

In this work, we use LoRA as an adapter to perform lightweight fine-tuning of the LLM backbone as shown in Figure 5. The core idea of the LoRA fine-tuning method is to dynamically adjust the model weights by adding low-rank matrices to the pre-trained model’s weight matrices. This fine-tuning method allows effective adjustment of the model’s performance for specific tasks using only a small number of additional parameters, while keeping the pre-trained model’s weight parameters unchanged.
We use Llama3-Instruct-8b as the LLM backbone, which consists of 32 transformer layers and has excellent semantic understanding and logical reasoning capabilities. In this work, LoRA is added to each linear layer of Llama3-8B-Instruct, including the multi-head attention (MHA) layer and feed-forward network (FFN) layer. Input tokens R L first enter the MHA layer. In this layer, LoRA is added to each attention matrix, which includes the query matrix Q L ,key matrix K L ,value matrix V L .LoRA is a bypass module containing a down-projection matrix A and an up-projection matrix B. During fine-tuning, both the input and output dimensions of the MHA layer remain unchanged, and the linear transformation matrices of each attention matrix are frozen. Only the additionally introduced down-projection matrix A and up-projection matrix B are trained, where A is initialized using a random Gaussian distribution and B is initialized with a zero matrix. The entire MHA mechanism computation process is as follows:
MultiHead ( Q L , K L , V L ) = Concat ( head 1 , . . . . . . , head h ) W 0
head h = Attn ( Q L , K L , V L ) = softmax ( Q L K L T d k )
w h e r e Q L = ( W h q + B h q A h q ) Y L K L = ( W h k + B h k A h k ) Y L V L = ( W h v + B h v A h v ) Y L
where W h q R ( d m d k ) , W h k R ( d m d k ) , W h ν R ( d m d k ) are the linear transformation matrices for each attention head Q L , K L , V L . A h q R ( d llama , r ) , A h k R ( d llama , r ) are the corresponding down-projection matrices. B h q R ( r , d k ) , B h k R ( r , d k ) , B h v R ( r , d k ) are the corresponding up-projection matrices. W 0 represents the transformed output matrix. Then, after a residual connection and layer normalization, we obtain E L :
E L = RMSNorm ( H L + MultiHead ( Q L , K L , V L ) )
Next, E L passes through the feed-forward network layer rFNN ( · ) , which includes two fully connected layers and an activation function SwiGLU ( · ) . LoRA is added to both fully connected layers, using the same training method as above: freezing the weight matrices of the two fully connected layers and training only the down-projection matrix A and up-projection matrix B. After another residual connection and layer normalization, we obtain the output of this transformation layer H ˜ L :
H ˜ L = RMSNorm ( E L + rFFN ( E L ) )
where H ˜ L serves as input tokens for next transformer layer and above operations are repeated. Finally, after passing through L = 32 layers of transformers with added LoRA, predicted tokens Z L of LLM backbone are obtained:
Z L = L = 1 32 ( T r a n s f o m e r ( H ˜ L ) ) L
where Z L contains all reasoning information for the current task.

3.2.3. Driver Behavior Intent Decoding

After obtaining the predicted token Z L generated by the LLM backbone, Llama3’s default detokenizer f μ ( · ) is used to decode Z L into human-readable language. This decoding method improves the transparency of Cockpit-Llama throughout the entire prediction process while ensuring the accuracy of the prediction results. The entire decoding process is represented as follows:
h t o b s + 1 = f μ ( Z L )
where, h t o b s + 1 represents the driver’s interaction behavior in the next frame, and is output in natural language form.

3.3. Model Fine-Tuning Strategy

Cockpit-Llama is fine-tuned using the instruction dataset constructed in this work to achieve driver behavior prediction. Considering that Cockpit-Llama’s output is in natural language, cross-entropy loss L ( θ ) is used to supervise the model’s inference output. The cross-entropy loss is defined as follows:
L ( θ ) = i N p θ ( h t o b s + l ) l o g ( p θ ( h t o b s + l X T X I ) )
By minimizing this cross-entropy loss function, we can find the optimal parameters, allowing the model to better interpret or fit the data. As shown in the Figure 6, the cross-entropy curve during Cockpit-Llama fine-tuning converges around a loss value of 0.02.
During fine-tuning, to significantly reduce training costs, the LoRA method was used to effectively optimize the model parameter size while keeping the weights of the image tokenizer and the LLM backbone fixed. By minimizing cross-entropy loss, the LoRA weights were updated, enabling Cockpit-Llama to accurately infer future driver behavior.

4. Experiment

This section evaluates the performance of Cockpit-Llama through extensive experiments on the multi-attribute cockpit dataset. First, the experimental setup and evaluation metrics are introduced. Next, this method is compared with other baseline methods, demonstrating its superior performance in intent prediction. This work also designs a comparative experiment involving in-context learning and freeze fine-tuning to highlight the necessity of model fine-tuning and the efficiency of LoRA fine-tuning.

4.1. Experimental Setup

4.1.1. Implementation Details

We split the entire cockpit dataset into a 4:1 ratio for the training and testing sets, fine-tuning the model on the training set and evaluating its performance on the testing set. To ensure a fair comparison with baseline methods, the models compared against Cockpit-Llama in the experiments were evaluated using the same test set. The initial learning rate was set to 5 × 10 5 and gradually decayed to zero using cosine annealing. Considering the large number of parameters in the model, the LoRA fine-tuning method was used to optimize parameter size and reduce training costs. The rank r of the down-projection matrix was set to 8, targeting all linear layers in the LLM backbone of Cockpit-Llama. We performed 3 epochs of fine-tuning on four L20 (48 GB) GPUs, with a batch size of 2 for each GPU. The entire fine-tuning process was conducted on these GPUs.

4.1.2. Evaluation Metrics

To quantitatively evaluate Cockpit-Llama’s performance in intent prediction, different evaluation metrics are used. For the intent prediction task, this study uses BLEU-4, Rouge-N, and Rouge-L as evaluation metrics. These are commonly used automatic evaluation metrics for assessing the similarity and quality of generated text in generative tasks (such as machine translation, text summarization, etc.) compared to reference texts. The specific roles and principles of these metrics are as follows:
1. BLEU - 4 : BLEU-4 [39] is a common metric for evaluating machine translation quality, measuring the similarity between source and target sentences. Its formula is:
B L E U 4 = B P · e x p ( i N 1 w 4 l o g P 4 )
where B P represents the brevity penalty, w 4 represents the weight proportion of the current word n-gram and P 4 represents the 4-gram precision.
2. ROUGE - N : ROUGE-N [40] measures the overlap of words between the model output and reference text, with higher scores indicating greater similarity:
R o u g e N = C o u n t m a t c h ( g r a m n ) C o u n t o u t p u t ( g r a m n )
where C o u n t m a t c h ( . ) is the number of n-grams in Cockpit-Llama’s output that match the test set, and C o u n t o u t p u t ( . ) is the number of n-grams in the model’s output. In this experiment, we use ROUGE-1 and ROUGE-2 to evaluate performance.
3. ROUGE - L :
R O U G E L = ( 1 + λ 2 ) I l c s K l c s I l c s + λ 2 K l c s
where I l c s represents the longest common subsequence match count, K l c s represents the longest common subsequence length proportion, and λ 2 is the adjustment factor.

4.2. Comparison with Baseline Methods

To demonstrate the superior performance of Cockpit-Llama in intent prediction, we compared it with the advanced baseline methods ChatGPT-4 and ChatGLM-3, which have shown outstanding performance in text understanding and logical reasoning. After fine-tuning with the constructed dataset, Cockpit-Llama significantly outperformed ChatGPT-4 and ChatGLM-3 on all metrics for the specific task of predicting driver behavior. This experiment demonstrates the superior performance of Cockpit-Llama. As shown in Table 1, Cockpit-Llama’s BLEU-4 score improved by 92.34% over ChatGPT-4 and 135.14% over ChatGLM-3. This score indicates a high similarity between Cockpit-Llama’s predicted driver behavior and the human-annotated driver behavior, far surpassing ChatGPT-4 and ChatGLM-3, demonstrating its accuracy and effectiveness in predicting driver behavior.
In the ROUGE-1 evaluation, Cockpit-Llama outperformed ChatGPT-4 by 183.61% and ChatGLM-3 by 211.41%. For ROUGE-2, Cockpit-Llama’s score was 95.54% higher than ChatGPT-4’s and 459.80% higher than ChatGLM-3’s. The ROUGE-n metric counts the number of overlapping units between the model’s predicted output text and the human-annotated output text, as well as the proportion of overlapping units in the human-annotated text. According to the experimental data, Cockpit-Llama achieved significantly higher recall rates compared to ChatGPT-4 and ChatGLM-3.
In terms of ROUGE-L, Cockpit-Llama outperformed ChatGPT-4 by 201.27% and ChatGLM-3 by 196.41%, indicating a high similarity between the longest common subsequences of the model’s predicted output text and the human-annotated text, significantly surpassing the two comparison models. Based on these four evaluation metrics, it is evident that Cockpit-Llama surpasses the other two models in both token repetition and longest common subsequence length, demonstrating its accuracy and reliability in driver behavior prediction tasks.
In summary, compared to the other two models, our Cockpit-Llama achieves superior results in driver intent prediction tasks. We attribute this to two main factors: On the one hand, Cockpit-Llama has 8 billion parameters and has been pre-trained on large-scale internet data, which provides it with prior knowledge about the world, enabling it to better understand and process language tasks. On the other hand, we fine-tune the model using a multi-attribute cockpit dataset specifically designed for driver intent prediction, which provides various contextual information, including driver historical interactions, driver emotional states, vehicle states, and external environments, to assist the model in making more accurate predictions.

4.3. In-Context Learning and Freeze vs. Cockpit-Llama (LoRA)

To further verify the compatibility and optimized design of our method, this experiment compared the lightweight fine-tuning method used in our model (LoRA) with Freeze fine-tuning and in-context learning.
As shown in Table 2, compared to the in-context learning method, Cockpit-Llama achieved an 80.36% higher BLEU-4 score, 120.75% higher ROUGE-1 score, 74.67% higher ROUGE-2 score, and 92.52% higher ROUGE-L score. Compared to Freeze fine-tuning, Cockpit-Llama achieved a 52.13% higher BLEU-4 score, 61.79% higher ROUGE-1 score, 41.54% higher ROUGE-2 score, and 60.27% higher ROUGE-L score.
It can be seen that using the LoRA fine-tuning method is more effective than using in-context learning or Freeze fine-tuning. Typically, Freeze fine-tuning only fine-tunes the fully connected layer parameters in the last few layers of the transformer while freezing all other parameters, which may lead to insufficient training and forgetting previously acquired capabilities. In-context learning, on the other hand, simply provides the model with a few examples, allowing the model to learn from these examples in context before making predictions, without changing the model parameters. This results in poor performance when dealing with complex scenarios like the intelligent cockpit environment.
The LoRA method, however, is a bypass adapter approach that decomposes the pre-trained model’s weights into low-rank matrices and adjusts the parameters of these matrices. This method reduces the training load while retaining training parameters, allowing the model to effectively adapt to complex scenarios.

4.4. Dataset Analysis and Visualization

To demonstrate the accuracy of Cockpit-Llama across multiple scenarios, Figure 7 presents four typical scenarios: commute work scenario, daily scenario, far driving scenario, and travel scenario. The figure visualizes the driver’s historical interactions and the model’s predicted future behavior. In the commute work scenario (Figure 7a), Cockpit-Llama accurately predicts that the driver will initiate navigation based on five interactions. In the daily scenario (Figure 7b), Cockpit-Llama predicts that the driver will likely lower the air conditioning temperature, also based on five interactions. Similarly, Figure 7c shows the far driving scenario where Cockpit-Llama predicts the driver will start the navigation. Finally, in the weekend travel scenario (Figure 7d), Cockpit-Llama predicts that the driver’s next action will be to open the trunk lid.
In summary, through quantitative analysis of Cockpit-Llama’s results in multiple experiments, its superior prediction performance has been demonstrated; through prediction examples and qualitative analysis in various typical scenarios, its superior generalization ability has been further demonstrated.

5. Conclusions

In this study, we propose a large language model specifically designed for predicting driver behavior, named Cockpit-Llama. This model is applied in a smart cockpit environment, integrating contextual information and historical interactions to provide accurate predictions of future driver intentions. To support the fine-tuning of Cockpit-Llama, we also construct a novel multi-attribute cockpit dataset, which includes extensive historical driver interactions and multi-attribute states. During the fine-tuning process, we adopt the LoRA method to optimize the model parameters and reduce training costs. Extensive experimental results demonstrate that Cockpit-Llama excels in predicting driver intentions, outperforming baseline models. Therefore, the LLM-based intelligent cockpit intent prediction method proposed in this work drives the transformation of human–cockpit interaction from isolated single-task to collaborative multi-task, from explicit passive to implicit proactive, promoting the cockpit’s evolution towards higher-level intelligence and significantly enhancing the user experience. Moreover, this method can also help drivers better respond to potential hazardous situations, improving driving safety.
  • Limitations and Future Work: Due to the limited size of the dataset, only about 60,000 samples were used to fine-tune Cockpit-Llama, and the inference time is relatively long due to the constraints of parameter size, making it challenging to meet the real-time requirements of commercial driving applications. Additionally, the current model may not fully capture the complexity of driver behavior and the various influencing factors. Future work should focus on constructing larger-scale datasets aiming to provide prior knowledge for the model. It should also explore using large language models to guide smaller ones to optimize inference time. Furthermore, integrating real-time data and additional contextual variables could further enhance the model’s predictive accuracy.

Author Contributions

Conceptualization, Y.C.; methodology, C.L.; software, Q.Y.; validation, Y.F.; formal analysis, Y.L.; investigation, J.L.; resources, F.G.; writing—original draft preparation, Y.C.; writing—review and editing, R.Z.; visualization, X.G.; funding acquisition, R.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Science and Technology Development Project of Jilin province under Grant 20230301008ZD, and the National Science Foundation of China under Grant 52202495 and Grant 52202494.

Institutional Review Board Statement

Not applicable for studies not involving humans or animals.

Informed Consent Statement

This study does not involve humans.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author.

Conflicts of Interest

On behalf of all the authors, the corresponding author states that there are no conflicts of interest.

References

  1. Li, W.; Cao, D.; Tan, R.; Shi, T.; Gao, Z.; Ma, J.; Guo, G.; Hu, H.; Feng, J.; Wang, L. Intelligent cockpit for intelligent connected vehicles: Definition, taxonomy, technology and evaluation. IEEE Trans. Intell. Veh. 2023, 9, 3140–3153. [Google Scholar] [CrossRef]
  2. Sun, X.; Chen, H.; Shi, J.; Guo, W.; Li, J. From hmi to hri: Human-vehicle interaction design for smart cockpit. In Proceedings of the Human-Computer Interaction. Interaction in Context: 20th International Conference, HCI International 2018, Las Vegas, NV, USA, 15–20 July 2018; Proceedings, Part II 20. Springer: Cham, Switzerland, 2018; pp. 440–454. [Google Scholar]
  3. Son, Y.S.; Kim, W.; Lee, S.H.; Chung, C.C. Robust multirate control scheme with predictive virtual lanes for lane-keeping system of autonomous highway driving. IEEE Trans. Veh. Technol. 2014, 64, 3378–3391. [Google Scholar] [CrossRef]
  4. Rosekind, M.R.; Gander, P.H.; Gregory, K.B.; Smith, R.M.; Miller, D.L.; Oyung, R.; Webbon, L.L.; Johnson, J.M. Managing fatigue in operational settings 1: Physiological considerations and counter-measures. Hosp. Top. 1997, 75, 23–30. [Google Scholar] [CrossRef] [PubMed]
  5. Large, D.R.; Burnett, G.; Antrobus, V.; Skrypchuk, L. Driven to discussion: Engaging drivers in conversation with a digital assistant as a countermeasure to passive task-related fatigue. IET Intell. Transp. Syst. 2018, 12, 420–426. [Google Scholar] [CrossRef]
  6. Wong, P.N.; Brumby, D.P.; Babu, H.V.R.; Kobayashi, K. Voices in self-driving cars should be assertive to more quickly grab a distracted driver’s attention. In Proceedings of the 11th International Conference on Automotive User Interfaces and Interactive Vehicular Applications, Utrecht, The Netherlands, 21–25 September 2019; pp. 165–176. [Google Scholar]
  7. Politis, I.; Langdon, P.; Adebayo, D.; Bradley, M.; Clarkson, P.J.; Skrypchuk, L.; Mouzakitis, A.; Eriksson, A.; Brown, J.W.; Revell, K.; et al. An evaluation of inclusive dialogue-based interfaces for the takeover of control in autonomous cars. In Proceedings of the 23rd International Conference on Intelligent User Interfaces, Tokyo, Japan, 7–11 March 2018; pp. 601–606. [Google Scholar]
  8. Kaboli, M.; Long, A.; Cheng, G. Humanoids learn touch modalities identification via multi-modal robotic skin and robust tactile descriptors. Adv. Robot. 2015, 29, 1411–1425. [Google Scholar] [CrossRef]
  9. Kaboli, M.; Mittendorfer, P.; Hügel, V.; Cheng, G. Humanoids learn object properties from robust tactile feature descriptors via multi-modal artificial skin. In Proceedings of the 2014 IEEE-RAS International Conference on Humanoid Robots, Madrid, Spain, 18–20 November 2014; pp. 187–192. [Google Scholar]
  10. Kaboli, M.; Walker, R.; Cheng, G. In-hand object recognition via texture properties with robotic hands, artificial skin, and novel tactile descriptors. In Proceedings of the 2015 IEEE-RAS 15th International Conference on Humanoid Robots (Humanoids), Seoul, Republic of Korea, 3–5 November 2015; pp. 1155–1160. [Google Scholar]
  11. Braun, A.; Neumann, S.; Schmidt, S.; Wichert, R.; Kuijper, A. Towards interactive car interiors: The active armrest. In Proceedings of the 8th Nordic Conference on Human-Computer Interaction: Fun, Fast, Foundational, Helsinki, Finland, 26–30 October 2014; pp. 911–914. [Google Scholar]
  12. Endres, C.; Schwartz, T.; Müller, C.A. Geremin 2D microgestures for drivers based on electric field sensing. In Proceedings of the 16th International Conference on Intelligent User Interfaces, Palo Alto, CA, USA, 13–16 February 2011; pp. 327–330. [Google Scholar]
  13. Pfleging, B.; Schneegass, S.; Schmidt, A. Multimodal interaction in the car: Combining speech and gestures on the steering wheel. In Proceedings of the 4th International Conference on Automotive User Interfaces and Interactive Vehicular Applications, Portsmouth, UK, 17–19 October 2012; pp. 155–162. [Google Scholar]
  14. Lee, S.H.; Yoon, S.O. User interface for in-vehicle systems with on-wheel finger spreading gestures and head-up displays. J. Comput. Des. Eng. 2020, 7, 700–721. [Google Scholar] [CrossRef]
  15. Li, B.; Cao, D.; Tang, S.; Zhang, T.; Dong, H.; Wang, Y.; Wang, F.Y. Sharing traffic priorities via cyber–physical–social intelligence: A lane-free autonomous intersection management method in metaverse. IEEE Trans. Syst. Man Cybern. Syst. 2022, 53, 2025–2036. [Google Scholar] [CrossRef]
  16. Ma, Y.; Cao, Y.; Sun, J.; Pavone, M.; Xiao, C. Dolphins: Multimodal language model for driving. arXiv 2023, arXiv:2312.00438. [Google Scholar]
  17. Xu, Z.; Zhang, Y.; Xie, E.; Zhao, Z.; Guo, Y.; Wong, K.Y.K.; Li, Z.; Zhao, H. Drivegpt4: Interpretable end-to-end autonomous driving via large language model. IEEE Robot. Autom. Lett. 2024, 9, 8186–8193. [Google Scholar] [CrossRef]
  18. Yuan, J.; Sun, S.; Omeiza, D.; Zhao, B.; Newman, P.; Kunze, L.; Gadd, M. Rag-driver: Generalisable driving explanations with retrieval-augmented in-context learning in multi-modal large language model. arXiv 2024, arXiv:2402.10828. [Google Scholar]
  19. Zhang, W.; Zhang, H. Research on distracted driving identification of truck drivers based on simulated driving experiment. IOP Conf. Ser. Earth Environ. Sci. 2021, 638, 012039. [Google Scholar] [CrossRef]
  20. Tran, D.; Manh Do, H.; Sheng, W.; Bai, H.; Chowdhary, G. Real-time detection of distracted driving based on deep learning. IET Intell. Transp. Syst. 2018, 12, 1210–1219. [Google Scholar] [CrossRef]
  21. Fasanmade, A.; He, Y.; Al-Bayatti, A.H.; Morden, J.N.; Aliyu, S.O.; Alfakeeh, A.S.; Alsayed, A.O. A fuzzy-logic approach to dynamic bayesian severity level classification of driver distraction using image recognition. IEEE Access 2020, 8, 95197–95207. [Google Scholar] [CrossRef]
  22. Riaz, F.; Khadim, S.; Rauf, R.; Ahmad, M.; Jabbar, S.; Chaudhry, J. A validated fuzzy logic inspired driver distraction evaluation system for road safety using artificial human driver emotion. Comput. Netw. 2018, 143, 62–73. [Google Scholar] [CrossRef]
  23. Teyeb, I.; Snoun, A.; Jemai, O.; Zaied, M. Fuzzy logic decision support system for hypovigilance detection based on CNN feature extractor and WN classifier. J. Comput. Sci. 2018, 14, 1546–1564. [Google Scholar] [CrossRef]
  24. Xiao, H.; Li, W.; Zeng, G.; Wu, Y.; Xue, J.; Zhang, J.; Li, C.; Guo, G. On-road driver emotion recognition using facial expression. Appl. Sci. 2022, 12, 807. [Google Scholar] [CrossRef]
  25. Li, W.; Xue, J.; Tan, R.; Wang, C.; Deng, Z.; Li, S.; Guo, G.; Cao, D. Global-local-feature-fused driver speech emotion detection for intelligent cockpit in automated driving. IEEE Trans. Intell. Veh. 2023, 8, 2684–2697. [Google Scholar] [CrossRef]
  26. Oehl, M.; Siebert, F.W.; Tews, T.K.; Höger, R.; Pfister, H.R. Improving human-machine interaction–a non invasive approach to detect emotions in car drivers. In Proceedings of the Human-Computer Interaction. Towards Mobile and Intelligent Interaction Environments: 14th International Conference, HCI International 2011, Orlando, FL, USA, 9–14 July 2011; Proceedings, Part III 14. Springer: Berlin/Heidelberg, Germany, 2011; pp. 577–585. [Google Scholar]
  27. Hu, J.; Yang, B.; Song, H. Data-driven Prediction of Vehicle Air Conditioner Set Temperature. Mech. Sci. Technol. Aerosp. Eng. 2022, 41, 134–142. [Google Scholar] [CrossRef]
  28. Ferenc, G.; Timotijević, D.; Tanasijević, I.; Simić, D. Towards Enhanced Autonomous Driving Takeovers: Fuzzy Logic Perspective for Predicting Situational Awareness. Appl. Sci. 2024, 14, 5697. [Google Scholar] [CrossRef]
  29. Li, Y.; Metzner, M.; Schwieger, V. Driving environment inference from POI of navigation map: Fuzzy logic and machine learning approaches. Sensors 2023, 23, 9156. [Google Scholar] [CrossRef] [PubMed]
  30. Ye, M.; Yin, P.; Lee, W.C.; Lee, D.L. Exploiting geographical influence for collaborative point-of-interest recommendation. In Proceedings of the 34th International ACM SIGIR Conference on Research and Development in Information Retrieval, Beijing, China, 25–29 July 2011; pp. 325–334. [Google Scholar]
  31. Mukhopadhyay, S.; Kumar, A.; Parashar, D.; Singh, M. Enhanced Music Recommendation Systems: A Comparative Study of Content-Based Filtering and K-Means Clustering Approaches. Rev. D’Intelligence Artif. 2024, 38, 365–376. [Google Scholar]
  32. Ouyang, L.; Wu, J.; Jiang, X.; Almeida, D.; Wainwright, C.; Mishkin, P.; Zhang, C.; Agarwal, S.; Slama, K.; Ray, A.; et al. Training language models to follow instructions with human feedback. Adv. Neural Inf. Process. Syst. 2022, 35, 27730–27744. [Google Scholar]
  33. Achiam, J.; Adler, S.; Agarwal, S.; Ahmad, L.; Akkaya, I.; Leoni Aleman, F.; Almeida, D.; Altenschmidt, J.; Altman, S.; et al.; OpenAI GPT-4 Technical Report. arXiv 2023, arXiv:2303.08774. [Google Scholar]
  34. Chowdhery, A.; Narang, S.; Devlin, J.; Bosma, M.; Mishra, G.; Roberts, A.; Barham, P.; Chung, H.W.; Sutton, C.; Gehrmann, S.; et al. Palm: Scaling language modeling with pathways. J. Mach. Learn. Res. 2023, 24, 11324–11436. [Google Scholar]
  35. Zheng, L.; Chiang, W.L.; Sheng, Y.; Zhuang, S.; Wu, Z.; Zhuang, Y.; Lin, Z.; Li, Z.; Li, D.; Xing, E.; et al. Judging llm-as-a-judge with mt-bench and chatbot arena. Adv. Neural Inf. Process. Syst. 2023, 36, 46595–46623. [Google Scholar]
  36. Touvron, H.; Lavril, T.; Izacard, G.; Martinet, X.; Lachaux, M.A.; Lacroix, T.; Rozière, B.; Goyal, N.; Hambro, E.; Azhar, F.; et al. Llama: Open and efficient foundation language models. arXiv 2023, arXiv:2302.13971. [Google Scholar]
  37. Touvron, H.; Martin, L.; Stone, K.; Albert, P.; Almahairi, A.; Babaei, Y.; Bashlykov, N.; Batra, S.; Bhargava, P.; Bhosale, S.; et al. Llama 2: Open foundation and fine-tuned chat models. arXiv 2023, arXiv:2307.09288. [Google Scholar]
  38. Zhang, J.; Pu, J.; Xue, J.; Yang, M.; Xu, X.; Wang, X.; Wang, F.Y. HiVeGPT: Human-machine-augmented intelligent vehicles with generative pre-trained transformer. IEEE Trans. Intell. Veh. 2023, 8, 2027–2033. [Google Scholar] [CrossRef]
  39. Papineni, K.; Roukos, S.; Ward, T.; Zhu, W.J. Bleu: A method for automatic evaluation of machine translation. In Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics, Philadelphia, PA, USA, 6–12 July 2002; pp. 311–318. [Google Scholar]
  40. Lin, C.Y. Rouge: A package for automatic evaluation of summaries. In Proceedings of the Text Summarization Branches Out, Barcelona, Spain, 25–26 July 2004; pp. 74–81. [Google Scholar]
  41. Zeng, A.; Xu, B.; Wang, B.; Zhang, C.; Yin, D.; Rojas, D.; Feng, G.; Zhao, H.; Lai, H.; Yu, H.; et al. ChatGLM: A Family of Large Language Models from GLM-130B to GLM-4 All Tools. arXiv 2024, arXiv:2406.12793. [Google Scholar]
Figure 1. The overview of Cockpit-Llama.
Figure 1. The overview of Cockpit-Llama.
Sensors 25 00064 g001
Figure 2. Example of multi-attribute cockpit dataset.
Figure 2. Example of multi-attribute cockpit dataset.
Sensors 25 00064 g002
Figure 3. Statistical analysis of dataset intents.
Figure 3. Statistical analysis of dataset intents.
Sensors 25 00064 g003
Figure 4. The overall architecture of Cockpit-Llama is divided into three parts: modal text data encoding, lightweight fine-tuning of LLM backbone using adapter, and driver behavior intent decoding.
Figure 4. The overall architecture of Cockpit-Llama is divided into three parts: modal text data encoding, lightweight fine-tuning of LLM backbone using adapter, and driver behavior intent decoding.
Sensors 25 00064 g004
Figure 5. Lightweight fine-tuning with adapter for LLM backbone.
Figure 5. Lightweight fine-tuning with adapter for LLM backbone.
Sensors 25 00064 g005
Figure 6. The loss curve variation during Cockpit-Llama training.
Figure 6. The loss curve variation during Cockpit-Llama training.
Sensors 25 00064 g006
Figure 7. Visualization examples of Cockpit-Llama’s prediction process and interpretative language captions: (a) commute work scenario, (b) daily scenario, (c) far driving scenario, (d) weekend travel scenario.
Figure 7. Visualization examples of Cockpit-Llama’s prediction process and interpretative language captions: (a) commute work scenario, (b) daily scenario, (c) far driving scenario, (d) weekend travel scenario.
Sensors 25 00064 g007
Table 1. Baseline methods vs. Cockpit-Llama.
Table 1. Baseline methods vs. Cockpit-Llama.
MethodTypeEvaluation
Bleu-4 Rouge-1 Rouge-2 Rouge-L
ChatGPT-4 [33]LLM37.0828.2139.3227.03
ChatGLM-3 [41]LLM30.3325.6913.7427.47
Cockpit-LlamaLLM71.3280.0176.8981.42
Table 2. In-context learning and Freeze vs. Cockpit-Llama (LoRA).
Table 2. In-context learning and Freeze vs. Cockpit-Llama (LoRA).
MethodEvaluation
Bleu-4 Rouge-1 Rouge-2 Rouge-L
In-Context39.5436.2444.0242.29
Freeze46.8849.4554.3250.80
Cockpit (Lora)71.3280.0176.8981.42
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, Y.; Li, C.; Yuan, Q.; Li, J.; Fan, Y.; Ge, X.; Li, Y.; Gao, F.; Zhao, R. Cockpit-Llama: Driver Intent Prediction in Intelligent Cockpit via Large Language Model. Sensors 2025, 25, 64. https://doi.org/10.3390/s25010064

AMA Style

Chen Y, Li C, Yuan Q, Li J, Fan Y, Ge X, Li Y, Gao F, Zhao R. Cockpit-Llama: Driver Intent Prediction in Intelligent Cockpit via Large Language Model. Sensors. 2025; 25(1):64. https://doi.org/10.3390/s25010064

Chicago/Turabian Style

Chen, Yi, Chengzhe Li, Qirui Yuan, Jinyu Li, Yuze Fan, Xiaojun Ge, Yun Li, Fei Gao, and Rui Zhao. 2025. "Cockpit-Llama: Driver Intent Prediction in Intelligent Cockpit via Large Language Model" Sensors 25, no. 1: 64. https://doi.org/10.3390/s25010064

APA Style

Chen, Y., Li, C., Yuan, Q., Li, J., Fan, Y., Ge, X., Li, Y., Gao, F., & Zhao, R. (2025). Cockpit-Llama: Driver Intent Prediction in Intelligent Cockpit via Large Language Model. Sensors, 25(1), 64. https://doi.org/10.3390/s25010064

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop