Next Article in Journal
An Efficient and Robust Dual-Channel Signal Gluing Method for Atmospheric Lidar
Previous Article in Journal
Intelligent Soft Sensor for Spindle Convective Heat Transfer Coefficient Under Varying Operating Conditions Using Improved Grey Wolf Optimization Algorithm
Previous Article in Special Issue
Research on Lateral Stability Control of Four-Wheel Independent Drive Electric Vehicle Based on State Estimation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

From Concept to Representation: Modeling Driving Capability and Task Demand with a Multimodal Large Language Model

1
Graduate School of Informatics, Nagoya University, Furo-cho, Chikusa-ku, Nagoya 464-8603, Japan
2
Graduate School of Engineering, Gifu University, 1-1 Yanagido, Gifu City 501-1193, Japan
3
Institute of Innovation for Future Society, Nagoya University, Furo-cho, Chikusa-ku, Nagoya 464-8601, Japan
4
Tier IV Inc., Nagoya University Open Innovation Center, 1-3, Mei-eki 1-Chome, Nakamura-ku, Nagoya 450-6610, Japan
5
Institute of Physical and Chemical Research (RIKEN) Center for Advanced Intelligence Project, 1-5, Yamadaoka, Suita 565-0871, Japan
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(18), 5805; https://doi.org/10.3390/s25185805
Submission received: 7 July 2025 / Revised: 20 August 2025 / Accepted: 15 September 2025 / Published: 17 September 2025

Abstract

Driving safety hinges on the dynamic interplay between task demand and driving capability, yet these concepts lack a unified, quantifiable formulation. In this work, we present a framework based on a multimodal large language model that transforms heterogeneous driving signals—scene images, maneuver descriptions, control inputs, and surrounding traffic states—into low-dimensional embeddings of task demand and driving capability. By projecting both embeddings into a shared latent space, the framework yields an interpretable measurement of task difficulty that alerts to capability shortfalls before unsafe behavior arises. Built upon a customized BLIP 2 backbone and fine-tuned on diverse simulated driving scenarios, the model respects consistency within tasks, captures impairment-related capability degradation, and can transfer to real-world motorway data without additional training. These findings endorse the framework as a concise yet effective step toward proactive, explainable risk assessment in intelligent vehicles.

1. Introduction

Traffic accidents remain a critical social issue. According to a recent report from the World Health Organization (WHO), road accidents cause more than 1.2 million fatalities per year globally [1], and human error is estimated to contribute to roughly 94% of accidents [2]. Various active safety systems have been developed for decades—ranging from fundamental technologies such as Anti-lock Braking Systems (ABS) and Electronic Stability Control (ESC), to Advanced Driver-Assistance Systems (ADAS) integrating functions like Adaptive Cruise Control (ACC), Autonomous Emergency Braking (AEB), and Lane Keeping Assist (LKA), and more recently, to assisted and autonomous driving systems [3]—all aiming to reduce and finally prevent traffic accidents. However, even the latest automated driving technologies have not eliminated collisions. An accident occurred in 2019 in Delray Beach when a Tesla in Autopilot mode collided with a truck, resulting in the death of the driver [4]. In 2023, a Cruise autonomous taxi in San Francisco struck a pedestrian and dragged her for approximately 20 feet [5]. These cases illustrate that state-of-the-art driving systems, much like human drivers, still encounter situations beyond their capabilities, leading to accidents.
Previous research efforts [6,7] have pointed out a fundamental cause among these crashes: a misalignment between driving capability and task demand. As shown in Figure 1, in Fuller’s Task–Capability Interface (TCI) model, safe driving is maintained only when the driving capability meets or exceeds the momentary demands of the driving task. If the required task demand exceeds the capability of the human driver or the autonomous system, the risk of an accident increases dramatically, further leading to a collision if unfortunately. In practice, this capability–demand misalignment can occur due to various factors: a human driver might be fatigued, distracted, or unskilled in a challenging scenario, or an autonomous driving system might face an edge case outside its operational design domain.
Recognizing this misalignment in time or even in advance could enable preventive interventions to mitigate the associated risks. For example, alerting a distracted driver to refocus, engaging emergency braking, or prompting a Level-3 automated vehicle to hand control back to a more capable driver. Unfortunately, accomplishing this in practice has been challenging, in part because driving capability and task demand are not formally defined or directly measurable in existing frameworks. These concepts are multifaceted and dynamic, influenced by vehicle state, environmental complexity, and human factors, and there is no universally accepted quantitative model for these terms [8]. As a result, current driver-assistance systems lack a rigorous basis to evaluate whether a driver (human or AI) can cope with an upcoming situation, making further intervention difficult to justify or execute.
Multimodal Large Language Models (MLLMs) offer a promising new avenue to address this challenge. Recent breakthroughs in artificial intelligence have produced large models that learn from massive amounts of multimodal data, enabling an unprecedented level of general scene understanding and reasoning. In the driving domain, such a model could analyze traffic scenes (vehicles, pedestrians, road signs, weather conditions, etc.) while also considering the descriptive context (traffic rules, learned common sense about how scenarios typically unfold) [9]. Leveraging these advantages, this work proposes a novel MLLM-based approach to formally model driving capability and task demand. By harnessing the generalization power and multimodal perception of MLLM, we aim to bridge the definitional gap and develop a data-driven framework for identifying driving capability shortfalls before they lead to accidents.
The main contributions of this paper can be summarized as follows.
  • We adapt the architecture of BLIP-2 [10] with a novel custom design aligned with the TCI model, enabling formal modeling of task demand and driving capability. This tailored architecture provides a principled framework for capturing and analyzing alignments in dynamic driving scenarios.
  • We propose a new method to explicitly distill low-dimensional semantic vectors for driving capability and task demand. By imposing a dimensionality constraint on the MLLM’s output and integrating a multi-label contrastive learning objective, our approach yields concise yet expressive representations that characterize the driving capability and the task demand within a shared latent space.
  • We design a comparator module to measure the distance between capability vectors and between capability and demand vectors. This enables the effective detection of mismatches between driving capability and task demand, thereby facilitating early intervention and risk awareness. Validation and analysis across multiple datasets demonstrate the practical feasibility and interpretability of the proposed method.
The remainder of this paper is organized as follows. In Section 2, we provide a review of the literature of prior work relevant to this study. Section 3 details the research problem addressed in this study and the methodologies employed. The dataset we constructed to support this work is introduced in Section 4. In Section 5, we present the evaluation, analysis, and ablation study of the proposed method. Following in Section 6, we discuss the key findings, challenges encountered, and potential future research directions. Finally, Section 7 concludes this paper.

2. Related Works

2.1. Task–Capability Interface Model

The Task–Capability Interface (TCI) model, which was proposed by Fuller in 2000, conceptualizes driving as the “management of task difficulty” where task difficulty correlates with the probability of risk occurrence [6]. Within this context, task difficulty is determined by two key components: task demand and driving capability. The task difficulty arises from the dynamic discrepancy between these two components [11]. When driving capability significantly exceeds task demand, the task is perceived as easy; conversely, when task demand approaches or exceeds the driving capability, the task is considered difficult, resulting in the driver’s inability to maintain complete control of the vehicle, which may consequently result in an accident [7]. In the framework of the TCI model, task demand refers to the objective complexity of the current driving scenario, primarily influenced by the driving environment, surrounding vehicles, other traffic participants, and the driver’s intentioned maneuvers such as trajectory and speed selection. In comparison, driving capability denotes the momentary delivery of competence by the driver during the process. Here, competence broadly encompasses driving knowledge and skills such as vehicle control, road observation, and anticipation.
Subsequently, the TCI model has been employed as a theoretical tool in studies of driving behaviors in various scenarios. From the perspective of driving capability, different groups of drivers (e.g., drivers differing in driving experience [12,13], age [12,14,15], or gender [13]), or drivers operating in different physical [15,16,17,18] or mental [19,20,21] states, exhibit different driving capabilities, which consequently lead to variations in driving behaviors. Regarding task demand, attention focuses primarily on how the behavior of similar drivers varies when faced with different speeds [19,22] and levels of environmental complexity [15,23]. The TCI model provides researchers with a robust theoretical framework for better understanding and explaining the causal relationships underlying driver behaviors.
When employing the TCI model to infer behaviors and evaluate driving performance, however, the situation becomes somewhat more complex, primarily due to the lack of concretely specifying its internal mechanisms [24,25]. Driving capability and task demand lack standardized numeric definitions and, since they operate on different conceptual dimensions, direct quantitative comparisons are not feasible. Moreover, since inherent driving competence can only set an upper limit on the driving ability delivered, actual driving capability must be assessed dynamically during driving [26]. These factors further complicate the functional implementation of the TCI model. Although some studies attempt to regard task difficulty as driving risk [27] or driver fitness [28] in order to circumvent the need for direct use of driving capability and task demand, it is believed that explicit modeling of driving capability and task demand has significant potential to improve real-time safety interventions and adaptive driving assistance.

2.2. Impaired Driving

Impaired driving refers to the driving operations of a driver whose ability is compromised. The impairment may manifest itself as reduced reaction time, distracted attention, improper decision making, or decreased motor coordination [1]. Within the framework of the TCI model, impaired driving can be understood as a state in which temporary human factors cause the delivered driving capability to fall below the expectation, resulting in a mismatch with task demand and consequently a risk of losing control and accident.
Among all forms of impaired driving, the most common is Driving Under the Influence (DUI) of alcohol or drugs, especially under the influence of alcohol [29]. For example, in Japan, DUI-related accidents accounted for 5.5% of all personal injury traffic accidents in 2017 [30]. DUI is manifested through various observable signals such as body movements [31,32,33] and facial expressions [34,35,36], but the primary impact lies in the driver’s operation of vehicle controls, such as the input of the steering and pedals [37]. For example, intoxicated drivers tend to make more frequent steering corrections compared to their normal state [38]. Furthermore, studies have shown that alcohol consumption can alter a driver’s personal driving style [39], further affecting the decisions and actual capability delivered while driving.
Other frequently observed forms include drowsiness, distraction, and extreme emotions. Drowsiness is considered to produce effects similar to those of alcohol [40], such as reducing the ability to control steering and braking and increasing the variability of driving speed [41,42,43]. Distraction, either due to events within [44,45] or outside [46,47] the vehicle, leads to allocation of the attentional resource to a secondary activity, thereby increasing the probability of missing critical events during driving [48]. Extreme emotional states such as anger, anxiety, or sadness can affect the judgment of a driver and result in the adoption of more aggressive or erratic driving styles, such as sudden acceleration, tailgating, or risky overtaking maneuvers [49,50,51].
In summary, while the causes and manifestations of impaired driving may vary, a consistent feature is that the driver’s performance under impairment is, to some extent, degraded compared to their normal state. Under equivalent task demand, this degradation can be interpreted as a reduction in the driving capability delivered by the driver, resulting in a lower output relative to their unimpaired condition.

2.3. Multimodal Large Language Model in Driving

With recent advances in computational hardware, the implementation of Large Language Models (LLM) based on the Transformer architecture [52] has become feasible, as exemplified by models such as PaLM [53], LLaMA [54], GPT-4 [55], etc. These models, trained on massive amounts of internet data, can capture complex semantic patterns and perform at or beyond human levels in tasks such as language understanding, reasoning, and text generation [56,57,58]. An increasing number of applications [59,60,61] are now exploring the development of intelligent agents powered by LLMs to assist or even replace human involvement, including in the field of driving.
In driving scenarios, vehicles typically rely on a variety of sensors such as cameras, LiDARs, radars, and Controller Area Network (CAN) bus systems to perceive the surrounding environment and monitor internal states. Consequently, intelligent agents designed for driving must be LLMs that are capable of processing multimodal data, i.e., the Multimodal Large Language Model (MLLM) [62]. One of the most representative classes of such models is the Vision-Language Model (VLM), which takes natural language and images as input and performs tasks such as image captioning and visual question answering (VQA), such as CLIP [63], ViLBERT [64], BLIP-2 [10], and Flamingo [65]. In addition to these, audio [66,67] and pointcloud [68] are also frequently used modalities in driving applications.
MLLMs offer significant advantages in driving, particularly in their ability to fuse multiple modalities and generalize to novel, unseen situations [63,65,69]. Currently, research on applying MLLMs in driving focuses mainly on two core tasks: perception and planning/control. Discriminative perception models such as HiLMD [70] and Talk2BEV [71] utilize video and natural language input to generate visuo-linguistic responses related to driving, while generative models such as GAIA-1 [72] and UniSim [73] can anticipate environmental dynamics based on previous driving maneuvers. In the domain of planning and control, the goal extends beyond the understanding of the scenario: It involves using language instructions to recommend future waypoints to accomplish the current driving task [74,75], or even directly generating control signals to operate the vehicle [76,77]. Although still in the early stages, research into MLLMs in driving is rapidly gaining momentum, with more promising developments expected in the near future.

3. Methodology

3.1. Problem Statement

As shown in Figure 1, in the Task–Capability Interface (TCI) model, task demand is defined by the driving environment and the intended driving maneuvers, while driving capability is constrained by driver maximum driving competence and influenced by various human-related factors, reflected in actual driving behaviors. In the previous study, an important observation is that drivers in any case always attempt to dynamically allocate only the most appropriate level of capability in response to the specific driving context, to maintain perceived task difficulty at a preferred level [26]. And empirical evidence also suggests that the intended maneuvers of a driver and the maneuvers ultimately executed are, in practice, highly congruent. As a result, we define a unified input X as a pair of driving environment X e and driver behavior X b :
X = ( X e , X b )
Based on this input, here two functions f d ( · ) , f c ( · ) are defined to map the driving context to a latent representation to formally capture task demand and driving capability:
d = f d ( X )
c = f c ( X )
where d , c R n are semantic embeddings that lie in a comparable space, enabling the subsequent computation of task difficulty Δ d as a function of the discrepancy between task demand d and driving capability c , formally defined as:
Δ d = g ( d , c )
where Δ d R denotes the task difficulty, g ( · , · ) quantifies the discrepancy between demand and capability.
To obtain meaningful embeddings of demand and capability, the following properties are imposed as conceptual constraints or regularization objectives. These properties ensure that the obtained embeddings d and c align with intuitive and theoretical expectations:
Property 1
(Task Consistency). Task-demand embeddings d and the corresponding delivered driving capability embeddings c should remain consistent in similar scenarios X i , X j , while differentiating clearly between semantically dissimilar scenarios X i , X k .
f d ( X i ) f d ( X j ) f d ( X i ) f d ( X k )
f c ( X i ) f c ( X j ) f c ( X i ) f c ( X k )
Property 2
(Intra-task Capability Ordering). For a given task X i e , the embeddings of the different delivered driving capabilities c i ( j ) should be orderable, reflecting their relative adequacy in fulfilling the task demand.
c i ( 1 ) c i ( 2 ) c i ( m )
c i ( m ) = f c ( ( X i e , X m b ) )
Property 3
(Task–Capability Comparability). For a given scenario X i , the embeddings of task demand d i and the driving capability delivered c i should be comparable in a shared latent space. The comparison result should reflect the principle in the TCI model: task difficulty increases when the demand exceeds the capability, and remains low when the capability meets or exceeds the demand.
Δ d i = g ( d i , c i )
c i d i Δ d i 0 , c i < d i Δ d i

3.2. Objective and Framework Overview

The goal of this research is to obtain descriptors that can explicitly represent driving capability and task demand, and to assess the task difficulty in the current driving situation by comparison between the two components. Specifically, the study consists of two main objectives:
  • Learning and modeling the functions f d ( · ) , f c ( · ) , g ( · , · ) introduced in Section 3.1.
  • Validating the success of the modeling by examining whether the properties described in Section 3.1 are satisfied.
To this end, a novel framework is proposed in this study to model both task demand and driving capability through data-driven representation learning. The general structure of the framework is illustrated in Figure 2. The proposed framework unifies the driving environment and the driver behavior as the input, using a multimodal large language model (MLLM) to encode the input into latent embeddings of task demand and driving capability. These embeddings are fed into a lightweight comparator that projects both vectors into a shared, comparable space, yielding scalar scores that quantify their relative magnitudes. By contrasting the demand score with the capability score, the framework finally estimates the task difficulty for the current driving scenario.

3.3. Task–Capability Large Model

In this study, we employ an MLLM-based task–capability large model to model the function f d ( · ) and f c ( · ) described in Section 3.1. The model takes a multimodal input of the driving environment and the driver behavior, and outputs latent embeddings for both task demand and driving capability. The architecture of the network is illustrated in Figure 3.
The task–capability large model is designed based on the BLIP-2 [10] architecture, utilizing a pre-trained visual encoder and a pre-trained large language model (LLM) with the corresponding text tokenizer to reduce the dependence on large-scale data and computational resources. In addition to scene image, natural language description of the current driving scenario, and task prompt also used in BLIP-2, the task–capability large model also incorporates modalities including driver behavior data (such as manipulations of the steering wheel and the accelerator or brake pedals) and surrounding vehicle information (such as position and speed). In the implementation, the surrounding encoder and the behavior encoder first employ a lightweight three-layer Multi-Layer Perceptron (MLP) to extract features from each frame. The resulting feature vectors are then partitioned into patches along the temporal dimension and integrated by a six-layer transformer encoder to capture spatio-temporal representations. Subsequently, a projector composed of fully connected layers aligns the output with the input dimension of the LLM. The capability and demand pipelines share an identical architecture; only certain modules differ in their parameter settings. Each modality is encoded into tokens and aligned with LLM’s input space by its respective encoder and projector, and the natural language description is used to query a subset of key visual tokens via a BLIP-2-style Q-former module.
Different modalities of tokens are subsequently formed for LLM input. The structure of the series of LLM input tokens and the token number for each part are shown in Figure 4. For each image frame in a data sample, the Q-Former selects four key representative tokens, concatenates them with a behavior token and a surrounding token of the corresponding frame, and a special flag token is prefixed to mark the beginning of the frame-wise feature sequence. Frame-wise feature sequences are first stitched together, after which the instruction tokens are added to the front and the prompt tokens to the end. The entire token sequence is then fed into the LLM to obtain hidden states, which are decoded by a three-layer MLP de-tokenizer into the corresponding representation embeddings.
To obtain meaningful representations and explicitly endow them with Property 1, we extend the previous Supervised Contrastive learning (SupCon) method [78] to train the model in a multi-label manner. In accordance with previous works [38,39], scene category and action category are used as supervisory labels for task demand, while driver ID and driver state are also employed to supervise driving capability. For each backpropagation step, a batch of normalized representations Z = [ z 1 , z 2 , , z B ] is produced and a Boolean relation matrix R { 0 , 1 } B × B is generated for each label, where R i j = 1 means sample i and j share the same attribute on this label. First, a normalized cosine similarity matrix S is calculated for every pair of samples in the batch. Each element S i j of the matrix is calculated as follows:
S i j = z i z j τ
τ is a temperature hyperparameter to scale the similarity. For every anchor i, a categorical distribution is computed over other samples in the batch, the logarithmic probability log p i j is given by:
log p i j = S i j log k i e S i k , j i
In addition, the supervised contrastive loss L is calculated as:
L = 1 i , j R i j i , j R i j log p i j
With multiple relation matrices R ( k ) , the final object L total is:
L total = k λ k L Z , R ( k )
where λ k weighs different attributes. This loss encourages samples that share a larger number of labels to be closer in the embedding space while automatically generating negative pairs from the remaining samples in the batch.

3.4. Embedding Comparator

After obtaining the representations of task demand d and driving capability c , we further model g ( · , · ) with an embedding comparator. The structure of the embedding comparator is illustrated in Figure 5. Specifically, d and c are projected into a scalar space with the respective three-layer MLP to produce d score and c score . The d score and c score are designed to enhance interpretability, which are obtained through unsupervised learning to represent the relative magnitudes of task demand and driving capability within a category. The difference between the scores is then calculated, and this difference is subsequently mapped to the final difficulty measure Δ d .
As mentioned in Property 3 and illustrated in Figure 6, task difficulty should increase when demand exceeds capability and remain low when capability meets or exceeds demand. Consequently, a learnable function g score ( · ) is introduced within the embedding comparator to fit the target function, defined as follows:
g score = log ( 1 + α e β ( d score c score ) )
where α and β are trainable parameters, α controls the growth scale, and β controls the growth rate. Compared with simple ReLU-like functions, the proposed function can more smoothly capture task difficulty when demand and capability are close, while producing a sharp increase in task difficulty when the gap is large, better reflecting how task difficulty is perceived in real driving scenarios. To exhibit Property 3, the predicted task difficulty Δ d is compared with the ground-truth task difficulty Δ g t to ensure a meaningful comparability, and the regression loss is computed as follows, where N denotes the number of samples:
L reg = 1 N i = 1 N ( Δ d i Δ g t i ) 2
To ensure that the extracted capability exhibits Property 2, we introduce the following assumption to impose an order relation on certain pairs of driving capabilities:
Assumption 1.
For a given task, a driver always exhibits a higher driving capability in a normal state compared to an impaired state.
Based on this assumption, we pair driving capability samples [ d normal , d impaired ] obtained from the same driver and the same scenario in different states, and employ a pairwise logistic loss L pair [79], defined as:
L pair = ( i , j ) log ( 1 + exp ( p d ( d impaired ) p d ( d normal ) ) )
This pairwise logistic loss L pair automatically determines the appropriate margin of the data with respect to the ordering constraint, thereby enforcing an ordinal relationship among different capabilities. The total loss L total for the embedding comparator is the sum of the regression loss L reg , the pairwise logistic loss L pair :
L total = L reg + L pair

4. Dataset

4.1. Overview

As discussed in Section 3.4, obtaining comparable representations of driving capability requires a dataset that contains multiple instances of driving behavior under similar tasks and admits a partial order of these instances with respect to the capability of the driver. To meet this requirement, we leverage a dataset which was built and described in detail in our previous works [38,39]. The dataset was constructed with the driving simulator provided by the Advanced Research and Innovation Center, DENSO CORPORATION, as illustrated in Figure 7. For each driving task, the same driver performed both normal and drunk sessions, enabling direct, task-consistent comparisons.
In this study, following the guidance of previous research [6,7,80], we investigated several representative driving tasks in urban scenarios. These tasks were segmented into approximately 32,000 samples, each 10 s long, totaling approximately 5300 min of driving data. Each data sample contains multimodal information, including scenario and driver metadata, natural language task descriptions, camera images, driver operation signals, and the behaviors of related surrounding traffic participants. This dataset was used for finetuning the model and for conducting part of the experimental validation.

4.2. Scenario Categorization

For both driving capability and task demand, the differences between scenarios are not merely quantitative but qualitative in nature. For example, during car-following on a straight road, the driver primarily needs to control the speed and maintain the distance using the accelerator and brake pedals, while at intersections involving turns, steering control becomes equally essential. As a result, categorizing the appropriate driving scenarios becomes essential.
In the dataset, a two-step procedure is adopted to categorize the scenarios. The first step involves segmentation based on the speed of the ego vehicle. In the TCI model, the task demand is positively correlated with the vehicle speed: when the speed is zero, the task demand is also considered zero [6]. In such cases, since no active driving effort is required, the scene does not contribute meaningful information to the driving ability and is therefore excluded from further analysis. Subsequently, the segmented driving clips were classified according to the geometry of the road (non-intersection, branch zone, merge zone, intersection) and vehicle behavior (going straight, lane change, turning) according to the criteria defined in [80]. In the dataset, it primarily considers vehicle behaviors of going straight and turning, and road structures of intersections and non-intersections (including both straight and curve).
Figure 8 presents the result of the categorization of scenarios. In this study, two driving routes (Route A and Route B) are utilized, both constructed based on real urban traffic environments in Japan. Route A shown in Figure 8a features moderate traffic, with the ego vehicle primarily engaged in car following behavior. In comparison, Route B in Figure 8b has lower traffic density but includes several pedestrian-related events, such as crossing pedestrian crosswalks. The route shown in Figure 8a contains a tunnel segment which, due to its markedly different environment, is treated as a separate segment (NA2) in our analysis.

4.3. Data Modality

After obtaining the categorized scenarios, it is further segmented into samples of 10 s duration following the methodology of previous studies [38,39]. Each sample comprises various modalities, including scenario and driver metadata, natural language descriptions of the current scene and task, in-cabin camera images, driver operations, and information about surrounding vehicles.
The metadata includes information such as the data collection time, driver identity, driver condition, scenario category, and estimated task difficulty. It served as the key supervision for multi-label contrastive learning in MLLM as previously described in Section 3.3, and the rank learning and task difficulty estimation of the task–capability comparator as mentioned in Section 3.4.
Each segment of categorized scenarios is accompanied by a set of natural language descriptions to summarize the traffic context and the expected behavior of the ego vehicle. For example, in the scenario “IA2” shown in Figure 8a, one of the descriptions is “Stay in the left-turn lane and make a left at this urban junction. Ensure a safe gap from the vehicle in front.” Each set contains five semantically similar descriptions, from which one is randomly sampled when loading a data instance.
In-cabin camera images are captured at a resolution of 1920 × 1080 pixels and sampling rate of 30 Frames Per Second (FPS). The raw video data are downsampled to 2 FPS, resulting in 20 images per data sample. Each image is cropped to remove irrelevant or occluded regions and then resized to a resolution of 672 × 224 pixels, as shown in Figure 9. Subsequently, the images are normalized using CLIP [63] parameters and converted into tensors.
Driver operation data are collected from the simulator’s Controller Area Network (CAN) bus. In line with previous studies [38,39], we use six features to characterize the motion state of the ego vehicle and the driver’s manipulations, as shown in Figure 10. The data are downsampled to 10 frames per second and standardized using z-score normalization.
For information on surrounding vehicles, a set with a maximum length of 11 is constructed following the guidelines proposed in [80]. Each element in the set represents a surrounding vehicle and is described using five features: relative longitudinal and lateral positions, relative yaw angle, the surrounding vehicle’s speed, and the ego vehicle’s speed. As with driver operation data, surrounding vehicle information is also obtained directly from the simulator, downsampled to 10 FPS, and normalized using z-score normalization.

5. Experimental Evaluation

5.1. Experiment Setup

In the experiment, the whole framework is implemented with PyTorch [81] (Version 2.7.0). The task–capability large model is implemented on the LAVIS [82] (commit 506965b, accessed 7 May 2025) framework, while the visual encoder, Large Language Model (LLM), and the corresponding text tokenizer are realized using the transformers [83] library (Version 4.33.2). The visual encoder is based on CLIP [63], using the publicly released ViT-B/32 checkpoint (Hugging Face Hub ID: openai/clip-vit-base-patch32, accessed on 7 May 2025). The LLM and the corresponding text tokenizer are initialized with the publicly released LLaVA-v1.6-Vicuna-7B model (Hugging Face Hub ID: liuhaotian/llava-v1.6-vicuna-7b, accessed on 7 May 2025) from previous research [84].
Following previous studies [38,39], we partition the entire dataset into training, validation, and test sets with a ratio of 7:2:1 at the driving record level. The entire framework was trained in two stages:
Stage 1: task–capability large model. In the implementation, we treated the task-demand pipeline and the driving capability pipeline as two individual models. They share the same inputs except for the prompt, which are detailed in the Appendix A. For multi-label supervised contrastive learning, the capability pipeline uses status, category, and driver as supervisory labels with respective weights of 1, 0.5, and 0.25 to differentiate driving capabilities across different statuses, scenarios, and drivers. The demand pipeline, in contrast, employs category as its sole label. The hyperparameter settings for training were determined in accordance with BLIP-2 [10]. Both capability and demand pipelines were trained on the training set for 1000 epochs using the AdamW optimizer, with the standard “linear warm-up → cosine annealing” schedule. The learning rate started from 2 × 10 7 , warming for 1000 steps to reach 2 × 10 5 . The cosine decay progressively reduced the learning rate to 2 × 10 6 . This training stage was executed in parallel on four NVIDIA RTX A6000 GPUs (NVIDIA Corporation, Santa Clara, CA, USA), each handling a batch of 16 samples.
Stage 2: embedding comparator. Since task difficulty lacks a precise quantitative definition, we adopt the Anomalous Proportion (AnoP) metric introduced in previous work [39] as the supervisory signal during training. AnoP represents the proportion of anomalous behavior within a given time window, and we posit that it is positively correlated with task difficulty. For the learnable function g score ( · ) , the initial values of α and β are set to 0.01 and 5 to fit the distribution of AnoP. The embedding comparator was also trained on the training set for 1000 epochs, using Adam as the optimizer. The initial learning rate was set to 8 × 10 4 with a decay rate of 0.95 every 250 epochs. This stage was performed on a single NVIDIA RTX A6000 GPU.

5.2. Overall Evaluation Results

To evaluate whether the proposed framework produces meaningful representations of task demand d and driving capability c , we examined its performance on the test set and verified compliance with the properties defined in Section 3.1. The remainder of this section presents the experimental results in detail.

5.2.1. Task Consistency

In the proposed framework, an MLLM-based task–capability large model is employed to learn f d ( · ) and f c ( · ) . The model extracts the task demand d and the driver’s delivered driving capability c from the driving environment and the driving behavior inputs, representing each as a 16-dimensional feature vector. As noted in Property 1 in Section 3.1, d and c should exhibit task consistency, i.e., the distance between d or c in similar scenarios should be smaller than that in dissimilar scenarios. To validate it, we evaluated the embeddings obtained from the test set. For each driving trip, we compute the average d and c for every scenario. The embeddings are further projected onto a two-dimensional plane using t-SNE, as shown in Figure 11. In the figure, each point represents one trip–scenario instance, and different colors correspond to different scenario categories.
In Figure 11, it is observed that after the projection of t-SNE [85], the embeddings corresponding to different scenario categories are clearly separated, indicating that the model learns task-dependent demand and capability representations that are consistent within each scenario type. Moreover, the spatial pattern of task-demand embeddings in Figure 11a closely mirrors the pattern of driving capability embeddings in Figure 12b. Since drivers always attempt to allocate only the most appropriate capability level to the current driving context, the similarity observed here also suggests that the task–capability large model has learned demand and capability embeddings that sit in nearly the same latent space. To further quantify the inter-scenario separation, we compute the Euclidean distances between the centroids of all category clusters and visualize the results as a distance matrix as shown in Figure 12.
The distance matrix in Figure 12 is divided into three regions: green, yellow, and red. Overall, the green and yellow cells are visibly lighter than those in the red region, indicating that inter-class embedding distances (green/yellow) are larger than intra-class distances (red). Table 1 lists the mean distances for each region in detail. Since demand and capability cluster in a similar way, their distance statistics also line up closely. In particular, the distances between embeddings of the same scenario type are markedly larger than those between embeddings of different types, with values ranging roughly from 0.05 to 0.07. Within the same scenario class, the average distance for non-intersection scenes is lower than that for intersection scenes, with values around 0.02, suggesting that intersections are inherently more complex and variable, which is consistent with human intuition.
Table 2 dives deeper, listing how far the embedding of each scenario lies from the embeddings of a similar/different scenario group. Overall, the average distance between each scenario and similar scenarios exceeds that of the dissimilar ones, thereby satisfying Property 1. Figure 12 and Table 2 reveal that several scenes (IA1, IB1, IB2) exhibit embedding distances to scenarios of the same type that are nearly as large as their distances to scenes of different types. A common feature of these scenes is the absence of other vehicles or pedestrians. Under such circumstances, the intersection behaves more like a non-intersection curve than a car-following or pedestrian-containing intersection.

5.2.2. Intra-Task Capability Ordering

After generating the embedding of the driving capability c , the framework employs an embedding comparator to map it to a scalar space, resulting in a capability score c score . In this work, we introduce the Assumption 1, which states that a driver in a drunk state invariably exhibits lower capability than in the normal state. Experiments are adopted on the test set to verify whether the results comply with Property 2. Figure 13 illustrates the distribution of the driving capability scores under different driver states across all driving scenarios. A z-score normalization is applied to linearly rescale the values to the range [ 0 , 1 ] .
In Figure 13, the blue boxes correspond to the normal state and the orange boxes to the drunk state. In almost every scenario, the drunk distribution shifts downward, resulting in systematically lower medians than the normal state, which supports Assumption 1 and satisfies Property 2 that impaired drivers exhibit reduced capability within the same task context. To further our analysis, we paired each driver’s driving capability scores before and after alcohol consumption within the same scenario and calculated the post-drinking reduction Δ . The score differences were classified into five categories: marked decrease ( Δ 0.1 ), slight decrease ( 0.05 Δ < 0.1 ), almost no change ( 0.05 < Δ < 0.05 ), slight increase ( 0.1 < Δ 0.05 ), and marked increase ( Δ 0.1 ). For each scenario, we calculated the proportion of pairs that fall into each category. The results are summarized in Table 3.
In general, only about 17% of the cases show an increase in the driving capability score of at least 0.05, while 54% exhibit a decrease of that magnitude, and the remaining 28% stay roughly unchanged. The intersection scenarios are the most vulnerable, where the “marked decrease” shares reach 50% in IA2, IA3, and all IB scenarios. In contrast, NA2, NA4, and several NB scenes have loss rates below 25% and high proportions of “almost no change”, suggesting that non-interaction road segments leave less room for model-detectable degradation. Figure 13 and Table 3 confirm that the degradation of driving capability after drinking dominates across scenarios, especially at intersections, reinforcing Property 2 of the study.

5.2.3. Task–Capability Comparability

After obtaining the task-demand score and the driving capability score, the embedding comparator uses the learned g score ( · ) function to predict task difficulty. In our experiments, we fitted g score ( · ) to obtain task difficulty using AnoP, an objective metric proposed in previous work [38], as a ground-truth reference. Experiments are adopted in the test set to verify whether the results comply with the Property 3. The results are summarized in Table 4.
Table 4 indicates that the predicted task difficulty is consistently lower than the reference value in most cases and exhibits a relatively large mean absolute error (MAE) in several scenarios. However, the predicted difficulty for the drunk driving condition remains higher than that for the normal condition, which aligns with our expectations and gives the results a measure of practical significance. In general, the latent demand–capability representations can indeed be mapped onto a common, interpretable axis. However, the current results are suboptimal. Given that the other properties are well satisfied, the discrepancy may arise from factors such as: (i) the functional form of g score ( · ) may be inadequate to capture the true demand–capability–difficulty relationship; (ii) the reference metric may not coincide perfectly with the definition of task difficulty adopted in the TCI framework; or (iii) information may be lost when high-dimensional embeddings are compressed into scalar scores. In the following Section 5.3, we will investigate these possibilities in greater depth.

5.3. Ablation Study

To gain a deeper understanding of the proposed framework, we conduct a series of ablation experiments on different aspects. The results are detailed in the following sections.

5.3.1. Embedding Dimension

In the experiments of Section 5.2, the embedding dimensionality of task demand and driving capability was set to 16. To further investigate how this hyperparameter influences the learned demand and capability representations, we additionally evaluate two comparison settings with 8 and 32 dimensions. All other hyperparameters are kept identical across these runs.
Following the same procedure in Section 5.2.1, we visualize the learned embeddings using t-SNE as shown in Figure 14. Figure 14a,b show the 8-dimensional embeddings of the task demand and driving capability embeddings, respectively, Figure 14c,d depict the corresponding 32-dimensional embeddings. Comparison of the two plots shows that the t-SNE clusters are markedly more separable at 32 dimensions, whereas at 8 dimensions the clusters overlap heavily.
Similarly, the distances between scenarios of the same category and those of different categories are calculated. The results are summarized in Table 5. Increasing the latent dimension from 8 to 32 raises both intra-class and inter-class distances for demand and capability alike, which means that the clusters become more distinct; however, their internal cohesion also weakens. Currently, the dimension of 16 offers a balanced compromise. However, this is largely because our current scenario taxonomy is not highly fine-grained. We anticipate that once the scenarios are categorized at a more detailed level, a higher-dimensional embedding will yield better performance.
Table 6 presents the effect of the dimension of embedding of driving capability on the distribution of the reduction of the post-drinking driving capability score. As the embedding dimension increases, the ratio of the driving capability score reduction also increases, with the improvement particularly pronounced when the dimension increases from 8 to 16. However, when the dimensionality is further increased to 32, only marginal additional gain is offered, and the rate of “marked increase” also increases. In general, increasing embedding dimensionality enhances the sensitivity of the framework to capability degradation, but with tapering benefits beyond 16 dimensions under the current level of scenario granularity.
Table 7 summarizes how varying the dimensionality of demand and capability embedding spaces affects the accuracy of task-difficulty estimation. Although higher-dimensional embeddings can, in principle, encode richer information, the MAE remains nearly constant across the tested dimensions. We attribute this performance plateau to the final projection step: both embeddings are reduced to a single scalar before being passed to g score ( · ) , creating a bottleneck that discards much of the structure that distinguishes one driving episode from another. Together with the findings in Section 5.2.3, these results suggest that compressing demand and capability to a single scalar and modeling g score ( · ) with a simple functional form is overly restrictive. A more expressive scoring mechanism, such as a small neural network operating on the full vector-valued embeddings, might preserve more latent information and may yield more accurate task-difficulty estimates. We leave this exploration in the following sections.

5.3.2. Task-Difficulty Reference

As mentioned in Section 5.2.3, we fitted g score ( · ) to obtain task difficulty using AnoP as a reference to ground truth. In our previous experiments, we adopted the AnoP computed with the high-tolerance configuration reported in earlier work [38]. Consequently, its values clustered near 0, which may have hindered the ability of our function g score ( · ) = log ( 1 + α e β ( d score c score ) ) to achieve a good fit. Consequently, we also include the reference obtained under a lower tolerance configuration in both training and evaluation, and investigate how this choice affects the fitting of g score ( · ) . The results are presented in Table 8.
The results show that replacing the training reference with a stricter tolerance setting broadens the spread of the predicted task-difficulty scores. Under the stringent-tolerance reference, the ME even becomes positive, indicating that the model now overestimates the difficulty relative to the reference. However, the MAE remains comparatively large, reflecting limited accuracy. Employing both relaxed- and stringent-tolerance references in a joint-training scheme likewise widens the predicted range but fails to yield any meaningful reduction in MAE. In short, while the choice of task-difficulty reference metric clearly influences the numerical results, it is not the main cause of the suboptimal performance of the model. The following section investigates alternative explanations in more depth.

5.3.3. Task-Difficulty Function

Considering that neither embedding dimensionality nor the choice of difficulty reference appears to be the decisive factor behind the suboptimal results, we next investigate the influence of the g score function itself. Specifically, we replace the previous parametric formulation with a four-layer MLP and, in an extended setting, provide the raw demand and capability embeddings as additional inputs to the MLP. The corresponding results are summarized in Table 9.
In particular, simply replacing the original g score function with an MLP actually degrades the accuracy. This decline may arise from low-dimensional input, which makes it difficult for the MLP to approximate the desired nonlinear mapping. However, the adoption of an MLP allows the embeddings of task demand and driving capability to be incorporated as part of the input. In this configuration, the MAE is significantly reduced, with the stringent-tolerance AnoP reference; the error is cut by about half compared with the baseline. The experimental results indicate that, compared to scalar scores, embeddings capture a more comprehensive set of information and, therefore, achieve higher accuracy in predicting task difficulty.

5.4. Generalization to Real-World Data

As noted in Section 4.1, our framework requires a dataset that contains multiple instances of driving behavior under similar tasks, with drivers exhibiting different levels of capability in those instances, in order to produce meaningful experimental results. To this end, we select the UAH-Driveset [86] to test the performance of our framework on real-world data. UAH-Driveset comprises six different drivers, two routes (motorway, secondary), and three distinct driving states (normal, drowsy, aggressive). The dataset offers multiple modalities that characterize both the driving environment and driver behavior, including in-car camera video and accelerometer signals.
For our experiments, we select the motorway route and extract only the car-following segments to align with our own data. The data format is also formed to fit our data format. For the various modalities, we proceed as follows:
  • Image: Frames are sampled from the in-car camera footage.
  • Description: A set of natural language descriptions is constructed to depict the car-following scenario; one sentence is randomly selected for each sample.
  • Surroundings: Only the vehicle information directly ahead is used.
  • Behavior: Because the angle of the steering wheel and the pedal positions are unavailable, we represent driver behavior using only vehicle speed, yaw rotation speed, and lateral deviation from the center of the lane.
  • Prompt: The prompt is kept identical to that in our original dataset.
Since real-world traffic is highly dynamic, it is impossible to pair similar driving scenarios to what we have done through a simulator. To evaluate the proposed framework, we therefore calculate, for every sample, the ratio of the capability score to the demand score r, which reflects the moment-by-moment level of driver involvement. The experimental results are presented in Table 10.
As shown in Table 10, almost half of the normal samples lie in the balanced zone ( 0.95 r 1.05 ), with only 11.0% indicating a capability shortfall. When drivers become drowsy, the shortfall share almost doubles to 17.7%, while the proportion of marked surpluses ( r 1.10 ) drops to 8.6%, revealing a clear left shift caused by fatigue. In contrast, aggressive driving produces the opposite pattern: balanced episodes fall to 37.2%, whereas capability surpluses rise to 51.3%, reflecting frequent over-control associated with risk-taking behavior.

6. Discussion

This work takes an initial step toward the overarching goal of quantitatively modeling the Task–Capability Interface (TCI) so that misalignments between task demand and driving capability can be detected early and addressed in time to maintain safety. We have proposed a Multimodal Large Language Model (MLLM) framework that explicitly disentangles and quantifies the two TCI components. First, MLLM performs representation learning to obtain latent embeddings for task demand and driving capability. These embeddings are then projected into a common, orderable scalar space, enabling two complementary comparisons: the instantaneous gap between demand and capability, which flags potential risk moments; and the relative capability levels exhibited when the same task is performed by different drivers or by the same driver at different moments, providing a dynamic view of capability fluctuations. The experiments confirm that the learned representations satisfy the conceptual constraints introduced in Section 3.1: Property 1 and Property 2 are satisfied, while Property 3 is only partially met. Property 1 (task consistency) and Property 2 (intra-task capability ordering) are well met, as demonstrated by the larger inter-scenario embedding distances compared with intra-scenario distances and the monotonic degradation patterns observed after alcohol consumption. Property 3 (task–capability comparability) is partially satisfied: Although predicted difficulty preserves expected ordering between normal and impaired conditions, the mean absolute error (MAE) remains non-negligible. Finally, the framework generalizes to external data. In the subset of UAH-Driveset motorways, we observed clear state-dependent patterns: the capability-to-demand ratio shifted left under fatigue, indicating larger demand–capability gaps; and shifted right during aggressive maneuvers, reflecting the deployment of surplus capability. These trends align with established traffic psychology findings and underscore their interpretability.
Despite the encouraging results, there are still issues that remain open and deserve to be addressed. The first is a reliable estimate of task difficulty. Although the framework yields meaningful demand and capability embeddings, predicting task difficulty remains a challenge. Our experiments suggest that information loss in the embedding-to-scalar projection, together with the limited expressiveness of the current mapping function, contributes to this issue. However, a more fundamental obstacle is the lack of a precise and quantitative definition of task difficulty. AnoP used in the experiments might not be a perfect metric for task difficulty, and when alternative metrics are adopted, there is still a noticeable gap. It is necessary in the future to establish a reliable definition or representation of task difficulty. Another issue is establishing cross-task comparability. Because task demands are not yet comparable between heterogeneous driving tasks, capability comparisons can be performed only within similar tasks. Establishing cross-task comparability would “connect the dots,” linking isolated task-specific subgraphs into a unified demand–capability graph.
Building on this work, a practical and meaningful future direction is to generate natural language explanations alongside the extraction of task demand, driving capability, and task difficulty. Currently, in the proposed framework, the LLM is used to produce hidden states that are converted into demand and capability embeddings through a custom de-tokenizer. This procedure leverages the multimodal fusion of MLLM and the extensive latent knowledge in the LLM. However, an additional advantage of LLMs lies in their ability to generate coherent text. If the model could simultaneously output explanatory sentences, the resulting interpretations would enable more reasonable and situationally appropriate interventions.

7. Conclusions

This study proposed a data-driven route to operationalize the Task–Capability Interface (TCI) model by unifying visual, behavioral, and textual signals within a customized BLIP-2 architecture. The resulting low-dimensional embeddings preserve scenario semantics across diverse urban situations while capturing moment-to-moment capability variations induced by driver impairment. Empirically, the embeddings clustered tightly within similar tasks, the capability score consistently declined after alcohol consumption, and the scalar comparator reproduced the expected ordering between normal and impaired states, though residual errors suggest that a richer mapping could further improve accuracy. The framework also generalized to the real-world UAH Driveset without additional finetuning, revealing sensible capability—demand ratios during drowsy and aggressive maneuvers. In general, the proposed MLLM framework constitutes a promising step toward an explainable proactive risk assessment in intelligent vehicles, bridging the conceptual gap between task demand and driving capability and bringing us closer to co-drivers who can anticipate hazards before operational limits are exceeded.

Author Contributions

Conceptualization, H.Z.; methodology, H.Z.; software, H.Z.; validation, H.Z.; formal analysis, H.Z.; investigation, H.Z.; resources, K.T.; data curation, H.Z.; writing—original draft preparation, H.Z.; writing—review and editing, H.Z., A.C., and K.F.; visualization, H.Z.; supervision, A.C., K.F., and K.T.; project administration, A.C. and K.T.; funding acquisition, A.C. and K.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Japan Science and Technology Agency (JST) SPRING program under Grant Number JPMJSP2125.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset utilized in this project was collected from the driving simulator in the Advanced Research and Innovation Center, DENSO CORPORATION. This dataset has been specifically provided for the purpose of conducting various experiments within this project.

Acknowledgments

Our sincere thanks go to the “THERS Make New Standards Program for the Next Generation Researchers” for their invaluable support, and to DENSO CORPORATION for providing the equipment and collecting the dataset.

Conflicts of Interest

Alexander Carballo and Kazuya Takeda are employed by Tier IV Inc. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest. The funder was not involved in the study design, collection, analysis, interpretation of data, the writing of this article, or the decision to submit it for publication.

Abbreviations

The following abbreviations are used in this manuscript:
WHOWorld Health Organization
ABSAnti-lock Braking Systems
ESCElectronic Stability Control
ADASAdvanced Driver-Assistance Systems
ACCAdaptive Cruise Control
AEBAutonomous Emergency Braking
LKALane Keeping Assist
TCITask–Capability Interface
MLLMMultimodal Large Language Model
SupConSupervised Contrastive learning
MLPMulti-Layer Perceptron
FPSFrames Per Second
CANController Area Network
LLMLarge Language Model
AnoPAnomalous Proportion
MAEMean Absolute Error

Appendix A. Prompt of the Task–Capability Large Model

In the task–capability large model, the prompt comprises two components. The first component provides a general description of the data; for instance, when we use the simulator dataset whose driving scenes are modeled after Japanese roadway, it is “The driving environment assumes left-side traffic, so turns, lane changes, and yield rules are consistent with left-hand driving.” And for the evaluation on the UAH-Driveset, which was recorded on Spanish roads, the General Description section of the prompt is adjusted to emphasize right-side traffic.
The second component specifies the task itself. For the task-demand pipeline, the prompts are as follows:
  • Describe the overall task demand presented by the driving environment in this scene.
  • Based on the input data, what are the primary environmental and situational challenges the driver faces during this period?
  • Summarize the situational demands posed by this scene, including traffic, road structure, and potential hazards.
  • Based on the provided information, estimate the level of attentional and control effort required to navigate this scene.
  • How do the current traffic and road conditions impact the complexity of the driving task?
And the driving capability prompts are defined as follows:
  • Describe the driving capability demonstrated by the driver in this scene.
  • Evaluate the driver’s performance based on the scene and driving behavior provided.
  • Based on the input data, what can you infer about the driver’s driving skill and behavior during this period?
  • Given the visual and sensor data, extract indicators of driver control quality and risk awareness.
  • Summarize the driver-delivered capability using the combined perception from image and sensor data.
Every time the model is called, one prompt configuration is randomly selected from the pool of available combinations and used alongside the current input for inference.

References

  1. World Health Organization. Global Status Report on Road Safety 2023; Technical Report; World Health Organization: Geneva, Switzerland, 2023. [Google Scholar]
  2. Favarò, F.M.; Nader, N.; Eurich, S.O.; Tripp, M.; Varadaraju, N. Examining accident reports involving autonomous vehicles in California. PLoS ONE 2017, 12, e0184952. [Google Scholar] [CrossRef] [PubMed]
  3. SAE International. Taxonomy and Definitions for Terms Related to Driving Automation Systems for On-Road Motor Vehicles. SAE Standard J3016_2021. 2021. Available online: https://www.sae.org/standards/content/j3016_202104/ (accessed on 8 June 2025).
  4. Jatavallabha, A. Tesla’s Autopilot: Ethics and Tragedy. arXiv 2024, arXiv:2409.17380. [Google Scholar]
  5. Koopman, P. Lessons from the cruise robotaxi pedestrian dragging mishap. IEEE Reliab. Mag. 2024, 1, 54–61. [Google Scholar] [CrossRef]
  6. Fuller, R. The task-capability interface model of the driving process. Rech. 2000, 66, 47–57. [Google Scholar] [CrossRef]
  7. Fuller, R. Towards a general theory of driver behaviour. Accid. Anal. Prev. 2005, 37, 461–472. [Google Scholar] [CrossRef]
  8. Wong, J.T.; Huang, S.H. Modeling Driver Mental Workload for Accident Causation and Prevention. In Proceedings of the Eastern Asia Society for Transportation Studies, Surabaya, Indonesia, 16–19 November 2009; p. 365. [Google Scholar]
  9. Wu, J.; Gao, B.; Gao, J.; Yu, J.; Chu, H.; Yu, Q.; Gong, X.; Chang, Y.; Tseng, H.E.; Chen, H.; et al. Prospective role of foundation models in advancing autonomous vehicles. Research 2024, 7, 0399. [Google Scholar] [CrossRef]
  10. Li, J.; Li, D.; Savarese, S.; Hoi, S. Blip-2: Bootstrapping language-image pre-training with frozen image encoders and large language models. In Proceedings of the International Conference on Machine Learning, PMLR, Honolulu, HI, USA, 23–29 July 2023; pp. 19730–19742. [Google Scholar]
  11. Brookhuis, K.A.; de Waard, D. Assessment of drivers’ workload: Performance and subjective and physiological indexes. In Stress, Workload, and Fatigue; CRC Press: Boca Raton, FL, USA, 2000; pp. 321–333. [Google Scholar]
  12. Vlakveld, W. Hazard Anticipation of Young Novice Drivers: Assessing and Enhancing the Capabilities of Young Novice Drivers to Anticipate Latent Hazards in Road and Traffic Situations. Ph.D. Thesis, University of Groningen, Groningen, The Netherlands, 2011. [Google Scholar]
  13. Cestac, J.; Paran, F.; Delhomme, P. Young drivers’ sensation seeking, subjective norms, and perceived behavioral control and their roles in predicting speeding intention: How risk-taking motivations evolve with gender and driving experience. Saf. Sci. 2011, 49, 424–432. [Google Scholar] [CrossRef]
  14. Körber, M.; Gold, C.; Lechner, D.; Bengler, K. The influence of age on the take-over of vehicle control in highly automated driving. Transp. Res. Part Traffic Psychol. Behav. 2016, 39, 19–32. [Google Scholar] [CrossRef]
  15. Yan, Y.; Zhong, S.; Tian, J.; Song, L. Driving distraction at night: The impact of cell phone use on driving behaviors among young drivers. Transp. Res. Part Traffic Psychol. Behav. 2022, 91, 401–413. [Google Scholar] [CrossRef]
  16. Teh, E.; Jamson, S.; Carsten, O.; Jamson, H. Temporal fluctuations in driving demand: The effect of traffic complexity on subjective measures of workload and driving performance. Transp. Res. Part Traffic Psychol. Behav. 2014, 22, 207–217. [Google Scholar] [CrossRef]
  17. Engström, J.; Markkula, G.; Victor, T.; Merat, N. Effects of cognitive load on driving performance: The cognitive control hypothesis. Hum. Factors 2017, 59, 734–764. [Google Scholar] [CrossRef] [PubMed]
  18. Li, X.; Oviedo-Trespalacios, O.; Rakotonirainy, A.; Yan, X. Collision risk management of cognitively distracted drivers in a car-following situation. Transp. Res. Part Traffic Psychol. Behav. 2019, 60, 288–298. [Google Scholar] [CrossRef]
  19. Yang, Y.; Chen, Y.; Wu, C.; Easa, S.M.; Lin, W.; Zheng, X. Effect of highway directional signs on driver mental workload and behavior using eye movement and brain wave. Accid. Anal. Prev. 2020, 146, 105705. [Google Scholar] [CrossRef] [PubMed]
  20. Summala, H. Towards understanding motivational and emotional factors in driver behaviour: Comfort through satisficing. In Modelling Driver Behaviour in Automotive Environments: Critical Issues in Driver Interactions with Intelligent Transport Systems; Springer: London, UK, 2007; pp. 189–207. [Google Scholar]
  21. Foy, H.J.; Chapman, P. Mental workload is reflected in driver behaviour, physiology, eye movements and prefrontal cortex activation. Appl. Ergon. 2018, 73, 90–99. [Google Scholar] [CrossRef]
  22. Delmas, M.; Camps, V.; Lemercier, C. Should my automated car drive as I do? Investigating speed preferences of drivengers in various driving conditions. PLoS ONE 2023, 18, e0281702. [Google Scholar] [CrossRef]
  23. Sun, Z.; Xu, J.; Gu, C.; Xin, T.; Zhang, W. Investigation of Car following and Lane Changing Behavior in Diverging Areas of Tunnel–Interchange Connecting Sections Based on Driving Simulation. Appl. Sci. 2024, 14, 3768. [Google Scholar] [CrossRef]
  24. Kolekar, S.; De Winter, J.; Abbink, D. Human-like driving behaviour emerges from a risk-based driver model. Nat. Commun. 2020, 11, 1–13. [Google Scholar] [CrossRef]
  25. Saifuzzaman, M.; Zheng, Z.; Haque, M.M.; Washington, S. Revisiting the Task–Capability Interface model for incorporating human factors into car-following models. Transp. Res. Part Methodol. 2015, 82, 1–19. [Google Scholar] [CrossRef]
  26. Delhomme, P.; Meyer, T. Control motivation and young drivers’ decision making. Ergonomics 1998, 41, 373–393. [Google Scholar] [CrossRef]
  27. Yu, S.Y.; Malawade, A.V.; Muthirayan, D.; Khargonekar, P.P.; Al Faruque, M.A. Scene-graph augmented data-driven risk assessment of autonomous vehicle decisions. IEEE Trans. Intell. Transp. Syst. 2021, 23, 7941–7951. [Google Scholar] [CrossRef]
  28. de Winkel, K.N.; Christoph, M.; van Nes, N. Towards a framework of driver fitness: Operationalization and comparative risk assessment. Transp. Res. Interdiscip. Perspect. 2024, 23, 101030. [Google Scholar] [CrossRef]
  29. Rezapour, M.; Ksaibati, K. Identification of factors associated with various types of impaired driving. Humanit. Soc. Sci. Commun. 2022, 9, 1–11. [Google Scholar] [CrossRef]
  30. Nishitani, Y. Alcohol and traffic accidents in Japan. IATSS Res. 2019, 43, 79–83. [Google Scholar] [CrossRef]
  31. Shiferaw, B.A.; Crewther, D.P.; Downey, L.A. Gaze entropy measures detect alcohol-induced driver impairment. Drug Alcohol Depend. 2019, 204, 107519. [Google Scholar] [CrossRef] [PubMed]
  32. Wang, J.; Chai, W.; Venkatachalapathy, A.; Tan, K.L.; Haghighat, A.; Velipasalar, S.; Adu-Gyamfi, Y.; Sharma, A. A survey on driver behavior analysis from in-vehicle cameras. IEEE Trans. Intell. Transp. Syst. 2021, 23, 10186–10209. [Google Scholar] [CrossRef]
  33. Koch, K.; Maritsch, M.; Van Weenen, E.; Feuerriegel, S.; Pfäffli, M.; Fleisch, E.; Weinmann, W.; Wortmann, F. Leveraging driver vehicle and environment interaction: Machine learning using driver monitoring cameras to detect drunk driving. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems, Hamburg, Germany, 23–28 April 2023; pp. 1–32. [Google Scholar]
  34. Chatterjee, I.; Sharma, A. Driving Fitness Detection: A Holistic Approach for Prevention of Drowsy and Drunk Driving using Computer Vision Techniques. In Proceedings of the 2018 South-Eastern European Design Automation, Computer Engineering, Computer Networks and Society Media Conference (SEEDA_CECNSM), Kastoria, Greece, 22–24 September 2018; pp. 1–6. [Google Scholar] [CrossRef]
  35. Ki, M.; Cho, B.; Jeon, T.; Choi, Y.; Byun, H. Face identification for an in-vehicle surveillance system using near infrared camera. In Proceedings of the 2018 15th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Auckland, New Zealand, 27–30 November 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–6. [Google Scholar]
  36. Varghese, R.R.; Jacob, P.M.; Jacob, J.; Babu, M.N.; Ravikanth, R.; George, S.M. An integrated framework for driver drowsiness detection and alcohol intoxication using machine learning. In Proceedings of the 2021 International Conference on Data Analytics for Business and Industry (ICDABI), Sakheer, Bahrain, 25–26 October 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 531–536. [Google Scholar]
  37. Dai, J.; Teng, J.; Bai, X.; Shen, Z.; Xuan, D. Mobile phone based drunk driving detection. In Proceedings of the 2010 4th International Conference on Pervasive Computing Technologies for Healthcare, Munich, Germany, 22–25 March 2010; pp. 1–8. [Google Scholar] [CrossRef]
  38. Zhou, H.; Carballo, A.; Yamaoka, M.; Yamataka, M.; Fujii, K.; Takeda, K. DUIncoder: Learning to Detect Driving Under the Influence Behaviors from Various Normal Driving Data. Sensors 2025, 25, 1699. [Google Scholar] [CrossRef]
  39. Zhou, H.; Carballo, A.; Yamaoka, M.; Yamataka, M.; Takeda, K. A Self-Supervised Approach for Detection and Analysis of Driving Under Influence. In Proceedings of the 2024 IEEE 27th International Conference on Intelligent Transportation Systems (ITSC), Edmonton, AB, Canada, 24–27 September 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 4130–4137. [Google Scholar]
  40. Lowrie, J.; Brownlow, H. The impact of sleep deprivation and alcohol on driving: A comparative study. BMC Public Health 2020, 20, 1–9. [Google Scholar] [CrossRef]
  41. Saxby, D.J.; Matthews, G.; Warm, J.S.; Hitchcock, E.M.; Neubauer, C. Active and passive fatigue in simulated driving: Discriminating styles of workload regulation and their safety impacts. J. Exp. Psychol. Appl. 2013, 19, 287. [Google Scholar] [CrossRef]
  42. Jackson, M.L.; Croft, R.J.; Kennedy, G.; Owens, K.; Howard, M.E. Cognitive components of simulated driving performance: Sleep loss effects and predictors. Accid. Anal. Prev. 2013, 50, 438–444. [Google Scholar] [CrossRef]
  43. Zhang, X.; Zhao, X.; Du, H.; Rong, J. A study on the effects of fatigue driving and drunk driving on drivers’ physical characteristics. Traffic Inj. Prev. 2014, 15, 801–808. [Google Scholar] [CrossRef]
  44. Oviedo-Trespalacios, O.; Haque, M.M.; King, M.; Washington, S. Self-regulation of driving speed among distracted drivers: An application of driver behavioral adaptation theory. Traffic Inj. Prev. 2017, 18, 599–605. [Google Scholar] [CrossRef]
  45. Turnbull, P.R.; Khanal, S.; Dakin, S.C. The effect of cellphone position on driving and gaze behaviour. Sci. Rep. 2021, 11, 7692. [Google Scholar] [CrossRef]
  46. McEvoy, S.P.; Stevenson, M.R.; Woodward, M. The impact of driver distraction on road safety: Results from a representative survey in two Australian states. Inj. Prev. 2006, 12, 242–247. [Google Scholar] [CrossRef]
  47. Sheykhfard, A.; Haghighi, F. Driver distraction by digital billboards? Structural equation modeling based on naturalistic driving study data: A case study of Iran. J. Saf. Res. 2020, 72, 1–8. [Google Scholar] [CrossRef]
  48. Hughes, G.M.; Rudin-Brown, C.M.; Young, K.L. A simulator study of the effects of singing on driving performance. Accid. Anal. Prev. 2013, 50, 787–792. [Google Scholar] [CrossRef]
  49. Deffenbacher, J.L.; Deffenbacher, D.M.; Lynch, R.S.; Richards, T.L. Anger, aggression, and risky behavior: A comparison of high and low anger drivers. Behav. Res. Ther. 2003, 41, 701–718. [Google Scholar] [CrossRef]
  50. Hu, T.Y.; Xie, X.; Li, J. Negative or positive? The effect of emotion and mood on risky driving. Transp. Res. Part Traffic Psychol. Behav. 2013, 16, 29–40. [Google Scholar] [CrossRef]
  51. Eboli, L.; Mazzulla, G.; Pungillo, G. The influence of physical and emotional factors on driving style of car drivers: A survey design. Travel Behav. Soc. 2017, 7, 43–51. [Google Scholar] [CrossRef]
  52. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 5998–6008. [Google Scholar]
  53. Chowdhery, A.; Narang, S.; Devlin, J.; Bosma, M.; Mishra, G.; Roberts, A.; Barham, P.; Chung, H.W.; Sutton, C.; Gehrmann, S.; et al. Palm: Scaling language modeling with pathways. J. Mach. Learn. Res. 2023, 24, 1–113. [Google Scholar]
  54. Touvron, H.; Martin, L.; Stone, K.; Albert, P.; Almahairi, A.; Babaei, Y.; Bashlykov, N.; Batra, S.; Bhargava, P.; Bhosale, S.; et al. Llama 2: Open foundation and fine-tuned chat models. arXiv 2023, arXiv:2307.09288. [Google Scholar] [CrossRef]
  55. Achiam, J.; Adler, S.; Agarwal, S.; Ahmad, L.; Akkaya, I.; Aleman, F.L.; Almeida, D.; Altenschmidt, J.; Altman, S.; Anadkat, S.; et al. Gpt-4 technical report. arXiv 2023, arXiv:2303.08774. [Google Scholar] [CrossRef]
  56. Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language models are few-shot learners. Adv. Neural Inf. Process. Syst. 2020, 33, 1877–1901. [Google Scholar]
  57. Ouyang, L.; Wu, J.; Jiang, X.; Almeida, D.; Wainwright, C.; Mishkin, P.; Zhang, C.; Agarwal, S.; Slama, K.; Ray, A.; et al. Training language models to follow instructions with human feedback. Adv. Neural Inf. Process. Syst. 2022, 35, 27730–27744. [Google Scholar]
  58. Wei, J.; Wang, X.; Schuurmans, D.; Bosma, M.; Xia, F.; Chi, E.; Le, Q.V.; Zhou, D. Chain-of-thought prompting elicits reasoning in large language models. Adv. Neural Inf. Process. Syst. 2022, 35, 24824–24837. [Google Scholar]
  59. Ahn, M.; Brohan, A.; Brown, N.; Chebotar, Y.; Cortes, O.; David, B.; Finn, C.; Fu, C.; Gopalakrishnan, K.; Hausman, K.; et al. Do as i can, not as i say: Grounding language in robotic affordances. arXiv 2022, arXiv:2204.01691. [Google Scholar] [CrossRef]
  60. Wang, G.; Xie, Y.; Jiang, Y.; Mandlekar, A.; Xiao, C.; Zhu, Y.; Fan, L.; Anandkumar, A. Voyager: An open-ended embodied agent with large language models. arXiv 2023, arXiv:2305.16291. [Google Scholar] [CrossRef]
  61. Huang, W.; Wang, C.; Zhang, R.; Li, Y.; Wu, J.; Fei-Fei, L. Voxposer: Composable 3d value maps for robotic manipulation with language models. arXiv 2023, arXiv:2307.05973. [Google Scholar] [CrossRef]
  62. Cui, C.; Ma, Y.; Cao, X.; Ye, W.; Zhou, Y.; Liang, K.; Chen, J.; Lu, J.; Yang, Z.; Liao, K.D.; et al. A survey on multimodal large language models for autonomous driving. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2024; pp. 958–979. [Google Scholar]
  63. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 18–24 July 2021; pp. 8748–8763. [Google Scholar]
  64. Lu, J.; Batra, D.; Parikh, D.; Lee, S. Vilbert: Pretraining task-agnostic visiolinguistic representations for vision-and-language tasks. Adv. Neural Inf. Process. Syst. 2019, 13–23. [Google Scholar]
  65. Alayrac, J.B.; Donahue, J.; Luc, P.; Miech, A.; Barr, I.; Hasson, Y.; Lenc, K.; Mensch, A.; Millican, K.; Reynolds, M.; et al. Flamingo: A visual language model for few-shot learning. Adv. Neural Inf. Process. Syst. 2022, 35, 23716–23736. [Google Scholar]
  66. Huang, P.Y.; Xu, H.; Li, J.; Baevski, A.; Auli, M.; Galuba, W.; Metze, F.; Feichtenhofer, C. Masked autoencoders that listen. Adv. Neural Inf. Process. Syst. 2022, 35, 28708–28720. [Google Scholar]
  67. Georgescu, M.I.; Fonseca, E.; Ionescu, R.T.; Lucic, M.; Schmid, C.; Arnab, A. Audiovisual masked autoencoders. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–3 October 2023; pp. 16144–16154. [Google Scholar]
  68. Guo, Z.; Zhang, R.; Zhu, X.; Tang, Y.; Ma, X.; Han, J.; Chen, K.; Gao, P.; Li, X.; Li, H.; et al. Point-bind & point-llm: Aligning point cloud with multi-modality for 3d understanding, generation, and instruction following. arXiv 2023, arXiv:2309.00615. [Google Scholar]
  69. Tsimpoukelli, M.; Menick, J.L.; Cabi, S.; Eslami, S.; Vinyals, O.; Hill, F. Multimodal few-shot learning with frozen language models. Adv. Neural Inf. Process. Syst. 2021, 34, 200–212. [Google Scholar]
  70. Ding, X.; Han, J.; Xu, H.; Zhang, W.; Li, X. Hilm-d: Towards high-resolution understanding in multimodal large language models for autonomous driving. arXiv 2023, arXiv:2309.05186. [Google Scholar]
  71. Choudhary, T.; Dewangan, V.; Chandhok, S.; Priyadarshan, S.; Jain, A.; Singh, A.K.; Srivastava, S.; Jatavallabhula, K.M.; Krishna, K.M. Talk2bev: Language-enhanced bird’s-eye view maps for autonomous driving. In Proceedings of the 2024 IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–17 May 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 16345–16352. [Google Scholar]
  72. Hu, A.; Russell, L.; Yeo, H.; Murez, Z.; Fedoseev, G.; Kendall, A.; Shotton, J.; Corrado, G. Gaia-1: A generative world model for autonomous driving. arXiv 2023, arXiv:2309.17080. [Google Scholar] [CrossRef]
  73. Yang, M.; Du, Y.; Ghasemipour, K.; Tompson, J.; Schuurmans, D.; Abbeel, P. Learning interactive real-world simulators. arXiv 2023, 1, 6. [Google Scholar]
  74. Chen, L.; Sinavski, O.; Hünermann, J.; Karnsund, A.; Willmott, A.J.; Birch, D.; Maund, D.; Shotton, J. Driving with llms: Fusing object-level vector modality for explainable autonomous driving. In Proceedings of the 2024 IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–17 May 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 14093–14100. [Google Scholar]
  75. Fu, D.; Li, X.; Wen, L.; Dou, M.; Cai, P.; Shi, B.; Qiao, Y. Drive like a human: Rethinking autonomous driving with large language models. In Proceedings of the 2024 IEEE/CVF Winter Conference on Applications of Computer Vision Workshops (WACVW), Waikoloa, HI, USA, 1–6 January 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 910–919. [Google Scholar]
  76. Xu, Z.; Zhang, Y.; Xie, E.; Zhao, Z.; Guo, Y.; Wong, K.Y.K.; Li, Z.; Zhao, H. Drivegpt4: Interpretable end-to-end autonomous driving via large language model. IEEE Robot. Autom. Lett. 2024. [Google Scholar] [CrossRef]
  77. Shao, H.; Hu, Y.; Wang, L.; Song, G.; Waslander, S.L.; Liu, Y.; Li, H. Lmdrive: Closed-loop end-to-end driving with large language models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 15120–15130. [Google Scholar]
  78. Khosla, P.; Teterwak, P.; Wang, C.; Sarna, A.; Tian, Y.; Isola, P.; Maschinot, A.; Liu, C.; Krishnan, D. Supervised Contrastive Learning. In Advances in Neural Information Processing Systems; Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., Lin, H., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2020; Volume 33, pp. 18661–18673. [Google Scholar]
  79. Burges, C.; Shaked, T.; Renshaw, E.; Lazier, A.; Deeds, M.; Hamilton, N.; Hullender, G. Learning to rank using gradient descent. In Proceedings of the 22nd International Conference on Machine Learning, Bonn, Germany, 7–11 August 2005; pp. 89–96. [Google Scholar]
  80. Sectional Committee of AD Safety Evaluation, Automated Driving Subcommittee; Japan Automobile Manufacturers Association, Inc. Automated Driving Safety Evaluation Framework Ver. 3.0: Guidelines for Safety Evaluation of Automated Driving Technology; Technical Report; Japan Automobile Manufacturers Association: Tokyo, Japan, 2022. [Google Scholar]
  81. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. Pytorch: An imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 2019, 8026–8037. [Google Scholar]
  82. Li, D.; Li, J.; Le, H.; Wang, G.; Savarese, S.; Hoi, S.C. LAVIS: A One-stop Library for Language-Vision Intelligence. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 3: System Demonstrations), Toronto, Canada, 9–14 July 2023; pp. 31–41. [Google Scholar]
  83. Wolf, T.; Debut, L.; Sanh, V.; Chaumond, J.; Delangue, C.; Moi, A.; Cistac, P.; Rault, T.; Louf, R.; Funtowicz, M.; et al. Transformers: State-of-the-Art Natural Language Processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, Online, 16–20 November 2020; pp. 38–45. [Google Scholar]
  84. Sima, C.; Renz, K.; Chitta, K.; Chen, L.; Zhang, H.; Xie, C.; Luo, P.; Geiger, A.; Li, H. DriveLM: Driving with Graph Visual Question Answering. arXiv 2023, arXiv:2312.14150. [Google Scholar] [CrossRef]
  85. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  86. Romera, E.; Bergasa, L.M.; Arroyo, R. Need data for driver behaviour analysis? Presenting the public UAH-DriveSet. In Proceedings of the 2016 IEEE 19th international conference on intelligent transportation systems (ITSC), Rio de Janeiro, Brazil, 1–4 November 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 387–392. [Google Scholar]
Figure 1. Illustration of the TCI model. Reproduce from [7], © 2005 by Elsevier, with permission.
Figure 1. Illustration of the TCI model. Reproduce from [7], © 2005 by Elsevier, with permission.
Sensors 25 05805 g001
Figure 2. Overall structure of the proposed framework.
Figure 2. Overall structure of the proposed framework.
Sensors 25 05805 g002
Figure 3. Network structure of the task–capability large model. A flame marks modules whose parameters are trainable, while a snowflake marks those kept frozen. Color coding denotes the pipeline in which each component is used: green for task demand, orange for driving capability, and gray for components shared by both.
Figure 3. Network structure of the task–capability large model. A flame marks modules whose parameters are trainable, while a snowflake marks those kept frozen. Color coding denotes the pipeline in which each component is used: green for task demand, orange for driving capability, and gray for components shared by both.
Sensors 25 05805 g003
Figure 4. Illustration of the structure of the LLM input token series. Each color corresponds to a specific modality of tokens. n indicates how many image frames are included in the input.
Figure 4. Illustration of the structure of the LLM input token series. Each color corresponds to a specific modality of tokens. n indicates how many image frames are included in the input.
Sensors 25 05805 g004
Figure 5. Network structure of the embedding comparator.
Figure 5. Network structure of the embedding comparator.
Sensors 25 05805 g005
Figure 6. Illustration of the target (red) and the learned (blue) demand–capability comparison function.
Figure 6. Illustration of the target (red) and the learned (blue) demand–capability comparison function.
Sensors 25 05805 g006
Figure 7. The driving simulator for collecting driving data. Reproduced from [38], distributed under the terms of the Creative Commons Attribution License (CC BY 4.0).
Figure 7. The driving simulator for collecting driving data. Reproduced from [38], distributed under the terms of the Creative Commons Attribution License (CC BY 4.0).
Sensors 25 05805 g007
Figure 8. Illustration of the categorized scenarios: Circles and diamonds denote the start and end points, respectively; Arrows indicate the travel direction. Scenario categories are color-coded: green for non-intersection, blue for non-intersection (tunnel), and orange for intersections. Each scenario is represented using a code: NA for non-intersection in Route A; IB for intersection in Route B. (a) Route A—urban road with moderate traffic; (b) Route B—urban road with light traffic and pedestrian presence.
Figure 8. Illustration of the categorized scenarios: Circles and diamonds denote the start and end points, respectively; Arrows indicate the travel direction. Scenario categories are color-coded: green for non-intersection, blue for non-intersection (tunnel), and orange for intersections. Each scenario is represented using a code: NA for non-intersection in Route A; IB for intersection in Route B. (a) Route A—urban road with moderate traffic; (b) Route B—urban road with light traffic and pedestrian presence.
Sensors 25 05805 g008
Figure 9. Examples of the in-cabin camera image modality. Reproduced from [38], distributed under the terms of the Creative Commons Attribution License (CC BY 4.0).
Figure 9. Examples of the in-cabin camera image modality. Reproduced from [38], distributed under the terms of the Creative Commons Attribution License (CC BY 4.0).
Sensors 25 05805 g009
Figure 10. Example of the driver operation modality: (a) vehicle velocity (km/h); (b) steering angle ([−1, 1]); (c) throttle position ([0, 1]); (d) brake position ([0, 1]); (e) lane center deviation (m); (f) yaw rotation speed (rad/s). Reproduced from [38], distributed under the terms of the Creative Commons Attribution License (CC BY 4.0).
Figure 10. Example of the driver operation modality: (a) vehicle velocity (km/h); (b) steering angle ([−1, 1]); (c) throttle position ([0, 1]); (d) brake position ([0, 1]); (e) lane center deviation (m); (f) yaw rotation speed (rad/s). Reproduced from [38], distributed under the terms of the Creative Commons Attribution License (CC BY 4.0).
Sensors 25 05805 g010
Figure 11. Scenario-wise t-SNE visualization of task demand and driving capability: (a) task demand; (b) driving capability.
Figure 11. Scenario-wise t-SNE visualization of task demand and driving capability: (a) task demand; (b) driving capability.
Sensors 25 05805 g011
Figure 12. Inter-scenario Euclidean distance matrices of task demand and driving capability: (a) task demand; (b) driving capability. The colored frames highlight three region types: green—distances within intersection scenarios; yellow—distances within non-intersection scenarios; red—distances between intersection and non-intersection scenarios.
Figure 12. Inter-scenario Euclidean distance matrices of task demand and driving capability: (a) task demand; (b) driving capability. The colored frames highlight three region types: green—distances within intersection scenarios; yellow—distances within non-intersection scenarios; red—distances between intersection and non-intersection scenarios.
Sensors 25 05805 g012
Figure 13. Distribution of normalized driving capability scores for normal and drunk drivers across different scenarios.
Figure 13. Distribution of normalized driving capability scores for normal and drunk drivers across different scenarios.
Sensors 25 05805 g013
Figure 14. Impact of embedding dimensionality on cluster separability in the learned latent spaces: (a) task demand (dimension = 8); (b) driving capability (dimension = 8); (c) task demand (dimension = 32); (d) driving capability (dimension = 32).
Figure 14. Impact of embedding dimensionality on cluster separability in the learned latent spaces: (a) task demand (dimension = 8); (b) driving capability (dimension = 8); (c) task demand (dimension = 32); (d) driving capability (dimension = 32).
Sensors 25 05805 g014
Table 1. Average Euclidean distance between latent embeddings of intersection and non-intersection scenario groups.
Table 1. Average Euclidean distance between latent embeddings of intersection and non-intersection scenario groups.
I2I ↓N2N ↓I2N ↑
Demand0.2070.1880.260
Capability0.2090.1900.263
I2I = Intersection-to-Intersection; N2N = Non-intersection-to-Non-intersection; I2N = Intersection-to-Non-intersection. ↑ larger value signifies better results. ↓ smaller value signifies better results.
Table 2. Scenario-level Euclidean distances to intersection and non-intersection scenario groups.
Table 2. Scenario-level Euclidean distances to intersection and non-intersection scenario groups.
ScenarioDemandCapability
Similar ↓Different ↑AverageSimilar ↓Different ↑Average
IA10.2380.2330.2340.2410.2360.238
IA20.1800.2650.2360.1810.2670.237
IA30.1800.2720.2400.1820.2760.243
IA40.1970.2930.2400.2010.3010.266
IB10.2330.2280.2300.2350.2310.233
IB20.2070.2030.2050.2090.2040.206
IB30.2130.3230.2840.2150.3290.289
IB40.2230.2700.2540.2250.2740.256
NA10.2610.3980.3090.2630.4020.312
NA20.2000.2930.2330.2040.3020.238
NA30.2150.2780.2370.2170.2800.239
NA40.1960.2510.2150.1970.2550.217
NA50.1460.2430.1800.1480.2470.183
NA60.1490.2450.1830.1510.2490.185
NA70.1940.2190.2030.1950.2220.205
NB10.1740.2360.1960.1770.2410.199
NB20.1670.2780.2060.1680.2820.208
NB30.1450.2320.1750.1460.2340.177
NB40.1610.2160.1800.1620.2190.182
NB50.1640.2650.1990.1650.2670.201
↑ larger value signifies better results. ↓ smaller value signifies better results.
Table 3. Scenario-wise distribution (%) of post-drinking driving capability score reduction.
Table 3. Scenario-wise distribution (%) of post-drinking driving capability score reduction.
Scenario ( , 0.1 ] ( 0.1 , 0.05 ] ( 0.05 , 0.05 ) [ 0.05 , 0.1 ) [ 0.1 , )
IA116.7016.716.750.0
IA25.6033.333.327.8
IA35.611.127.85.650.0
IA416.711.161.15.65.6
IB19.13.021.212.154.5
IB211.817.623.517.629.4
IB311.82.923.52.958.8
IB410.55.231.610.542.1
NA15.611.138.916.727.8
NA210.010.060.015.05.0
NA35.35.321.126.342.1
NA46.69.821.313.149.2
NA59.78.323.620.837.5
NA68.32.816.730.641.7
NA715.112.127.29.136.3
NB105.947.15.941.2
NB28.112.937.112.929.0
NB315.72.013.723.545.1
NB45.714.331.417.131.4
NB55.95.947.111.829.4
Overall9.47.828.415.938.5
↑ larger value signifies better results. ↓ smaller value signifies better results.
Table 4. Scenario-wise accuracy of the learned task-difficulty estimator.
Table 4. Scenario-wise accuracy of the learned task-difficulty estimator.
ScenarioNormalDrunk
M.RefM.PredMAE ↓M.RefM.PredMAE ↓
IA10.0680.0020.0650.1300.0800.100
IA20.0720.0340.0410.1130.0810.054
IA30.0380.0110.0340.0310.0740.063
IA40.0680.0160.0540.1010.0220.090
IB10.0700.0320.0500.1610.1500.104
IB20.0420.0060.0390.0980.0070.092
IB30.0510.0140.0440.0960.0310.076
IB40.0420.0130.0310.0520.0250.040
NA100.0010.0010.0030.0010.002
NA20.0020.0060.0060.0050.0270.025
NA30.0410.0220.0360.0900.1260.124
NA40.0150.0020.0140.0420.0170.026
NA50.0130.0010.0130.0330.0050.032
NA60.0250.0020.0230.0460.0180.041
NA70.0060.0030.0060.0680.0320.046
NB10.00700.0070.0060.0020.006
NB20.0080.0010.0080.0230.0070.018
NB30.0080.0020.0090.0190.0050.017
NB40.0210.0060.0170.0420.0320.045
NB50.0430.0390.0520.0540.0590.061
Overall0.0290.0080.0260.0550.0310.045
M.Ref = mean reference task difficulty; M.Pred = mean predicted task difficulty; MAE = mean absolute error. ↓ smaller value signifies better results.
Table 5. Effect of embedding dimensionality on intra- and inter-category centroid distances.
Table 5. Effect of embedding dimensionality on intra- and inter-category centroid distances.
DimensionDemandCapability
Similar ↓Different ↑Similar ↓Different ↑
80.1420.1770.1770.181
160.2090.2600.1940.264
320.2700.3490.2670.344
↑ larger value signifies better results. ↓ smaller value signifies better results.
Table 6. Effect of embedding dimensionality on the distribution (%) of post-drinking driving capability score reduction.
Table 6. Effect of embedding dimensionality on the distribution (%) of post-drinking driving capability score reduction.
Dimension ( , 0.1 ] ( 0.1 , 0.05 ] ( 0.05 , 0.05 ) [ 0.05 , 0.1 ) [ 0.1 , )
89.710.234.918.326.8
169.47.828.415.938.5
3211.48.624.815.239.9
↑ larger value signifies better results. ↓ smaller value signifies better results.
Table 7. Effect of embedding dimensionality on the accuracy of task-difficulty estimation.
Table 7. Effect of embedding dimensionality on the accuracy of task-difficulty estimation.
DimensionNormalDrunk
MEMAE ↓MEMAE ↓
8−0.0150.025−0.0330.044
16−0.0220.026−0.0240.045
32−0.0250.027−0.0440.047
ME = mean error; MAE = mean absolute error. ↓ smaller value signifies better results.
Table 8. Effect of AnoP-tolerance configuration on the accuracy of the task-difficulty estimation.
Table 8. Effect of AnoP-tolerance configuration on the accuracy of the task-difficulty estimation.
ToleranceNormalDrunk
RelaxedStringentMEMAE ↓MEMAE ↓
−0.0220.026−0.0240.045
0.0550.110−0.1950.209
Δ −0.0120.026−0.0370.047
Δ 0.0630.114−0.1780.196
ME = mean error; MAE = mean absolute error. ▲ means it is used in both training and evaluation; Δ means it is only used during training. ↓ smaller value signifies better results.
Table 9. Effect of task-difficulty function on the accuracy of the task-difficulty estimation.
Table 9. Effect of task-difficulty function on the accuracy of the task-difficulty estimation.
MethodToleranceNormalDrunk
RelaxedStringentMEMAE ↓MEMAE ↓
baseline −0.0220.026−0.0240.045
w/ NN −0.0190.027−0.0440.051
w/ NN & EMB −0.0140.021−0.0160.037
baseline 0.0550.110−0.1950.209
w/ NN −0.0110.083−0.2760.283
w/ NN & EMB −0.0120.061−0.0630.102
ME = mean error; MAE = mean absolute error. NN means a neural network served as g score ( · ) . EMB means using the embeddings of task demand and driving capability together with scores as the input of g score ( · ) . ▲ means it is used in both training and evaluation. ↓ smaller value signifies better results.
Table 10. Distribution (%) of capability-to-demand score ratios under different driver states on the UAH-Driveset motorway subset.
Table 10. Distribution (%) of capability-to-demand score ratios under different driver states on the UAH-Driveset motorway subset.
Status ( 0 , 0.9 ] ( 0.9 , 0.95 ] ( 0.95 , 1.05 ) [ 1.05 , 1.1 ) [ 1.1 , 1.2 ) [ 1.2 , )
Normal4.86.248.623.110.66.7
Drowsy6.910.846.926.85.33.3
Aggressive5.16.437.234.19.47.8
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, H.; Carballo, A.; Fujii, K.; Takeda, K. From Concept to Representation: Modeling Driving Capability and Task Demand with a Multimodal Large Language Model. Sensors 2025, 25, 5805. https://doi.org/10.3390/s25185805

AMA Style

Zhou H, Carballo A, Fujii K, Takeda K. From Concept to Representation: Modeling Driving Capability and Task Demand with a Multimodal Large Language Model. Sensors. 2025; 25(18):5805. https://doi.org/10.3390/s25185805

Chicago/Turabian Style

Zhou, Haoran, Alexander Carballo, Keisuke Fujii, and Kazuya Takeda. 2025. "From Concept to Representation: Modeling Driving Capability and Task Demand with a Multimodal Large Language Model" Sensors 25, no. 18: 5805. https://doi.org/10.3390/s25185805

APA Style

Zhou, H., Carballo, A., Fujii, K., & Takeda, K. (2025). From Concept to Representation: Modeling Driving Capability and Task Demand with a Multimodal Large Language Model. Sensors, 25(18), 5805. https://doi.org/10.3390/s25185805

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop