Next Article in Journal
Eye Movement Impairment in Women Undergoing Chemotherapy
Previous Article in Journal
A Review of Digital Eye Strain: Binocular Vision Anomalies, Ocular Surface Changes, and the Need for Objective Assessment
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Interpretable Quantification of Scene-Induced Driver Visual Load: Linking Eye-Tracking Behavior to Road Scene Features via SHAP Analysis

1
School of Automotive and Traffic Engineering, Jiangsu University, Zhenjiang 212013, China
2
Changzhou Engineering and Technology Institute, Jiangsu University, Zhenjiang 213000, China
*
Author to whom correspondence should be addressed.
J. Eye Mov. Res. 2025, 18(5), 40; https://doi.org/10.3390/jemr18050040
Submission received: 14 August 2025 / Revised: 1 September 2025 / Accepted: 5 September 2025 / Published: 9 September 2025

Abstract

Road traffic accidents remain a major global public health concern, where complex urban driving environments significantly elevate drivers’ visual load and accident risks. Unlike existing research that adopts a macro perspective by considering multiple factors such as the driver, vehicle, and road, this study focuses on the driver’s visual load, a key safety factor, and its direct source—the driver’s visual environment. We have developed an interpretable framework combining computer vision and machine learning to quantify how road scene features influence oculomotor behavior and scene-induced visual load, establishing a complete and interpretable link between scene features, eye movement behavior, and visual load. Using the DR(eye)VE dataset, visual attention demand is established through occlusion experiments and confirmed to correlate with eye-tracking metrics. K-means clustering is applied to classify visual load levels based on discriminative oculomotor features, while semantic segmentation extracts quantifiable road scene features such as the Green Visibility Index, Sky Visibility Index and Street Canyon Enclosure. Among multiple machine learning models (Random Forest, Ada-Boost, XGBoost, and SVM), XGBoost demonstrates optimal performance in visual load detection. SHAP analysis reveals critical thresholds: the probability of high visual load increases when pole density exceeds 0.08%, signage surpasses 0.55%, or buildings account for more than 14%; while blink duration/rate decrease when street enclosure exceeds 38% or road congestion goes beyond 25%, indicating elevated visual load. The proposed framework provides actionable insights for urban design and driver assistance systems, advancing traffic safety through data-driven optimization of road environments.

1. Introduction

Urban traffic accidents remain a critical public health issue, and improving urban road safety requires multifaceted approaches. Researchers in urban planning and management suggest that urban landscapes and built environments are closely related to traffic safety, as drivers’ perception and feedback on their driving environment affect the overall safety of urban traffic [1,2]. Kaygisiz et al. (2017) discussed the impact of road alignment, building coverage, lane count, land use types, and the number of bus stops on urban traffic safety [3]. Asadi et al. (2022) analyzed the relationship between multiscale built environment factors—such as the year of housing construction, land use type, road hierarchy, cycling activities and infrastructure, proximity characteristics, and neighboring variables—and the number of traffic accidents [4]. With advancements in street view maps and the rapid development of deep learning in computer vision, discovering the quantitative relationship between urban landscape elements and traffic safety has become an effective method for assessing urban traffic safety, providing a solid basis for improving the urban road environment [5,6,7,8]. However, current studies often rely on subjective assessments [9] or historical traffic accident data [10,11,12], and there is insufficient research on the mechanisms by which urban environmental elements influence drivers and consequently contribute to traffic accidents.
Scholars in the field of driver assistance systems aim to improve traffic safety by studying the impact of driving scenarios on drivers’ physiological load and attention. Compared to highways and rural roads, urban environments are often more complex. Variables such as intersections, dynamic changes at bus stops, pedestrians and non-motorized vehicles, parked vehicles, traffic signs, and billboards significantly increase the amount of visual information drivers need to process [13], thereby increasing their driving load. Visual load is a critical factor affecting drivers’ ability to process visual information [14]. Sustained high-intensity driving loads can lead to driver fatigue, delayed reactions, more frequent operational errors, and poor decision-making, all of which impair driving performance and are a major cause of urban traffic accidents [15,16,17,18]. Therefore, by monitoring drivers’ physiological indicators, facial expressions, and eye movements, constructing multimodal feature fusion models to identify high-load states and provide warnings can effectively reduce the risk of traffic accidents [19,20,21]. Nevertheless, while existing approaches can monitor visual load through physiological signals, they rarely deconstruct the combined effects of road scene elements or provide actionable optimization suggestions for urban design.
Building upon these findings, this study proposes an interpretable analytical framework that quantifies scene-induced visual load by establishing correlations between eye-tracking behavior and road scene features through SHAP analysis. Specifically, the research will: (1) classify visual load levels using eye movement indicators; (2) identify high-impact urban scene elements and their dynamic effects on visual load; and (3) elucidate the mechanisms through which scene modifications influence eye movements and consequently alter visual load. This approach is expected to provide a novel data-driven paradigm for assessing urban traffic safety based on scene-visual load interactions.

2. Literature Review

2.1. Review of the Impact of Urban Landscape and Built Environment on Traffic Safety

In recent years, with the rapid pace of urbanization and advancements in computer vision technology, the impact of urban scene elements and built environments on traffic safety has become a research hotspot. Using technologies such as remote sensing and street view images, researchers have aimed to uncover the complex relationships between road environments, driving behavior, and traffic safety. Liu et al. (2023), utilizing high-resolution remote sensing and street view images combined with deep learning-based semantic segmentation techniques, analyzed the visual landscape features of urban expressways [1]. Their results established a model linking road landscape features with driver fatigue, showing that landscape richness can reduce fatigue and improve traffic safety. Similarly, Cai et al. (2022) applied deep neural networks to extract visual environment features from Google Street View images to study the impact of these visual features on speeding-related accidents [6]. They found that tree coverage along roadsides was linked to fewer speeding accidents, while highly complex visual environments increased collision risks.
In the study of built environments and pedestrian safety, González et al. (2023) combined collision prediction with built environment detection using panoramic street view images [22]. They showed that built environment elements strongly influence pedestrian collision risks. Bustos et al. (2021) proposed an automated method that used classification, segmentation, and explainability techniques to capture dangerous urban landscapes [23]. Their study found that reducing the space between roads and buildings while increasing green open spaces could improve pedestrian safety. Additionally, Tanprasert et al. (2022) identified road accident hotspots with street view image recognition and semantic segmentation techniques [24]. They discovered that the size and proximity of surrounding objects were key factors influencing accident susceptibility. Qin et al. (2020), using graph convolutional networks, analyzed the relationship between urban road networks and traffic congestion [25]. They found that physical, functional, and topological features of the built environment are critical in shaping congestion, offering strategies to prevent jams and enhance road safety.
In summary, existing literature has demonstrated that urban scene semantics and built environment features profoundly affect driver behavior and traffic safety. Factors such as the complexity of road environments and the layout of greenery can significantly impact driver performance, accident rates, and pedestrian safety.

2.2. Review of Scenarios and Driving Visual Load

Different driving scenarios require drivers to process varying degrees of visual information, which is a key factor influencing driving visual load. By studying the characteristics of these scenarios, we can gain a deeper understanding of the sources of driving load and optimize driving conditions to reduce driver strain. Foy et al. (2018) designed four road types in a simulated environment with different scene elements [26]. They found that when the road environment shifted from a simple two-lane road with a median strip to the more complex “city center” road, drivers’ psychological cognitive load increased significantly. Liu et al. (2011) used a driving simulator where drivers wore eye-tracking devices to examine the effects of varying information levels on highway signage in a visual recognition experiment [27]. They reported that when a sign contained more than five elements, drivers’ eye movement patterns changed markedly, and recognition effectiveness declined. He et al. (2017) utilized a non-invasive eye-tracking system to record drivers’ eye movement parameters as they drove through highway tunnels under two lighting conditions [28]. They showed that reduced brightness, lower uniformity, and flicker inside tunnels enlarged pupil diameter and increased visual load. Ahlström et al. (2018) studied two different driving scenarios: a winding, low-traffic rural environment and a suburban road with higher traffic density and denser roadside environments, finding that driver demands and visual load varied between the two [29]. Fiolić et al. (2023) investigated the relationship between road markings and signs with different visibility levels in nighttime conditions and driver behavior and cognitive load [30]. They discovered that improved visibility positively impacted cognitive load, with optimal cognitive load occurring under high visibility. These studies, while providing partial insights into the impact of environmental factors on visual workload, predominantly focus either on isolated scene elements or generalized scenario comparisons. They fail to deliver either comprehensive quantitative analysis of all elements within specific driving contexts or establish an interpretable correlation model connecting scene features, oculomotor behavior, and visual load dynamics.

2.3. Review of Driver Attention Studies

When drivers face excessive attention demands, the driving visual load correspondingly increases, which not only heightens psychological and physiological stress but may also negatively impact driving safety. The study by Cvahte et al. (2019) [31] points out that visual and cognitive attention demands directly affect drivers’ ability to perceive changes in the traffic environment. When drivers’ attention is distracted, they are more prone to making driving errors, thereby increasing accident risk. Gaze occlusion, eye-tracking, and physiological measurements are three effective methods for assessing driver attention. Gaze occlusion can identify when and for how long drivers need to observe the traffic environment, while also estimating residual attention. Kircher et al. (2020) found that attention demands on highways increase as following distance decreases and traffic density rises, which reduced occlusion frequency and duration [32]. Liu et al. (2021) set up urban road traffic scenarios involving multiple traffic variables and conducted gaze occlusion experiments with 30 participants to explore the traffic factors that influence attention demands in urban environments [13].
In the field of eye-tracking, Zeng et al. (2023) examined prolonged driving through simulation and eye-tracking [33]. They found that longer driving reduced fixation durations but increased saccade amplitudes, indicating impaired visual sensitivity and risk assessment. Regarding physiological indicators, Aminosharieh et al. (2023) used EEG, ECG, and EDA signals in a simulated environment to assess drivers’ attention states [34]. Liu et al. (2023) further confirmed through EEG studies that brain activity differs significantly under varying cognitive loads, particularly in high-cognitive-load tasks where drivers’ attention is more easily disrupted [35]. Given the limitations of quantifying driving visual load through physiological indicators, this study introduces visual attention demand as an indicator to explore the correlation between drivers’ visual metrics and attention demands. Based on this, driving visual load is quantified, aiming to provide a more reasonable and reliable method for predicting and evaluating driving visual load.

3. Methodology

3.1. Methodological Framework

The overall framework of this study is illustrated in Figure 1, in which * is used as a locator identifier. First, using the DR(eye)VE dataset [36], we filtered and selected scenes related to urban environments, extracted the corresponding in-vehicle video and eye-tracking data, and segmented the samples to construct the urban driving dataset for this research.
Next, based on the eye movement event types identified from raw eye-tracking data, we calculated the corresponding visual metrics. For each scene sample, we extracted fixation metrics (fixation duration and fixation rate), blink metrics (blink duration and blink rate), and saccade metrics (saccade duration, saccade frequency, and average saccade amplitude). For the in-vehicle video data, we conducted a standardized visual occlusion experiment to quantify attention demand: the driver’s attention demand in each scene sample was characterized by the frequency of visual occlusion operations performed by the experimenter, covering typical urban scenarios such as curves, straight roads, and intersections. Using a pre-trained DeeplabV3+ semantic segmentation pipeline, we extracted scene element information from each video frame, calculating pixel-wise proportions of key elements (e.g., vehicles, pedestrians, traffic signs) to represent environmental characteristics. Frame-level features were then averaged to generate scene-level representations.
Finally, Pearson correlation analysis was performed to examine the relationship between visual metrics and attention demand, thereby identifying eye movement features significantly associated with scene-induced visual load. Based on these significantly correlated visual metrics, K-means clustering was applied. By integrating the distribution of attention demand intensity with spatial mapping of cluster centroids, we achieved quantitative calibration of visual load levels. Following calibration, interpretable predictive modeling was implemented: machine learning models were constructed to simultaneously predict visual load levels and eye movement metrics from scene features. SHAP value analysis was then conducted to quantify the directional effects of specific scene features on visual load and eye movement behavior, providing an actionable decision-making basis for optimizing urban road environments to reduce visual load.

3.2. Data Source

In this study, we use the DR(eye)VE dataset as the basis and filter the portions related to urban driving scenarios, including driving videos captured in urban environments at 1080 p/25 fps resolution by a GARMIN VirbX camera mounted on the roof of the vehicle, and the raw eye-tracking data of the driver collected by an SMI ETG 2w commercial eye-tracking glasses. These are used to construct the urban driving dataset. The eye-tracking device contains two built-in cameras dedicated to tracking the pupil at 60 fps, providing information on the driver’s gaze, including fixation, saccadic movements, and blinks. We aim to combine the driver’s eye-tracking data with the onboard camera video to analyze the driver’s visual attention demand and driving visual load in complex urban environments. A total of 800 urban driving scenes were selected from the dataset as analysis samples, each lasting 5 s. The raw data consists of urban driving videos recorded by the onboard camera and corresponding driver eye movement parameters.

3.3. Attention Demand Experiment

Driving attention demand refers to the amount of traffic information a driver needs to process per unit of time or distance [37]. The vision occlusion technique can be used to assess when and where a driver needs time to gather traffic information [38]. The default state of vision occlusion can be either occluded or non-occluded; the duration of occlusion can be controlled by the driver or set by the experiment. Some scholars have set the default state of vision occlusion as non-occluded, allowing the driver to choose when to occlude their vision and for how long, to study the characteristics of attention demand during high-speed driving [39]. Vision occlusion provides an effective method for studying attention demand in different driving environments by simulating restricted vision. By analyzing drivers’ occlusion behavior in various scenes, we can better understand their attention demand and driving visual load.
As shown in Figure 2, urban driving videos were played on a monitor, and the occlusion of the forward view was controlled by a keyboard in this study. When the participant felt they no longer needed to observe the traffic environment, they pressed the space bar to occlude the view ahead, simulating the driver’s gaze moving away from the road. The occluding rectangle in the forward view automatically disappeared after a period of time.
Some studies indicate that with longer durations of occlusion, occlusion behavior tends to concentrate on specific, highly complex segments of the route, whereas brief, 1 s occlusions promote a more uniform distribution of behavior across the entire experimental route [40]. Additionally, related driving simulator experiments found that under a 1 s occlusion condition, vehicle longitudinal and lateral control performance was satisfactory and more closely resembled actual driving conditions; in contrast, when the occlusion duration reached 2 s, driver uncertainty increased significantly [41]. Consequently, to achieve a more uniform distribution of occlusion behavior across the test course, avoid the systematic bias introduced by extended occlusion periods clustered in a few highly complex segments, and preserve ecological validity, the occlusion duration was fixed at 1 s. All experimental scenes came from the DR(eye)VE dataset, labeled as urban driving scenes. We collected the frame positions corresponding to when the participant pressed the space bar to match the different samples in the urban driving dataset established in this study.
In addition, to approximate real-world driving conditions as closely as possible, the present study emphasized the following aspects in its experimental design:
(1)
Participant selection: All participants held a valid People’s Republic of China motor vehicle driver’s license and possessed extensive driving experience, ensuring sufficient familiarity with and responsiveness to urban traffic environments.
(2)
Visual environment simulation: A laboratory display of appropriate size was carefully selected for the driving video presentation, with the aim of reproducing drivers’ natural visual environment as realistically as possible.
(3)
Driving state induction: Prior to the formal experiment, a non-experimental urban driving video was shown to help participants gradually adapt to the simulated driving environment. During the experiment, the laboratory was kept quiet and lighting was dimmed to facilitate immersion in the driving task.
(4)
Task logic aligned with reality: The occlusion task was designed to be actively triggered by participants, thereby simulating the self-initiated gaze shifts that occur in real driving and better reflecting the natural mechanisms of attentional allocation.
(5)
Although certain differences between laboratory conditions and real-world driving are unavoidable, the above design measures ensured that, under controlled conditions, the visual task characteristics and decision-making processes of urban driving were reproduced to the greatest extent possible. This provided a solid experimental foundation for the quantification of attentional demands.
Given that the occlusion duration was fixed, we consider that more frequent occlusions in a driving scene indicate more remaining attention, reflecting a lower attention demand. Therefore, in this study, we calculated the occlusion frequency (the number of occlusions per sample divided by the sample duration) for different urban driving samples to inversely represent the driver’s attention demand in various urban driving scenes.

3.4. Visual Metrics Extraction

In urban driving scenarios, attention demand and driving visual load are key factors affecting driving safety and efficiency. High driving visual load typically results in changes in visual metrics, such as increased fixation duration and fixation rate, altered blink rate, and changes in saccade frequency.
There is a strong correlation between visual metrics such as fixation, saccade, and blink, and attention demand and driving visual load. Fixation duration and fixation rate are important metrics for evaluating the driver’s attention to critical information in the environment. When the driver encounters complex traffic situations or needs to process a large amount of information, fixation duration and fixation rate typically increase to better capture and process the information [42,43]. On the other hand, saccade duration and saccade frequency reflect the driver’s efficiency in visual search and environment scanning. Under high driving visual load conditions, the saccade duration and saccade frequency may increase, indicating that the driver requires more time to process information [44,45].
Blink duration and blink rate are also effective indicators of attention and fatigue. In high-load driving environments, blink rate and blink duration may decrease, indicating changes in the driver’s attention under stress or fatigue [46,47].
In summary, changes in eye movement metrics can reflect the driver’s attention demand and driving visual load, serving as an important tool for assessing driving visual load [48]. This provides a significant reference for designing driving safety and assistance systems. Therefore, in this study, we selected fixation metrics (fixation duration, fixation rate), saccade metrics (saccade duration, average saccade duration, saccade frequency), and blink metrics (blink duration, blink rate) to analyze the correlation between visual metrics and attention demand (driving visual load).
The DR(eye)VE dataset contains driver eye movement parameters collected by eye-tracking equipment with a sampling frequency of 60 fps, continuously recording the driver’s gaze points and eye movement event types (including fixation, saccade, and blink). Based on the raw eye-tracking data from the dataset, we developed a program to calculate seven eye movement metric characteristics for each sample. Fixation duration is the number of fixation events divided by the sampling frequency of the eye tracker, and fixation rate is the number of fixations per sample divided by the total sample time. Saccade duration is the number of saccade events divided by the sampling frequency of the eye tracker, average saccade duration is the total saccade duration divided by the number of saccades in the sample, and saccade frequency is the number of saccades per sample divided by the total sample time. Blink duration is the number of blink events divided by the sampling frequency, and blink rate is the number of blinks per sample divided by the total sample time. The results of these calculations were used for subsequent correlation analysis between visual metrics and attention demand (driving visual load), as well as for calibrating the driver’s attention demand level (or driving visual load level).

3.5. Scene Features Extraction

3.5.1. Semantic Segmentation Model

Most of the external environmental information during driving is obtained visually by the driver [49]. In this study, we use the Deeplabv3+ model, pre-trained on the ADE20K dataset, from the MMsegmentation module in OpenMMLab [50] to extract semantic information from environmental images to represent the driving environment. Each pixel of the image after semantic segmentation is assigned to a specific element category and marked with a different color. The ADE20K dataset contains 150 categories, as shown in Figure 3. The pre-trained Deeplabv3+ model can recognize elements including roads, trees, and the sky, covering almost all scene elements in urban driving environments.

3.5.2. Element Characteristics Extraction

The urban driving scene videos in the dataset are processed by extracting each frame, which is then input into the pre-trained Deeplabv3+ model for semantic segmentation. After the semantic segmentation is completed, the proportion of each scene element in each frame is calculated. The proportion of scene elements is the ratio of the pixels of a specific element to the total pixels of the scene. The formula is as follows:
f r a m e _ p r o p i j = k = 1 n p i x e l i k i = 1 150 k = 1 n p i x e l i k              
where f r a m e _ p r o p i j represents the proportion of the i-th scene element in the j-th frame ( i = 1, 2, …, 150), and p i x e l i k refers to the k-th pixel of the i-th element.
Once the scene element proportions for each frame are calculated, the average of the proportions for all frames in a scene sample is taken to represent the scene element characteristics. The formula is as follows:
    s c e n e _ p r o p i =   j m f r a m e _ p r o p i j m
where s c e n e _ p r o p i is the average proportion of the i-th scene element in the scene sample (i = 1, 2, …, 150), and m is the total number of frames in the sample.
After extracting the scene element features through semantic segmentation for all scene samples in this study, we found that 30 elements, such as walls, buildings, and trees, accounted for more than 99% of the total. Thus, these 30 elements are considered sufficient to represent the urban driving environment. Therefore, the proportions of these 30 scene elements were selected to quantify the urban driving environment characteristics, which were used as feature variables for subsequent analysis of urban driving visual load based on attention demand.
In addition, based on the results of scene semantic segmentation, this study further established the Green Visibility Index (GVI), Sky Visibility Index (SVI), and Road Congestion (RC), among other indicators, to further analyze the element characteristics of urban driving scenes.
The GVI reflects the visibility of urban vegetation, which has a positive effect on the driver’s psychological comfort and relaxation. The SVI is an important indicator for evaluating the esthetic quality of the urban environment, which may affect the driver’s environmental cognition and processing. The RC directly affects driving behavior and decision-making processes, while Road Width (RW) is related to speed control and driving safety. The Street Canyon Enclosure (SCE) and Building View Index (BVI) are closely related to the driver’s environmental perception and sense of security. The Traffic Mix (TM) affects drivers’ cognitive processing of the environment, thereby influencing their driving behavior and decision-making processes, which is related to driving safety.
This study particularly focuses on the above scene features to analyze their impact on driving visual load and visual metrics based on attention demand, with the calculation formulas of the indicators shown in Table 1.

3.6. Machine Learning Methods

3.6.1. Clustering Algorithm for Driving Visual Load Calibration

In this study, the K-Means clustering algorithm was used to calibrate the driver’s driving visual load levels based on attention demand. K-Means is a widely used unsupervised learning algorithm that partitions data points into K clusters by minimizing the distance between the data points and the cluster centroids. Before applying K-Means, the selected visual metrics were standardized to eliminate the influence of different dimensions and magnitudes. The optimal number of clusters was determined using the silhouette coefficient and DBI index. The clustering results helped identify patterns of different driving visual load levels, reflecting the driver’s attention demand under various driving scenarios.
After completing the K-Means clustering, we preliminarily calibrated the driving visual load levels for each group based on the occlusion frequency. We further analyzed the centroid values of each group, which represent the center of the cluster in the visual metric feature space. To validate the rationality of the driving visual load level calibration after clustering, we combined the correlation analysis results between visual metrics and occlusion frequency. This verified whether the visual metric changes in each group’s centroids followed a consistent trend (positive correlation if driving visual load levels increased with higher visual metrics, negative correlation otherwise). Ultimately, the driving visual load levels for each group were determined.

3.6.2. Predictive Modeling for Visual Load and Eye Movement Metrics

In this study, Random Forest (RF), Adaptive Boosting (AdaBoost), eXtreme Gradient Boosting (XGBoost), and Support Vector Machine (SVM) were used to construct detection models for driving visual load and visual metrics. These algorithms are well-suited for handling complex datasets and capturing non-linear relationships. These four algorithms were chosen for their proven performance in classification and regression tasks, with each employing different principles. RF, AdaBoost, and XGBoost are ensemble algorithms based on decision trees, with RF using a bagging strategy and the latter two using boosting strategies [51,52,53]. Additionally, SVM was included for comparison as it classifies by finding the maximum-margin hyperplane.
The driving visual load level and driver’s eye movement metrics were taken as outputs, and the feature variables of urban driving scene extracted by semantic segmentation technology were taken as inputs to train various machine learning models, so as to seek the influence relationship between scene features, driving visual load and driver’s eye movement metrics.

3.6.3. SHAP-Based Interpretability Analysis

SHAP (SHapley Additive exPlanations) is a method for explaining machine learning model outputs. Based on the concept of Shapley values from cooperative game theory, it quantifies each feature’s contribution to model predictions. SHAP was introduced by Lundberg and Lee in 2017 [54], with the primary goal of providing a clear and accurate explanation for each prediction, revealing the contribution of individual features to the model’s output.
The Shapley value, derived from cooperative game theory, measures each participant’s contribution to the total value in a cooperative setting. In machine learning, this concept is applied to features, calculating each feature’s contribution to the model’s predictions. For a given prediction
f(x) and feature vector x, the SHAP value ϕ i provides a contribution metric for each feature x i , calculated as:
ϕ i   =   S     N     { i } S ! N     S     1 ! N !     f S   { i }     f S  
where N is the set of all features, S is a subset of the feature set N excluding x i , and f(S) is the model’s prediction using the feature subset S. The absolute values of S and N represent the sizes of subset S and the universal set of features N , respectively.
The formula represents the weighted average of the marginal contribution of feature x i to the prediction across all possible subsets of features. In this study, SHAP values are used to explain the impact of each feature on the classification labels in the classification models.

4. Results

4.1. Attention Demand and Visual Metrics

4.1.1. Correlation Analysis

Using various visual metrics as independent variables and occlusion frequency across different scenarios as the dependent variable, a Spearman correlation analysis was conducted to examine the relationship between visual metrics and the driver’s attention demand (or driving visual load). Significant visual metrics were selected as a basis for establishing driving visual load levels based on attention demand.
The results are shown in Table 2, five metrics related to blinking and saccades—blink duration, blink rate, saccade duration, average saccade duration, and saccade frequency—are significantly associated with attention demand, while fixation duration and fixation rate were not. Saccade duration, saccade frequency, and average saccade duration were positively correlated with attention demand: higher values for these metrics corresponded to higher attention demand and higher driving visual load. Conversely, blink rate and blink duration were negatively correlated with attention demand: lower blink rate and shorter blink duration indicated higher attention demand and driving visual load.
In previous driving research, fixation duration and fixation rate have often been employed as key indicators for assessing driving load and attentional allocation [43,55]. However, the sensitivity of these measures varies across tasks and experimental paradigms. On the one hand, prolonged fixations may not necessarily reflect increased information-processing demands from the external environment, but rather internal, non-driving-related cognitive activities [56]. On the other hand, prior studies have suggested that dynamic measures such as blinks and saccades exhibit greater sensitivity to changes in cognitive load in complex scenarios [57]. Furthermore, given that the scenarios in this study primarily involved multi-lane urban roads and intersections, drivers were required to frequently switch their gaze across multiple areas of interest (rear-view mirrors, side windows, pedestrians, traffic signals). Such frequent shifts often interrupted fixations, leading to numerous short fixations or mixed scan–fixation patterns. This, in turn, may have reduced the sensitivity of conventional fixation-based metrics (duration/rate) to overall attentional demands, resulting in the lack of a significant association between fixation-related measures and attention demand in this study.

4.1.2. K-Means Clustering Based on Visual Metrics

From the results of the correlation analysis, five visual metrics—blink duration, blink rate, saccade duration, average saccade duration, and saccade frequency—were significantly correlated with attention demand. These five metrics were used to determine the driver’s driving visual load levels based on attention demand.
First, the eye movement metrics were standardized using the z-score method. Then, the K-Means clustering algorithm was applied to label the driving visual load levels. The number of clusters was set from 1 to 10, and the clustering results were evaluated using the silhouette coefficient and DBI index. As shown in Figure 4, it was found that when the number of clusters was 3, the silhouette coefficient was the highest (0.394) and the DBI value was the lowest (0.881).
Moreover, according to Wickens’ Multiple Resource Theory (MRT) model [58], increased workload does not invariably lead to performance decline. Both excessively high and excessively low workload levels can impair performance. Excessive workload may cause errors, difficulties, or performance deterioration, whereas sustained underload may result in boredom, loss of situational awareness, and reduced vigilance. This reflects the three workload states of underload, optimal load, and overload. Furthermore, studies such as Huang et al. (2024) have successfully employed an N-back task in simulated driving to construct a three-level workload classification model (low, medium, high), thereby validating the effectiveness of a three-class structure for identifying drivers’ mental workload [59]. Accordingly, the present study adopts a three-class framework to categorize drivers’ visual workload states, which is theoretically grounded and statistically justified.
After clustering, the average occlusion frequencies of the three groups were 0.3352, 0.2805, and 0.2132, respectively. Variance analysis on occlusion frequencies across the groups revealed an F-value of 6.155 and a p-value of 0.002 (p < 0.05), indicating significant differences between groups. Thus, the groups were labeled as low-load, medium-load, and high-load based on attention demand.
Additionally, in Table 3, combining the correlation analysis results of visual metrics and attention demand (where “↓” indicates a negative correlation and “↑” indicates a positive correlation) with the centroids of the clusters, it was observed that the average visual load states of the groups were consistent with the relationships between metrics and attention demand, confirming the rationality of this classification.
In conclusion, the three clustered groups were labeled as low-load, medium-load, and high-load, with respective sample sizes of 148, 438, and 214. The results were included in Table 4.

4.2. Driving Visual Load and Scene Features

4.2.1. Construction and Evaluation of the Visual Load Detection Model

With the driving visual load levels as labels, 35 characteristics variables (excluding the proportions of the sky and road elements, as they were represented by SVI and RW characteristics) were used as inputs. The dataset was randomly divided into training and testing sets in an 8:2 ratio. Considering the imbalance between low, medium, and high load samples, Synthetic Minority Oversampling Technique (SMOTE) was employed to rebalance the training set. SMOTE generates synthetic samples to bridge the gap between minority and majority classes, resulting in a more even distribution in the feature space, which improves the model’s performance in identifying minority classes.
Five-fold cross-validation was implemented to train four models based on different methods, aiming to prevent overfitting and ensure the model’s reliability in detecting driving load in new datasets. Grid search combined with five-fold cross-validation was used to traverse all combinations of key hyperparameters to identify the optimal configuration for the best driving load detection model.
The models were evaluated using precision, recall, and F1-score, as summarized in Table 5. It was observed that the Random Forest model and the XGBoost model achieved the best performance on the original and balanced datasets, respectively. Moreover, all four models performed better on the balanced datasets than on the original datasets, confirming that using SMOTE can effectively alleviate the issue of data imbalance and improve classification performance.
Given that the Random Forest model performed best on the original dataset and exhibited good performance on the balanced dataset, the trained Random Forest model was used for subsequent SHAP analysis to examine the influence of different scene features on driving visual load.

4.2.2. Effect Analysis of Scene Features on Drivers’ Visual Load

Figure 5a shows the SHAP importance rankings of all features, with the top ten characteristics being SCE, tree proportion, RW, pole proportion, railing proportion, SVI, car proportion, BVI, signboard proportion, and plant flora proportion. The results indicate that these urban driving environment characteristics have a significant impact on the driving visual load state.
Since high load conditions may increase driver fatigue, potentially leading to traffic accidents and reduced road safety, maintaining a low load condition for drivers becomes a key objective. Therefore, further SHAP importance rankings for high and low load states were plotted and analyzed, as shown in Figure 5b,c.
As shown in Figure 5b, in high-load states, static elements such as poles, trees, signboards, and buildings have the most significant influence. The proportion of poles, signboards, and buildings is positively correlated with the probability of a high-load state, meaning that the larger the feature value, the higher the likelihood of the high-load state occurring. In contrast, the proportion of trees is negatively correlated with the probability of a high-load state, so a higher feature value corresponds to a lower probability of a high-load state. Among dynamic elements, cars and people have the most significant impact. The proportion of people is positively correlated with the probability of a high-load state, with a higher feature value leading to a greater likelihood of the high-load state occurring. The relationship between car proportions and high-load states is more complex. Among comprehensive scene indicators, SCE and SVI have the most significant impact on the occurrence of high-load states. The SCE is positively correlated with the probability of a high-load state, meaning the higher the index, the more likely a high-load state will occur. On the contrary, SVI is negatively correlated with the probability of a high-load state, with higher values corresponding to a reduced likelihood of a high-load state.
As shown in Figure 5c, in low-load states, static elements such as railings, trees, and plant flora have the most significant influence. The proportions of railings and plant flora are negatively correlated with the probability of a low-load state, meaning that as their values increase, the probability of a low-load state decreases. The relationship between tree proportions and low-load states is more complex. Among dynamic elements, cars and people still have the most significant impact. The proportions of both cars and people are negatively correlated with the probability of a low-load state, with higher feature values leading to a decreased probability of a high-load state. Among comprehensive scene indicators, SCE, RW, and road RC have the most significant influence on the occurrence of low-load states. The SCE and RC are negatively correlated with the probability of a low-load state, meaning that as their values increase, the likelihood of a low-load state decreases. However, RW is positively correlated with the probability of a low-load state, with larger values leading to an increased likelihood of a low-load state.
To further analyze how different scenes affect drivers in both high and low visual states, this study generated SHAP dependence plots. These plots visually show how changes in feature values affect the driver’s visual load state, providing a quantitative analysis of the feature’s influence:
(1)
Among static elements, as shown in Figure 6a, when the proportion of poles exceeds 0.08%, the probability of the driver being in a high-load state increase. In urban environments, elements such as traffic sign poles, traffic signal poles, and utility poles are categorized as pole elements. The increase in pole elements signifies a rise in environmental complexity. More poles may indicate additional traffic signs or indicative boards, which draw the driver’s attention and elevate attention demand, thereby increasing visual load. As shown in Figure 6b,j, tree elements have different effects on the driver’s high-load and low-load states. When the proportion of trees exceeds 35%, the probability of the driver being in a high-load state significantly decreases. This may be because trees create a more relaxing environment and simplify the visual field, thus reducing attention demand and alleviating visual load. In contrast, under low-load conditions, the impact of tree elements is more complex. When the proportion of trees is between 11% and 26%, the probability of a low-load state increases. However, when tree coverage exceeds 26%, the probability decreases, showing an initial rise followed by a decline. This suggests that a moderate proportion of trees reduces visual load, while proportions that are too low or too high increase complexity and raise attention demand. As shown in Figure 6c,d, when the proportion of signboard and building elements exceeds 0.55% and 14%, respectively, the probability of a high-load state increases. Signboard elements, which include signs and billboards, tend to distract drivers. Therefore, an excess of signboards in the scene heightens attention demand. Similarly, a high concentration of buildings suggests a more complex driving environment, thereby increasing driver attention demand and consequently raising visual load. As shown in Figure 6i,k, the increase in railing and plant flora elements both decrease the probability of the driver being in a low-load state. Railings are typically located in road repair areas, accident-prone zones, or places where boundaries need to be emphasized. These conditions require constant awareness, thereby increasing attention demand and visual load. When the proportion of plant flora exceeds 1.2%, it may cause additional distractions, thereby somewhat increasing the driving visual load.
(2)
Among dynamic elements, car and person have the most significant impact on the driving visual load, which is an evident result. As shown in Figure 6e,l, as the proportion of cars increases, the probability of a high-load state initially decreases. However, when the proportion reaches 24%, the probability of a high-load state rises sharply, with SHAP values turning positive. This indicates that an excessive number of cars in the scene increases the driver’s attention demand, leading to a rise in driving load. This is also confirmed under the low-load condition: when the proportion of cars exceeds 13%, the probability of the driver being in a low-load state drops suddenly, triggering an increase in driving load. As the most unpredictable factor in the traffic system, person has an even more pronounced impact on driver load. When pedestrians appear in the scene, the probability of a high-load state rises, while the probability of a low-load state decreases. This shows that pedestrians add instability to the environment. Drivers must monitor not only the road ahead but also the behavior of roadside pedestrians to prevent accidents. Such demands increase attention and raise driving visual load. As dynamic elements in urban driving environments, car and person significantly influence the driver’s distraction and attention. Ensuring traffic safety thus requires effective management of vehicles and pedestrians within the traffic system.
(3)
Among comprehensive scene indicators, as shown in Figure 6g,n, SCE has the most significant impact on both high-load and low-load visual states. When the SCE exceeds 39%, the probability of a high-load state increases, and when it exceeds 36%, the probability of a low-load state decrease. A higher SCE indicates a more spatially enclosed environment, which narrows the driver’s field of view and raises environmental complexity. This adds distracting elements, makes it harder for drivers to perceive their surroundings, and increases attention demand, thereby raising visual load. As shown in Figure 6h, under high-load conditions, when SVI reaches 16%, the probability of the driver being in a high-load state decreases. A higher SVI suggests a wider field of view with fewer elements in the forward vision, reducing complexity and lowering attention demand. This helps relieve the driver’s high-load state. As shown in Figure 6o, when RW reaches 38%, the probability of a low-load state suddenly increases. An increase in RW implies a higher proportion of road elements, which means fewer attention-demanding elements such as vehicles and pedestrians. It also provides a broader field of view, reducing environmental complexity and thus lowering the driver’s load. Additionally, RC is another important factor affecting driver load. As shown in Figure 6p, when RC reaches 40%, the probability of a low-load state decrease. A higher RC reflects more vehicles on the road. An excessive number of dynamic elements increases cognitive load, raises attention demands, and consequently elevates driving visual load.

4.3. Visual Metrics and Scene Features

4.3.1. Model Construction for Analysis

Based on the correlation analysis in Section 4.1.1, the driver’s visual metrics, including blink metrics (blink duration and blink rate) and saccade metrics (saccade duration, average time per saccade, and saccade frequency), show a certain degree of correlation with the driver’s attention demand. Therefore, this study selects these five visual metrics as dependent variables and the 35 extracted scene feature variables as independent variables to analyze how driving scenes affect changes in visual metrics and further influence the driver’s load.
In this study, five different machine learning regression models were trained based on the five visual metrics. All samples were randomly divided into training and test sets in an 8:2 ratio. Additionally, 5-fold cross-validation and grid search were used to traverse all combinations of key hyperparameters across the four models to determine the optimal hyperparameter combination, thereby obtaining the best regression models.
Among the five models, the XGBoost method consistently achieved the best performance, with R 2 values of 0.94, 0.85, 0.75, 0.73, and 0.78, respectively. Thus, the trained XGBoost models were used in the subsequent SHAP analysis to evaluate the influence of different scene characteristics on visual metrics.

4.3.2. Effect Analysis of Scene Features on Driver’s Eye Movement Behavior

(1) Blink Metrics
Figure 7a,c list the ten scene characteristics that have the greatest influence on blink duration and blink rate, respectively. Among them, door, SCE, RW, SVI, and RC have the greatest impact on blink duration, while RW, SVI, palm, RC, and SCE have the greatest impact on blink rate. It can be observed that the four comprehensive indicators—SCE, RW, SVI, and RC—have a more significant effect on blink metrics compared to other simple scene feature ratios. Therefore, this study further plots SHAP dependence diagrams for these four indicators to analyze their impact on blink metrics.
As blink duration and blink rate are both blink-related metrics, scene characteristics have similar effects on them. As shown in Figure 8a,e, when SVI exceeds 25%, the SHAP value rises and becomes greater than zero, leading to an increase in blink duration and blink rate. According to the previous correlation analysis, blink metrics are negatively correlated with the driver’s attention demand, suggesting that a high level of SVI may reduce the driver’s attention demand, thus positively contributing to reducing driving load. As shown in Figure 8b,f, when RW is between 35% and 47%, the SHAP value is greater than zero, indicating that too narrow or too wide a road width decreases blink duration and blink rate, thereby increasing the driver’s load. As shown in Figure 8c,g, when SCE reaches 40%, the SHAP value for blink duration turns from positive to negative, and similarly, when SCE reaches 38%, the SHAP value for blink rate also turns from positive to negative. This indicates that a high level of SCE increases the complexity of the scene and the driving pressure, requiring the driver to focus more on acquiring environmental information, thereby reducing blink duration and blink rate and increasing driving load. As shown in Figure 8d,h, when RC reaches 25%, the SHAP value for blink duration becomes negative, and it remains negative after reaching 35%. For blink rate, the SHAP value turns from positive to negative when RC reaches 40%. As RC increases, the excessive number of vehicles forces the driver to constantly monitor the movement of other vehicles to ensure driving safety. This increases the time required to gather information, reducing blink duration and blink rate, which leads to an increase in driving visual load.
(2) Saccade Metrics
Figure 9a,c,e list the top ten scene characteristics that most significantly affect the driver’s saccade duration, average saccade duration, and saccade frequency, respectively. BVI, TM, door, and SVI have the greatest impact on the driver’s saccade duration; building, SCE, pole, and tree have the greatest influence on the average saccade duration; and SCE, TM, door, and truck have the greatest impact on saccade frequency. SHAP dependency diagrams were plotted for these scene characteristics to provide further analysis.
As shown in Figure 10a,b, when BVI and TM reach 13% and 0.02%, respectively, the SHAP values for saccade duration turn positive. High BVI increases the complexity of static elements in the scene, while high TM indicates the presence of more non-motorized vehicles and pedestrians, which undoubtedly adds to the scene’s information load. Since saccadic behavior is the driver’s way of acquiring information from the scene, increased saccade duration results in higher driving load. According to Figure 10c,k, the presence of door elements in the scene at a certain proportion reduces both saccade duration and frequency. A possible explanation for this interesting phenomenon is that when the proportion of door elements is high, the driver’s attention is more likely to be focused on the door elements, reducing the time spent scanning and searching for unrelated information, which in turn decreases saccadic behavior to some extent. As shown in Figure 10d, when RC reaches 20%, the driver’s saccade duration decreases, reducing the complexity of the driving scene and alleviating driving load.
In Figure 10e–g, when the proportions of building, SCE, and the proportions of pole elements reach 11%, 39%, and 0.1%, respectively, the average saccade duration increases. This implies that the driver’s saccadic searches become more detailed and cover a larger area. An increase in building elements likely requires the driver to process more building-related information, increasing the visual complexity of the scene. Meanwhile, high SCE may require the driver to spend more time evaluating spatial relationships and navigation paths. Additionally, although the proportion of pole elements is relatively small, their presence suggests the existence of more signs and other elements that could distract the driver’s visual search, thus increasing the time per saccade and raising driving load. As shown in Figure 10h, when the proportion of tree elements reaches 36%, the SHAP value for the average saccade duration remains negative. This indicates that a higher proportion of tree elements can effectively reduce the driver’s saccadic search time. A possible reason is that a high proportion of tree elements may obscure certain visual elements along the roadside, simplifying the visual scene on either side of the driver’s field of view, thus reducing the time and space required to search the environment and lowering driving load.
According to Figure 10i,j, when SCE and TM reach 40% and 0.02%, respectively, the driver’s saccade frequency increases. Similarly, this is a result of increased complexity in both static and dynamic elements in the scene. High proportions of buildings, facilities, pedestrians, and non-motorized vehicles increase the amount of information in the scene, leading to more frequent visual searches and higher driving load. As shown in Figure 10l, the presence of truck elements decreases the driver’s saccade frequency. This may be because truck elements, being particularly hazardous in the traffic system, are more likely to capture the driver’s attention compared to other elements in the scene, leading to more focused attention and, consequently, a reduction in the driver’s saccadic behavior.

5. Discussion

5.1. Comparison with Existing Studies

SHAP interpretability analysis revealed that the presence of billboards or signage and increased pedestrian or vehicular traffic elevates drivers’ visual workload, whereas vegetation has a mitigating effect. These findings are consistent with those of White and Shah (2019), who highlighted the restorative characteristics of natural scenes [60]. Additionally, drivers’ visual workload increases with building complexity, which aligns with experimental results reported by Edquist et al. (2012) [61] and Foy et al. (2018). [26]. The SHAP analysis further indicated that a tree coverage proportion of 11–26% favors the occurrence of low workload states, whereas increases in building proportion and street canyon enclosure (SCE) exacerbate drivers’ visual workload. This is similar to Cai’s findings [6], where a certain proportion of trees reduced the number of speeding-related collisions, while increased building proportion and visual complexity were associated with a rise in such incidents—possibly because drivers face difficulties handling heightened visual complexity in emergency situations [62].
In prior road safety research, including studies by Cai, Kaygisiz, and Asadi [3,4,6], the focus has largely been on macro-level factors such as traffic data, road structure, land use attributes, and socio-demographic characteristics to establish their association with traffic accidents. However, the occurrence of accidents is influenced by a complex interplay of factors that extend beyond the driver to the environment and vehicle, making such approaches relatively coarse-grained. In contrast, this study innovatively centers on drivers’ visual workload—a critical safety factor—and emphasizes a detailed analysis and quantitative mapping of its direct source: the visual environment. This perspective not only deepens understanding of the intrinsic mechanisms of driving safety but also provides a novel pathway for human-factor-based interventions in traffic safety.
Moreover, existing studies have primarily focused on the independent associations between road environment features and driving performance [61] or eye movement behavior [63]. Some studies have applied eye-tracking technology to detect driving workload [64,65], but these largely remain at the binary “feature–behavior” or “feature–workload” level. In contrast, this study employs SHAP analysis to not only reveal the mechanisms through which urban driving environment features influence drivers’ visual workload, but also further elucidate their effects on various eye movement parameters, a key mediating factor. This establishes a systematic scene–eye–workload explanatory chain, overcoming the limitations of previous studies that insufficiently captured the dynamic human-factor response process.

5.2. Response to Research Objectives

First, this study employed a visual occlusion method to quantify visual attention demands across different urban scenarios, confirming a significant correlation between eye-movement indicators and visual attention requirements. Based on this, the K-Means clustering algorithm was applied to classify drivers’ visual load.
Subsequently, using machine learning techniques combined with SHAP interpretability analysis, the study identified key elements of urban scenes and their dynamic effects on visual workload. For example, when street canyon enclosure (SCE) exceeded 39%, the probability of drivers being in a high-load state increased significantly, whereas when it exceeded 36%, the probability of a low-load state decreased markedly. Moreover, SHAP analysis further elucidated how different scene features influence the mediating changes in drivers’ eye-movement indicators, revealing the mechanism through which urban scenes affect visual load. For instance, excessively high SCE reduces blink metrics and increases saccade metrics, thereby resulting in higher visual load.

5.3. Practical Implications

Based on SHAP-based interpretations of how specific urban scene features affect drivers’ visual workload and eye-movement behavior, this study translates its analytical findings into actionable strategies for urban planners and traffic safety policymakers. These measures aim to reasonably alleviate visual workload and enhance road safety:
(1) Regulation of street canyon visual complexity
The study found that excessively high street canyon enclosure (SCE) strongly influences reduced blink indicators and increased saccade indicators. When SCE exceeds 39%, the probability of high workload states rises significantly. Therefore, planning and design in high-risk urban areas (e.g., intersections, ramps) should consider controlling SCE. Measures may include establishing building setback regulations, managing street block height-to-width ratios, and introducing periodic visual-open buffer zones to break continuous enclosure and maintain sky visibility.
(2) Visual optimization of green landscapes
SHAP analysis results showed that tree coverage between 11–26% facilitates low workload states, whereas coverage exceeding 35% significantly reduces the probability of high workload. Accordingly, targeted roadside greening programs should be implemented in urban streetscape design. In areas with high SCE (e.g., dense urban canyons), tree planting should be prioritized to leverage its workload-reducing effects, while in open areas, tree density should be kept below 26% to avoid adding visual complexity and to maintain visual comfort.
(3) Management of visual interference from street facilities
The presence of poles (>0.08%) and billboards (>0.55%) significantly increases the probability of high visual workload. Thus, optimizing the layout and density of street furniture is necessary to reduce visual clutter. Specifically, planning regulations can limit unnecessary billboards, encourage the use of large integrated signs instead of multiple small ones, and in high-workload areas, consider overhead gantry structures as alternatives to roadside poles to minimize horizontal line-of-sight interference.
(4) Optimization of visual openness and road geometry
Sky view index (SVI > 16%) and roadway width (RW > 38%) are significantly associated with reduced visual workload. Where feasible, restrictions on overhead structures and the adoption of stepped building designs in narrow blocks should be encouraged to enhance sky visibility. During design and planning, particularly for segments with high cognitive demand, maintaining or increasing roadway width above 38% can provide a broader field of view and reduce feelings of spatial confinement.
(5) Dynamic traffic management strategies
Road congestion (RC > 40%) and the presence of pedestrians and vehicles—particularly when vehicle density exceeds 24%—substantially increase workload. Intelligent traffic management systems should be deployed to monitor traffic density in real time. When congestion exceeds critical thresholds, variable message signs can be used to recommend alternative routes. Signal timing optimization can reduce vehicle–pedestrian conflicts, and in high pedestrian-density areas, pedestrianization or segregated pedestrian systems should be considered to minimize sudden visual disturbances for drivers.

5.4. Limitations and Future Directions

Despite the extensive scope of this study, several limitations remain:
(1)
The quantification of scene features primarily relied on pixel-based measures, which did not fully capture geometric attributes such as line structures and shapes, dynamic motion information, or semantic elements such as road markings, thus limiting the richness of feature representation.
(2)
Assessment of drivers’ visual workload was primarily based on eye-movement indicators, lacking integration with multimodal physiological data (e.g., EEG, ECG), which constrains the comprehensive reflection of drivers’ cognitive and physiological states.
(3)
This study focused on urban driving environments from the DR(eye)VE dataset, covering typical routes in several European cities. The homogeneity in road types, traffic culture, and climatic conditions limits the generalizability of the model to rural, highway, or other regional contexts, and the model’s applicability in non-urban environments (e.g., highways, rural roads, or extreme weather conditions) remains unverified.
For future research, feature extraction can be enhanced by incorporating advanced computer vision models, such as attention-based semantic segmentation networks or graph neural networks, to integrate geometric, dynamic, and contextual semantic information, thereby providing a more comprehensive representation of driving scenes. In terms of workload assessment, multimodal physiological signals (e.g., EEG, ECG, galvanic skin response) should be integrated to construct a more robust and holistic evaluation system for drivers’ visual workload. Regarding model generalization, large-scale validation across diverse road types and cross-cultural driving scenarios is necessary, alongside exploration of domain adaptation and transfer learning techniques, to improve the model’s applicability and reliability in complex real-world environments.

6. Conclusions

During driving, the visual environment serves as the primary source of drivers’ visual load and is a key basis for predicting it. Drivers’ eye movement parameters effectively reflect their visual load levels, and different scene elements exert distinct influences on visual load and eye-movement behavior. By employing SHAP analysis, these effects can be clearly elucidated, enabling the construction of a systematic explanatory pathway linking “scene–eye movements–load.” This study not only provides a scientific method and tools for understanding drivers’ visual load in urban driving environments, but also offers targeted strategies and recommendations for urban planning and traffic management, thereby reducing visual load and enhancing road safety.

Author Contributions

J.N.: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Resources, Writing—Original Draft, Writing—Review and Editing, Supervision, Validation. Y.S.: Data curation, Formal analysis, Investigation, Methodology, Resources, Writing—Original Draft, Writing—Review and Editing, Validation. Y.G. (Yiwen Guo): Data curation, Resources, Writing—Review and Editing, Validation. Y.G. (Yongqi Gu): Data curation, Resources, Supervision, Writing—Review and Editing. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (No. 51905224) and the Natural Science Foundation of Jiangsu Province of China (BK20241785).

Institutional Review Board Statement

The study did not need ethical commission approval.

Informed Consent Statement

Informed consent for participation is not required as per local legislation [Ethical Review Measures for Life Sciences and Medical Research Involving Humans, National Health Commission (NHC) et al. (jointly issued), 2023, Article 32 (1) (2)].

Data Availability Statement

The dataset used in this study is publicly available from the DR(eye)VE Project at http://imagelab.ing.unimore.it/dreyeve (accessed on 11 April 2024) and was originally introduced in Palazzi et al. (2019) [36].

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Liu, L.; Gao, Z.; Luo, P.; Duan, W.; Hu, M.; Mohd Arif Zainol, M.R.R.; Zawawi, M.H. The influence of visual landscapes on road traffic safety: An assessment using remote sensing and deep learning. Remote Sens. 2023, 15, 4437. [Google Scholar] [CrossRef]
  2. Iamtrakul, P.; Chayphong, S.; Mateo-Babiano, D. The Transition of Land Use and Road Safety Studies: A Systematic Literature Review (2000–2021). Sustainability 2023, 15, 8894. [Google Scholar] [CrossRef]
  3. Kaygisiz, Ö.; Senbil, M.; Yildiz, A. Influence of urban built environment on traffic accidents: The case of Eskisehir (Turkey). Case Stud. Transp. Policy 2017, 5, 306–313. [Google Scholar] [CrossRef]
  4. Asadi, M.; Ulak, M.B.; Geurs, K.T.; Weijermars, W.; Schepers, P. A comprehensive analysis of the relationships between the built environment and traffic safety in the Dutch urban areas. Accid. Anal. Prev. 2022, 172, 106683. [Google Scholar] [CrossRef]
  5. Sun, H.; Xu, H.; He, H.; Wei, Q.; Yan, Y.; Chen, Z.; Li, X.; Zheng, J.; Li, T. A Spatial Analysis of Urban Streets under Deep Learning Based on Street View Imagery: Quantifying Perceptual and Elemental Perceptual Relationships. Sustainability 2023, 15, 14798. [Google Scholar] [CrossRef]
  6. Cai, Q.; Abdel-Aty, M.; Zheng, O.; Wu, Y. Applying machine learning and google street view to explore effects of drivers’ visual environment on traffic safety. Transp. Res. Part C Emerg. Technol. 2022, 135, 103541. [Google Scholar] [CrossRef]
  7. Yu, X.; Ma, J.; Tang, Y.; Yang, T.; Jiang, F. Can we trust our eyes? Interpreting the misperception of road safety from street view images and deep learning. Accid. Anal. Prev. 2024, 197, 107455. [Google Scholar] [CrossRef] [PubMed]
  8. Rita, L.; Peliteiro, M.; Bostan, T.C.; Tamagusko, T.; Ferreira, A. Using deep learning and Google Street View imagery to assess and improve cyclist safety in London. Sustainability 2023, 15, 10270. [Google Scholar] [CrossRef]
  9. Von Stülpnagel, R.; Binnig, N. How safe do you feel?–A large-scale survey concerning the subjective safety associated with different kinds of cycling lanes. Accid. Anal. Prev. 2022, 167, 106577. [Google Scholar] [CrossRef] [PubMed]
  10. Hu, Z.; Zhou, J.; Zhang, E. Improving traffic safety through traffic accident risk assessment. Sustainability 2023, 15, 3748. [Google Scholar] [CrossRef]
  11. Parsa, A.B.; Movahedi, A.; Taghipour, H.; Derrible, S.; Mohammadian, A.K. Toward safer highways, application of XGBoost and SHAP for real-time accident detection and feature analysis. Accid. Anal. Prev. 2020, 136, 105405. [Google Scholar] [CrossRef] [PubMed]
  12. Intini, P.; Berloco, N.; Fonzone, A.; Fountas, G.; Ranieri, V. The influence of traffic, geometric and context variables on urban crash types: A grouped random parameter multinomial logit approach. Anal. Methods Accid. Res. 2020, 28, 100141. [Google Scholar] [CrossRef]
  13. Liu, Z.; Zhao, X.; Wu, F. Influencing Factors of Driving Attention Demand in Urban Road Scenario. J. Transp. Syst. Eng. Inf. Technol. 2021, 21, 132. [Google Scholar]
  14. Gershon, P.; Sita, K.R.; Zhu, C.; Ehsani, J.P.; Klauer, S.G.; Dingus, T.A.; Simons-Morton, B.G. Distracted driving, visual inattention, and crash risk among teenage drivers. Am. J. Prev. Med. 2019, 56, 494–500. [Google Scholar] [CrossRef] [PubMed]
  15. Useche, S.A.; Gómez Ortiz, V.; Cendales, B.E. Stress-related psychosocial factors at work, fatigue, and risky driving behavior in bus rapid transport (BRT) drivers. Accid. Anal. Prev. 2017, 104, 106–114. [Google Scholar] [CrossRef]
  16. Zokaei, M.; Jafari, M.J.; Khosrowabadi, R.; Nahvi, A.; Khodakarim, S.; Pouyakian, M. Tracing the physiological response and behavioral performance of drivers at different levels of mental workload using driving simulators. J. Saf. Res. 2020, 72, 213–223. [Google Scholar] [CrossRef]
  17. Paschalidis, E.; Choudhury, C.F.; Hess, S. Combining driving simulator and physiological sensor data in a latent variable model to incorporate the effect of stress in car-following behaviour. Anal. Methods Accid. Res. 2019, 22, 100089. [Google Scholar] [CrossRef]
  18. Ni, J.; Chen, J.; Xie, W.; Shao, Y. The impacts of the traffic situation, road conditions, and driving environment on driver stress: A systematic review. Transp. Res. Part F Traffic Psychol. Behav. 2024, 103, 141–162. [Google Scholar] [CrossRef]
  19. Mou, L.; Zhou, C.; Zhao, P.; Nakisa, B.; Rastgoo, M.N.; Jain, R.; Gao, W. Driver stress detection via multimodal fusion using attention-based CNN-LSTM. Expert Syst. Appl. 2021, 173, 114693. [Google Scholar] [CrossRef]
  20. Siam, A.I.; Gamel, S.A.; Talaat, F.M. Automatic stress detection in car drivers based on non-invasive physiological signals using machine learning techniques. Neural Comput. Appl. 2023, 35, 12891–12904. [Google Scholar] [CrossRef]
  21. Hu, S.; Gao, Q.; Xie, K.; Wen, C.; Zhang, W.; He, J. Efficient detection of driver fatigue state based on all-weather illumination scenarios. Sci. Rep. 2024, 14, 17075. [Google Scholar] [CrossRef]
  22. González, C.; Ayobi, N.; Escallón, F.; Baldovino-Chiquillo, L.; Wilches-Mogollón, M.; Pasos, D.; Ramírez, N.; Pinzón, J.; Sarmiento, O.; Quistberg, D.A.; et al. STRIDE: Street View-based Environmental Feature Detection and Pedestrian Collision Prediction. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), Paris, France, 2–6 October 2023; IEEE: New York, NY, USA, 2023; pp. 3222–3234. [Google Scholar]
  23. Bustos, C.; Rhoads, D.; Solé-Ribalta, A.; Masip, D.; Arenas, A.; Lapedriza, A.; Borge-Holthoefer, J. Explainable, automated urban interventions to improve pedestrian and vehicle safety. Transp. Res. Part C Emerg. Technol. 2021, 125, 103018. [Google Scholar] [CrossRef]
  24. Tanprasert, T.; Siripanpornchana, C.; Surasvadi, N.; Thajchayapong, S. Recognizing traffic black spots from street view images using environment-aware image processing and neural network. IEEE Access 2020, 8, 121469–121478. [Google Scholar] [CrossRef]
  25. Qin, K.; Xu, Y.; Kang, C.; Kwan, M.P. A graph convolutional network model for evaluating potential congestion spots based on local urban built environments. Trans. GIS 2020, 24, 1382–1401. [Google Scholar] [CrossRef]
  26. Foy, H.J.; Chapman, P. Mental workload is reflected in driver behaviour, physiology, eye movements and prefrontal cortex activation. Appl. Ergon. 2018, 73, 90–99. [Google Scholar] [CrossRef]
  27. Liu, B.; Sun, L.; Jian, R. Driver’s visual cognition behaviors of traffic signs based on eye movement parameters. J. Transp. Syst. Eng. Inf. Technol. 2011, 11, 22–27. [Google Scholar] [CrossRef]
  28. He, S.; Liang, B.; Pan, G.; Wang, F.; Cui, L. Influence of dynamic highway tunnel lighting environment on driving safety based on eye movement parameters of the driver. Tunn. Undergr. Space Technol. 2017, 67, 52–60. [Google Scholar] [CrossRef]
  29. Ahlström, C.; Anund, A.; Fors, C.; Åkerstedt, T. Effects of the road environment on the development of driver sleepiness in young male drivers. Accid. Anal. Prev. 2018, 112, 127–134. [Google Scholar] [CrossRef] [PubMed]
  30. Fiolić, M.; Babić, D.; Babić, D.; Tomasović, S. Effect of road markings and road signs quality on driving behaviour, driver’s gaze patterns and driver’s cognitive load at night-time. Transp. Res. Part F Traffic Psychol. Behav. 2023, 99, 306–318. [Google Scholar] [CrossRef]
  31. Cvahte Ojsteršek, T.; Topolšek, D. Influence of drivers’ visual and cognitive attention on their perception of changes in the traffic environment. Eur. Transp. Res. Rev. 2019, 11, 45. [Google Scholar] [CrossRef]
  32. Kircher, K.; Kujala, T.; Ahlström, C. On the difference between necessary and unnecessary glances away from the forward roadway: An occlusion study on the motorway. Hum. Factors 2020, 62, 1117–1131. [Google Scholar] [CrossRef] [PubMed]
  33. Zeng, Q.; Chen, Y.; Zheng, X.; Zhang, M.; Li, D.; Hu, Q. Exploring the visual attention mechanism of long-distance driving in an underground construction cavern: Eye-tracking and simulated driving. Sustainability 2023, 15, 9140. [Google Scholar] [CrossRef]
  34. Aminosharieh Najafi, T.; Affanni, A.; Rinaldo, R.; Zontone, P. Driver attention assessment using physiological measures from EEG, ECG, and EDA signals. Sensors 2023, 23, 2039. [Google Scholar] [CrossRef]
  35. Liu, R.; Qi, S.; Hao, S.; Lian, G.; Luo, Y. Using electroencephalography to analyse drivers’ different cognitive workload characteristics based on on-road experiment. Front. Psychol. 2023, 14, 1107176. [Google Scholar] [CrossRef]
  36. Palazzi, A.; Abati, D.; Calderara, S.; Solera, F.; Cucchiara, R. Predicting the driver’s focus of attention: The DR(eye)VE project. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 40, 2891–2904. [Google Scholar] [CrossRef]
  37. Kujala, T.; Mäkelä, J.; Kotilainen, I.; Tokkonen, T. The attentional demand of automobile driving revisited: Occlusion distance as a function of task-relevant event density in realistic driving scenarios. Hum. Factors 2016, 58, 163–180. [Google Scholar] [CrossRef]
  38. Saffarian, M.; de Winter, J.C.F.; Senders, J.W. Measuring drivers’ visual information needs during braking: A simulator study using a screen-occlusion method. Transp. Res. Part F Traffic Psychol. Behav. 2015, 33, 48–65. [Google Scholar] [CrossRef]
  39. Kircher, K.; Ahlstrom, C. Evaluation of methods for the assessment of attention while driving. Accid. Anal. Prev. 2018, 114, 40–47. [Google Scholar] [CrossRef]
  40. Liu, Z.; Ahlström, C.; Forsman, Å.; Kircher, K. Attentional demand as a function of contextual factors in different traffic scenarios. Hum. Factors 2020, 62, 1171–1189. [Google Scholar] [CrossRef] [PubMed]
  41. Liu, Z.; Yuan, W.; Ma, Y. Drivers’ attention strategies before eyes-off-road in different traffic scenarios: Adaptation and anticipation. Int. J. Environ. Res. Public Health 2021, 18, 3716. [Google Scholar] [CrossRef] [PubMed]
  42. Recarte, M.A.; Nunes, L.M. Mental workload while driving: Effects on visual search, discrimination, and decision making. J. Exp. Psychol. Appl. 2003, 9, 119–137. [Google Scholar] [CrossRef]
  43. Victor, T.W.; Harbluk, J.L.; Engström, J.A. Sensitivity of eye-movement measures to in-vehicle task difficulty. Transp. Res. Part F Traffic Psychol. Behav. 2005, 8, 167–190. [Google Scholar] [CrossRef]
  44. Di Stasi, L.L.; Antolí, A.; Cañas, J.J. Main sequence: An index for detecting mental workload variation in complex tasks. Appl. Ergon. 2011, 42, 807–813. [Google Scholar] [CrossRef]
  45. Underwood, G.; Chapman, P.; Bowden, K.; Crundall, D. Visual search while driving: Skill and awareness during inspection of the scene. Transp. Res. Part F Traffic Psychol. Behav. 2002, 5, 87–97. [Google Scholar] [CrossRef]
  46. Schleicher, R.; Galley, N.; Briest, S.; Galley, L. Blinks and saccades as indicators of fatigue in sleepiness warnings: Looking tired? Ergonomics 2008, 51, 982–1010. [Google Scholar] [CrossRef] [PubMed]
  47. Stern, J.A.; Boyer, D.; Schroeder, D. Blink rate: A possible measure of fatigue. Hum. Factors 1994, 36, 285–297. [Google Scholar] [CrossRef]
  48. Wickens, C.D.; Helton, W.S.; Hollands, J.G.; Banbury, S. Engineering Psychology and Human Performance; Routledge: London, UK, 2021. [Google Scholar]
  49. Wood, J.M. Vision impairment and on-road driving. Annu. Rev. Vis. Sci. 2022, 8, 195–216. [Google Scholar] [CrossRef]
  50. MMSegmentation Contributors. MMSegmentation: OpenMMLab Semantic Segmentation Toolbox and Benchmark. GitHub. 2020. Available online: https://github.com/open-mmlab/mmsegmentation (accessed on 24 April 2024).
  51. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  52. Freund, Y.; Schapire, R.E. A decision-theoretic generalization of on-line learning and an application to boosting. J. Comput. Syst. Sci. 1997, 55, 119–139. [Google Scholar] [CrossRef]
  53. Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  54. Lundberg, S.M.; Lee, S.-I. A unified approach to interpreting model predictions. Adv. Neural Inf. Process. Syst 2017, 30, 4765–4774. [Google Scholar]
  55. Liang, Y.; Lee, J.D. Combining cognitive and visual distraction: Less than the sum of its parts. Accid. Anal. Prev. 2010, 42, 881–890. [Google Scholar] [CrossRef]
  56. Recarte, M.A.; Nunes, L.M. Effects of verbal and spatial-imagery tasks on eye fixations while driving. J. Exp. Psychol. Appl. 2000, 6, 31–43. [Google Scholar] [CrossRef]
  57. Cardona, G.; Quevedo, N. Blinking and driving: The influence of saccades and cognitive workload. Curr. Eye Res. 2014, 39, 239–244. [Google Scholar] [CrossRef]
  58. Wickens, C. Processing resources and attention. In Varieties of Attention; Parasuraman, R., Davis, D., Eds.; Academic Press: Cambridge, MA, USA, 1984. [Google Scholar]
  59. Huang, J.; Zhang, Q.; Zhang, T.; Wang, T.; Tao, D. Assessment of drivers’ mental workload by multimodal measures during auditory-based dual-task driving scenarios. Sensors 2024, 24, 1041. [Google Scholar] [CrossRef]
  60. White, H.; Shah, P. Attention in urban and natural environments. Yale J. Biol. Med. 2019, 92, 115. [Google Scholar]
  61. Edquist, J.; Rudin-Brown, C.M.; Lenné, M.G. The effects of on-street parking and road environment visual complexity on travel speed and reaction time. Accid. Anal. Prev. 2012, 45, 759–765. [Google Scholar] [CrossRef]
  62. Lee, S.C.; Kim, Y.W.; Ji, Y.G. Effects of visual complexity of in-vehicle information display: Age-related differences in visual search task in the driving context. Appl. Ergon. 2019, 81, 102888. [Google Scholar] [CrossRef] [PubMed]
  63. Qin, Z. Eye Movement Analysis on Driving Behavior for Urban Road Environment. In Proceedings of the International Conference on Mechatronics and Intelligent Robotics, Kunming, China, 20–21 May 2017; Springer International Publishing: Cham, Switzerland, 2017; pp. 486–491. [Google Scholar]
  64. Liang, Y.; Reyes, M.L.; Lee, J.D. Real-time detection of driver cognitive distraction using support vector machines. IEEE Trans. Intell. Transp. Syst. 2007, 8, 340–350. [Google Scholar] [CrossRef]
  65. Stapel, J.; El Hassnaoui, M.; Happee, R. Measuring driver perception: Combining eye-tracking and auto-mated road scene perception. Hum. Factors 2022, 64, 714–731. [Google Scholar] [CrossRef]
Figure 1. Overall methodological framework.
Figure 1. Overall methodological framework.
Jemr 18 00040 g001
Figure 2. Visual occlusion experiment.
Figure 2. Visual occlusion experiment.
Jemr 18 00040 g002
Figure 3. Semantic segmentation extracts scene features.
Figure 3. Semantic segmentation extracts scene features.
Jemr 18 00040 g003
Figure 4. Clustering performance evaluation. A larger silhouette coefficient and a smaller DBI (Davies–Bouldin Index) indicate greater separation between clusters and higher compactness within clusters, thereby reflecting better clustering performance.
Figure 4. Clustering performance evaluation. A larger silhouette coefficient and a smaller DBI (Davies–Bouldin Index) indicate greater separation between clusters and higher compactness within clusters, thereby reflecting better clustering performance.
Jemr 18 00040 g004
Figure 5. Ranking of the importance of scene features on visual load. (a) Overall ranking of SHAP value importance for scene features influencing low/medium/high visual workload states; (b) ranking of SHAP value importance for scene features influencing the high visual workload state; (c) ranking of SHAP value importance for scene features influencing the low visual workload state. In the figure, the color of each point represents the magnitude of the feature value, while its position indicates the SHAP value of the feature for the corresponding load state. A positive SHAP value indicates that the feature increases the probability of that load state occurring, whereas a negative SHAP value indicates a decreased probability.
Figure 5. Ranking of the importance of scene features on visual load. (a) Overall ranking of SHAP value importance for scene features influencing low/medium/high visual workload states; (b) ranking of SHAP value importance for scene features influencing the high visual workload state; (c) ranking of SHAP value importance for scene features influencing the low visual workload state. In the figure, the color of each point represents the magnitude of the feature value, while its position indicates the SHAP value of the feature for the corresponding load state. A positive SHAP value indicates that the feature increases the probability of that load state occurring, whereas a negative SHAP value indicates a decreased probability.
Jemr 18 00040 g005
Figure 6. SHAP dependence plots of key scene features under high and low load states. The X-axis represents the value range of the corresponding feature, and the Y-axis represents the SHAP value of that feature, indicating the change in probability it contributes to the model output.
Figure 6. SHAP dependence plots of key scene features under high and low load states. The X-axis represents the value range of the corresponding feature, and the Y-axis represents the SHAP value of that feature, indicating the change in probability it contributes to the model output.
Jemr 18 00040 g006
Figure 7. Ranking of the importance of scene features on blink-related indicators. (a,c) show the SHAP value importance ranking of scene features influencing blink duration and blink rate, respectively; (b,d) present SHAP feature density scatter plots of scene features influencing blink duration and blink rate. In the scatter plots, the color of each point represents the magnitude of the feature value, while its position indicates the SHAP value of the feature for the corresponding indicator. A positive SHAP value indicates that the feature increases the indicator value, whereas a negative SHAP value indicates a decrease.
Figure 7. Ranking of the importance of scene features on blink-related indicators. (a,c) show the SHAP value importance ranking of scene features influencing blink duration and blink rate, respectively; (b,d) present SHAP feature density scatter plots of scene features influencing blink duration and blink rate. In the scatter plots, the color of each point represents the magnitude of the feature value, while its position indicates the SHAP value of the feature for the corresponding indicator. A positive SHAP value indicates that the feature increases the indicator value, whereas a negative SHAP value indicates a decrease.
Jemr 18 00040 g007
Figure 8. SHAP dependence plots of key scene features for blink duration and blink rate. The X-axis represents the value range of the corresponding feature, and the Y-axis represents the SHAP value of that feature, indicating the magnitude of change it contributes to the model output.
Figure 8. SHAP dependence plots of key scene features for blink duration and blink rate. The X-axis represents the value range of the corresponding feature, and the Y-axis represents the SHAP value of that feature, indicating the magnitude of change it contributes to the model output.
Jemr 18 00040 g008
Figure 9. Ranking of the importance of scene features on blink-related indicators. (a,c,e) show the SHAP value importance rankings of scene features influencing saccade duration, average saccade duration and saccade frequency, respectively; (b,d,f) present SHAP feature density scatter plots of scene features influencing saccade duration, average saccade duration and saccade frequency. In the scatter plots, the color of each point represents the magnitude of the feature value, while its position indicates the SHAP value of the feature for the corresponding indicator. A positive SHAP value indicates that the feature increases the indicator value, whereas a negative SHAP value indicates a decrease.
Figure 9. Ranking of the importance of scene features on blink-related indicators. (a,c,e) show the SHAP value importance rankings of scene features influencing saccade duration, average saccade duration and saccade frequency, respectively; (b,d,f) present SHAP feature density scatter plots of scene features influencing saccade duration, average saccade duration and saccade frequency. In the scatter plots, the color of each point represents the magnitude of the feature value, while its position indicates the SHAP value of the feature for the corresponding indicator. A positive SHAP value indicates that the feature increases the indicator value, whereas a negative SHAP value indicates a decrease.
Jemr 18 00040 g009
Figure 10. SHAP dependence plots of key scene features for saccade duration, average saccade duration, and saccade frequency. The X-axis represents the value range of the corresponding feature, and the Y-axis represents the SHAP value of that feature, indicating the magnitude of change it contributes to the model output.
Figure 10. SHAP dependence plots of key scene features for saccade duration, average saccade duration, and saccade frequency. The X-axis represents the value range of the corresponding feature, and the Y-axis represents the SHAP value of that feature, indicating the magnitude of change it contributes to the model output.
Jemr 18 00040 g010
Table 1. Comprehensive features of urban driving scenes.
Table 1. Comprehensive features of urban driving scenes.
IndexImplication
Green Visibility Index (GVI)The ratio of the number of vegetation pixels to the total number of pixels in the image
Sky Visibility Index (SVI)The ratio of the number of sky pixels to the total number of pixels in the image
Road Width (RW)The ratio of the number of road pixels to the total number of pixels in the image
Street Canyon Enclosure (SCE)The ratio of the number of pixels of buildings, vegetation, traffic signs, fences, walls, and poles to the total number of pixels in the image
Building View Index (BVI)The ratio of the number of building pixels to the total number of pixels in the image
Road Congestion (RC)The ratio of the number of vehicle pixels to the number of road pixels
Traffic Mix (TM)The ratio of the number of non-motor vehicle and pedestrian pixels to the number of motor vehicle pixels
Table 2. Spearman correlation analysis of drivers’ visual metrics and attention demand.
Table 2. Spearman correlation analysis of drivers’ visual metrics and attention demand.
Visual IndexFixation
Duration (s)
Fixation Rate
(Times/s)
Blink Duration (s)Blink Rate (Times/s)Saccade
Duration
(s)
Average Saccade Duration
(ms)
Saccade
Frequency
(Times/s)
Correlation coefficient
(p-value)
−0.019
(0.849)
−0.186
(0.064)
0.239
(0.017 *)
0.262
(0.008 **)
−0.272
(0.006 **)
−0.357
(0.000 **)
−0.234
(0.019 *)
Note: ** and * represent 1% and 5% significance levels, respectively.
Table 3. Calibration of cluster centroid value and driving load class.
Table 3. Calibration of cluster centroid value and driving load class.
Centroid of Each GroupBlink
Duration (↓)
Blink
Rate
(↓)
Saccade
Duration (↑)
Average Saccade
Duration (↑)
Saccade
Frequency (↑)
Tentative Load
Status
Group 11.48531.5881−0.9653−0.1188−1.2978Low-load
Group 2−0.3322−0.3350−0.3318−0.49980.0354Medium-load
Group 3−0.3406−0.39981.16870.93780.7384High-load
Note: “↓” indicates that a lower value corresponds to a higher load; “↑” indicates that a higher value corresponds to a higher load.
Table 4. Driving visual load classification results.
Table 4. Driving visual load classification results.
GroupLoad ConditionSample SizeProportion
Group 1Low-load14818.50%
Group 2Medium-load43854.75%
Group 3High-load21426.75%
Table 5. Model performance evaluation.
Table 5. Model performance evaluation.
IndexOriginal DatasetBalanced Dataset
RFXGBAdaSVMRFXGBAdaSVM
PrecisionLow-load86.4179.8278.5081.2593.6095.6994.0493.77
Medium-load91.9888.6290.0084.4892.1394.6992.9691.64
High-load86.4087.0983.6587.5787.1588.8682.4688.44
Macro average88.2685.1884.0584.4390.9693.0889.8291.28
Weighted average87.9986.2784.5785.6190.9393.0689.9091.25
RecallLow-load80.9179.0976.3682.7393.0394.2490.9195.76
Medium-load81.8781.3279.1280.7790.1193.1387.0990.38
High-load93.1091.0989.6689.0889.6691.6790.5287.93
Macro average85.2983.8381.7184.1990.9393.0189.5091.36
Weighted average87.8186.2584.3885.6390.8892.9989.4491.27
F1-ScoreLow-load83.5779.4577.4281.9893.3194.9692.4588.44
Medium-load86.6384.8184.2182.5891.1193.9189.9391.01
High-load89.6389.0486.5588.3288.3990.2486.3088.18
Macro average86.6184.4482.7384.3090.9493.0489.5691.32
Weighted average87.7386.1984.3185.6090.9093.0289.5291.25
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ni, J.; Shao, Y.; Guo, Y.; Gu, Y. Interpretable Quantification of Scene-Induced Driver Visual Load: Linking Eye-Tracking Behavior to Road Scene Features via SHAP Analysis. J. Eye Mov. Res. 2025, 18, 40. https://doi.org/10.3390/jemr18050040

AMA Style

Ni J, Shao Y, Guo Y, Gu Y. Interpretable Quantification of Scene-Induced Driver Visual Load: Linking Eye-Tracking Behavior to Road Scene Features via SHAP Analysis. Journal of Eye Movement Research. 2025; 18(5):40. https://doi.org/10.3390/jemr18050040

Chicago/Turabian Style

Ni, Jie, Yifu Shao, Yiwen Guo, and Yongqi Gu. 2025. "Interpretable Quantification of Scene-Induced Driver Visual Load: Linking Eye-Tracking Behavior to Road Scene Features via SHAP Analysis" Journal of Eye Movement Research 18, no. 5: 40. https://doi.org/10.3390/jemr18050040

APA Style

Ni, J., Shao, Y., Guo, Y., & Gu, Y. (2025). Interpretable Quantification of Scene-Induced Driver Visual Load: Linking Eye-Tracking Behavior to Road Scene Features via SHAP Analysis. Journal of Eye Movement Research, 18(5), 40. https://doi.org/10.3390/jemr18050040

Article Metrics

Back to TopTop