Next Article in Journal
Application of Neural Networks to Analyse the Spatial Distribution of Bicycle Traffic Before, During and After the Closure of the Mill Road Bridge in Cambridgeshire, United Kingdom
Previous Article in Journal
Soil Porosity Detection Method Based on Ultrasound and Multi-Scale Feature Extraction
Previous Article in Special Issue
Hardware Acceleration-Based Privacy-Aware Authentication Scheme for Internet of Vehicles Using Physical Unclonable Function
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Driver Steering Intention Prediction for Human-Machine Shared Systems of Intelligent Vehicles Based on CNN-GRU Network

1
Information and Control Engineering, China University of Mining and Technology, Xuzhou 221000, China
2
Fujian Key Laboratory of Green Intelligent Drive and Transmission for Mobile Machinery, Huaqiao University, Xiamen 361021, China
3
Advanced Institute of Nano Technology, Sungkyunkwan University, Suwon 16228, Republic of Korea
4
Institute of Industrial Science, The University of Tokyo, Tokyo 113-8654, Japan
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(10), 3224; https://doi.org/10.3390/s25103224
Submission received: 31 March 2025 / Revised: 15 May 2025 / Accepted: 19 May 2025 / Published: 20 May 2025

Abstract

:
In order to mitigate human-machine conflicts and optimize shared control strategy in advance, it is essential for the shared control system to understand and predict driver behavior. This paper proposes a method for predicting driver steering intention with a CNN-GRU hybrid machine learning model. The convolutional neural network (CNN) layer extracts features from the stochastic driver behavior, which is input to the gated-recurrent-unit (GRU) layer. And the driver’s steering intention is forecasted based on the GRU model. Our study was conducted using a driving simulator to observe the lateral control behaviors of 18 participants in four different driving circumstances. Finally, the efficiency of the suggested prediction approach was evaluated employing long-short-term-memory, GRU, CNN, Transformer, and back propagation networks. Experimental results demonstrated that the proposed CNN-GRU model performs significantly better than baseline models. Compared with the GRU network, the CNN-GRU network reduced the RMSE, MAE, and MAPE of the driver’s input torque by 33.22%, 32.33%, and 35.86%, respectively. The proposed prediction method also possesses adaptability to different driver behaviors.

1. Introduction

New types of automated driving techniques bring enormous opportunities and challenges to the automotive industry [1,2]. According to various factors such as legal and ethical considerations, a fully autonomous vehicle is difficult to achieve in the short term. In addition to the challenge of determining driver preferences regarding vehicle operation, it is currently not feasible for drivers to detach themselves completely from the driving process [3]. Therefore, human-machine shared control is necessary to transition from semi-automation to full automation [4,5].
Human-machine shared steering control methods can be classified into two types: haptic shared control and indirect shared steering [6,7]. Compared with the indirect shared steering system, the haptic shared steering system enables the driver with haptic torque feedback from the automation system, via the steering wheel [8,9]. This method reduces the burden on the driver by continuously adjusting the torque on the driver [10]. Drivers can also exert greater force on the steering wheel to overturn the decisions made by the automation system whenever they choose [11]. Therefore, the haptic shared steering system has been extensively applied in vehicles [12].
Haptic shared steering system essentially predicts the driver’s steering intention [13]. The haptic steering system uses the predictions to optimize shared control strategies and mitigate human-machine conflicts [14]. Predicting driver intention also smoothens the transfer of control from the vehicle to the driver [15]. For example, in emergency situations, the haptic steering system can determine the capability of the driver to take over by predicting driver intention [16].
There are many driver intention prediction tasks in different environments. This study addresses the driver’s intention prediction during the steering process, in which the driver’s behavior is continuously changing. This behavior is directly influenced by the driver’s steering intention. Since the steering torque applied by the driver varies depending on the changes in behavior, the haptic shared steering system can estimate the intention by predicting the steering torque. There are two main categories of methods used to predict driver intention: model-based methods and machine learning methods. Model-based methods have notable interpretability and a high level of security [17]. Researchers have combined a linear quadratic regulator and a Kalman filter to predict the steering torque based on a pre-assumed driver model in [18]. Others have utilized an unscented Kalman filter to predict steering torque with a constant turning rate and acceleration mode in [19]. To ensure the prediction accuracy, these methods require previous information and detailed mathematical models of the system [20]. However, obtaining precise physical parameters is often difficult due to the stochastic human driving behavior, especially in highly dynamic driving scenarios. Dong et al. [21] validated that the stochastic variations of driver behaviors can significantly affect the accuracy of driver behavior modeling. Therefore, it is urgent to develop an efficient approach that can predict human driver behavior accurately.
In recent years, machine learning methods have been drawing broad attention due to their freeform model, such as Hidden Markov Model, Bayesian network, artificial neural network, and reinforcement learning [22]. The Hidden Markov Model is capable of predicting drivers’ behaviors through probabilistic inference. However, it requires predefined parameters of different driver states, which limits its generalization ability [23]. A filter based on dynamic Bayesian networks is proposed to predict driver behavior, but this method has a high computational complexity and still relies on pre-set assumptions [24]. Moreover, the artificial neural network can automatically extract features from the original input data with an end-to-end modeling approach. However, it is difficult for artificial neural networks to handle multidimensional inputs owing to their simple structures [25]. Reinforcement learning aims to maximize long-term rewards through interactions with the environment, which is more suitable for optimizing long-term decisions. Nonetheless, it may struggle to react quickly and adapt to complex driving scenarios when predicting the driver’s short-term intentions. Additionally, the personalized nature of driver behaviors makes the design of the reinforcement learning reward function more challenging.
Although past modeling methods are capable of predicting driver intentions to some extent, it is challenging to predict multi-step intentions in nonlinear human-machine systems [26]. In the shared steering systems of intelligent vehicles, high nonlinearity stems from the pronounced haptic interaction between the driver and the automation system [27,28]. Nevertheless, the long-short-term-memory (LSTM) network can predict longer steps in nonlinear situations because LSTM networks are specifically designed to analyze temporal data. Therefore, the LSTM network suitably encodes the behavioral characteristics of the driver to predict in the time domain [29,30]. Although the LSTM network can finish the prediction of time series data, it is difficult for the system to achieve fast prediction speed with a complex architecture. This is achieved with gate mechanisms that increase memory storage through the elimination of dedicated memory units [31]. Unlike the LSTM network, the GRU network is capable of efficiently predicting driver intention with a simplified architecture [32].
In human-machine shared control systems, the driver’s output torque exhibits strong nonlinearity due to the intervention of haptic guidance torque. Moreover, the input includes the haptic guidance torque and multi-source information, such as vehicle states and driver near and far points, resulting in a multi-dimensional feature set. Due to its strong ability to resist interference and process multidimensional information during feature extraction, a convolutional neural network (CNN) performs well in handling complex input data. As a result, a front-end feature extractor is incorporated with a CNN in this study. Researchers have demonstrated the CNN as an effective feature extractor in various fields [33,34]. On one hand, the convolutional layers of the CNN model are sparsely connected to the input with individual filters to extract the features effectively [35]. On the other hand, a pooling layer assists the CNN with retaining all features through reduced parameters [36].
Inspired by previous work, a CNN-GRU network was designed in this study to anticipate drivers’ steering intentions. The CNN model extracted features from the driver behavior data, whereas the GRU model predicted the driver’s steering intention. Experiments were conducted using a driving simulator, involving 18 participants, to analyze the lateral control behaviors exhibited by the drivers. Four experimental conditions were conducted to account for different driver steering tendencies. The experimental results verified the efficacy of the CNN-GRU network in different haptic torque scenarios during driver interactions.
The main contributions of the proposed methods can be summarized as follows:
  • The highly nonlinear and multidimensional input features (the haptic guidance torque, vehicle state, and driver near and far points) of the driver steering intention prediction model are extracted based on the CNN module.
  • The CNN and GRU networks are combined to predict the steering intention of drivers accurately by considering the haptic interaction between the driver and the automation system. Moreover, the proposed driver intention prediction strategy is validated through driving simulator experiments.
This publication provides further information that supports this conclusion: Section 2 provides a detailed description of the experimental platform, equipment, and situations. Section 3 explains the method for predicting driver steering intentions using the CNN-GRU model. In Section 4, we discuss the experimental results. In Section 5, the conclusions are finally drawn.

2. Experimental Design

The current study carried out driving simulator experiments to capture the driver behavior data. The proposed method for predicting steering intentions was formulated using the gathered data. What follows are descriptions of the test subjects, equipment, and experimental conditions involved in the data collection.

2.1. Participants

To obtain the driver steering behaviors, 18 participants consisting of 16 males and 2 females (mean age = 22 years, standard deviation (SD) = 1.8 years) were recruited for the experiments. Each participant owned a valid Japanese driver’s license with driving experience (mean = 2.7 years, SD =2.6 years). These participants drove about 6000 km each year on average. In addition, these participants usually drove on urban roads and expressways. While the gender ratio in the sample is uneven, it is reasonable because there are few significant gender differences in lane-changing behaviors, which are mainly influenced by drivers’ personal habits and response abilities. The University of Tokyo’s Ethics and Safety Office for Life Sciences sanctioned the experiment (No. 12 in 2017).

2.2. Apparatus

To guarantee the precision of the experimental results, the experiment utilized a driving simulator [32] to simulate a real-life road setting, as shown in Figure 1. This driving simulator can be divided into two parts: the D3sim software platform (Ver.6) and actual hardware modules. The D3sim software platform on the host computer was used to simulate the driving scenarios and vehicle movement, and it was developed by Mitsubishi Precision Co., Ltd. (Tokyo, Japan) The actual hardware modules consisted of a steering wheel, brake pedal, accelerator pedal, instrument panel, stereo speakers, 140 degrees viewer, and 6 degrees motion platform. The 140-degree viewer can be used to provide drivers with a broad driving view. The 6 degrees motion platform was utilized to replicate driving conditions by generating high-frequency vibrations [37]. This motion platform enabled drivers to have a better feeling of the road and vehicle status. In addition, the sound of the engine was simulated for drivers through two stereo speakers. These hardware modules and D3sim software with realistic driving scenarios make drivers approach real behaviors, which strengthens the driving simulator. In the driving simulator experiments, the driver’s input command with the real steering wheel, brake pedal, and accelerator pedal is transferred to the host computer [38]. Then the vehicle is manipulated from the driver’s command on the D3sim software. Meanwhile, the human-machine shared controller calculates and offers precise haptic torque input to drivers. Finally, the human-machine shared steering system is realized in the driving simulator.

2.3. Experimental Conditions and Scenario

Table 1 lists four experimental conditions accounting for different driver states and haptic guidance configurations. The experiment was conducted under attentive and distracted states with two forms of haptic guidance: fixed gain (HG-Fixed) and adaptive gain (HG-Adaptive).
An adaptive gain haptic guidance system was designed to adjust the allocation of human-machine authority. The torque for haptic guidance is calculated using the following formula:
T h = K ( a 1 e y ( n e a r ) + a 2 e θ ( f a r ) )
where e y ( n e a r ) , e θ ( f a r ) represents the visual information from the driver; a 1 and a 2 are the constant gains for the e y ( n e a r ) and e θ ( f a r ) , and their values are respectively set to 0.19 and 3.8. K is a constant factor that determines the magnitude of the haptic guiding torque.
This system uses a Myo armband to collect surface electromyography signals from the driver’s forearm. To prepare the surface electromyography signals for further processing, it calculates the root mean square of the activation signals from the stainless-steel armband sensors. The driver’s grip force is normalized against each participant’s maximum surface electromyographic signals. Then, based on the normalized surface Electromyography signals, it adjusts the proportion between the total steering torque and the tactile guidance torque. When the grip force increases, the system reduces the proportion of the haptic guidance torque to the total steering torque. The fixed gain is set as 25% of the total steering torque in the experiment (HG-Fixed, K = 0.25).
In order to replicate the distracted state of the driver in the experiments, an additional task called rhythmic auditory serial addition was introduced. The researchers displayed a number to each driver every 30 s, and then the drivers were asked to add the current number and the previous number.
During the experiment, participants were required to complete a predefined double lane change (DLC) task. A straight segment is simulated to provide consistent driving conditions in the experiment. Participants were instructed to drive along a 300-m straight lane first, and they were instructed to maintain their position within the lane. The vehicle speed was regulated automatically by the driving simulator without driver intervention. The vehicle began to accelerate uniformly at a rate of 1.8 m/s2 until reaching the speed of 50 km/h, after which it maintained a constant velocity. This speed setup was designed to encourage participants to focus on their steering behaviors, particularly under the influence of haptic guidance during the lane change maneuvers.

2.4. Data Analysis

For the sake of determining whether there was any significant difference among the prediction models, the experimental data were statistically analyzed using one-way repeated measures analysis of variance. If the p-value is less than 0.05, the null hypothesis, which states that there is no significant difference, would be rejected. A p-value ranging from 0.05 to 0.1 would suggest a tendency towards statistical significance.

3. Predictive Model of Driver Steering Intention

This section introduces a hybrid CNN-GRU model that is designed to forecast the driver’s steering intention for human-machine shared systems of intelligent vehicles. The overall structure of the driver steering intention prediction system is shown in Figure 2.
This section firstly discusses and defines the input and output features of the steering intention prediction model. Then, the CNN module and GRU module are integrated to predict the steering intention of drivers. Finally, we discuss the selection of the evaluation measures of the hybrid CNN-GRU model to validate accuracy.

3.1. Feature Learning

A total of 18 participants provided driving data under four conditions shown in Table 1. Information collected from six drivers formed the test data set, while the remainder served as training data. Both training and test data were randomized. Simple random sampling was applied to select data from 6 drivers to serve as the test dataset. To make multi-step predictions in the future, the data format was configured. Sequential data segmentation was performed using 10 data units, covering both the forecast and historical horizons. The structure of input and output is shown in Figure 3.
S t represents the anticipated sequential steering torque at time t :
S t = ( s ^ t + 1 , s ^ t + 2 , s ^ t + 3 , s ^ t + k )
where s ^ t + k is the predicted steering torque at time step t + k , t is the time in sequences, and k refers to the time step, with each step spaced 16.7 ms apart. In this paper, the value of k is set to 10.
Table 2 displays the input and output features of the driver steering prediction model.
Note that φ and SWA are used to reflect the vehicle state. T h represents the driver’s tactile interaction with the automated system. In addition, the output signal is the driver input torque T d . The driver’s intention can be recognized by categorizing the steering prediction based on the variation of T d . Therefore, the input I n _ c t for the model can be characterized in the following manner:
I n _ c t = e y ( t k ) , e θ ( t k ) , φ ( t k ) , S W A ( t k ) , T h ( t k ) , e y ( t k + 1 ) , e θ ( t k + 1 ) , , S W A ( t ) , T h ( t )
The input data are normalized to stabilize the training process and eliminate the influence of outliers. Then the data are packaged in 10-step intervals according to the format in Figure 3 to enable prediction of the next 10 steps. Afterwards, the data are fed into the network.

3.2. The Hybrid CNN-GRU Model

A hybrid CNN-GRU model is constructed by combining CNN and GRU networks to enhance the accuracy of predicting the driver’s steering intention.
First, the CNN is responsible for extracting the characteristics of time series data. The CNN feature extraction module is composed of the following three layers: average pooling, convolutional, and ReLU. In one-dimensional data, a convolutional kernel is considered to be a feature detector and filter for removing outliers. The convolutional kernel is applied to different regions of the input data to identify the time-related interdependencies within the data. The convolutional layer produces feature maps that are sensitive to the positions of the input features. Increasing the number of feature maps elevates the dimensionality of the data, and the equation of the convolution layer is presented as:
y p ( l + 1 ) = f ( y p ( l ) r o t 180 ( ω p ( l ) + b p ( l ) ) )
where y p ( l ) is the activation value of the l -th layer, ω p ( l ) and b p ( l ) are the weight and bias matrices, respectively, and r o t 180 ( ) means to rotate the input by 180°.
To reduce the spatial size of the feature mappings, while preserving important features, a pooling layer is used to sample the feature maps after the convolutional layer. The average pooling ensures the integrity of the information. The ReLU activation function is designed to improve the learning capacity of the model in complex environments and to address the issue of vanishing gradients. The equation of the pooling layer is presented as:
y p ( l + 1 ) = p o o l i n g ( y p ( l ) )
After the CNN finishes feature extraction, a GRU network adjusts the level of information updates through update and forget gates. Then the GRU propagates memory information through the unit’s hidden state h t j . At time t, the update of h t j is composed of the previous hidden state h t 1 j and the candidate state h ˜ t j . Meanwhile, the update of the unit hidden state is controlled by an update gate. The procedure can be described as follows:
h t j = ( 1 z t j ) h t 1 j + z t j h ˜ t 1 j
The update gate z t j calculates the amount of information that should be preserved from the previous hidden state. If z t j = 1, the h t 1 j ignores the previous content completely, and is computed based on the current input x t . The calculation for the update gate is as follows:
    z t j = σ ( W z x t + U z h t 1 ) j
The candidate hidden state h ˜ t j summarizes the information from the new input x t and hidden state h t 1 j :
h ˜ t j = tanh ( W z x t + U ( r t h t + 1 ) ) j
The reset gate plays a crucial role in determining the amount of information that is forgotten about the previous hidden state. If r t j = 1, the previous hidden state h t 1 j is updated and stored in the new memory. The reset gate can be obtained through the formula as follows:
    r t j = σ ( W r x t + U r h t 1 ) j
where W r , z , U r , z are learnable parameters, σ is the Sigmoid activation function.
We employ the GRU network to enhance the precision of prediction outcomes. This section starts as a prediction module with two layers of GRUs and intersperses two dropout layers to avoid overfitting. The purpose of having the last layer as a fully connected layer is to combine the characteristic features of the steering process extracted by the preceding layers. MATLAB (R2024b) was employed for the execution of the suggested model. The output channels of the convolutional layer are configured to 10 in order to improve feature extraction. Both hidden layers of the GRU are set to 180 channels, whereas the FC layer has 10 channels. The regularization parameter is set to 0.001, while the activation functions for the GRU and convolutional layers are ReLUs. Cross-entropy error served as the output of the loss function.

3.3. Model Evaluation

The evaluation of the prediction performance of driver steering intention involves the use of various metrics, including the root mean square error (RMSE), mean absolute error (MAE), and mean absolute percentage error (MAPE) [39]. The MAE can be described as:
MAE = 1 n i = 1 n | m ^ i m i |
The MAPE can be written as:
MAPE = 1 n i = 1 n | m ^ i m i | m ^ i × 100 %
The RMSE can be obtained as:
RMSE = 1 n i = 1 n ( m ^ i m i ) 2
where n is the total number of testing sequences, i is the i -th data point, m ^ i and m i indicates the predicted value and true value, respectively.
The prediction outcomes are subjected to statistical analysis utilizing a one-way repeated measures analysis of variance with the Post Hoc test, so as to ascertain the presence of any significant disparities across the driving models. A significance threshold of p < 0.05 is established to reject the null hypothesis, which posits that there is no statistically significant difference. The presence of a p-value ranging from 0.05 to 0.1 suggests a tendency towards a noteworthy distinction.

4. Experimental Results

To validate the effectiveness of the suggested CNN-GRU model, we used the model to predict driver intention using experimental data involving human-machine collaboration. For comparison, a propagation (BP) network, a CNN network, an LSTM network, a GRU network, and a Transformer were also evaluated. Four experimental conditions in Table 1 were applied to obtain predictions. The five modeling methods and their parameters are defined as follows:
  • LSTM Network: The LSTM network was structured with a 5-180-1 three-layer configuration, and the Adam optimization solver was employed with a learning rate of 0.005.
  • BP Network: The BP network consisted of 5-6-1 three layers, and the target training accuracy was set to 0.001, which ensures model convergence and the minimum prediction error.
  • CNN Network: The CNN network was made up of a 3-layer architecture, in which the convolution layers are a 1 × 10 filter size, and the fully connected layers were configured with 128 neurons for feature extraction. The learning rate of the Adam optimizer was set to 0.005 to achieve training.
  • GRU Network: The GRU network was structured with three layers, arranged in a 5-180-1 configuration. The Adam optimizer was applied with a learning rate of 0.005.
  • Transformer network: The Transformer network was structured with three layers, an Embedding layer, an Encoder layer, and a Decoder layer. The Adam optimizer was applied with a learning rate of 0.001.

4.1. Steering Torque Prediction of Attentive Drivers

The predicted input torque for the 18th driver under experimental Condition 1 is shown in Figure 4. It is evident that the CNN-GRU model outperforms other models in terms of tracking the observed driver input torque. Its predictions exhibit the closest alignment with the actual data, showcasing higher accuracy and better tracking capabilities. Other models, such as GRU, LSTM, CNN, Transformer, and BP, deviate more significantly from the observed values, particularly in regions of rapid fluctuations and transitions.
The experimental results of the final six drivers are shown in Table 3 to showcase the predictive accuracy of the proposed CNN-GRU model. The average values of RMSE, MAE, and MAPE for the steering torque prediction of the last six drivers are presented in Figure 5. According to Table 3, the CNN-GRU model achieved the least RMSE, MAE, and MAPE under Conditions 1 and 2. Based on these findings, it is clear that the CNN-GRU model is the most effective for predicting attentive driving states, as opposed to the other methods. Traditional models like GRU and LSTM primarily focus on capturing temporal dependencies but lack the ability to extract and leverage spatial features from raw input data effectively, which limits their performance for predicting driver steering intention. Due to vanishing gradients and slow convergence, it is harder for the BP network to capture both spatial and temporal features together. In addition, the GRU, LSTM, and BP methods show no significant variance from each other relative to the same performance metrics. As opposed to the BP, LSTM, GRU, and CNN networks, the CNN-GRU network significantly reduces the average RMSE of the driver input torque by 31.93%, 21.13%, 31.78%, 24.95%, and 30.65%, respectively. The CNN-GRU network reduces the average RMSE, MAE, and MAPE by 31.78%, 29.52%, and 30.02%, respectively, compared to the GRU network in Condition 2. The pairwise comparison results among different models indicate that the CNN-GRU model significantly surpasses its three counterparts in terms of RMSE, MAE, and MAPE.

4.2. Steering Torque Prediction of Distracted Drivers

The predicted input torque for the 14th driver in a distracted state was shown in Figure 6. The CNN-GRU model demonstrated superior accuracy in predicting driver input torque compared to the other models. This higher prediction accuracy reflects the advantage of the CNN-GRU model.
Table 4 shows experimental prediction results for the steering torque prediction of the last six distracted drivers. The average values of RMSE, MAE, and MAPE are shown in Figure 7. Table 4 indicates that the prediction errors of the BP, LSTM, GRU, CNN, and Transformer models surpass the CNN-GRU model under experimental Conditions 3 and 4. Therefore, the proposed CNN-GRU model emerges as the most precise forecaster of driver input torque. On the other hand, the other models have no significant difference in terms of RMSE, MAE, and MAPE.
The CNN-GRU network, on the other hand, achieves a significant reduction in the average RMSE of the driver input torque compared to the BP, LSTM, GRU, CNN, and Transformer networks. Specifically, it decreases the average RMSE by 28.94%, 35.21%, 33.22%, 21.72%, and 27.65%, respectively. Compared with the GRU network under Condition 3, the CNN-GRU network decreases RMSE, MAE, and MAPE of the driver input torque by 33.22%, 32.33%, and 35.86%, respectively. The pairwise comparison results demonstrate that the CNN-GRU model outperforms the others by a significant margin.
The proposed CNN-GRU model was implemented on the MATLAB platform and tested on a system with an Intel Core i5—10,200H CPU (2.4 GHz) and 16 GB RAM. On average, the prediction time per sample was within 4.8–6.1 ms, which supports real-time computing in a human-machine shared control system. In comparison, baseline models such as GRU LSTM, CNN, and Transformer require approximately 8.3 ms, 10.5 ms, 11.4 ms, and 10.7 ms per sample, respectively. Specifically, our model improves the computing speed by 34.30%, 48.09%, 52.02%, and 49.06%, respectively. These results demonstrate that the proposed CNN-GRU model achieves the fastest solving speed. The automated system can conduct model training offline and directly apply model predictions to future time steps in practical applications.
In addition to superior predictive accuracy, the proposed CNN-GRU network effectively predicts intention across different driver states and haptic interactions. The model thus demonstrates strong adaptation and generalization. By combining the CNN’s capability to extract local spatial features and the GRU’s proficiency in modeling temporal dependencies, the CNN-GRU network represents the intricate inherent dynamics more comprehensively in human-machine interaction. As a result, the CNN-GRU network achieves more accurate and robust prediction of driver intentions.
To evaluate the sensitivity of the proposed model, a detailed analysis of the input window size and the number of GRU hidden units affecting prediction accuracy was conducted. The analysis data originated from a 1 h driver under distracted conditions and the HG-Adaptive scenario, and the sensitivity analysis results are shown in Table 5. It indicates that the shortest input window achieved the best performance with an RMSE of 0.1802 N·m and an MAE of 0.0997 N·m. It is obvious that the prediction error also increased with the increase in the window size. And the poorest performance is accompanied by the 20-frame window, which may be caused by the noise and redundant information of compromised feature extraction. Regarding the number of GRU hidden units, the configuration with 180 units offered an optimal balance between prediction accuracy and generalization ability. In contrast, the configuration with 60 units resulted in underfitting due to limited representation capacity, while overfitting is shown with 240 units. These findings demonstrate that the proposed CNN-GRU model is sensitive to both the temporal length of input data and the network capacity. Hence, the optimal tuning of these parameters is essential to ensure accurate prediction of driver steering intentions.

5. Conclusions

This article proposes a novel CNN-GRU model for predicting driver steering intention in a human-machine shared system. The proposed model extracts data features using a CNN layer and predicts the driver’s steering intention using the GRU layer. To confirm the efficacy of the suggested approach, driving simulation results are presented under four different experimental conditions. The BP, LSTM, GRU, CNN, and Transformer networks are used as performance benchmarks.
Based on the experimental results, it is evident that the proposed CNN-GRU network outperformed the other models in terms of accuracy. For example, compared with the GRU network in experimental Condition 4, the CNN-GRU network reduced RMSE, MAE, and MAPE by 27.56%, 27.34%, and 28.73%, respectively. A further benefit was the ability of the CNN-GRU network to be readily applied to differing experimental conditions, thus demonstrating strong adaptation and generalization capabilities.
However, the experiments only considered attentive and distracted states of drivers. Therefore, future research will include more intricate driver states in order to enhance the validation of the suggested prediction model’s generalizability. An additional constraint pertains to the exclusive use of a simulator platform for the execution of these investigations, and thus, we plan to conduct actual vehicle experiments to more accurately assess the effectiveness of our method.

Author Contributions

Conceptualization, C.Z.; methodology, C.Z.; software, F.Z. and E.J.C.N.; validation, F.Z.; formal analysis, E.J.C.N.; investigation, Z.W.; resources, Z.W.; writing—original draft preparation, F.Z.; writing—review and editing, F.-X.X.; visualization, F.Z.; supervision, F.-X.X.; project administration, F.-X.X.; funding acquisition, F.-X.X. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 52475143, in part by Jiangsu Provincial Natural Science Foundation under Grant BK20241638, in part by Xuzhou City Scientific and Technological Achievements Transformation Project under Grant KC23369, in part by the Open Foundation of Fujian Key Laboratory of Green Intelligent Drive and Transmission for Mobile Machinery under Grant 202304, and in part by National Key Research and Development Program under Grant 2023YFC2907600.

Institutional Review Board Statement

The study was approved by the Office for Life Science Research Ethics and Safety, Graduate School of Interdisciplinary Information Studies, University of Tokyo (protocol code No. 12 in 2017).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available from the corresponding author upon request.

Acknowledgments

We would like to thank Chuanwang Shen from Inspur Zhuoshu Big Data Industry Development Co., Ltd., Wuxi, China, for his contributions in curating the data of this research.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Song, R.; Li, B. Surrounding vehicles’ lane change maneuver prediction and detection for intelligent vehicles: A comprehensive review. IEEE Trans. Intell. Transp. Syst. 2022, 23, 6046–6062. [Google Scholar] [CrossRef]
  2. Cao, Y.; Zhang, B.; Hou, X.; Gan, M.; Wu, W. Human-Centric Spatial Cognition Detecting System Based on Drivers’ Electroencephalogram Signals for Autonomous Driving. Sensors 2025, 25, 397. [Google Scholar] [CrossRef] [PubMed]
  3. Lu, Y.; Huang, L.; Yao, J.; Su, R. Intention Prediction-Based Control for Vehicle Platoon to Handle Driver Cut-In. IEEE Trans. Intell. Transp. Syst. 2023, 24, 5489–5501. [Google Scholar] [CrossRef]
  4. Teng, S.; Hu, X.; Deng, P.; Li, B.; Li, Y.; Ai, Y.; Yang, D.; Li, L.; Xuan, Y.Z.; Zhu, F.; et al. Motion planning for autonomous driving: The state of the art and future perspectives. IEEE Trans. Intell. Veh. 2023, 8, 3692–3711. [Google Scholar] [CrossRef]
  5. Chen, T.; Chen, Y.; Li, H.; Gao, T.; Tu, H.; Li, S. Driver Intent-Based Intersection Autonomous Driving Collision Avoidance Reinforcement Learning Algorithm. Sensors 2022, 22, 9943. [Google Scholar] [CrossRef]
  6. Oudainia, M.R.; Sentouh, C.; Nguyen, A.-T.; Popieul, J.-C. Adaptive cost function-based shared driving control for cooperative lane-keeping systems with user-test experiments. IEEE Trans. Intell. Veh. 2024, 9, 304–314. [Google Scholar] [CrossRef]
  7. Marcano, M.; Díaz, S.; Pérez, J.; Irigoyen, E. A review of shared control for automated vehicles: Theory and applications. IEEE Trans. Human-Mach. Syst. 2020, 50, 475–491. [Google Scholar] [CrossRef]
  8. Yang, L.; Babayi Semiromi, M.; Xing, Y.; Lv, C.; Brighton, J.; Zhao, Y. The Identification of Non-Driving Activities with Associated Implication on the Take-Over Process. Sensors 2022, 22, 42. [Google Scholar] [CrossRef]
  9. Zhang, J.; Guo, X.; Fu, Z.; Liu, Y.; Ding, Y. Noncooperative-game-theory-based driver-automation shared steering control considering driver steering behavior characteristics. IEEE Internet Things J. 2024, 11, 28465–28479. [Google Scholar] [CrossRef]
  10. Wang, Y.; Wang, C.; Zhao, W.; Xu, C. Decision-making and planning method for autonomous vehicles based on motivation and risk assessment. IEEE Trans. Veh. Technol. 2021, 70, 107–120. [Google Scholar] [CrossRef]
  11. Guo, H.; Shi, W.; Liu, J.; Guo, J.; Meng, Q.; Cao, D. Game-theoretic human-machine shared steering control strategy under extreme conditions. IEEE Trans. Intell. Veh. 2024, 9, 2766–2779. [Google Scholar] [CrossRef]
  12. Lazcano, A.M.R.; Niu, T.; Akutain, X.C.; Cole, D.; Shyrokau, B. Mpc-based haptic shared steering system: A driver modeling approach for symbiotic driving. IEEE/ASME Trans. Mechatron. 2021, 26, 1201–1211. [Google Scholar] [CrossRef]
  13. Lee, H.; Kim, H.; Choi, S. Driving skill modeling using neural networks for performance-based haptic assistance. IEEE Trans. Hum. Mach. Syst. 2021, 51, 198–210. [Google Scholar] [CrossRef]
  14. Noubissie Tientcheu, S.I.; Du, S.; Djouani, K.; Liu, Q. Data-Driven Controller for Drivers’ Steering-Wheel Operating Behaviour in Haptic Assistive Driving System. Electronics 2024, 13, 1157. [Google Scholar] [CrossRef]
  15. Fang, Z.; Wang, J.; Liang, J.; Yan, Y.; Pi, D.; Zhang, H.; Yin, G. Authority allocation strategy for shared steering control considering human-machine mutual trust level. IEEE Trans. Intell. Veh. 2024, 9, 2002–2015. [Google Scholar] [CrossRef]
  16. Okada, K.; Sonoda, K.; Wada, T. Transferring from automated to manual driving when traversing a curve via haptic shared control. IEEE Trans. Intell. Veh. 2021, 6, 266–275. [Google Scholar] [CrossRef]
  17. Bhattacharyya, R.; Jung, S.; Kruse, L.A.; Senanayake, R.; Kochenderfer, M.J. A hybrid rule-based and data-driven approach to driver modeling through particle filtering. IEEE Trans. Intell. Transp. Syst. 2022, 23, 13055–13068. [Google Scholar] [CrossRef]
  18. Schubert, R.; Richter, E.; Wanielik, G. Comparison and evaluation of advanced motion models for vehicle tracking. In Proceedings of the 2008 11th International Conference on Information Fusion, Cologne, Germany, 30 June–3 July 2008; IEEE: Piscataway, NJ, USA, 2008; pp. 1–6. [Google Scholar]
  19. Xing, Y.; Lv, C. Dynamic state estimation for the advanced brake system of electric vehicles by using deep recurrent neural networks. IEEE Trans. Ind. Electron. 2020, 67, 9536–9547. [Google Scholar] [CrossRef]
  20. Mozaffari, S.; Al-Jarrah, O.Y.; Dianati, M.; Jennings, P.; Mouzakitis, A. Deep learning-based vehicle behavior prediction for autonomous driving applications: A review. IEEE Trans. Intell. Transp. Syst. 2022, 23, 33–47. [Google Scholar] [CrossRef]
  21. Dong, H.; Shi, J.; Zhuang, W.; Li, Z.; Song, Z. Analyzing the impact of mixed vehicle platoon formations on vehicle energy and traffic efficiencies. Appl. Energy 2024, 377, 124448. [Google Scholar] [CrossRef]
  22. Koochaki, F.; Najafizadeh, L. A data-driven framework for intention prediction via eye movement with applications to assistive systems. IEEE Trans. Neural Syst. Rehabil. Eng. 2021, 29, 974–984. [Google Scholar] [CrossRef] [PubMed]
  23. Deng, Q.; Soffker, D. A review of HMM-based approaches of driving behaviors recognition and prediction. IEEE Trans. Intell. Veh. 2022, 7, 21–31. [Google Scholar] [CrossRef]
  24. Xie, G.; Gao, H.; Qian, L.; Huang, B.; Li, K.; Wang, J. Vehicle Trajectory Prediction by Integrating Physics- and Maneuver-Based Approaches Using Interactive Multiple Models. IEEE Trans. Ind. Electron. 2018, 65, 5999–6008. [Google Scholar] [CrossRef]
  25. Lee, S.; Khan, M.Q.; Husen, M.N. Continuous Car Driving Intent Detection Using Structural Pattern Recognition. IEEE Trans. Intell. Transp. Syst. 2021, 22, 1001–1013. [Google Scholar] [CrossRef]
  26. Ou, C.; Karray, F. Deep learning-based driving maneuver prediction system. IEEE Trans. Veh. Technol. 2022, 69, 1328–1340. [Google Scholar] [CrossRef]
  27. Lupberger, S.; Degel, W.; Odenthal, D.; Bajcinca, N. Nonlinear control design for regenerative and hybrid antilock braking in electric vehicles. IEEE Trans. Control Syst. Technol. 2022, 30, 1375–1389. [Google Scholar] [CrossRef]
  28. Okamoto, K.; Tsiotras, P. Data-driven human driver lateral control models for developing haptic-shared control advanced driver assist systems. Robot. Auton. Syst. 2019, 114, 155–171. [Google Scholar] [CrossRef]
  29. Cheng, K.; Sun, D.; Qin, D.; Cai, J.; Chen, C. Deep learning approach for unified recognition of driver speed and lateral intentions using naturalistic driving data. Neural Netw. 2024, 179, 106569. [Google Scholar] [CrossRef]
  30. Jahan, I.A.; Huq, A.S.; Mahadi, M.K.; Jamil, I.A.; Shahriar, M.Z. RoadSense: A Framework for Road Condition Monitoring using Sensors and Machine Learning. IEEE Trans. Intell. Veh. 2024; early access. [Google Scholar] [CrossRef]
  31. Dey, R.; Salem, F.M. Gate-variants of Gated Recurrent Unit (GRU) neural networks. In Proceedings of the 2017 IEEE 60th International Midwest Symposium on Circuits and Systems (MWSCAS), Boston, MA, USA, 6–9 August 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1597–1600. [Google Scholar]
  32. Xu, F.-X.; Feng, S.-Y.; Nacpil, E.J.C.; Wang, Z.; Wang, G.-Q.; Zhou, C. Modeling lateral control behaviors of distracted drivers for haptic-shared steering system. IEEE Trans. Intell. Transp. Syst. 2023, 24, 14772–14782. [Google Scholar] [CrossRef]
  33. Choi, G.; Lim, K.; Pan, S.B. Driver Identification System Using 2D ECG and EMG Based on Multistream CNN for Intelligent Vehicle. IEEE Sens. Lett. 2022, 6, 6001904. [Google Scholar] [CrossRef]
  34. Nguyen, D.-L.; Putro, M.D.; Jo, K.-H. Lightweight CNN-Based Driver Eye Status Surveillance for Smart Vehicles. IEEE Trans. Ind. Inf. 2024, 20, 3154–3162. [Google Scholar] [CrossRef]
  35. Cura, A.; Küçük, H.; Ergen, E.; Öksüzoğlu, İ.B. Driver profiling using long short term memory (LSTM) and convolutional neural network (CNN) methods. IEEE Trans. Intell. Transp. Syst. 2021, 22, 6572–6582. [Google Scholar] [CrossRef]
  36. Uddin, M.Z.; Hassan, M.M. Activity Recognition for Cognitive Assistance Using Body Sensors Data and Deep Convolutional Neural Network. IEEE Sens. J. 2019, 19, 8413–8419. [Google Scholar] [CrossRef]
  37. Krasniuk, S.; Toxopeus, R.; Knott, M.; McKeown, M.; Crizzle, A.M. The effectiveness of driving simulator training on driving skills and safety in young novice drivers: A systematic review of interventions. J. Saf. Res. 2024, 91, 20–37. [Google Scholar] [CrossRef]
  38. Wang, Z.; Suga, S.; Nacpil, E.J.C.; Yang, B.; Nakano, K. Effect of Fixed and sEMG-Based Adaptive Shared Steering Control on Distracted Driver Behavior. Sensors 2021, 21, 7691. [Google Scholar] [CrossRef]
  39. Xu, F.-X.; Zhou, C.; Liu, X.-H.; Wang, J. GRNN inverse system based decoupling control strategy for active front steering and hydro-pneumatic suspension systems of emergency rescue vehicle. Mech. Syst. Signal Process. 2022, 167, 108595. [Google Scholar] [CrossRef]
Figure 1. Driving simulator [32].
Figure 1. Driving simulator [32].
Sensors 25 03224 g001
Figure 2. Overall structure of the driver steering intention prediction system.
Figure 2. Overall structure of the driver steering intention prediction system.
Sensors 25 03224 g002
Figure 3. Data structure framework of input and output.
Figure 3. Data structure framework of input and output.
Sensors 25 03224 g003
Figure 4. Continuous steering torque prediction results by Driver 18.
Figure 4. Continuous steering torque prediction results by Driver 18.
Sensors 25 03224 g004
Figure 5. Torque prediction results for attentive driver under HG-Fixed driving conditions across four modeling methods: (a) RMSE, (b) MAE, (c) MPAE.
Figure 5. Torque prediction results for attentive driver under HG-Fixed driving conditions across four modeling methods: (a) RMSE, (b) MAE, (c) MPAE.
Sensors 25 03224 g005aSensors 25 03224 g005b
Figure 6. Continuous steering torque prediction results by Driver 14.
Figure 6. Continuous steering torque prediction results by Driver 14.
Sensors 25 03224 g006
Figure 7. Results of torque prediction under four modeling methods during HG-adaptive driving conditions: (a) RMSE, (b) MAE, (c) MPAE.
Figure 7. Results of torque prediction under four modeling methods during HG-adaptive driving conditions: (a) RMSE, (b) MAE, (c) MPAE.
Sensors 25 03224 g007
Table 1. Experimental conditions.
Table 1. Experimental conditions.
ConditionDriver StateHaptic Guidance
1AttentiveHG-Fixed
2AttentiveHG-Adaptive
3DistractedHG-Fixed
4DistractedHG-Adaptive
Table 2. Input and output features of the driver steering prediction model.
Table 2. Input and output features of the driver steering prediction model.
DataVariableDescription
Input e y ( n e a r ) Lateral error at the near point (m)
e θ ( f a r ) Yaw error at the far point (deg)
φ Angle between the vehicle’s
longitudinal axis and the current lane (deg)
SWASteering wheel angle (deg)
T h Haptic guidance torque (N⋅m)
Output T d Driver input torque (N⋅m)
Table 3. Experimental prediction results for the steering torque prediction of the last six attentive drivers.
Table 3. Experimental prediction results for the steering torque prediction of the last six attentive drivers.
VariableCNN-GRU
(Ours)
(1)
M(SD)
GRU
(2)
M(SD)
LSTM
(3)
M(SD)
BP
(4)
M(SD)
CNN
(5)
M(SD)
Transformer
(6)
M(SD)
p 1–21–31–41–51–62–32–43–4
Condition 1RMSE
(N·m)
0.2172
(0.0257)
0.3184
(0.0261)
0.2754
(0.034)
0.3191
(0.0587)
0.2894
(0.0461)
0.3132
(0.0258)
0.000**********0.6511.0000.421
MAE
(N·m)
0.1724
(0.0194)
0.2446
(0.0059)
0.2192
(0.032)
0.2436
(0.062)
0.2366
(0.0512)
0.2460
(0.0674)
0.000*******0.8451.0001.000
MAPE
(%)
11.26
(0.544)
16.09
(0.341)
14.63
(0.96)
15.44
(1.631)
15.61
(3.265)
13.14
(7.747)
0.000*******1.0001.0001.000
Condition 2RMSE
(N·m)
0.2287
(0.0174)
0.3119
(0.0212)
0.3552
(0.023)
0.3810
(0.020)
0.2942
(0.032)
0.3278
(0.0437)
0.000********1.000*1.000
MAE
(N·m)
0.1860
(0.0142)
0.2559
(0.0177)
0.2785
(0.017)
0.3003
(0.016)
0.2291
(0.039)
0.2665
(0.045)
0.000*******0.293*1.000**1.000
MAPE
(%)
12.83
(1.173)
17.47
(1.303)
18.38
(1.428)
19.62
(1.324)
15.21
(3.563)
17.67
(3.742)
0.000**0.1897**0.061*1.0000.1971.000
* p < 0.05, ** p < 0.01, and *** p < 0.001; M denotes mean value.
Table 4. Experimental prediction results for the steering torque prediction of the last six distracted drivers.
Table 4. Experimental prediction results for the steering torque prediction of the last six distracted drivers.
VariableCNN-GRU
(Ours)
(1)
M(SD)
GRU
(2)
M(SD)
LSTM
(3)
M(SD)
BP
(4)
M(SD)
CNN
(5)
M(SD)
Transformer
(6)
M(SD)
p1–21–31–41–51–62–32–43–4
Condition 3RMSE
(N·m)
0.2054
(0.0158)
0.3076
(0.0095)
0.3170
(0.0338)
0.2892
(0.0139)
0.2624
(0.0387)
0.2839
(0.0354)
0.000******1.0001.0001.000
MAE
(N·m)
0.1595
(0.0139)
0.2357
(0.0122)
0.2294
(0.0163)
0.2242
(0.0101)
0.2037
(0.0296)
0.2286
(0.0338)
0.000****0.1130.161.0001.0001.000
MAPE
(%)
10.82
(2.44)
16.87
(1.75)
16.17
(3.38)
14.66
(1.93)
13.70
(2.41)
15.12
(2.15)
0.000*****0.0951.0000.5261.000
Condition 4RMSE
(N·m)
0.2413
(0.0316)
0.3331
(0.0288)
0.3171
(0.0329)
0.3684
(0.0461)
0.3159
(0.0731)
0.3231
(0.0629)
0.000******1.0001.0000.6147
MAE
(N·m)
0.1941
(0.0251)
0.2670
(0.0227)
0.2423
(0.0230)
0.3013
(0.0380)
0.2457
(0.0654)
0.2651
(0.0553)
0.000***0.057**0.42180.60230.2667
MAPE
(%)
13.52
(1.95)
18.97
(1.45)
17.26
(1.67)
20.23
(2.15)
16.82
(4.61)
18.27
(4.49)
0.000*******0.36471.0000.4931
* p < 0.05, ** p < 0.01.
Table 5. Sensitivity analysis results of the CNN-GRU model for the 18th driver.
Table 5. Sensitivity analysis results of the CNN-GRU model for the 18th driver.
VariableRMSE
(N·m)
MAE
(N·m)
Windows size of input data50.18020.0997
100.21720.1205
200.33270.2017
Number of GRU hidden units600.28720.1665
1800.21720.1205
2400.35340.2413
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, C.; Zhang, F.; Nacpil, E.J.C.; Wang, Z.; Xu, F.-X. Driver Steering Intention Prediction for Human-Machine Shared Systems of Intelligent Vehicles Based on CNN-GRU Network. Sensors 2025, 25, 3224. https://doi.org/10.3390/s25103224

AMA Style

Zhou C, Zhang F, Nacpil EJC, Wang Z, Xu F-X. Driver Steering Intention Prediction for Human-Machine Shared Systems of Intelligent Vehicles Based on CNN-GRU Network. Sensors. 2025; 25(10):3224. https://doi.org/10.3390/s25103224

Chicago/Turabian Style

Zhou, Chen, Fan Zhang, Edric John Cruz Nacpil, Zheng Wang, and Fei-Xiang Xu. 2025. "Driver Steering Intention Prediction for Human-Machine Shared Systems of Intelligent Vehicles Based on CNN-GRU Network" Sensors 25, no. 10: 3224. https://doi.org/10.3390/s25103224

APA Style

Zhou, C., Zhang, F., Nacpil, E. J. C., Wang, Z., & Xu, F.-X. (2025). Driver Steering Intention Prediction for Human-Machine Shared Systems of Intelligent Vehicles Based on CNN-GRU Network. Sensors, 25(10), 3224. https://doi.org/10.3390/s25103224

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop