Next Article in Journal
ECG Electrode Localization: 3D DS Camera System for Use in Diverse Clinical Environments
Next Article in Special Issue
Conditional Generative Models for Dynamic Trajectory Generation and Urban Driving
Previous Article in Journal
Microwave Gas Sensors Based on Electrodeposited Polypyrrole–Nickel Phthalocyanine Hybrid Films
Previous Article in Special Issue
Deep Deterministic Policy Gradient-Based Autonomous Driving for Mobile Robots in Sparse Reward Environments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sensor-Based Classification of Primary and Secondary Car Driver Activities Using Convolutional Neural Networks

1
Department of Biosensors and Processing of Biomedical Signals, Faculty of Biomedical Engineering, Silesian University of Technology, Roosevelta 40, 41-800 Zabrze, Poland
2
Institute of Medical Informatics, University of Lübeck, Ratzeburger Allee 160, 23562 Lübeck, Germany
3
Department of Information Technology, University of the Punjab, Lahore 54000, Pakistan
4
Department of Knowledge Engineering, University of Economics in Katowice, Bogucicka 3, 40-287 Katowice, Poland
*
Authors to whom correspondence should be addressed.
Sensors 2023, 23(12), 5551; https://doi.org/10.3390/s23125551
Submission received: 25 April 2023 / Revised: 6 June 2023 / Accepted: 9 June 2023 / Published: 13 June 2023

Abstract

:
To drive safely, the driver must be aware of the surroundings, pay attention to the road traffic, and be ready to adapt to new circumstances. Most studies on driving safety focus on detecting anomalies in driver behavior and monitoring cognitive capabilities in drivers. In our study, we proposed a classifier for basic activities in driving a car, based on a similar approach that could be applied to the recognition of basic activities in daily life, that is, using electrooculographic (EOG) signals and a one-dimensional convolutional neural network (1D CNN). Our classifier achieved an accuracy of 80% for the 16 primary and secondary activities. The accuracy related to activities in driving, including crossroad, parking, roundabout, and secondary activities, was 97.9%, 96.8%, 97.4%, and 99.5%, respectively. The F1 score for secondary driving actions (0.99) was higher than for primary driving activities (0.93–0.94). Furthermore, using the same algorithm, it was possible to distinguish four activities related to activities of daily life that were secondary activities when driving a car.

1. Introduction

To drive safely, the driver must be sufficiently aware of his/her surroundings, pay constant attention to the road and traffic, and be alert enough to react to unexpected circumstances [1,2,3,4]. Tasks that are directly related to maneuvering a vehicle are called basic driving activities [5].
The lack of concentration of drivers remains one of the crucial factors that contribute to serious accidents and deaths on the road and continues to be a problem for international road safety measures, as they affect not only the driver but also everyone else on the road [6,7]. Approximately 324,000 people were injured due to driver inattention in the United States in 2020 and more than 3000 lost their lives, representing 8.1% of all fatal accidents in the country [8,9].
The term “driver fatigue” refers to a particular type of inattention that occurs when a driver removes his/her focus from basic vehicle navigation tasks to focus on another activity [10]. These distractions may come from common activities, such as talking to other passengers and eating, as well as using mobile phones and systems [11]. These activities can have different effects on drivers. From the point of view of support by measurement technologies, existing research indicates two main areas:
  • Detect anomalies in driver behavior to prevent an accident, with personalized behavior measures of driving style through face detection or Internet of Things (IoT) technologies [12,13,14,15,16];
  • Monitor correct cognitive and safe driver behaviors with intelligent sensors and IoT to monitor the face, eyes, or movements of a driver’s entire body for a novel driver education process [17].
Although there are two distinct categories, they have a lot in common; for example, they are recorded and classified within the framework of currently available technologies, and many activities can be assigned into more than one of them. The use of a device requires participation in all of these distractions, also known as secondary driving activities. The cognitive distraction that occurs in the driver’s brain is the most difficult to identify. This phenomenon is also known as “looking but not seeing”. Attention requirements for distracting work and the prevalence of multitasking among drivers are two fundamental elements of the problem of distributed driving safety [18,19,20].
Task demand is the total amount of visual, physical, and cognitive resources required to perform the activity. The second issue is the frequency with which the drivers perform the task. Even a task that is small, but performed frequently, can pose a safety concern [2,3,19,21,22].
According to [23,24], the results suggest that activities that require the driver to look away from the road or perform manual tasks significantly increase the probability of a collision. The risk of a traffic accident increases by 2.05 when using a mobile phone, especially when dialing (×12) and sending messages (×6).
The long time spent looking away from the road also has a significant impact. According to some studies, removing your eyes from the road for more than two seconds significantly increases the probability of safety-critical events [25]. In fact, the U.S. Department of Transportation advises against taking your eyes off the road repeatedly in 12 s while operating a motor vehicle [26]. Recognition of human activity based on preconstructed groups of activities is a commonly used approach [27,28,29,30].
There are many well-described activities, mainly related to basic needs and daily life, e.g., breathing, eating, sleeping, and walking [31,32,33,34,35]. Among the recognition of these activities, some were divided into even more detailed (complex) activities, e.g., food was divided into food preparation, and food preparation was even more separated for the preparation of breakfast, lunch, and dinner. Using this convention, we decided to analyze and recognize the activities and scenarios that accompany driving a vehicle.
To explore more deeply the research problem, our objective was two-fold: to identify the prevailing road conditions during a trip and to determine whether individuals exhibited improved parking skills after the journey. Accomplishing this required the acquisition of a substantial volume of data. This paper outlines the integration of data and fundamental principles of physics into sensors embedded within JINS MEME ES_R glasses, as well as the methodology employed to acquire and analyze the collected data for classification purposes [3,36,37,38,39,40].
To summarize, we make the following contributions:
  • We investigate the use of JINS MEME ES_R (smart glasses) sensor data and develop a state-of-the-art machine learning model that learns patterns related to the primary and secondary activities of drivers and classifies them into their respective classes.
  • We perform a comparative analysis of wearable sensor data consisting of nine activities of the first driver and four activities of the second driver.
  • We provide a brief review of related approaches.
The rest of the article is structured as follows: Section 2 presents the current state of the art in the field of recognition of vehicle driver activities. Section 3 describes the materials and methods used to analyze signals to assess these activities. Section 4 presents the experimental results. Section 5 provides a discussion and, finally, Section 6 concludes this work.

2. Related Works

When looking for examples of similar studies to compare, it should be noted that, in a ratio of four to one, articles were found dedicated to searching for anomalies such as drowsiness, fatigue, lack of driver concentration, and external factors associated with vehicle damage and atmospheric factors associated with driving conditions [41,42,43].
In another study based on data tracking the head and eyes in driving simulation conditions, the activity of 73 people who performed various secondary tasks while driving was recorded. The results of this research improved performance classification through the development of new functions, in particular to assess the context of autonomous driving [44]. Algorithms for the classification of eye movements were divided into methods based on statistical thresholds and probabilistic methods. Algorithms based on static thresholds are usually selected for the classification of tasks assigned to the person who performs them; in other words, they are limited in quantity.
Probabilistic methods were introduced to meet the challenge of automatic adaptation of many people as a result of various behaviors, for example, individual visual cues. Drowsiness while driving is a critical issue in the context of road safety. Several approaches have been developed to reduce the risk of driver drowsiness. Fatigue and drowsiness detection techniques are divided into three broad strategies, namely vehicle-based, physiological, and driver-based approaches. This article discusses the latest research on diagnosing driver drowsiness based on behavior, in particular changes in eye movements and facial features.

2.1. Drowsiness in Drivers

Another research project turned to a traffic surveillance system developed to detect and warn the driver of a degree of drowsiness or stress [45,46,47]. A smartphone with a mobile application, using the Android operating system, was used to implement a human–computer interaction system. To detect drowsiness, the most important visual indicators that reflect the driver’s condition are the behavior of the eyes, the side and front of the head, and yawning. The system works well under natural light conditions and regardless of the use of accessories supplied by the driver, such as glasses, hearing aids, or a cap.
Due to the large number of road accidents in which drivers fall asleep, this project was implemented to develop methods to prevent napping by providing a non-invasive system that is easy to operate and without the need to purchase additional specialized equipment. This method was able to detect drowsiness with an efficiency of 93.4% [48].
Another significant educational research experiment evaluated how an educational program affected the fatigue and conduct of teenage and adult drivers, as well as their performance and behavior during simulated driving at night. A 4-week sleep program and a 4-week driving program were randomly assigned to 34 volunteers (aged 18 to 26). The findings imply that the educational program increases people’s awareness of sleepiness. Sleep and driving instruction can reduce the risk that young drivers become fatigued and suffer accidents related to fatigue, but this requires a more comprehensive evaluation of their real driving abilities [49].

2.2. Wireless Sensor Networks

Next, we consider a second group of studies related to eliminating typical driver behavior and IoT-based traffic management to increase road safety. IoT is an innovative design paradigm designed as a network of billions to trillions of tiny sensors communicating with each other to offer innovative solutions to problems in real time [50]. These sensors form a network called a wireless sensor network (WSN) to monitor the physical environment and distribute the collected data back to the base station via multiple hops.
WSN has the ability to collect and report data for a specific application. Location information plays an important role in various wireless sensor network applications. Therefore, such systems can improve driving safety. However, real-time monitoring of driving behavior and conditions is linked to various issues, including dizziness caused by long journeys, drastic changes in lighting, and reflections in a driver’s glasses.

2.3. Deep Learning and Driver’s Gaze

A deep learning approach was presented in [51,52,53,54] to address this problem, where the authors used a near-infrared (NIR) camera sensor to detect glances, as well as head and eye movements, without the need for user calibration at first. The proposed system was evaluated on a dedicated database, as well as on Columbia’s open dataset (The Face Tracer CAVE-DB database).
A comprehensive solution was introduced in previous works [51,52,53,54] to address the aforementioned issue by employing deep learning models. This approach used a near-infrared (NIR) camera sensor to accurately identify glances, head movements, and eye movements, all without the need for initial user calibration. The efficacy of the proposed system was assessed on a specialized database and additionally validated using Columbia University’s publicly accessible dataset, known as The Face Tracer CAVE-DB database.
The driver’s gaze turned out to be an excellent way to create a system for driving intelligent vehicles. Due to the fashion for highly autonomous vehicles, the driver’s view can be useful in determining the time of transmission of the gesture from the driver to the traffic management system. Although there have been significant improvements in the personalization of driver vision assessment systems, a universal generalized system that is immutable for different perspectives and scales has not yet been developed. We are taking a step towards this general system using convolutional neural networks (CNNs).
The utilization of the driver’s gaze has emerged as a promising avenue for developing intelligent driving systems. In the context of the rising popularity of highly autonomous vehicles, leveraging the driver’s perspective becomes crucial in accurately timing the transmission of gestures to the traffic management system. Despite notable advances in tailoring driver vision assessment systems to individual users, a universally applicable and adaptable system, capable of accommodating diverse perspectives and scales, remains an open problem. To address this challenge, we are progressing towards the development of a comprehensive framework using convolutional neural networks (CNNs), aiming to establish a generalized solution.
In [55,56] four prominent convolutional neural network (CNN) architectures specifically designed for this purpose were used to conduct detailed comparisons of their performance. Additionally, various modifications were applied to the input images and the influence of the image size on the effectiveness of the models was examined.
To facilitate network training and evaluation, a substantial dataset was collected comprising 11 extended driving activity recordings. This dataset encompassed the driving behaviors of 10 individuals in two distinct vehicles. The most successful models achieved a recognition accuracy of 95.2% during the comparative testing phase.
Subsequently, the highest performing model was subjected to a comparison with the publicly available Columbia Gaze dataset. This dataset consisted of images showing 56 individuals displaying various head positions and viewing directions. Interestingly, even without any specific training on this particular dataset, the model effectively interpreted different perspectives from disparate datasets [57].

3. Materials and Methods

This section presents details on the sensor modalities that were used for data acquisition, discusses the data acquisition process, and explains the experimental settings. Figure 1 shows all the steps in the process from data acquisition to evaluation, which has been extensively described in [40,58,59].

3.1. Data Acquisition

We acquired the dataset using JINS MEME smart glasses, which have a six-axis inertial measurement unit (IMU) that incorporates EOG, an accelerometer and a gyroscope [60,61]. Participants volunteered for the study and gave their informed consent.
The experiments were carried out in a simulated environment [11,62] as presented in Figure 2.
The simulator consists of the following components:
  • A central unit equipped with:
    An Intel Core i7 processor;
    XFX RADEON HD 5770 1 GB graphic card with NVIDIA processor and 3D VISION system;
    4 GB memory;
    Gigabyte’s Ultra Durable 3 motherboard;
  • A special construction made of steel;
  • A two-way adjustable seat;
  • A Logitech set: steering wheel, pedals, and gearbox;
  • Three LED 27 monitors suitable for long operation;
  • A sound system;
  • Dedicated software “Nauka jazdy” (English: Driver training).
The study consisted of two independent experiments that were conducted separately. Both were completed using the JINS MEME ES_R software with the default settings. The EOG sampling rate was 200 Hz, the accelerometer sampling frequency was 100 Hz, and the accelerometer measurement range was  ± 2  g. We synchronized all frequencies to 50 Hz. The signals were recorded simultaneously for each subject while they received voice commands during the driving simulation.
Nine subjects in total (five men and four women) volunteered to participate in the study. Six individuals, all graduate students in their 20s, four men and two women, performed the fundamental driving tasks. In total, we collected 1200 samples of primary driver activities, evenly divided into classes that represent a different activity. Half of the samples were created by one participant, while the remaining samples were evenly distributed among the other subjects.
For the secondary driver activities, we recorded 700 samples that were distributed equally among all classes. Four subjects, one male and three female, with ages ranging from 23 to 57 years, participated. One participant provided 100 samples, while the other participants each contributed 25. None of the subjects had vision problems. One subject participated in both data acquisitions (primary and secondary driver activities). All participants agreed to participate in this study and use the results for research purposes. In total, 2100 samples were collected for this investigation.

3.1.1. Scenarios

The tests consisted of scenarios that serve as good representations of basic and distracting driving behaviors. Primary activity scenarios were chosen as recommended by the local Driving Exam Center (WORD) and were evaluated while the driving test was administered.
As stated in Tables No. 2 and No. 7, Appendix No. 2 of the Regulation of the Minister of Infrastructure of the Republic of Poland [63], these activities include:
  • Passing through uncontrolled intersections (three- and four-way);
  • Passing through intersections marked with signs establishing priority of passage;
  • Drive through intersections with traffic lights;
  • Drive through intersections where traffic flows around a traffic island;
  • Perform one of the following parking maneuvers: perpendicular, angle, and parallel.

3.1.2. Basic Driving Activities

The driving simulator was used to carry out this experiment. To familiarize themselves with the machinery, each participant began with a test ride. Once they felt comfortable, a scenario was given and they were asked to complete the action while wearing JINS MEME ES_R Eyewear. To allow participants to concentrate solely on driving, the supervisor was in charge of managing the computer program and issuing voice commands. Three types of situation were created, each of which was performed in an appropriate setting. There were a total of 12 scenarios in this section. The first set of tasks was carried out in a roundabout. It involved making a left turn, a right turn, or going straight ahead, choosing the first, second, or third exit. The actions are illustrated in Figure 3. The second set of actions was executed at an intersection. The scenarios are similar to the roundabout. The second series of actions was carried out at a crossroad. The situations resemble those of a roundabout and are illustrated in Figure 4. The final set of situations comprises various parking methods, specifically, angle, parallel, and perpendicular parking. Each action was carried out twice, on each side of the street. All scenarios are illustrated in Figure 5.

3.1.3. Distracting Driving Activities

The second investigation focused on secondary or distracting driving activities. They represent all actions that are performed when operating a vehicle that are not related to actual driving. However, they affect performance quality. These actions were carried out in a setting similar to sitting behind a wheel because they do not require being in a vehicle. This section of the study introduced four scenarios: eating, drinking, turning, and bending. Actions are explained in detail in Table 1.

3.1.4. Data Format and Label Information

First, the data acquisition parameters are presented, followed by the header describing the content of each column that contains the sample number, the date in the format: dd.mm.rrrr:hh:mm:ss, and then the 3 channel accelerometer components:
  • ACC X , Y , Z : =  acceleration on the X, Y, and Z axes.
Followed by the EOG sensor components:
  • EOG L , R : =  raw EOG signal from the left and right eye, respectively;
  • EOG H : =  the difference between the left and right eye potential (EOG L − EOG R );
  • EOG V : =  negative arithmetical mean of the left and right eye potential − (EOG L  + EOG R )/2.
A list of dataframes comprising one sample signal is created by successively reading the data from the relevant path by folders. To accurately describe all signals, the rows containing the parameter specifications are removed, and the header is fixed. The labels for the primary activities are presented in Table 2.

3.2. Preprocessing

The data collected by smart glasses include signals from the four EOG channels (EOG L , EOG R , EOG H , and EOG V ), three axes of the accelerometer (ACC X , ACC Y , and ACC Z ) and three axes of the gyroscope (GYRO X , GYRO Y , and GYRO Z ).
The signals collected by these sensors are often contaminated by noise and artifacts. For example, EOG channels can pick up electrical signals from the surrounding environment, which can cause baseline drift and power line noise. Linear and angular acceleration can be affected by vibrations or other disturbances, which can cause errors in measurements. To address these issues, various preprocessing techniques were applied to the data, which involves applying mathematical operations to the signals to remove unwanted components.
The first step of preprocessing was to apply a low-pass filter to remove power line noise (50 or 60 Hz, depending on the country) and baseline wandering. The next step was to use a band-pass filter to remove DC components of the EOG signal caused by electrode polarization.
After preprocessing the data, they can finally be analyzed using the statistical analysis and machine learning technique. Clean data provide valuable associations of changes in the EOG signal in the recognition of human behavior or cognition.
The raw EOG signal presented in Figure 6 contains different types of artifacts that must first be filtered out. To reduce the noise from electricity lines and other potential types of noise, a second-order low-pass Butterworth filter is used to filter the EOG signal. It is applied to the signal twice: once forward and once backward. Such a filter has twice the order of the initial filter and zero phase. In addition, a slow unrelated alteration that is superimposed on the EOG signal, known as a baseline drift, might appear. It could be caused by a variety of things, including electrode polarization or interference with background signals [64]. To eliminate this effect, we have applied detrending by differentiating.
Linear acceleration signal in three axes undergoes preprocessing that consists of applying a median filter and a low-pass filter. The purpose of the median filter is to remove short irregular peaks. Since vigorous voluntary head rotations typically have frequencies below 20 Hz, a low-pass filter is applied to remove components with lower frequencies. However, this type of filter can introduce unwanted distortions while preserving low frequencies. To make the most of both techniques, they are combined by first applying the median filter and then passing the resulting signal through a low-pass filter with a Hamming window. A disadvantage of this approach is the potential weakening of values at the signal edges. However, these values were excluded due to the potential presence of noise caused by human control.
The entire dataset was then independently normalized using Z-score normalization. Z-score normalization helps distinguish the rest values and the values related to activities. The mean and standard deviation of each signal are calculated, and the samples values are replaced with the newly determined values using the following formula:
x = ( x μ ) / σ ,
where  μ  is the mean of the signal,  σ  is the standard deviation, x is the current value of a sample, and x is the new value, so that the new mean of all values is 0 and the standard deviation is 1.
A sliding time window technique was used to segment all normalized sensor signals, with a window length of 5.6 s (280 samples) and a 50% stride (140 samples). Final samples were eliminated if the signal length was not divisible by 140.
Completing some tasks required more time than for the others. Also, depending on the precision of the driver, the acquired signals had different lengths. The shortest one was obtained for secondary activity Turning back and lasted 103 samples (2.06 s); the longest one was for primary activity, while taking a left turn at a roundabout was 3013 samples (60.26 s) long. To train the model, the signal data were resampled at a rate of 3000. The results of signal preprocessing are shown in Figure 7.

3.3. Classification

In this study, we used 1D CNN for feature learning and classification. Multiple convolutional operators in CNN allow automatic recognition of important features from a higher-dimensional input [65,66,67,68]. Convolutions offer the advantage of taking into account the spatial organization of the data. In doing so, additional information about the position in relation to other samples is expected to be taken into account.
The 1D CNN can be used to analyze time series with one or more variables. The latter scenario involves combining two or more concurrent signals. On the basis of our previous experiments, we segmented the data using the sliding window segmentation technique (SWS). Different settings were tested to select the length T and stride size ΔS of a time window and the best values were chosen empirically. In 1D CNN, the only samples with an inherent ordering are those along the time dimension. The channels for the various variables do not have this, in contrast to the most popular 2D CNN.
The basic architecture of a CNN model is shown in Figure 8 and the parameters used in the 1D CNN are shown in Table 3. The first dimension of the input and output data is the batch size, the second dimension is the length of the sequence, and the third dimension is the number of features. The batch size was 32, the number of epochs was 100, and the learning rate (lr) was set to  2 × 10 4 .
In terms of functionality, the model can be divided into two parts. The first component, common for this type of network, acts as a feature extractor. It matches templates using convolutional filtering techniques. To create the so-called “feature maps”, it uses layers and functions that include a convolutional layer, a batch normalization layer, a ReLU activation function, and a pooling layer. The network can learn higher-level features by being trained on a large dataset using a suitable number of epochs and a learning rate.
The second component is the categorization into one of the output classes. The input vector values are first reshaped using the global average pooling layer, a further dropout layer to prevent the model from overfitting, and a dense layer with the “softmax” activation function, which assigns the final label representing the predicted class value by performing a matrix vector multiplication. This process results in a new vector at the output.

3.4. Evaluation

The performance of the classifier was expressed in the form of tables with with the numbers of accuracy, precision, recall, and F1 score and confusion matrix.
Accuracy presents the percentage of correct predictions relative to all predictions made.
Accuracy = t p + t n t p + t n + f p + f n × 100 % ,
where:
-
True positive ( t p )—correctly classified trials;
-
False positive ( f p )—incorrectly classified trials;
-
True negative ( t n )—correctly classified nonevent trials;
-
False negative ( f n )—incorrectly classified nonevent trials.
Precision is a metric that identifies the successful predictions of all predictions made in favor of the event.
Precision = t p t p + f p
Recall presents the fraction of correctly classified predictions of a particular activity with respect to all predictions made in favor of the activity.
Recall = t p t p + f n
The F1 score is a harmonic mean of precision and recall, which, compared to accuracy, should provide a more realistic model assessment in multiclass predictions with unbalanced classes.
F1-score = 2 × Precision × Recall Precision + Recall
Categorical cross-entropy loss measures the model performance by comparing the actual and predicted labels according to the formula:
CE = i = 1 N t i × l o g ( p i ) ,
where t is the true label, p the predicted label, and N the number of scalar values in the model output.
Linear acceleration and EOG signals that had already been analyzed were used to train and assess the network. A 9:1 ratio was used to divide the data into subsets for training and testing. A further division with 8:2 ratio was used on the training set to divide it into training and validation sets. Since the signals were sorted, the data had to be shuffled to train the model on signals from all possible classes.

4. Results

Accuracy and Loss While Training

Figure 9 shows the accuracy curve for training and validation, and loss of the model with respect to the number of epochs elapsed. The loss function is categorical cross-entropy. When the epoch reached 130, the training accuracy was found to be greater than 90%, providing a loss value of 0.2. The validation rate was 80% with a loss of 0.6. The model obtained the optimal parameters in 188 epochs.
Figure 10 shows how well the classes were segregated after 188 epochs in dimensions 2 and 3. A dimensionality reduction method known as principal component analysis (PCA) was used for visualization purposes. The correlation between different dimensions is used, and the goal is to provide as few variables as possible while preserving as much variation or information about the distribution of the original data as possible.
It can be seen that the distinction between primary and secondary driving activities is very apparent. The latter are also separated in such a way that they usually do not overlap. However, primary activities cover areas very close to each other, so the greatest misclassifications are anticipated.
The distinction between primary and secondary driving activities is readily apparent, as can be seen. The latter are divided in such a way that they do not primarily overlap. The greatest misclassifications were expected for groups of similar primary activities that covered areas that are relatively close to each other.
All activity predictions shown in Table 4 had a weighted average precision, recall, and F1 score of 0.83, 0.80, and 0.80, respectively. Drinking as a secondary activity is the category that has the best performance with all values equal to 1, while primary parallel parking on the left and perpendicular parking on the right are the categories that are mostly misclassified with F1 scores 0.42 and 0.44, respectively. In general, secondary driving actions performed better, all receiving F1 scores greater than 0.9.
The accuracy of the prediction of 15 driving actions is shown in Figure 11, where the accuracy of the prediction of each class is shown on the diagonal and inaccurate classifications are shown outside the diagonal.
Most misclassifications occurred in a group of activities that were related to each other. Parking activities show that the model had the most trouble detecting the difference between the same action being conducted on the left and right sides due to their similarity. The binary classification between secondary and primary activities has an accuracy rate of 99.5%. Although the latter was mistakenly classified as eating, the former was consistently assigned to the appropriate group.
An investigation of classifications for a particular collection of activities, including crossroad, parking, roundabout, and secondary activities, yielded accuracy ratings of 97.9%, 96.8%, 97.4%, and 99.5%, respectively. Table 5 displays these results together with the precision, recall, and F1 score values. Although the network received these actions as individual activities, it was still able to indicate patterns that differentiate the different types of action.

5. Discussion

Electrooculography (EOG) is a technique that is based on electrical features generated by the eye. By measuring the voltage difference between the cornea and the retina, it aims to capture the movements of the eyes [37]. JINS MEME ES_R Glasses (JINS Inc., Tokyo, Japan) are a smart glasses device that consists of a three-point electrooculography (EOG) and a six-axis inertial measurement unit (IMU) with a gyroscope and an accelerometer. They acquire ten channels: linear and angular acceleration on the X, Y, and Z axes, and four EOG channels: electric potentials on the electrodes on the right (EOG R ) and left (EOG L ), and the vertical (EOG V ) and horizontal (EOG H ) difference between them [61,69,70].
We have recognized road conditions based on electrooculograms acquired from drivers wearing JINS MEME ES_R smart glasses. The highest precision, recall, and F1 score for drinking (1.00 for each metric) were observed, whereas the lowest results were observed for parallel parking on the left side (precision of 0.44, recall of 0.4, and F1 score of 0.42).
Most misclassifications occurred in a group of activities that were related to each other, e.g., parking on the left and parking on the right side due to their similarity. The binary classification between secondary and primary activities has an accuracy rate of 99.5%. Although the latter was mistakenly classified as eating, the former was consistently assigned to the appropriate group.
In this study, the recognition of primary and secondary driver activities based on the processing of EOG signals with a convolutional network achieved excellent recognition performance, but there are still some limitations. The first limitation was to obtain the EOG signals in a simulated driving experiment. Although the experimental results showed that the turn or park condition was successfully induced and verified the effectiveness of the experimental scheme, it cannot be compared with the complexity of driving in real traffic. The second limitation was the limited number of experimental data segments. This setup could be used in future studies that do not expose volunteers to the dangers of real traffic. Classification models were trained on short signal samples. The third limitation was the use of only one time window width (5.6 s) to calculate the EOG characteristics without fully examining the impact of other time window divisions on the classification results.

6. Conclusions

In this paper, we introduced a CNN-based machine learning model to classify nine primary and four secondary car driver activities using physiological sensor data from JINS MEME ES_R (smart glasses):
  • We conducted a comparative analysis of wearable sensor data, including nine activities performed by the first driver and four activities performed by the second driver. Our proposed system achieves an impressive overall accuracy of 97% (±2) and an average F1 score of 95% (±2) in detecting these activities. Moreover, our model has the potential to prevent traffic accidents without requiring expensive safety equipment. To further validate our approach, future studies will involve acquiring additional data from real-world road conditions. Such an application would be beneficial for drivers, particularly older individuals or those with disabilities.
  • Research involves a comparative analysis of wearable sensor data obtained from different driving activities in various scenarios. By analyzing the signals collected from these sensors, researchers can assess the different activities performed by drivers, gaining insight into driver behavior and activity patterns in various driving scenarios. The developed system holds promise in preventing traffic accidents without the need for costly safety equipment.
  • Our investigation focuses on the utilization of sensor data from JINS MEME ES_R smart glasses and the development of an advanced machine learning model that can identify and classify primary and secondary activities of drivers. This state-of-the-art model learns the patterns associated with these activities and assigns them to their respective classes.
  • The use of JINS MEME ES_R sensor data involves analyzing and recognizing activities and scenarios associated with driving a vehicle. By integrating data from these wearable glasses, we created an efficient machine learning model that can learn activity patterns and accurately classify them into respective classes. This novel approach to the use of wearable sensor data offers valuable insights into driver behavior and activity recognition.
  • A notable contribution of this research is the realistic and noninvasive collection of data. The use of JINS MEME ES_R smart glasses provides a user-friendly and noninvasive method for gathering data during experiments. Unlike intrusive methods, these glasses capture data from the driver’s perspective without causing discomfort or interfering with the driving experience. This noninvasive approach ensures that the collected data closely resemble real-world driving scenarios, allowing for more precise analysis and classification of driver activities. By addressing the challenge of obtaining realistic data while prioritizing participant safety and comfort, this research underscores the importance of using such technology.
In summary, the main contributions of this research involve the utilization of JINS MEME ES_R sensor data, development of a machine learning model for activity recognition, comparative analysis of wearable sensor data, and a review of related approaches. These contributions improve understanding of driver behavior and activity recognition, potentially leading to improved driver safety and accident prevention when the time comes when autonomous car traffic with the participation of human drivers will become commonplace on the roads.

Author Contributions

Conceptualization, R.D. and J.K.; methodology, R.D.; software, J.K. and R.D.; validation, R.D. and J.K.; formal analysis, R.D., J.K. and M.T.I.; investigation, R.D., J.K. and M.T.I.; resources, R.D.; data curation, J.K. and R.D.; writing—original draft preparation, R.D. and J.K.; writing—review and editing, R.D., S.S., M.T.I., A.P. and M.G.; visualization, R.D., J.K., N.P., M.T.I., M.A.H. and A.P.; supervision, R.D., S.S., F.L., M.A.N. and M.G.; project administration, R.D. and M.G.; funding acquisition, R.D. and M.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received funding for car simulator from the pro-quality grant 57/2021 of the Rector of the Silesian University of Technology, Gliwice, Poland; decision number ZP/165980.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and the protocol was approved by the Bioethics Committee of the Medical University of Silesia on 16 October 2018 (KNW/0022/KB1/79/18).

Informed Consent Statement

All subjects gave their informed consent for inclusion before they participated in the study. The authors thank all participants in the experiment.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy concerns.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kinnear, N.; Stevens, A. The Battle for Attention: Driver Distraction—A Review of Recent Research and Knowledge; Technical Report; IAM: London, UK, 2015. [Google Scholar]
  2. Doniec, R.; Sieciński, S.; Piaseczna, N.; Mocny-Pachońska, K.; Lang, M.; Szymczyk, J. The Classifier Algorithm for Recognition of Basic Driving Scenarios. In Advances in Intelligent Systems and Computing; Springer International Publishing: Cham, Switzerland, 2020; pp. 359–367. [Google Scholar] [CrossRef]
  3. Doniec, R.J.; Sieciński, S.; Duraj, K.M.; Piaseczna, N.J.; Mocny-Pachońska, K.; Tkacz, E.J. Recognition of Drivers’ Activity Based on 1D Convolutional Neural Network. Electronics 2020, 9, 2002. [Google Scholar] [CrossRef]
  4. Al-Hussein, W.A.; Kiah, M.L.M.; Yee, P.L.; Zaidan, B.B. A systematic review on sensor-based driver behaviour studies: Coherent taxonomy, motivations, challenges, recommendations, substantial analysis and future directions. PeerJ Comput. Sci. 2021, 7, e632. [Google Scholar] [CrossRef] [PubMed]
  5. Ping, P.; Qin, W.; Xu, Y.; Miyajima, C.; Kazuya, T. Spectral clustering based approach for evaluating the effect of driving behavior on fuel economy. In Proceedings of the 2018 IEEE International Instrumentation and Measurement Technology Conference (I2MTC), Houston, TX, USA, 14–17 May 2018; pp. 1–6. [Google Scholar] [CrossRef]
  6. Jamroz, K.; Smolarek, L. Driver Fatigue and Road Safety on Poland’s National Roads. Int. J. Occup. Saf. Ergon. 2013, 19, 297–309. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Kurczynski, D.; Zuska, A. Analysis of the Impact of Invisible Road Icing on Selected Parameters of a Minibus Vehicle. Sensors 2022, 22, 9726. [Google Scholar] [CrossRef]
  8. National Center for Statistics and Analysis. Distracted Driving 2020; Technical Report; National Highway Traffic Safety Administration: Washington, DC, USA, 2022.
  9. Arevalo-Tamara, A.; Caicedo, A.; Orozco-Fontalvo, M.; Useche, S.A. Distracted driving in relation to risky road behaviors and traffic crashes in Bogota, Colombia. Saf. Sci. 2022, 153, 105803. [Google Scholar] [CrossRef]
  10. Brown, I.D. Driver Fatigue. Hum. Factors 1994, 36, 298–314. [Google Scholar] [CrossRef]
  11. Karthaus, M.; Wascher, E.; Getzmann, S. Distraction in the Driving Simulator: An Event-Related Potential (ERP) Study with Young, Middle-Aged, and Older Drivers. Safety 2021, 7, 36. [Google Scholar] [CrossRef]
  12. Moslemi, N.; Soryani, M.; Azmi, R. Computer vision-based recognition of driver distraction: A review. Concurr. Comput. Pract. Exp. 2021, 33, e6475. [Google Scholar] [CrossRef]
  13. Kashevnik, A.; Shchedrin, R.; Kaiser, C.; Stocker, A. Driver Distraction Detection Methods: A Literature Review and Framework. IEEE Access 2021, 9, 60063–60076. [Google Scholar] [CrossRef]
  14. Yan, Q.; Zhang, W.; Hu, W.; Cui, G.; Wei, D.; Xu, J. Gaze dynamics with spatiotemporal guided feature descriptor for prediction of driver’s maneuver behavior. Proc. Inst. Mech. Eng. Part D J. Automob. Eng. 2021, 235, 3051–3065. [Google Scholar] [CrossRef]
  15. Uma, S.; Eswari, R. Accident prevention and safety assistance using IOT and machine learning. J. Reliab. Intell. Environ. 2022, 8, 79–103. [Google Scholar] [CrossRef]
  16. Hozhabr Pour, H.; Li, F.; Wegmeth, L.; Trense, C.; Doniec, R.; Grzegorzek, M.; Wismuller, R. A Machine Learning Framework for Automated Accident Detection Based on Multimodal Sensors in Cars. Sensors 2022, 22, 3634. [Google Scholar] [CrossRef] [PubMed]
  17. Sun, W.; Si, Y.; Guo, M.; Li, S. Driver Distraction Recognition Using Wearable IMU Sensor Data. Sustainability 2021, 13, 1342. [Google Scholar] [CrossRef]
  18. Rizzo, M.; Hurtig, R. Looking but not seeing. Neurology 1987, 37, 1642. [Google Scholar] [CrossRef] [PubMed]
  19. Eraqi, H.M.; Abouelnaga, Y.; Saad, M.H.; Moustafa, M.N. Driver Distraction Identification with an Ensemble of Convolutional Neural Networks. J. Adv. Transp. 2019, 2019, 4125865. [Google Scholar] [CrossRef]
  20. Ariansyah, D.; Pardamean, B.; Caruso, G. The effect of visual advanced driver assistance systems on a following human driver in a mixed-traffic condition. Procedia Comput. Sci. 2023, 216, 221–229. [Google Scholar] [CrossRef]
  21. Böddeker, B.; von Wendorff, W.; Nguyen, N.; Diehl, P.; Meertens, R.; Johannson, R. Automated driving safety—The art of conscious risk taking—Minimum lateral distances to pedestrians. In Proceedings of the 2021 Design, Automation & Test in Europe Conference & Exhibition (DATE), Grenoble, France, 1–5 February 2021; pp. 1466–1471. [Google Scholar] [CrossRef]
  22. Lin, P.W.; Hsu, C.M. Innovative Framework for Distracted-Driving Alert System Based on Deep Learning. IEEE Access 2022, 10, 77523–77536. [Google Scholar] [CrossRef]
  23. Klauer, S.G.; Dingus, T.A.; Neale, V.L.; Sudweeks, J.; Ramsey, D.J. The Impact of Driver Inattention on Near-Crash/Crash Risk: An Analysis Using the 100-Car Naturalistic Driving Study Data; Technical Report; National Traffic Safety Administration, US Department of Transportation: Washington, DC, USA, 2006. [Google Scholar]
  24. Jomnonkwao, S.; Uttra, S.; Ratanavaraha, V. Analysis of a driving behavior measurement model using a modified driver behavior questionnaire encompassing texting, social media use, and drug and alcohol consumption. Transp. Res. Interdiscip. Perspect. 2021, 9, 100302. [Google Scholar] [CrossRef]
  25. Klauer, S.G.; Guo, F.; Simons-Morton, B.G.; Ouimet, M.C.; Lee, S.E.; Dingus, T.A. Distracted Driving and Risk of Road Crashes among Novice and Experienced Drivers. N. Engl. J. Med. 2014, 370, 54–59. [Google Scholar] [CrossRef] [Green Version]
  26. Goodman, M.J.; Tijerina, L.; Bents, F.D.; Wierwille, W.W. Using Cellular Telephones in Vehicles: Safe or Unsafe? Transp. Hum. Factors 1999, 1, 3–42. [Google Scholar] [CrossRef]
  27. Ishimaru, S.; Hoshika, K.; Kunze, K.; Kise, K.; Dengel, A. Towards reading trackers in the wild: Detecting reading activities by EOG glasses and deep neural networks. In Proceedings of the 2017 ACM International Joint Conference on Pervasive and Ubiquitous Computing and Proceedings of the 2017 ACM International Symposium on Wearable Computers, Maui, HI, USA, 11–15 September 2017; pp. 704–711. [Google Scholar] [CrossRef]
  28. Nisar, M.A.; Shirahama, K.; Li, F.; Huang, X.; Grzegorzek, M. Rank Pooling Approach for Wearable Sensor-Based ADLs Recognition. Sensors 2020, 20, 3463. [Google Scholar] [CrossRef] [PubMed]
  29. Uddin, M.Z.; Soylu, A. Human activity recognition using wearable sensors, discriminant analysis, and long short-term memory-based neural structured learning. Sci. Rep. 2021, 11, 16455. [Google Scholar] [CrossRef]
  30. Yazdansepas, D.; Niazi, A.H.; Gay, J.L.; Maier, F.W.; Ramaswamy, L.; Rasheed, K.; Buman, M.P. A Multi-featured Approach for Wearable Sensor-Based Human Activity Recognition. In Proceedings of the 2016 IEEE International Conference on Healthcare Informatics (ICHI), Chicago, IL, USA, 4–7 October 2016; pp. 423–431. [Google Scholar] [CrossRef]
  31. Bao, L.; Intille, S.S. Activity Recognition from User-Annotated Acceleration Data. In Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2004; pp. 1–17. [Google Scholar] [CrossRef]
  32. Chen, L.; Hoey, J.; Nugent, C.D.; Cook, D.J.; Yu, Z. Sensor-Based ActivityRecognition. IEEE Trans. Syst. Man Cybern. Part C Appl. Rev. 2012, 42, 790–808. [Google Scholar] [CrossRef]
  33. D’Souza, S.; Natarajan, S. Recognition of EOG based reading task using AR features. In Proceedings of the International Conference on Circuits, Communication, Control and Computing, Bangalore, India, 21–22 November 2014; pp. 113–117. [Google Scholar] [CrossRef]
  34. Huang, X.; Shirahama, K.; Irshad, M.T.; Nisar, M.A.; Piet, A.; Grzegorzek, M. Sleep Stage Classification in Children Using Self-Attention and Gaussian Noise Data Augmentation. Sensors 2023, 23, 3446. [Google Scholar] [CrossRef]
  35. Huang, X.; Shirahama, K.; Li, F.; Grzegorzek, M. Sleep stage classification for child patients using DeConvolutional Neural Network. Artif. Intell. Med. 2020, 110, 101981. [Google Scholar] [CrossRef]
  36. Doniec, R.; Piaseczna, N.; Li, F.; Duraj, K.; Hozhabr Pour, H.; Grzegorzek, M.; Mocny-Pachońska, K.; Tkacz, E. Classification of Roads and Types of Public Roads Using EOG Smart Glasses and an Algorithm Based on Machine Learning While Driving a Car. Electronics 2022, 11, 2960. [Google Scholar] [CrossRef]
  37. Bulling, A.; Roggen, D.; Tröster, G. Wearable EOG goggles: Seamless sensing and context-awareness in everyday environments. J. Ambient. Intell. Smart Environ. 2009, 1, 157–171. [Google Scholar] [CrossRef] [Green Version]
  38. Gruenewald, A.; Kroenert, D.; Poehler, J.; Brueck, R.; Li, F.; Littau, J.; Schnieber, K.; Piet, A.; Grzegorzek, M.; Kampling, H.; et al. Biomedical Data Acquisition and Processing to Recognize Emotions for Affective Learning. In Proceedings of the 2018 IEEE 18th International Conference on Bioinformatics and Bioengineering (BIBE), Taichung, Taiwan, 29–31 October 2018. [Google Scholar] [CrossRef]
  39. Lin, C.T.; Jiang, W.L.; Chen, S.F.; Huang, K.C.; Liao, L.D. Design of a Wearable Eye-Movement Detection System Based on Electrooculography Signals and Its Experimental Validation. Biosensors 2021, 11, 343. [Google Scholar] [CrossRef]
  40. Irshad, M.T.; Nisar, M.A.; Huang, X.; Hartz, J.; Flak, O.; Li, F.; Gouverneur, P.; Piet, A.; Oltmanns, K.M.; Grzegorzek, M. SenseHunger: Machine Learning Approach to Hunger Detection Using Wearable Sensors. Sensors 2022, 22, 7711. [Google Scholar] [CrossRef]
  41. Aksjonov, A.; Nedoma, P.; Vodovozov, V.; Petlenkov, E.; Herrmann, M. Detection and Evaluation of Driver Distraction Using Machine Learning and Fuzzy Logic. IEEE Trans. Intell. Transp. Syst. 2019, 20, 2048–2059. [Google Scholar] [CrossRef]
  42. Alotaibi, M.; Alotaibi, B. Distracted driver classification using deep learning. Signal Image Video Process. 2020, 14, 617–624. [Google Scholar] [CrossRef]
  43. Wang, J.; Xu, Y.; Tian, J.; Li, H.; Jiao, W.; Sun, Y.; Li, G. Driving Fatigue Detection with Three Non-Hair-Bearing EEG Channels and Modified Transformer Model. Entropy 2022, 24, 1715. [Google Scholar] [CrossRef] [PubMed]
  44. Braunagel, C.; Geisler, D.; Rosenstiel, W.; Kasneci, E. Online recognition of driver-activity based on visual scanpath classification. IEEE Intell. Transp. Syst. Mag. 2017, 9, 23–36. [Google Scholar] [CrossRef]
  45. Pansare, D.J.; Deshpande, R.; Shingare, S.; Deokar, H.; Manwar, P. Real-time Driver Drowsiness Detection with Android. Int. J. Res. Appl. Sci. Eng. Technol. 2022, 10, 441–445. [Google Scholar] [CrossRef]
  46. Mateos-García, N.; Gil-González, A.B.; Reboredo, A.d.L.; Pérez-Lancho, B. Driver Stress Detection in Simulated Driving Scenarios with Photoplethysmography. In Proceedings of the Distributed Computing and Artificial Intelligence, 19th International Conference, L’Aquila, Italy, 13–15 July 2022; Omatu, S., Mehmood, R., Sitek, P., Cicerone, S., Rodríguez, S., Eds.; Springer International Publishing: Cham, Switzerland, 2023; pp. 291–301. [Google Scholar]
  47. Jahan, I.; Uddin, K.M.A.; Murad, S.A.; Miah, M.S.U.; Khan, T.Z.; Masud, M.; Aljahdali, S.; Bairagi, A.K. 4D: A Real-Time Driver Drowsiness Detector Using Deep Learning. Electronics 2023, 12, 235. [Google Scholar] [CrossRef]
  48. Galarza, E.E.; Egas, F.D.; Silva, F.M.; Velasco, P.M.; Galarza, E.D. Real Time Driver Drowsiness Detection Based on Driver’s Face Image Behavior Using a System of Human Computer Interaction Implemented in a Smartphone. In Proceedings of the International Conference on Information Technology & Systems (ICITS 2018), Libertad City, Ecuador, 10–12 January 2018; Springer International Publishing: Berlin/Heidelberg, Germany, 2018; pp. 563–572. [Google Scholar] [CrossRef]
  49. Alvaro, P.K.; Burnett, N.M.; Kennedy, G.A.; Min, W.Y.X.; McMahon, M.; Barnes, M.; Jackson, M.; Howard, M.E. Driver education: Enhancing knowledge of sleep, fatigue and risky behaviour to improve decision making in young drivers. Accid. Anal. Prev. 2018, 112, 77–83. [Google Scholar] [CrossRef]
  50. AlZubi, A.A.; Abugabah, A.; Al-Maitah, M.; Ibrahim AlZobi, F. DL Multi-sensor information fusion service selective information scheme for improving the Internet of Things based user responses. Measurement 2021, 185, 110008. [Google Scholar] [CrossRef]
  51. Naqvi, R.A.; Arsalan, M.; Batchuluun, G.; Yoon, H.S.; Park, K.R. Deep learning-based gaze detection system for automobile drivers using a NIR camera sensor. Sensors 2018, 18, 456. [Google Scholar] [CrossRef] [Green Version]
  52. Zhang, Y.; Chen, Y.; Gao, C. Deep unsupervised multi-modal fusion network for detecting driver distraction. Neurocomputing 2021, 421, 26–38. [Google Scholar] [CrossRef]
  53. Vaegae, N.K.; Pulluri, K.K.; Bagadi, K.; Oyerinde, O.O. Design of an Efficient Distracted Driver Detection System: Deep Learning Approaches. IEEE Access 2022, 10, 116087–116097. [Google Scholar] [CrossRef]
  54. Cheng, S.; Yang, B.; Wang, Z.; Nakano, K. Spatio-Temporal Image Representation and Deep-Learning-Based Decision Framework for Automated Vehicles. IEEE Trans. Intell. Transp. Syst. 2022, 23, 24866–24875. [Google Scholar] [CrossRef]
  55. Ping, P.; Huang, C.; Ding, W.; Liu, Y.; Chiyomi, M.; Kazuya, T. Distracted driving detection based on the fusion of deep learning and causal reasoning. Inf. Fusion 2023, 89, 121–142. [Google Scholar] [CrossRef]
  56. Xing, Y.; Lv, C.; Wang, H.; Cao, D.; Velenis, E.; Wang, F.Y. Driver Activity Recognition for Intelligent Vehicles: A Deep Learning Approach. IEEE Trans. Veh. Technol. 2019, 68, 5379–5390. [Google Scholar] [CrossRef] [Green Version]
  57. Vora, S.; Rangesh, A.; Trivedi, M.M. Driver Gaze Zone Estimation Using Convolutional Neural Networks: A General Framework and Ablative Analysis. arXiv 2018, arXiv:180202690 Cs. [Google Scholar] [CrossRef]
  58. Li, F.; Shirahama, K.; Nisar, M.; Köping, L.; Grzegorzek, M. Comparison of Feature Learning Methods for Human Activity Recognition Using Wearable Sensors. Sensors 2018, 18, 679. [Google Scholar] [CrossRef] [Green Version]
  59. Irshad, M.T.; Nisar, M.A.; Gouverneur, P.; Rapp, M.; Grzegorzek, M. AI Approaches towards Prechtl’s Assessment of General Movements: A Systematic Literature Review. Sensors 2020, 20, 5321. [Google Scholar] [CrossRef]
  60. Banerjee, S.; Khadem, N.K.; Kabir, M.M.; Jeihani, M. Driver Behavior Post Cannabis Consumption: A Driving Simulator Study in Collaboration with Montgomery County Maryland. arXiv 2021, arXiv:2112.12026. [Google Scholar]
  61. JINS, Inc. JINS MEME Glasses Specifications. Available online: https://jins-meme.github.io/apdoc/en/ (accessed on 9 June 2023).
  62. Sun, Q.; Guo, Y.; Liu, Y.; Wang, C.; Gu, M.; Su, Y. Comparing the Effects of Visual Distraction in a High-Fidelity Driving Simulator and on a Real Highway. IEEE Trans. Intell. Transp. Syst. 2021, 24, 3738–3747. [Google Scholar] [CrossRef]
  63. Ministry of Infrastructure of the Republic of Poland. Rozporządzenie Ministra Infrastruktury z Dnia 28 Czerwca 2019 r. w Sprawie Egzaminowania Osób Ubiegających Się o Uprawnienia do Kierowania Pojazdami, Szkolenia, Egzaminowania i Uzyskiwania Uprawnień przez Egzaminatorów Oraz Wzorów Dokumentów Stosowanych w Tych Sprawacz, Dz.U. 2019 poz. 1206. 2019. Available online: https://www.dziennikustaw.gov.pl/D2019000120601.pdf (accessed on 9 June 2023).
  64. Gu, J.J.; Meng, M.Q.H.; Cook, A.; Faulkner, M.G. A study of natural eye movement detection and ocular implant movement control using processed EOG signals. In Proceedings of the 2001 ICRA. IEEE International Conference on Robotics and Automation (Cat. No.01CH37164), Seoul, Korea, 21–26 May 2001; Volume 2, pp. 1555–1560. [Google Scholar]
  65. Zhao, Z.; Zhou, N.; Zhang, L.; Yan, H.; Xu, Y.; Zhang, Z. Driver Fatigue Detection Based on Convolutional Neural Networks Using EM-CNN. Comput. Intell. Neurosci. 2020, 2020, 7251280. [Google Scholar] [CrossRef]
  66. Yan, C.; Coenen, F.; Zhang, B. Driving posture recognition by convolutional neural networks. IET Comput. Vis. 2016, 10, 103–114. [Google Scholar] [CrossRef]
  67. Shahverdy, M.; Fathy, M.; Berangi, R.; Sabokrou, M. Driver behavior detection and classification using deep convolutional neural networks. Expert Syst. Appl. 2020, 149, 113240. [Google Scholar] [CrossRef]
  68. Deng, T.; Yan, H.; Qin, L.; Ngo, T.; Manjunath, B.S. How Do Drivers Allocate Their Potential Attention? Driving Fixation Prediction via Convolutional Neural Networks. IEEE Trans. Intell. Transp. Syst. 2020, 21, 2146–2154. [Google Scholar] [CrossRef]
  69. Dhuliawala, M.; Lee, J.; Shimizu, J.; Bulling, A.; Kunze, K.; Starner, T.; Woo, W. Smooth eye movement interaction using EOG glasses. In Proceedings of the ICMI ’16: Proceedings of the 18th ACM International Conference on Multimodal Interaction, Tokyo, Japan, 12–16 November 2016; pp. 307–311. [Google Scholar] [CrossRef]
  70. Díaz, D.; Yee, N.; Daum, C.; Stroulia, E.; Liu, L. Activity Classification in Independent Living Environment with JINS MEME Eyewear. In Proceedings of the 2018 IEEE International Conference on Pervasive Computing and Communications (PerCom), Athens, Greece, 19–23 March 2018; pp. 1–9. [Google Scholar] [CrossRef]
Figure 1. Standard approach to developing a deep learning model. Each step in the chain should be optimized in parallel to achieve the best possible performance.
Figure 1. Standard approach to developing a deep learning model. Each step in the chain should be optimized in parallel to achieve the best possible performance.
Sensors 23 05551 g001
Figure 2. Driving simulator setup used for the data acquisition.
Figure 2. Driving simulator setup used for the data acquisition.
Sensors 23 05551 g002
Figure 3. Roundabout scenarios such as ‘right’, ‘straight’, and ‘left’, (ac), respectively.
Figure 3. Roundabout scenarios such as ‘right’, ‘straight’, and ‘left’, (ac), respectively.
Sensors 23 05551 g003
Figure 4. Crossroad scenarios such as ‘right’, ‘straight’, and ‘left’, (ac), respectively.
Figure 4. Crossroad scenarios such as ‘right’, ‘straight’, and ‘left’, (ac), respectively.
Sensors 23 05551 g004
Figure 5. Parking scenarios such as ‘angle left’, ‘parallel left’, ‘perpendicular left’, ‘angle right’, ‘parallel right’, and ‘perpendicular right’, (af), respectively.
Figure 5. Parking scenarios such as ‘angle left’, ‘parallel left’, ‘perpendicular left’, ‘angle right’, ‘parallel right’, and ‘perpendicular right’, (af), respectively.
Sensors 23 05551 g005
Figure 6. Raw EOG signal in channels EOG L , EOG R , EOG H , and EOG V  shown in (ad), respectively.
Figure 6. Raw EOG signal in channels EOG L , EOG R , EOG H , and EOG V  shown in (ad), respectively.
Sensors 23 05551 g006
Figure 7. Preprocessed EOG signal in channels EOG L , EOG R , EOG H , and EOG V  shown in (ad), respectively.
Figure 7. Preprocessed EOG signal in channels EOG L , EOG R , EOG H , and EOG V  shown in (ad), respectively.
Sensors 23 05551 g007
Figure 8. General architecture of a convolutional neural network for time series data classification. T represents time window, S represents the number of sensor channels, c represents a class, and n represents the number of classes.
Figure 8. General architecture of a convolutional neural network for time series data classification. T represents time window, S represents the number of sensor channels, c represents a class, and n represents the number of classes.
Sensors 23 05551 g008
Figure 9. Accuracy and loss curve of the CNN model during the training and validation phases.
Figure 9. Accuracy and loss curve of the CNN model during the training and validation phases.
Sensors 23 05551 g009
Figure 10. Scatter plot of the training and testing sets according to predicted labels: (a) 2D scatter plot of training set labels; (b) 3D scatter plot of training set labels; (c) 2D scatter plot of testing set labels; (d) 3D scatter plot of testing set labels.
Figure 10. Scatter plot of the training and testing sets according to predicted labels: (a) 2D scatter plot of training set labels; (b) 3D scatter plot of training set labels; (c) 2D scatter plot of testing set labels; (d) 3D scatter plot of testing set labels.
Sensors 23 05551 g010
Figure 11. Confusion matrix of the driving activities classification.
Figure 11. Confusion matrix of the driving activities classification.
Sensors 23 05551 g011
Table 1. Secondary driving activities scenarios.
Table 1. Secondary driving activities scenarios.
EatingDrinkingTurningBending
Taking a bite of any food.Taking a sip of water.Turning back and reaching to a passenger’s seat.Bending and picking up a fallen object.
Table 2. List of activities for which data were collected in this study (driving activities), using JINS MEME ES_R Smart Glasses.
Table 2. List of activities for which data were collected in this study (driving activities), using JINS MEME ES_R Smart Glasses.
LabelActivity
0P_Crossroad_Left
1P_Crossroad_Right
2P_Crossroad_Straight
3P_Parking_Diagonal_Left
4P_Parking_Diagonal_Right
5P_Parking_Parallel_Left
6P_Parking_Parallel_Right
7P_Parking_Perpendicular_Left
8P_Parking_Perpendicular_Right
9P_Roundabout_Left
10P_Roundabout_Right
11P_Roundabout_Straight
12S_Bending
13S_Drinking
14S_Eating
15S_Turning_Back
Table 3. CNN architecture with a fixed dropout rate of 0.4 and a minimum learning rate of  2 × 10 4 .
Table 3. CNN architecture with a fixed dropout rate of 0.4 and a minimum learning rate of  2 × 10 4 .
Layer NameNo. Kernels (Units)Kernel (Pool) SizeStride SizeActivation
Convolution12851ReLU
Batch norm----
Max pooling-3--
Convolution12851ReLU
Batch norm----
Convolution12851ReLU
Batch norm----
Global avg. pooling----
Dense2--Softmax
Table 4. Evaluation results (such as precision, recall, and F1 score) of driving activities.
Table 4. Evaluation results (such as precision, recall, and F1 score) of driving activities.
LabelActivityPrecisionRecallF1 Score
0P_Crossroad_Left1.000.670.80
1P_Crossroad_Right0.750.750.75
2P_Crossroad_Straight0.750.900.82
3P_Parking_Diagonal_Left0.670.750.71
4P_Parking_Diagonal_Right0.500.500.50
5P_Parking_Parallel_Left0.440.400.42
6P_Parking_Parallel_Right0.860.600.71
7P_Parking_Perpendicular_Left0.381.000.55
8P_Parking_Perpendicular_Right1.000.290.44
9P_Roundabout_Left1.000.880.93
10P_Roundabout_Right0.820.860.84
11P_Roundabout_Straight0.860.750.80
12S_Bending0.950.900.93
13S_Drinking1.001.001.00
14S_Eating0.900.950.93
15S_Turning_Back0.880.940.91
Table 5. Evaluation scores such as accuracy, precision, recall, and F1 scores, of the type-based classification.
Table 5. Evaluation scores such as accuracy, precision, recall, and F1 scores, of the type-based classification.
ActivityAccuracy [%]PrecisionRecallF1 Score
Crossroad97.90.970.910.94
Parking96.80.920.960.94
Roundabout97.40.940.920.93
Secondary99.50.991.000.99
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Doniec, R.; Konior, J.; Sieciński, S.; Piet, A.; Irshad, M.T.; Piaseczna, N.; Hasan, M.A.; Li, F.; Nisar, M.A.; Grzegorzek, M. Sensor-Based Classification of Primary and Secondary Car Driver Activities Using Convolutional Neural Networks. Sensors 2023, 23, 5551. https://doi.org/10.3390/s23125551

AMA Style

Doniec R, Konior J, Sieciński S, Piet A, Irshad MT, Piaseczna N, Hasan MA, Li F, Nisar MA, Grzegorzek M. Sensor-Based Classification of Primary and Secondary Car Driver Activities Using Convolutional Neural Networks. Sensors. 2023; 23(12):5551. https://doi.org/10.3390/s23125551

Chicago/Turabian Style

Doniec, Rafał, Justyna Konior, Szymon Sieciński, Artur Piet, Muhammad Tausif Irshad, Natalia Piaseczna, Md Abid Hasan, Frédéric Li, Muhammad Adeel Nisar, and Marcin Grzegorzek. 2023. "Sensor-Based Classification of Primary and Secondary Car Driver Activities Using Convolutional Neural Networks" Sensors 23, no. 12: 5551. https://doi.org/10.3390/s23125551

APA Style

Doniec, R., Konior, J., Sieciński, S., Piet, A., Irshad, M. T., Piaseczna, N., Hasan, M. A., Li, F., Nisar, M. A., & Grzegorzek, M. (2023). Sensor-Based Classification of Primary and Secondary Car Driver Activities Using Convolutional Neural Networks. Sensors, 23(12), 5551. https://doi.org/10.3390/s23125551

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop