Next Article in Journal
Timing and Performance Metrics for TWR-K70F120M Device
Next Article in Special Issue
Evaluating User Satisfaction Using Deep-Learning-Based Sentiment Analysis for Social Media Data in Saudi Arabia’s Telecommunication Sector
Previous Article in Journal
Stochastic Modeling for Intelligent Software-Defined Vehicular Networks: A Survey
Previous Article in Special Issue
Convolutional Neural Networks: A Survey
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Face Detection Using a Capsule Network for Driver Monitoring Application

Central Campus Győr, Széchenyi István University, 9026 Győr, Hungary
*
Authors to whom correspondence should be addressed.
Computers 2023, 12(8), 161; https://doi.org/10.3390/computers12080161
Submission received: 30 June 2023 / Revised: 4 August 2023 / Accepted: 10 August 2023 / Published: 12 August 2023

Abstract

:
Bus driver distraction and cognitive load lead to higher accident risk. Driver distraction sources and complex physical and psychological effects must be recognized and analyzed in real-world driving conditions to reduce risk and enhance overall road safety. The implementation of a camera-based system utilizing computer vision for face recognition emerges as a highly viable and effective driver monitoring approach applicable in public transport. Reliable, accurate, and unnoticeable software solutions need to be developed to reach the appropriate robustness of the system. The reliability of data recording depends mainly on external factors, such as vibration, camera lens contamination, lighting conditions, and other optical performance degradations. The current study introduces Capsule Networks (CapsNets) for image processing and face detection tasks. The authors’ goal is to create a fast and accurate system compared to state-of-the-art Neural Network (NN) algorithms. Based on the seven tests completed, the authors’ solution outperformed the other networks in terms of performance degradation in six out of seven cases. The results show that the applied capsule-based solution performs well, and the degradation in efficiency is noticeably smaller than for the presented convolutional neural networks when adversarial attack methods are used. From an application standpoint, ensuring the security and effectiveness of an image-based driver monitoring system relies heavily on the mitigation of disruptive occurrences, commonly referred to as “image distractions,” which represent attacks on the neural network.

1. Introduction

The need for safe public transport is becoming more demanding for drivers as traffic density increases, accompanied by a complex traffic mix and an influx of fast external stimuli, particularly prevalent in urban areas. Advanced driver-assistance systems (ADASs) can partly intervene in emergencies, e.g., braking [1], but the driver’s task and responsibility remains to control the vehicle safely. Even if ADASs’ goal is to maximize safety, it should be considered that these complex systems have side effects. The adoption and utilization are difficult; a learning process and trust are needed [2]. The analysis of selected papers reveals that ADASs are associated with increased secondary task engagement and improved performance. In contrast, some papers indicate a diversion of attention from driving tasks, emphasizing the critical role of human drivers alongside vehicle automation and the significance of user understanding for the appropriate and effective use of ADASs [3]. Analysis of driver attention and cognitive load provides an opportunity to optimize working conditions and improve accident prevention solutions.
The penetration of public road transport in Europe is still significant. Ensuring passenger safety is a priority as vehicles carry more and more passengers in increasing traffic density. Although buses (in addition to trams and trolleys) account for a decreasing share of all modes of transport, their share is still 7.4% in the EU, based on 2020 data [4]. The number of bus passenger-kilometers in the EU is more than 138 thousand million kilometers, after a 40% decrease due to COVID-19 (see Table 1) [5]. A review of historical data on accidents shows that the average number of fatalities in bus-related accidents per million inhabitants in the European Union countries is 1.3, based on 2017–2019 data. More analysts proved that with the increase in the number of motor vehicles, the number of traffic accidents with casualties also increase [6]. The victims of fatal accidents are mostly pedestrians, motorists, and bus passengers. Accident injury statistics (2010–2019) show that 2% of bus accidents involve serious injuries [4]. In the current study, Capsule Networks (CapsNets) were introduced as part of a professional camera-based Driver Monitoring System (DMS) for public transport. The study presented in this paper shows the initial results of a long-term research and development project.
Analysis of the data has identified many factors that influence the occurrence of accidents. Negligence of bus operators, driver failure, and external factors (weather and road conditions) were identified as the main causes [5]. Driver error can be broken down into several factors, typically depending on current mental and physical conditions, so monitoring driver behavior and later passive or active interventions can ensure safer bus transport [7]. Distractions are visual, auditory, physical, or cognitive stimuli that interfere with driving activities, as well as secondary activities, such as mobile phone use, communication with passengers, and daydreaming. If distraction can be defined as inattention, drowsiness can be defined as inattention related to the driver’s physiological response. This biological behavior is typically associated with insufficient sleep, poor health, or long periods of driving in monotonous environments [8].
Thus, boredom, fatigue, monotony, and sleep deprivation increase the risk of accidents, as reduced attention can lead to impaired information processing and the inability to make decisions, leaving the driver unable to react in an emergency [9]. Distracted driving is “any activity that diverts attention from driving, including talking or texting on your phone, eating and drinking, talking to people in your vehicle, fiddling with the stereo, entertainment, or navigation system—anything that takes your attention away from the task of safe driving” [10]. Most research on vigilance focuses on sleep deprivation [11,12,13]. However, accident data and simulated driving studies suggest that a loss of alertness can also occur during the day, especially on long, monotonous roads [14]. Driver errors are more common during monotonous driving, with low task demand and stimulus levels, reducing attention. Furthermore, the personality profile of drivers also influences the likelihood of crashes due to hypervigilance; extroverts and sensation seekers are at higher risk. Various monitoring systems are developed and used to observe or detect driver behavior. Different approaches use psychological tests and different physiological sensors, data collection, and analysis techniques, as follows: (1) vehicle-based measures, (2) behavioral measures, and (3) physiological measures [15].
Vehicle-based measures were utilized in more research studies in order to identify driver behavior. One method was developed for detecting driving patterns using vehicle sensor data from a single turn [16]. CAN-BUS data analyzed by a unique deep learning framework resulted in driving behavior identification [17]. A human–vehicle interaction monitoring system with principal component analysis is used to monitor fuel consumption, CO2 emissions, driving style, and driver’s health in real-time with high prediction levels [18]. High energy efficiency targets are also important in the case of rail vehicles, where recent research focuses on detecting and optimizing energy losses on the vehicle side [19].
Several behavioral measurement analyses—mainly using the Driver Behavior Questionnaire—have shown that professional drivers engage in less risky behavior than non-professional drivers but are more likely to be involved in traffic accidents due to extended driving periods [20]. Some studies proved that driver characteristics are related to driving circumstances (e.g., comfort level) and affect driving performance [21]. A great sample of inter-city bus drivers was analyzed by psychometric properties of the Multidimensional Driving Style Inventory (MDSI), Driver Behavior Questionnaire (DBQ), and Driver Anger Scale (DAS) questionnaires and highlighted the importance of diagnostics for distinguishing safe and unsafe drivers [22]. Results show that the correlation between the data collected from the vehicle and drivers’ signals (physiological sensors) is high, so it is expected that further studies based on the data collected from the vehicle will allow for high-accuracy inferences about driver behavior and driver identification [23].
Recently, it has been shown that the frequency and duration of fixation, the change in pupil diameter (pupillometry), the quantity of saccades and microsaccades, and the frequency and duration of blinking are closely related to cognitive load [24]. These measurements can be made with eye-tracking systems with high-quality cameras observing test participants’ eye movements. Some studies using driving scene videos have shown that fixation behavior and patterns correlate with driving experience [25,26]. In other studies, average pupil size gave a clear indication of divided attention while driving after performing tasks of varying difficulty or with different interfaces on the IVIS (e.g., physical button or touchscreen), and it was observed that average pupil size increased significantly [27,28,29].
Driver fatigue and distraction detection can be based on biometric signals, steering motion, or driver’s face monitoring [30]. Some researchers found that a higher average Heart Rate (HR) shows that the driver performs more demanding or secondary tasks [31]. Others performed heart rate variability analyses validated by electroencephalography for driver drowsiness detection [32]. More authors used wearable Galvanic Skin Responses (GSR) wristbands to identify the distraction of drivers during a driving experiment on-the-road with high accuracy and consistency [33]. Electroencephalograms (EEGs) are widely used for driver behavior detection, but their signals are not stationary, and the process is intrusive for testing participants, mainly when using multi-sensor systems [34]. Recently, single-channel EEG-based drowsiness detection systems were tested, where data analysis through small-sized time window datasets and a single feature computation enabled lower processing and storage capacities, leading to more straightforward implementation in embedded systems [35]. Other physiological sensors—like an electromyogram (EMG), which measures the driver’s muscle fatigue—were also used, and the appropriate measurement methodology was described [36]. Emotion classification with high prediction accuracies has also been investigated, where an electrocardiogram (ECG)—combined with GSR—was applied [37].
The most widespread solution for DMSs is the non-wearable, camera-based system using computer vision for face recognition. This approach is more convenient for drivers because it does not require additional wearable devices, allowing for a seamless and non-intrusive monitoring experience. It can be perfectly integrated into the mass production process and thus added to series production vehicles. “Attention warning in case of driver drowsiness or distraction” must be applied to all new European vehicles from 2024, according to EU Regulation 2019/2144 [38]. Visual sensors, RGB, IR, or other cameras facing the driver, with different viewing angles, are used to collect naturalistic driving (ND) data, including drivers’ interaction, behavior, and the surrounding (driver cabin). The following data categories are usually detected: hand movement, body pose estimation, face detection, and distraction/drowsiness [39]. Face detection methods are generally categorized as feature-based and learning-based methods [40]. Learning-based methods are usually more robust than feature-based methods but mostly take more computational resources. However, these methods can reach a more than 80% detection rate in laboratory conditions [41]. It is a challenge to test the applicability of the appropriate algorithm and to prove its benefits, but in the long term, optimized face recognition allows the development of a system with a higher confidence level and increases road safety. Data recording and processing reliability depend on the following factors: viewing angle; vibration; camera lens contamination/obscuration; masking or interfering objects; strong light or shadow effects; low illumination; and further optical performance degradation. It is essential to address these factors and eliminate them during the data acquisition and processing stage.
The current work focuses on the usability of capsule networks compared with state-of-the-art convolutional neural networks. In this case, the authors’ goal was to identify 15 keypoints of the face for further analysis. The authors compared the dynamic routing algorithm with the authors’ proposed routing method. The networks were also trained using different adversarial attack techniques, highlighting the applied capsule-based solution’s robustness. The research goal was that the authors’ CapsNets solution should perform well in professional applications like DMSs in public transport. The first phase of the current study, presented here, was to improve the robustness of the face detecting algorithm, and further investigation is needed to monitor the speed and computational need of the system.
The current paper is structured as follows. In Section 2, the authors present the theory of the capsule network, discussing the authors’ proposed optimization algorithm and activation function. The authors present the dataset used in their work and the network architectures developed. In Section 3, the authors clarify the conditions under which the networks were trained. The authors present the adversarial attack methods used in the training and the main parameters of the training process. The authors then present and explain the training results, comparing the different methods. In Section 4, the authors discuss the results and their importance, and finally, in Section 5, the authors summarize their work and give an outlook for future plans.

2. Methodology

The current study involves detecting specific parts of the human face. The primary aim is to find practical applications for the observation of drivers. In addition, the authors want to investigate the applicability and robustness of capsule networks in this area. This area is becoming more popular as technology develops, not only in research but also in an increasing number of practical applications, such as DMSs for bus drivers.
Verma et al. proposed a novel, real-time driver emotion monitoring system with face detection and facial expression analysis [42]. They used two VGG-16 neural networks. The first network extracts appearance features from the detected face image, and the second extracts geometrical features from the facial landmark points. Jain et al. proposed a capsule-network-based approach for detecting the distracted driver [43]. The proposed method performed well on real-world environment data, where the results were compared with convolutional-neural-network-based approaches.
Ali et al. generated a dataset for conducting different experiments on driver distraction [44]. They presented a novel approach that uses features based on facial points, especially those computed using motion vectors and interpolation to detect a special type of driver distraction. Liu et al. presented a comprehensive review of the research work on face recognition technology; they analyzed the difficulty of face recognition tasks, then introduced the main frameworks of face recognition from the geometric-feature-based, template-based, and model-based methods [40]. Their work compared different solutions and highlighted face recognition’s importance and current limitations.

2.1. Capsule Network Theory

CapsNets are similar to neural networks but include some improvements on the known problems of neural networks [45,46]. A convolutional neural network cannot spatially map what is learned. For example, if it only ever sees images of an object from the same viewpoint, it will not recognize it from another viewpoint. Convolutional neural networks cannot discover the relationship between the elements of an image. For example, when detecting a human face, if the image contains features of a human face, it will detect it as a human face, even if the components of the face are not in the correct positions. Various pooling layers are used in convolutional neural networks. These are generally efficient but tend to cause substantial loss of information by their nature. Capsule networks offer a more efficient solution to this by using a routing algorithm and vectors instead of scalar values, which allows them to store more information about a feature. The key difference between capsule and neural networks is the systems’ basic building blocks. While neural networks are made up of neurons, capsule networks contain so-called capsules. Table 2 shows the main differences between classical artificial neurons and capsules.
A capsule can be considered a group of interconnected neurons that perform much internal computation and encapsulate the results of the computations into an n -dimensional vector. Where the vector is the output of the capsule, it can be seen that the output of the capsule is not a probability value, as is usual for a neuron. The length of the output vector gives the probability value associated with the output. The direction of the n -dimensional vector can store different features related to a given task. For example, a network is supposed to detect a human face. In this case, low-level capsules are responsible for facial components, such as the eyes, nose, or mouth, and a high-level capsule is responsible for face recognition. The neurons in the low-level capsules encode the object’s internal properties, such as its position, orientation, color, texture, or shape.
In capsule networks, no pooling layer or similar solution is used for reduction between layers. In this approach, routing-by-agreement was used, where the output vector of any lower-level capsule is sent to all higher-level capsules. Each lower-level capsule output is compared with the actual output of the higher-level capsules. Routing aims to control the strength of the information flow between capsules. It is performed to achieve a stronger connection between related features.
Let i be a lower-level capsule and j a higher-level capsule. The u ^ input tensor for capsule j is calculated as:
u ^ ( j | i ) = W i j u i
where Wij is a weighting matrix, initialized with random values, and ui is a pose vector for the I-th capsule. The coupling coefficients are calculated with a simple softmax function, as in the following:
c i j = e x p ( b i j ) k e x p ( b i k )
where bij is the log probability of capsule I being coupled with capsule J and bij is initialized with zero values.
The total input to capsule j is a weighted sum over the prediction vectors, as in the following:
s j = i c i j u ^ j | i
Each layer of the capsule network gives output vectors. Therefore, the probability is determined by the length of the vector. A non-linear squashing activation function must be applied before determining the probability value. The squashing function is as follows:
v j = s q u a s h ( s j ) = s j 2 1 + s j 2 s j s j 2
The routing algorithm is used to determine the relationship between the capsule layers. The dynamic routing algorithm (Algorithm 1) updates the cij values and determines the output v j capsule vector.
Algorithm 1. Routing algorithm [46]
1:procedure   ROUTING   ( u ^ j | i ,   r ,   l )
2:     for   all   capsule   i   in   layer   l   and   capsule   j   in   layer   ( l + 1 ) :   b i j 0
3:  for r iterations do
4:       for   all   capsule   i   in   layer   l :   c i s o f t m a x ( b i )
5:       for   all   capsule   j   in   layer   ( l + 1 ) :   s j i c i j u ^ j | i
6:       for   all   capsule   j   in   layer   ( l + 1 ) :   v j s q u a s h ( s j )
7:       for   all   capsule   i   in   layer   l   and   capsule   j   in   layer   ( l + 1 ) :   b i j b i j + u ^ j | i v j
8:   return v j

2.2. Proposed Capsule Routing Mechanism

Present experiments on capsule networks theory have shown that the u ^ j | i input in the dynamic routing algorithm has too large an impact on the output tensor. When calculating the output vector v j , the formula includes the input u ^ j | i twice:
v j = sq ( i smax ( b i j + u ^ j | i v j ) u ^ j | i )
where sq ( · ) is the squash function and smax ( · ) is the softmax function. To be able to improve the routing mechanism between lower- and upper-level capsules, the following modifications to the routing algorithm are proposed:
v j = sq ( i smax ( b i j + j v j ) u ^ j | i )
This minimal modification makes the routing algorithm simpler and faster to compute. The other proposed change concerns the squashing function. In the last capsule layer, a modified squashing function was applied, as follows:
squash ( s ) = s e s s s + ε
where ε is a fine-tuning parameter. The authors of the article have carried out several experiments with different hyperparameters. As a result, ε = 1 × 10−7 gave the best performance.

2.3. Dataset

The Kaggle Facial Keypoints Detection Dataset [47] was used in this work to test the novel solution, which aims to detect the location of keypoints on face images. The dataset contains a total of 7049 samples of different human faces. The dataset contains human faces in various situations, not specifically of bus drivers. However, this does not affect the results because the images in the dataset were taken with settings that could be implemented on board a bus. The samples have 15-15 output keypoints, which represent the position of the following parts of the face: left eye center, right eye center, left eye inner corner, left eye outer corner, right eye inner corner, right eye outer corner, left eyebrow inner end, left eyebrow outer end, right eyebrow inner end, right eyebrow outer end, nose tip, mouth left corner, mouth right corner, mouth center top lip, and mouth center bottom lip. Unfortunately, the Kaggle dataset is incomplete; not all faces contain all 15 keypoints.
For this reason, only those samples were used that contain all the keypoints. Thus, a total of 2250 samples were used in the training process, with 1800 samples in the training set and 450 samples in the testset, where all images were 48 pixels in width and 48 pixels in height with 1 color channel. Figure 1 shows some samples from the dataset, where the keypoints are marked in red.

2.4. Network Architectures

The primary goal of the current study was to focus on the differences between the proposed routing algorithm and the dynamic routing algorithm by Sabour et al. Second, the authors wanted to compare the effectiveness of capsule-based solutions with convolutional-neural-network-based approaches. In this research, four network architectures were used: the Capsule Network (CapsNet) proposed by Sabour et al. [46], a modified version of the Capsule Network with the proposed routing algorithm and activation function, the Wide-ResNet network proposed by Zagoruyko et al. [48], and the SimpleCNN network proposed by An et al. [49].
The authors’ own capsule-based solution is based on the architecture of the original capsule network introduced by Sabour et al. [46]. The only differences are the routing algorithm and the activation function described in Section 2.1. The architecture of the authors’ capsule network solution for this task is illustrated in Figure 2.
The Wide-ResNet (Wide-Residual Network) architecture was introduced by Zagoruyko et al. [48]. The main building block of the Wide-ResNet network is the residual block:
x l + 1 = x l + ( x l ,   W l )
where xl is the input and xl+1 is the output of the l-th layer, F ( · ) is the residual function, and Wl is the parameter list of the residual block. The Wide-ResNet is based on a width parameter determined by factor k in the convolutional layers. In the original ResNet model, k = 1. However, in Wide-ResNet, the wideness of the network is determined by the k > 1 value. In this work, it has been shown that networks can be made more efficient by widening layers rather than creating deeper networks. In 2016, the Wide-ResNet architecture provided state-of-the-art results on several commonly used datasets, like Canadian Institute for Advanced Research-10 (CIFAR-10) [50], Canadian Institute for Advanced Research-100 (CIFAR-100) [50], Street View House Numbers (SVHN) [51], and Common Objects in Context (COCO) [52].
The Simple Convolutional Neural Network (SimpleCNN), introduced by An et al., consists of multiple convolution layers and a fully connected layer, where each convolution layer contains the following: 2D convolution, 2D batch normalization, and ReLU activation [49]. This neural network is made up of three sub-networks, where the sub-networks differ only in the convolutional layer kernel size. They used 3 × 3-, 5 × 5-, and 7 × 7-sized kernels. The final result of the three sub-networks is obtained by using majority voting. The SimpleCNN network is a state-of-the-art solution for image classification of the Modified National Institute of Standards and Technology (MNIST) dataset [53].

3. Training and Results

In the training process, adversarial attacking methods were applied to investigate the effectiveness of different neural and capsule networks. In this study, seven adversarial attacking methods were used on the presented two neural networks and two capsule networks.

3.1. Adversarial Attack Methods

The Fast Gradient Signed Method (FGSM) was proposed by Goodfellow et al. [54]. It was one of the first and most popular attacks to fool a neural network. The FGSM works by using the gradients of the neural network to create an adversarial example. The gradient value comes from the network’s loss, and the algorithm modifies the input image with the loss gradient to maximize the loss value of the network as follows:
x a d v = x + ϵ × sign ( x J ( θ ,   x ,   y ) )
where xadv is the adversarial image, x is the original input, y is the original label, ϵ is a parameter to ensure that the perturbations are small, θ is the model parameters, and J ( · ) is the loss function of the network.
Fast adversarial training using FGSM (FFGSM) was proposed by Wong and Rice et al. [55]. This method is very similar to the Fast Gradient Signed Method. The main difference is that this method uses random initialization, where random uniform distributed noise is added to the input image:
x = x + Uniform ( ϵ ,   ϵ )
where x is the input image, x is the noisy image, and ϵ is a parameter to ensure small perturbations.
Projected Gradient Descent (PGD) was proposed by Mądry et al. [56]. This approach is similar to the FGSM attack but iteratively executes the attack several times, as follows:
x t + 1 = ( x ,   ε ) { x t + α sign ( x t J ( f θ ( x ) , f θ ( x t ) ) ) }
where   ( x ,   ε ) refers to the projection to ( x ,   ε ) and N ( 0 n ,   I n ) is a normal distribution.
Two other variants of the PGD attack method were used. In the first case, noise generation was performed the same way as with the original PGD but with l2 loss (PGDL2). The next one was the TRADES’ Projected Gradient Descent (TPGD) proposed by Zhang et al. [57]. In this method, the attacked examples generated by Projected Gradient Descent (PGD) with Kullback–Leibler divergence loss are as follows:
x 0 = x + 0.001 × N ( 0 n ,   I n )
x t + 1 = ( x ,   ε ) { x t + α sign ( x t l K L ( f θ ( x ) , f θ ( x t ) ) ) }
where   ( x ,   ε ) refers to the projection to ( x ,   ε ) , N ( 0 n ,   I n ) is a normal distribution, and l K L ( · ) is a loss function [58].
The following attack method is based on a simple random noise generation. This approach adds Gaussian Noise (GN) to the input image. Here, the properties of the network, its current state, or the loss result are not used to generate the noise. The adversarial image is generated as follows:
x = x + σ N ( 0 n ,   I n )
where x is an input image, σ ( 0 ,   1 ) is a fine-tuning parameter, and N ( 0 n ,   I n ) is a normal distribution value.
The Basic Iterative Method (BIM) was proposed by Kurakin et al. This work introduced a straightforward way to extend the Fast Gradient Signed Method (FGSM) [59]. The FGSM attack was applied multiple times with a small step size and clip pixel values of intermediate results after each step to ensure that they were in an ϵ -neighbourhood of the original image. The attacked inputs were generated as follows:
x x + 1 = Clip x , ϵ { x n + α sign ( x J ( x n , y true ) ) }
where x 0 = x , x is the input image and
C l i p x , ϵ { x } ( u ,   v ,   z ) = m i n { 255 ,   x ( u ,   v ,   z ) + ϵ ,   m a x { 0 ,   x ( x ,   y ,   z ) ϵ ,   x ( u ,   v ,   z ) } }
is the clipping function, where x ( u ,   v ,   z ) is the value of channel z of the image x at coordinate (u, v).

3.2. Training Process

In this study, all four networks were trained in eight ways: in the first approach without using any adversarial attack technique, and then using the seven attack methods presented. In all cases, the training was performed under the same conditions with the same data. The training was carried out via the Paperspace cloud service, where all networks were trained under the same hardware conditions. An Nvidia RTX A4000 GPU was used in the training process. The Python programming language and the PyTorch deep learning framework were used for the implementation. The Adam (kingma2017) optimization algorithm was used for the training. The initial learning rate, based on empirical experience, was chosen to be 0.001. During the epochs, the learning rate was obtained as follows:
l r i + 1 = l r i × 0.96 1 2000
where lri is the learning rate in the i-th epoch and lr0 = 0.001 is the initial learning rate. In each case, all networks were trained over 50 epochs. In the study, the efficiency for different epoch values was investigated; however, experience shows that the difference between the networks is already well visible here and does not change significantly in further iterations. Batches of 16 sizes were used during the training and testing process. In this study, the following L1 loss was used. Let
l ( x ,   y ) = { l 1 ,   l 2 ,   ,   l n } T
where l is the loss of the given batch, x is the predicted output, y is the ground truth, and n is the size of the batch. Let
l i = { 0.5 ( x i y i ) 2 β ,     if   | x i y i | < β | x i y i | 0.5 β ,     otherwise

3.3. Results

The efficiencies measured during training are shown in Figure 3. In the no-attack case, it can be seen that the two capsule networks show a similar curve, but their approach shows a visibly smoother, more uniform learning. In the end, the authors’ own solution led to better results. The Wide-ResNet network performs slightly worse than the authors’ capsule-based solution. The SimpleCNN network converges more slowly initially but eventually outperforms all networks.
The results are a little different for the attack methods. The most considerable degradation was observed for the Wide-ResNet network. In several cases, the network failed to converge to an optimal state at all. The SimpleCNN network typically performed well, tending to be more affected by noise only at the beginning of the learning phase. Capsule-based solutions generally performed well. In the solution of Sabour et al., the training process started well in some cases, but by the end, the efficiency of the network had decreased. The current proposed capsule-based approach also typically performed well. A similar phenomenon can be observed as with the original capsule solution. For some attack methods, minimal degradation was observed by the end of the learning process. However, in all cases, the authors’ capsule-based solution with their routing mechanism performed better than the original with the dynamic routing algorithm.
Table 3 summarizes the best results achieved during the training. It can be seen that the SimpleCNN network, as a state-of-the-art solution, performed outstandingly in most cases. However, the currently proposed method performed best for three of the seven attack methods tested. Only for the capsule networks, the authors’ method always performs better.
If someone looks at how the effectiveness of each network has changed compared to itself under different attack methods, different results can be obtained. Table 4 shows by what percentage the efficiency decreased for each attack method compared to the case without attack. It can be seen that the capsule networks always performed best in this case. The authors’ solution performed better in most attack methods tested.

4. Discussion

The authors have presented their proposed optimization algorithm for capsule networks in this study. To support the validity of the authors’ own algorithm, they conducted tests where the authors’ solution was compared with capsule networks trained with the dynamic routing algorithm, as well as with several state-of-the-art convolutional neural network solutions [54,55,56,57,58,59]. In the comparison, different adversarial attack techniques were used to test the robustness of the presented systems. The results show that capsule-based solutions can compete with state-of-the-art convolutional neural-network-based solutions in this task. It can be said that when adversarial attack methods are used, capsule networks perform well. The authors’ results show that the degradation in efficiency is noticeably smaller for capsule networks than for the presented convolutional neural networks. For the case of capsule networks, the authors’ proposed method achieved better performance than the dynamic routing algorithm in all cases. Regarding performance degradation, the authors’ solution performed better in six out of seven cases. In terms of effectiveness, the proposed method was more effective in three out of seven cases.
From an application perspective, given the growing need for safety in public transport, a critical point for a sufficiently secure image-based driver monitoring system is the elimination of distracting or offensive incidents, the attacks. The reliability of data recording depends mainly on external factors, such as vibration, camera lens contamination, lighting conditions, and other optical performance degradations. The authors’ CapsNets method reduces the uncertainty of data recording to increase data processing accuracy.
This new approach has further potential for improvement. Besides effectiveness, speed and computational demand should also be examined to minimize hardware needs. All on-board DMSs require high-speed, real-time data processing, low power consumption, flexible connectivity, and robustness, for both software and hardware. The current research will further deepen the investigations into the algorithm’s applicability while striving to design a modular system, thus aiding the feasibility of implementation.
However, the current solution is not yet perfect, and its practical application is not recommended in its current state. The authors’ long-term goal is to have an efficient and reliable solution as part of the custom driver monitoring system, which is currently being implemented. To be able to achieve this, it is necessary to improve further the current capsule-based solution to achieve even higher efficiency. It will require further testing on an even wider range of datasets and testing the robustness of the network with additional attack methods. Finally, measuring the effectiveness and validating the system on real data from the end-user environment is necessary.

5. Conclusions

As wearable monitoring devices are unsuitable for monitoring drivers in public transport, reliable software solutions must be developed for face recognition.
The current study focuses on developing a capsule-network-based approach for face detection to support DMS applications for safer road traffic. This paper presents the proposed new routing algorithm for capsule network optimization. We have shown that the introduced method is suitable for detecting keypoints of the human face. Different adversarial attack techniques were used to test the robustness of the systems. The study results show that the efficiency of capsule networks is, in many cases, higher than the neural networks presented, and the authors’ proposed routing method achieved better performance than the dynamic routing algorithm in all cases. Regarding performance degradation, the authors’ solution performed better than most.
The algorithm’s outstanding performance warrants further investigation to develop an application-level implementation for a driver monitoring system for use in public road transportation. The authors’ computational vision method’s proven robustness can eliminate the disruption of external factors, such as vibration, camera lens contamination, lighting conditions, and other optical performance degradations.
This paper presents the preliminary findings of an extensive, ongoing research and development endeavor. In the following research phase, we want to build on these results. We want to validate the effectiveness of the networks on additional datasets. We would like to generate our own dataset of bus drivers, where drivers can be observed in real traffic situations. This will allow us to investigate specific edge cases and make the network capable of detecting them. We would like to optimize further our current solution pending further studies. We would like to compare the results obtained with other approaches on a broader scale, also in the application of further adversarial attack techniques.

Author Contributions

Conceptualization, J.H. and V.N.; methodology, J.H. and V.N.; software, J.H.; validation, J.H. and V.N.; formal analysis, J.H. and V.N.; investigation, J.H. and V.N.; resources, J.H. and V.N.; data curation, J.H. and V.N.; writing—original draft preparation, J.H., Á.B., G.K., S.F. and V.N.; writing—review and editing, J.H., Á.B., G.K., S.F. and V.N.; visualization, J.H.; supervision, J.H., Á.B., G.K., S.F. and V.N.; project administration, J.H., S.F., and V.N.; funding acquisition, J.H., S.F. and V.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

All data used in this research are presented in the article.

Acknowledgments

The authors wish to acknowledge the technical support received from the Vehicle Industry Research Centre and Széchenyi István University, Győr. The research was carried out as part of the Cooperative Doctoral Program supported by the National Research, Development and Innovation Office and the Ministry of Culture and Innovation.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Blades, L.; Douglas, R.; Early, J.; Lo, C.Y.; Best, R. Advanced Driver-Assistance Systems for City Bus Applications; SAE Technical Papers; Sage: Thousand Oaks, CA, USA, 2020. [Google Scholar] [CrossRef] [Green Version]
  2. Nylen, A.B.; Reyes, M.L.; Roe, C.A.; McGehee, D.V. Impacts on Driver Perceptions in Initial Exposure to ADAS Technologies. Transp. Res. Rec. 2019, 2673, 354–360. [Google Scholar] [CrossRef]
  3. Hungund, A.P.; Pai, G.; Pradhan, A.K. Systematic Review of Research on Driver Distraction in the Context of Advanced Driver Assistance Systems. In Transportation Research Record; SAGE Publications Ltd: Thousand Oaks, CA, USA, 2021; Volume 2675, pp. 756–765. [Google Scholar]
  4. Slootmans, F. European Road Safety Observatory—Facts and Figures—Light Trucks; European Commission: Brussels, Belgiumcon, 2021. [Google Scholar]
  5. Eurostat. Passenger Transport by Buses and Coaches by Type of—Vehicles Registered in the Reporting Country. Available online: https://ec.europa.eu/eurostat/databrowser/view/ROAD_PA_BUSCOA__custom_1210091/bookmark/table?lang=en&bookmarkId=d8cf5c80-4d26-4dfd-bda5-e94de54b8d49 (accessed on 30 March 2023).
  6. Damjanović, M.; Stević, Ž.; Stanimirović, D.; Tanackov, I.; Marinković, D. Impact of the Number of Vehicles on Traffic Safety: Multiphase Modeling. Facta Univ. Ser. Mech. Eng. 2022, 20, 177–197. [Google Scholar] [CrossRef]
  7. Goh, K.; Currie, G.; Sarvi, M.; Logan, D. Factors Affecting the Probability of Bus Drivers Being At-Fault in Bus-Involved Accidents. Accid. Anal. Prev. 2014, 66, 20–26. [Google Scholar] [CrossRef] [PubMed]
  8. Ferreira, S.; Kokkinogenis, Z.; Couto, A. Using Real-Life Alert-Based Data to Analyse Drowsiness and Distraction of Commercial Drivers. Transp. Res. Part F Traffic Psychol. Behav. 2019, 60, 25–36. [Google Scholar] [CrossRef]
  9. Young, K.; Regan, M.; Hammer, M. Driver Distraction: A Review of the Literature (Report); Monash University Accident Research Centre: Clayton, Australia, 2003; p. 66. [Google Scholar]
  10. National Highway Traffic Safety Administration. Distracted Driving; National Highway Traffic Safety Administration (NHTSA): Washington, DC, USA, 2021. [Google Scholar]
  11. Otmani, S.; Rogé, J.; Muzet, A. Sleepiness in Professional Drivers: Effect of Age and Time of Day. Accid. Anal. Prev. 2005, 37, 930–937. [Google Scholar] [CrossRef]
  12. Huhta, R.; Hirvonen, K.; Partinen, M. Prevalence of Sleep Apnea and Daytime Sleepiness in Professional Truck Drivers. Sleep Med. 2021, 81, 136–143. [Google Scholar] [CrossRef]
  13. Felix, M.; Intriago Alvarez, M.B.; Vanegas, E.; Farfán Bajaña, M.J.; Sarfraz, Z.; Sarfraz, A.; Michel, J.; Cherrez-Ojeda, I. Risk of Obstructive Sleep Apnea and Traffic Accidents among Male Bus Drivers in Ecuador: Is There a Significant Relationship? Ann. Med. Surg. 2022, 74, 103296. [Google Scholar] [CrossRef]
  14. Thiffault, P.; Bergeron, J. Monotony of Road Environment and Driver Fatigue: A Simulator Study. Accid. Anal. Prev. 2003, 35, 381–391. [Google Scholar] [CrossRef]
  15. Sahayadhas, A.; Sundaraj, K.; Murugappan, M. Detecting Driver Drowsiness Based on Sensors: A Review. Sensors 2012, 12, 16937–16953. [Google Scholar] [CrossRef] [Green Version]
  16. Hallac, D.; Sharang, A.; Stahlmann, R.; Lamprecht, A.; Huber, M.; Roehder, M.; Sosič, R.; Leskovec, J. Driver Identification Using Automobile Sensor Data from a Single Turn. In Proceedings of the IEEE Conference on Intelligent Transportation Systems, Rio de Janeiro, Brazil, 1–4 November 2016; pp. 953–958. [Google Scholar] [CrossRef] [Green Version]
  17. Zhang, Z.; Tang, Y.; Zhao, S.; Zhang, X. Real-Time Surface EMG Pattern Recognition for Hand Gestures Based on Support Vector Machine. In Proceedings of the IEEE International Conference on Robotics and Biomimetics, ROBIO, Dali, China, 6–8 December 2019; pp. 1258–1262. [Google Scholar] [CrossRef]
  18. Campos-Ferreira, A.E.; Lozoya-Santos, J.d.J.; Tudon-Martinez, J.C.; Mendoza, R.A.R.; Vargas-Martínez, A.; Morales-Menendez, R.; Lozano, D. Vehicle and Driver Monitoring System Using On-Board and Remote Sensors. Sensors 2023, 23, 814. [Google Scholar] [CrossRef]
  19. Fischer, S.; Szürke, S.K. Detection Process Of Energy Loss In Electric Railway Vehicles. Facta Univ. Ser. Mech. Eng. 2023, 21, 81–99. [Google Scholar] [CrossRef]
  20. Maslać, M.; Antić, B.; Lipovac, K.; Pešić, D.; Milutinović, N. Behaviours of Drivers in Serbia: Non-Professional versus Professional Drivers. Transp. Res. Part F Traffic Psychol. Behav. 2018, 52, 101–111. [Google Scholar] [CrossRef]
  21. Fancello, G.; Daga, M.; Serra, P.; Fadda, P.; Pau, M.; Arippa, F.; Medda, A. An Experimental Analysis on Driving Behaviour for Professional Bus Drivers. Transp. Res. Procedia 2020, 45, 779–786. [Google Scholar] [CrossRef]
  22. Karimi, S.; Aghabayk, K.; Moridpour, S. Impact of Driving Style, Behaviour and Anger on Crash Involvement among Iranian Intercity Bus Drivers. IATSS Res. 2022, 46, 457–466. [Google Scholar] [CrossRef]
  23. Bonfati, L.V.; Mendes Junior, J.J.A.; Siqueira, H.V.; Stevan, S.L. Correlation Analysis of In-Vehicle Sensors Data and Driver Signals in Identifying Driving and Driver Behaviors. Sensors 2023, 23, 263. [Google Scholar] [CrossRef] [PubMed]
  24. Biondi, F.N.; Saberi, B.; Graf, F.; Cort, J.; Pillai, P.; Balasingam, B. Distracted Worker: Using Pupil Size and Blink Rate to Detect Cognitive Load during Manufacturing Tasks. Appl. Ergon. 2023, 106, 103867. [Google Scholar] [CrossRef]
  25. Underwood, G.; Chapman, P.; Brocklehurst, N.; Underwood, J.; Crundall, D. Visual Attention While Driving: Sequences of Eye Fixations Made by Experienced and Novice Drivers. Ergonomics 2003, 46, 629–646. [Google Scholar] [CrossRef]
  26. Underwood, G.; Chapman, P.; Bowden, K.; Crundall, D. Visual Search While Driving: Skill and Awareness during Inspection of the Scene. Transp. Res. Part F Traffic Psychol. Behav. 2002, 5, 87–97. [Google Scholar] [CrossRef]
  27. Nagy, V.; Kovács, G.; Földesi, P.; Kurhan, D.; Sysyn, M.; Szalai, S.; Fischer, S. Testing Road Vehicle User Interfaces Concerning the Driver’s Cognitive Load. Infrastructures 2023, 8, 49. [Google Scholar] [CrossRef]
  28. Shi, Y.; Xu, Y.; Ferraioli, A.; Piga, B.; Mussone, L.; Caruso, G. A Method of Processing Eye-Tracking Data to Study Driver’s Visual Perception in Simulated Driving Scenarios. In Proceedings of the European Transport Conference, Online, 8–11 September 2020. [Google Scholar]
  29. Reilly, J.; Zuckerman, B.; Kelly, A.E. A Primer on Design and Data Analysis for Cognitive Pupillometry. Mod. Pupillometry 2021, 9–25. [Google Scholar] [CrossRef]
  30. Sigari, M.-H.; Pourshahabi, M.-R.; Soryani, M.; Fathy, M. A Review on Driver Face Monitoring Systems for Fatigue and Distraction Detection. Int. J. Adv. Sci. Technol. 2014, 64, 73–100. [Google Scholar] [CrossRef]
  31. Biondi, F.; Coleman, J.R.; Cooper, J.M.; Strayer, D.L. Average Heart Rate for Driver Monitoring Systems. Int. J. Hum. Factors Ergon. 2016, 4, 282–291. [Google Scholar] [CrossRef]
  32. Fujiwara, K.; Abe, E.; Kamata, K.; Nakayama, C.; Suzuki, Y.; Yamakawa, T.; Hiraoka, T.; Kano, M.; Sumi, Y.; Masuda, F.; et al. Heart Rate Variability-Based Driver Drowsiness Detection and Its Validation with EEG. IEEE Trans. Biomed. Eng. 2019, 66, 1769–1778. [Google Scholar] [CrossRef]
  33. Dehzangi, O.; Rajendra, V.; Taherisadr, M. Wearable Driver Distraction Identification On-the-Road via Continuous Decomposition of Galvanic Skin Responses. Sensors 2018, 18, 503. [Google Scholar] [CrossRef] [Green Version]
  34. Li, G.; Chung, W.Y. A Context-Aware EEG Headset System for Early Detection of Driver Drowsiness. Sensors 2015, 15, 20873–20893. [Google Scholar] [CrossRef] [PubMed]
  35. Balam, V.P.; Chinara, S. Development of Single-Channel Electroencephalography Signal Analysis Model for Real-Time Drowsiness Detection: SEEGDD. Phys. Eng. Sci. Med. 2021, 44, 713–726. [Google Scholar] [CrossRef]
  36. Rahman, N.A.A.; Mustafa, M.; Sulaiman, N.; Samad, R.; Abdullah, N.R.H. EMG Signal Segmentation to Predict Driver’s Vigilance State: Notes in Mechanical Engineering; Springer: Berlin/Heidelberg, Germany, 2022; pp. 29–42. [Google Scholar] [CrossRef]
  37. Bulagang, A.F.; Weng, N.G.; Mountstephens, J.; Teo, J. A Review of Recent Approaches for Emotion Classification Using Electrocardiography and Electrodermography Signals. Inform. Med. Unlocked 2020, 20, 100363. [Google Scholar] [CrossRef]
  38. European Parliament Regulation (EU) 2019/2144 of the European Parliament and of the Council. Off. J. Eur. Union 2019, L325, 32019R2144.
  39. Koay, H.V.; Chuah, J.H.; Chow, C.O.; Chang, Y.L. Detecting and Recognizing Driver Distraction through Various Data Modality Using Machine Learning: A Review, Recent Advances, Simplified Framework and Open Challenges (2014–2021). Eng. Appl. Artif. Intell. 2022, 115, 105309. [Google Scholar] [CrossRef]
  40. Liu, W.; Wang, X. Researches Advanced in Face Recognition. Highlights Sci. Eng. Technol. AMMSAC 2023, 49, 41. [Google Scholar] [CrossRef]
  41. Chaves, D.; Fidalgo, E.; Alegre, E.; Alaiz-Rodríguez, R.; Jáñez-Martino, F.; Azzopardi, G. Assessment and Estimation of Face Detection Performance Based on Deep Learning for Forensic Applications. Sensors 2020, 20, 4491. [Google Scholar] [CrossRef] [PubMed]
  42. Verma, B.; Choudhary, A. Deep Learning Based Real-Time Driver Emotion Monitoring. In Proceedings of the 2018 IEEE International Conference on Vehicular Electronics and Safety, ICVES 2018, Madrid, Spain, 12–14 September 2018; pp. 1–6. [Google Scholar] [CrossRef]
  43. Jain, D.K.; Jain, R.; Lan, X.; Upadhyay, Y.; Thareja, A. Driver Distraction Detection Using Capsule Network. Neural Comput. Appl. 2021, 33, 6183–6196. [Google Scholar] [CrossRef]
  44. Ali, S.F.; Hassan, M.T. Feature Based Techniques for a Driver’s Distraction Detection Using Supervised Learning Algorithms Based on Fixed Monocular Video Camera. KSII Trans. Internet Inf. Syst. 2018, 12, 3820–3841. [Google Scholar] [CrossRef]
  45. Hinton, G.E.; Krizhevsky, A.; Wang, S.D. Transforming Auto-Encoders; Springer: Berlin/Heidelberg, Germany, 2014. [Google Scholar] [CrossRef]
  46. Sabour, S.; Frosst, N.; Hinton, G.E. Dynamic Routing Between Capsules Sara. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; pp. 1–11. [Google Scholar]
  47. James, P.; Cukierski, W. Facial Keypoints Detection; Kaggle: San Francisco, CA, USA, 2017. [Google Scholar]
  48. Zagoruyko, S.; Komodakis, N. Wide Residual Networks. In Proceedings of the British Machine Vision Conference 2016 (BMVC 2016), York, UK, 19–22 September 2016; pp. 1–12. [Google Scholar] [CrossRef] [Green Version]
  49. An, S.; Lee, M.; Park, S.; Yang, H.; So, J. An Ensemble of Simple Convolutional Neural Network Models for MNIST Digit Recognition. arXiv 2020, arXiv:2008.10400v2. [Google Scholar]
  50. Krizhevsky, A. Learning Multiple Layers of Features from Tiny Images; University of Toronto: Toronto, ON, USA, 2009. [Google Scholar]
  51. Netzer, Y.; Wang, T.; Coates, A.; Bissacco, A.; Wu, B.; Ng, A.Y. Reading Digits in Natural Images with Unsupervised Feature Learning. In Proceedings of the NIPS Workshop on Deep Learning and Unsupervised Feature Learning 2011, Granada, Spain, 16 December 2011; pp. 1–9. [Google Scholar]
  52. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. In Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); 8693 LNCS; Springer: Cham, Switzerland, 2014; pp. 740–755. [Google Scholar] [CrossRef] [Green Version]
  53. LeCun, Y.; Cortes, C.; Christopher, J.C. Burges. The MNIST Database of Handwritten Digits. 2010. Available online: http://yann.lecun.com/exdb/mnist/ (accessed on 30 June 2023).
  54. Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and Harnessing Adversarial Examples. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015—Conference Track Proceedings, San Diego, CA, USA, 7–9 May 2015; pp. 1–11. [Google Scholar]
  55. Wong, E.; Rice, L.; Kolter, J.Z. Fast Is Better than Free: Revisiting Adversarial Training. arXiv 2020, arXiv:2001.03994. [Google Scholar]
  56. Madry, A.; Makelov, A.; Schmidt, L.; Tsipras, D.; Vladu, A. Towards Deep Learning Models Resistant to Adversarial Attacks. In Proceedings of the 6th International Conference on Learning Representations, ICLR 2018—Conference Track Proceedings, Vancouver, BC, Canada, 30 April–3 May 2018; pp. 1–28. [Google Scholar]
  57. Zhang, H.; Yu, Y.; Jiao, J.; Xing, E.P.; El Ghaoui, L.; Jordan, M.I. Theoretically Principled Trade-off between Robustness and Accuracy. In Proceedings of the 36th International Conference on Machine Learning, ICML, Long Beach, CA, USA, 10–15 June 2019; pp. 12907–12929. [Google Scholar]
  58. Joyce, J.M. Kullback-Leibler Divergence. International Encyclopedia of Statistical Science; Springer: Berlin/Heidelberg, Germany, 2011; pp. 720–722. [Google Scholar] [CrossRef]
  59. Kurakin, A.; Goodfellow, I.J.; Bengio, S. Adversarial Examples in the Physical World. In Proceedings of the 5th International Conference on Learning Representations, ICLR 2017—Workshop Track Proceedings, Toulon, France, 24–26 April 2017; pp. 1–14. [Google Scholar]
Figure 1. Samples from the Kaggle Facial Keypoints Detection Dataset [47].
Figure 1. Samples from the Kaggle Facial Keypoints Detection Dataset [47].
Computers 12 00161 g001
Figure 2. Capsule Network architecture.
Figure 2. Capsule Network architecture.
Computers 12 00161 g002
Figure 3. The effectiveness of networks in the training process under different attack methods (a—no attack; b—FGSM; c—FFGSM; d—TPGD; e—PGD; f—GN; g—BIM; h—PGDL2).
Figure 3. The effectiveness of networks in the training process under different attack methods (a—no attack; b—FGSM; c—FFGSM; d—TPGD; e—PGD; f—GN; g—BIM; h—PGDL2).
Computers 12 00161 g003aComputers 12 00161 g003b
Table 1. Motor coaches, buses, and trolley buses: European Union—27 countries [5].
Table 1. Motor coaches, buses, and trolley buses: European Union—27 countries [5].
Year201320142015201620172018201920202021
Percentage (%)10.410.010.010.09.69.59.57.4N/D
Passenger-kilometers (million km)238,368230,616246,661240,574224,813225,958231,099135,427138,262
Table 2. Differences between a capsule and a neuron.
Table 2. Differences between a capsule and a neuron.
CapsuleNeuron
Input Vector   ( u i ) Scalar   ( x i )
Affine transform u ^ j | i = W i j u i -
Weighting s j = i c i j u ^ j | i α j = i w i x i + b
Nonlinear activation v j = s j 2 1 + s j 2 s j s j 2 h j = f ( a j )
Output Vector   ( v j ) Scalar   ( h j )
Table 3. The effectiveness of each network under different attack methods (the bold numbers mean the highest values in each column).
Table 3. The effectiveness of each network under different attack methods (the bold numbers mean the highest values in each column).
No AttackFGSMFFGSMTPGDPGDGNBIMPGDL2
Wide-ResNet0.97070.95610.95280.95840.96610.96610.96160.9655
SimpleCNN0.97530.96180.82480.96230.97320.97240.96890.9689
CapsNet0.97030.95820.95830.96020.96780.96910.96350.9643
CapsNet (own, current)0.97140.95930.95990.96130.97120.97060.96930.9698
Table 4. The rate of degradation of network efficiency for different attack methods (the bold numbers mean the lowest percentages in each column).
Table 4. The rate of degradation of network efficiency for different attack methods (the bold numbers mean the lowest percentages in each column).
FGSMFFGSMTPGDPGDGNBIMPGDL2
Wide-ResNet1.50%1.84%1.27%0.47%0.47%0.93%0.53%
SimpleCNN1.39%15.43%1.34%0.22%0.30%0.66%0.66%
CapsNet1.24%1.24%1.04%0.25%0.12%0.70%0.61%
CapsNet (own, current)1.25%1.19%1.04%0.03%0.09%0.23%0.16%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hollósi, J.; Ballagi, Á.; Kovács, G.; Fischer, S.; Nagy, V. Face Detection Using a Capsule Network for Driver Monitoring Application. Computers 2023, 12, 161. https://doi.org/10.3390/computers12080161

AMA Style

Hollósi J, Ballagi Á, Kovács G, Fischer S, Nagy V. Face Detection Using a Capsule Network for Driver Monitoring Application. Computers. 2023; 12(8):161. https://doi.org/10.3390/computers12080161

Chicago/Turabian Style

Hollósi, János, Áron Ballagi, Gábor Kovács, Szabolcs Fischer, and Viktor Nagy. 2023. "Face Detection Using a Capsule Network for Driver Monitoring Application" Computers 12, no. 8: 161. https://doi.org/10.3390/computers12080161

APA Style

Hollósi, J., Ballagi, Á., Kovács, G., Fischer, S., & Nagy, V. (2023). Face Detection Using a Capsule Network for Driver Monitoring Application. Computers, 12(8), 161. https://doi.org/10.3390/computers12080161

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop