Next Article in Journal
CNN-Based Volume Flow Rate Prediction of Oil–Gas–Water Three-Phase Intermittent Flow from Multiple Sensors
Next Article in Special Issue
A High-Accuracy and Power-Efficient Self-Optimizing Wireless Water Level Monitoring IoT Device for Smart City
Previous Article in Journal
Industrial Control under Non-Ideal Measurements: Data-Based Signal Processing as an Alternative to Controller Retuning
Previous Article in Special Issue
SoC FPGA Accelerated Sub-Optimized Binary Fully Convolutional Neural Network for Robotic Floor Region Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Intelligent Brushing Monitoring Using a Smart Toothbrush with Recurrent Probabilistic Neural Network

Machine Intelligence and Automation Technology Lab, Department of Computer Science & Information Engineering, National Central University, No. 300, Zhongda Rd., Zhongli Dist., Taoyuan City 320, Taiwan
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(4), 1238; https://doi.org/10.3390/s21041238
Submission received: 4 November 2020 / Revised: 10 January 2021 / Accepted: 12 January 2021 / Published: 10 February 2021
(This article belongs to the Special Issue Perceptual Deep Learning in Image Processing and Computer Vision)

Abstract

:
Smart toothbrushes equipped with inertial sensors are emerging as high-tech oral health products in personalized health care. The real-time signal processing of nine-axis inertial sensing and toothbrush posture recognition requires high computational resources. This paper proposes a recurrent probabilistic neural network (RPNN) for toothbrush posture recognition that demonstrates the advantages of low computational resources as a requirement, along with high recognition accuracy and efficiency. The RPNN model is trained for toothbrush posture recognition and brushing position and then monitors the correctness and integrity of the Bass Brushing Technique. Compared to conventional deep learning models, the recognition accuracy of RPNN is 99.08% in our experiments, which is 16.2% higher than that of the Convolutional Neural Network (CNN) and 21.21% higher than the Long Short-Term Memory (LSTM) model. The model we used can greatly reduce the computing power of hardware devices, and thus, our system can be used directly on smartphones.

1. Introduction

The occurrence of chronic illnesses is a very common phenomenon in society, such as high blood pressure, diabetes, heart disease, etc., in adults and dental caries, periodontal disease and Gingivitis in both children and adults [1]. Most of these dental diseases are the result of bacterial deposition on the surfaces of teeth [2]. If a tooth is not brushed properly, then bacteria will accumulate on its surface, forming plaque, destroying the outermost layer of the tooth (enamel) and triggering gingivitis, which can lead to dental caries and gum disease. Thoroughly cleaning teeth can effectively reduce tooth plaque and prevent oral diseases.
Brushing one’s teeth every day is the primary method to prevent various oral diseases. The American Dental Association (ADA) and the Taiwan Dental Association both recommend using the Bass Brushing Technique to brush your teeth and to do it at least twice a day for two minutes [3,4]. However, studies have pointed out that most people’s brushing time is insufficient, and the average brushing time per person is between 30 s and 60 s [5,6,7,8]. For a brushing time of less than two minutes, each tooth’s brushing time is not enough [9,10]. The Bass Brushing Technique involves the following: (1) One must brush the outer side of the front row of teeth, the outer side of the left and right rows of teeth, and the inner side of the left and right rows of teeth. The toothbrush is at a 45-degree angle with the gums. These areas are brushed vertically with the toothbrush. Each tooth brushing starts from the gum line, starting with the tip of the tooth. (2) When brushing the chewing surface of the tooth, move the bristles of the toothbrush back and forth along the chewing surface of the tooth. (3) When brushing the inside of the front row of teeth, tilt the toothbrush vertically and move it up and down.
According to forecasts from UK market research organizations, more than 400 million smart wearable devices will be sold by 2020, with an estimated value of $34 billion [11]. Inertial sensors have been widely used in different wearable devices, such as smart watches, sports bracelets, smart sports shoes, smart glasses, Bluetooth headsets, etc., that are all embedded with Inertial Measurement Unit (IMU). It can detect the motion state of the device and identify various motion postures in different fields [12]. In the field of healthcare, inertial sensor data can be used to monitor the onset and recovery of diseases, such as assessing the recovery of patients during rehabilitation [13,14,15] and detecting and diagnosing diseases [16,17]. In the field of home care, a care system can provide monitoring or assistance to the occupants to ensure their health, safety, and good physical condition, such as a tracking and monitoring emergency help system [18] to assist patients with physical and mental disabilities [19,20,21]. In the field of sports and leisure, wearable sensor devices can be used to identify people’s sports and leisure activities and improve their quality of life, such as daily action recognition [22,23] and motion pose recognition [24,25].
In many of the above areas, the application of inertial sensors in the field of healthcare is particularly noteworthy. These studies show that there is broad development in the use of inertial sensors for human activity recognition (HAR) in healthcare applications. Space, the busy lives of people, and the increase of the elderly population make healthcare an urgent problem, and through the integration of digital technology, the healthcare industry can reduce the cost of care for humans and financial services and improve the quality of healthcare. Unlike these applications, toothbrushes are used to clean teeth. Brushing includes a series of movements for manipulating the toothbrush. The toothbrush and the user’s movements are closely combined. Therefore, the data extracted by the sensor needs to be used to accurately identify the user. For the operation of the toothbrush, even after learning the Bass Brushing Technique, there will be some differences in the brushing movements of different users. Due to the small space in the mouth, the range of the toothbrush can be small, and the classification of the brushing needs many regions, making it difficult to accurately classify small-scale actions. Establishing accurate models and effective monitoring is a very challenging topic.
On the topic of brushing motion recognition, studies have looked at how to use manual toothbrushes and smart watches to monitor the brushing quality of all tooth surfaces [26] and to capture brush movements and directions through a magnetometer attached to the toothbrush handle and a magnetic sensor in the watch. Based on the inertial sensing data from the watch, the brushing posture can be recognized, and the sound signal collected from the watch is used to assist the recognition. This method can be used with a smart watch, and the average recognition rate is only 85.6%. J.W. et al. [27] of Korea’s Konkuk University employed a three-axis accelerometer on the bottom of the toothbrush to measure the user’s brushing posture. This method only utilizes a three-axis accelerometer to measure motion information, and it cannot stably represent all brushing movements. This way is easy to cause misjudgment. In addition, research [28] took an accelerometer, magnetic sensor, and Euler angle as identification features and used the K-means algorithm to identify and classify 15 brushing areas but could not meet the correctness and integrity of brushing. Each research method has its own room for improvement.
The rise of deep learning in recent years has made deep learning widely applied in sequence learning-related research, such as image recognition [29,30,31], speech recognition [32,33], and sensor data analysis [34,35]. There are many studies that have begun to employ deep learning methods to sensor motion recognition [36]. Traditional human body recognition methods, such as the Hidden Markov Model (HMM) [37], specifically deal with the relationship between many random variables in the sequence data. The hidden Markov model uses the graph probability model to describe the problem, which has the Markov property. The current state is conditionally independent of the past state, and so, it can only learn the shallower features [38], resulting in a reduced effectiveness in unsupervised learning and limited classification accuracy and model promotion. Generally speaking, static data can only be used as a learning basis, and the features need to be extracted manually. However, in real life, motion data usually present a dynamic sequence, which requires stable and continuous learning.
Different from traditional human motion recognition methods, deep learning does not need to extract features manually. It can input unmarked data as features. Since the data collected by sensor data are time series, some studies have proposed using convolution. The Convolutional Neural Network (CNN) and the Long Short-Term Memory (LSTM) model are used to identify sensor data [39,40]. This method does not require preprocessing of the data. It uses raw data directly as a deep neural network input.
Although deep learning has many of the above advantages, its immediacy and computing resources are still not suitable for embedded systems. Deep learning in a neural network architecture refers to a neural network model with multiple hidden layers. There are many layers and nodes, which contain many parameters and require a large amount of training data to adjust the parameters. The model complexity is extremely high. To develop an effective deep learning model within a reasonable time, it is necessary to use the acceleration function of the Graphics Processing Unit (GPU). Therefore, the hardware cost required for typical deep learning is high, the computing resources are expensive, and it is not easy to apply to smart toothbrushes. Aiming at the shortcomings of existing smart toothbrushes, such as high cost, low recognition accuracy, and a lack of personal adaptation, this research plans an innovative deep neural network classifier that combines smart toothbrushes with nine-axis motion sensors and visualization software on a mobile phone, which can be applied to monitor the correctness and completeness of the brushing process for protecting oral health.

2. Formalization of Brushing Posture

This study is based on the Bass Brushing Technique’s brushing action guidelines, which divide the teeth into 15 brushing areas to describe the brushing area and posture. The fifteen brush cleaning areas are first defined, as shown in Figure 1.
In order to identify the brushing posture, we define the carrier coordinate frame of the toothbrush. The origin is located at the center of gravity of the carrier (toothbrush). The directions of the X, Y, and Z axes are usually located directly in the forward and lateral directions of the carrier and directly below the carrier. As shown in Figure 2a, the gravitational acceleration (G = 1 g) is used as the reference vector to calculate the attitude angle, and the gravitational acceleration direction is defined as the direction of the Z-axis coordinate. In rigid kinematics, the acceleration of a particle moving in space can be compared with a fixed coordinate system and a motion coordinate. Therefore, we used the North-East-Down Coordinate System (NED) as the fixed coordinate and the carrier coordinate (the body frame) as the moving coordinate (Figure 2b). The features used in the brushing area recognition are then defined as the angle of the toothbrush’s steering and brushing. We first use turning the toothbrush (Figure 2c) to classify the brushing posture and then use the attitude angle as the recognition area of the brushing area.
When we attach the nine-axis sensor to the toothbrush, we can measure the gravitational acceleration of the toothbrush. We use the gravitational acceleration measured by the toothbrush and the rotation angle of the Euler angle as the recognition features of the toothbrush steering. The elevation angle and the roll angle formula are as follows:
P i t c h = ρ = sin 1 ( A y )
R o l l = γ = sin 1 ( A x cos ρ )
The brushing system takes the coordinates on the brush holder as the origin, and the sensor will automatically correct the sensor value of the toothbrush on the toothbrush holder to be close to zero each time the toothbrush is used. When the user grabs the toothbrush, the brushing is started in front of the toothbrush. From brushing gesture recognition, Table 1 defines the range of toothbrush steering characteristic values for this project.
The attitude angle can indicate the direction of motion of the object in the three-dimensional space. When the teeth of different regions are brushed, the attitude angle will also change, and so, the attitude angle can be used as the distinguishing feature of the brushing area. Figure 3 illustrates the posture of the toothbrush when brushing different areas with the Bass Brushing Technique. This study classifies the brushing area by the attitude angle characteristics of different areas when brushing teeth.

3. Deep Learning for Brushing Posture Recognition

In the feedforward neural network, data are transmitted in one feedforward direction. Each input can be regarded as independent to the preceding input, and so, the network output only depends on the current input data. However, in the real world, the time series sensor data are generally streaming and time-dependent. The characteristics of such data depend on time, and thus, the input of the neural network is not only related to the input of the current moment but, also, related to the past time period. The input is related to the output.

3.1. Deep Neural Network Architecture and Learning Algorithms

In order to design a neural network for brushing gesture recognition, we first explored three mainstream deep neural networks: Convolutional Neural Networks (CNN), recurrent neural networks (RNN), and Long- and Short-Term Memory Models (LSTM). We will use these deep learning network architectures to design innovative neural networks for smart toothbrushes.

3.1.1. Convolutional Neural Network

The Convolutional Neural Network (CNN) is the most well-known neural network model in deep learning. It has excellent performance in image recognition and is widely used for identification.
The CNN neural network architecture is mainly composed of a Convolutional Layer, a Pooling Layer, and a Fully Connected Layer. The convolutional layer is used to filter the original image or is called a kernel (convolution) to extract the features of the image. The pooling layer usually uses the Max Pooling method, which is mainly used to reduce the size of the feature matrix, cut down the number of calculation parameters, avoid over-fitting of the neural network (overfitting), and retain important feature information. Part of the fully connected layer is used to classify and output the features extracted by the previous layers of the network.

3.1.2. Recurrent Neural Network

A recurrent neural network (RNN) is a kind of deep neural network that recurrently transmits information in its own neural network and accepts the input of time series data structure. Hence, it can be used to describe the behavior of dynamic time. Its architecture is shown in Figure 4.
In Figure 4, X is the input data at a certain moment; O is the output data at a certain moment; H is the hidden state; U is the weight from the input layer to the hidden layer, and the original input is abstracted and sent to the hidden layer; V is the weight from the hidden layer to the output layer, further abstracting the representation learned from the hidden layer; W is the weight from hidden layer to hidden layer, responsible for controlling and scheduling the memory of the network and X t represents the input at time t. The data are returned to the neural network operation. The general flow is X t , and the parameters U, W and H t 1 are calculated to produce H t , which will be stored in the memory of the neuron, and H t and the parameter V will be calculated to produce O t .
The recurrent neural network has a feedback mechanism. The hidden layer H is connected with the hidden layer H at the previous moment. At time t , the value of H t is the hidden layer value H t 1 of the previous moment and the input value X t of the current moment. The function of the composition, F ( x ) , is the excitation function of the hidden layer.
The right side of Figure 4 is the unfolding part of the hidden layer H , O 0 L ,   O t   is the label sequence, X 0 L ,     X t is the input sequence, O t is estimated by compressing all historical information in the past in H , and the parameters of the recurrent neural network are in the sequence data time. Sharing parameters can make the model less complex and present better promotion.

3.1.3. Long Short-Term Memory Model

In 1997 the recurrent neural network containing Long Short-Term Memory (LSTM) [41,42,43,44,45] was proposed by German researchers Sepp Hochreiter and Juergen Schmidhuber. The road evolved and was improved and promoted by Alex Graves in recent years. Although it is theoretically possible to retransmit the neural network to the long-term correlation of data sequences, the recurrent neural network (RNN) is prone to the problem of gradient disappearance. When the gradient disappears, the neural network only learns the short-term dependence of the data sequence—that is, when the number of layers of the neural network becomes greater, the deeper the hidden layer nodes are and the shallower the hidden layer nodes. With the perceived ability declining, LSTM adds a core element memory cell (Cell) to solve the problem of the disappearance of the recurrent neural network (RNN) gradient. Each memory in LSTM contains an input gate, output gate, forget gate, and LSTM block unit, as shown in Figure 5.
LSTM uses memory to strengthen the current decision and uses three control gates to determine the storage and use of memory. In addition to the predicted output, a memory branch is added, which is updated over time, and the current memory is represented by the Gt symbol. “Forget Gate” and “Input Gate” decide whether to update the memory.
Input Gate:
Input Gate (indicated by i t ): determines whether the current input and the newly generated memory cell are added to the long-term memory. Data can be written to memory cells when the input gate is turned on via a sigmoid transfer function:
i t = S i g m o i d ( P X )
Multiply i t and C u p d a t e to get i t   C u p d a t e .
If i t = 0 , then C u p d a t e means that the memory cells cannot be written; i t = 1 means C u p d a t e cannot write to the memory cells.
Output Gate:
Output Gate: determines whether the current word sentence is added to the output. This valve is also a Sigmoid function, indicating whether to add it or not and decide whether to read out the values in the memory cells, as in Formula (4):
O t = S i g m o i d ( R X )
Multiply A and B to get:
A = h t O t
If O t = 0 , then B cannot be read through the Output Gate; O t = 1   means that h t can be read through the Output Gate.
Forget Gate:
The Forget Gate is denoted by f . If the current sentence is a new topic or the opposite of the previous sentence, then the previous sentence will be filtered out by this valve; otherwise, it may continue to be retained in memory. Determine when you want to forget the cell memory:
f t = S i g m o i d ( Q X ) C t
f t and C t 1   are   multiplied to get f t   C t 1 .
For whether the long-term memory is added to the output (Output), the tanh function is usually used. The value falls between [–1, 1], and the −1 table removes the long-term memory. If f t = 0 , then the Forget Gate is closed, and C t 1 with the previous memory unit will be deleted; if f t = 1 , then the Forget Gate is enabled, and C t 1 with the previous memory unit will be deleted.
Memory cell update:
C t = f t C t 1 + i t C u p d a t e
C t   is the value of the most recent memory cell, and H ( x ) usually uses the Activation Function.
This study combines the nine-axis inertial sensing signal with the Euler angle eigenvalue as the input of the CNN and LSTM neural network. The weight of the CNN and LSTM neural network is then trained, and the long time series data of the LSTM network is established through the LSTM cyclic memory unit. This trained model can be used to predict the posture and position of brushing.

4. Recurrent Probabilistic Neural Network

With references to the concept and formalism of recurrent neural networks (RNN) and the Long-Term and Short-Term Memory (LSTM) model, when combined with the inference mechanism of a probabilistic neural network, we designed a recurrent probabilistic neural network suitable for toothbrush attitude recognition to support the Bass brushing technique.
We borrowed the long-term and short-term memory cell concept of RNN and LSTM, so that the signal input to the probabilistic neural network (PNN) can be feedback and remembered in the network. This allows the recurrent probabilistic neural network (RPNN) to correctly identify the time-dependent continuous brushing posture.
In 1988, DF Specht proposed the probabilistic neural network (PNN) [46]. PNN is a four-layer neural network architecture. Probabilistic neural networks are widely used, such as for object tracking and imaging. There are also related applications in the field of processing [47,48,49]. PNN belongs to the feedforward neural network architecture. The main theoretical basis is based on Bayesian classifiers.
The probabilistic neural network architecture is shown in Figure 6, which is the input layer, the hidden layer, the sum layer, and the output layer. The feature vector X = { X 1 , X 2 , , X N } of the probability-based neural network input layer can be any custom feature for classification, and the hidden layer is marked. The characteristic data, which are the characteristics of the input classification, are used to count the probability values of each classification through the summation unit. The summation unit corresponds to the output classification of the output layer. The number of output neurons of the output layer is the same as the number of classifications, and the output of the output layer is calculated by the summation unit. Therefore, the possibility of being classified is the highest.
The advantage of the probabilistic neural network is that its input vector size and type are not limited. It can be widely used in different types of problems. When the system is facing environmental changes and needs to add new training materials, only the newly marked training is needed. Data are added to the network or added to the corresponding new classification weights. There is no need to change the overall network architecture like other types of neural network architectures, and the weights are retrained through an iterative process. Therefore, network learning is very fast and suitable for use in real-time systems.
We use the probabilistic neural network as the core model of the neural network to identify the gesture of brushing. Combined with the RNN and LSTM models, we form a recurring probabilistic neural network, which allows the probabilistic neural network to have long-term and short-term memory functions. A continuous brushing motion sensing signal is taken from the toothbrush, and the recurrent probabilistic neural network continuously recognizes the brushing area by the motion attitude angle.
The recurrent probabilistic neural network we proposed is shown in Figure 7. The white square is the memory neuron. The output probability value is output as the neural network weight through the excitation function in the memory neuron. The memory unit updates the equation, such as:
P i _ u p d a t e ( t ) = ( 1 δ ) ( P i _ u p d a t e ( t 1 ) + P i * ( t ) ) ,   i f   P i ( t )   =   max i P ( t ) P i _ u p d a t e ( t ) = δ ( P i _ u p d a t e ( t 1 ) + P i * ( t ) ) , o t h e r w i s e
and
P i * ( t ) = { K P i ( t ) , i f t 0 0 , o t h e r w i s e
where δ is the forgetting factor of the memory unit, K is a parameter of the length of the memory, P ( t )   is the probability of dividing into the time, and P i u p d a t e ( t )   s the output value after the memory cell is updated.
The parameter δ in the probabilistic neural network is the Gaussian function smoothing coefficient. Each training sample can be regarded as a Gaussian function in the multidimensional feature space. The smoothness coefficient δ determines the breadth of its distribution. The larger the δ 2 value is, the wider the distribution is, and the higher the noise is that can be tolerated. The smaller the δ 2 value is, the narrower the distribution, the lower is the noise that can be tolerated, and δ is classified according to different classifications.
In order to make the neural network adaptive, we used the particle swarm optimization (PSO) algorithm to adjust the   δ parameters and the δ and K parameters in the recurrent neural unit to model the accuracy as a particle. The fitness function of the group optimization algorithm iterates the particles in the search space to the optimal solution in order to obtain a robust recognition performance.
Model migration.
The probabilistic neural network does not require a complicated training process to adjust the overall neural network architecture. However, when using the probabilistic neural network for identification, in order to improve the accuracy of model identification, the model is often complicated and large, so that the identification consumes a large amount of memory, resources, and computing time. Therefore, we used model migration to define the data directly related to the target task, called the Target Data. The data that were not directly related to the target task were called the Source Data, and the source data in the pretraining model belonged to the mark. The data X s , Y s , and the user’s brushing posture feature belonged to the unmarked target data X t . Since each user’s brushing posture is closely related to the user’s own brushing habits, each user’s brushing method may be slightly different. We sourced the identification model of the data domain that was transformed into the identification model of the target domain, as shown in Figure 8. It effectively reduced the complexity and calculation time of the model and improved the identification accuracy.

5. Evaluation of Hardware/Software Integrated Smart Toothbrush

We developed a smart toothbrush prototype system (Figure 9) to verify the identification performance of the proposed RPNN model. A nine-axis inertial sensor MPU-9255 and Bluetooth 4.2 module were integrated into the toothbrush. The inertial sensor was used to capture the brushing motion signal, and the Bluetooth signal was transmitted to the mobile phone instantly. On the mobile phone end, we designed software to implement RPNN-based Brushing Posture Recognition and included a graphical user interface for monitoring the integrity of Bayesian brushing.
We set up brushing data for at least 15 brushing people. The toothbrush holder was 30 to 80 cm away from the subject. Each subject was sitting in front of the toothbrush holder. After watching the shell brushing method, the inertial sensing was performed. After the calibration was completed, the subject picked up the toothbrush from the toothbrush holder and brushed the teeth at the start of the prescribed brushing sequence. When brushing the teeth, the nine-axis data continuously input were filtered by the Kalman filter, and then, the data were processed by the quaternion algorithm. The fusion produced the Euler angle and, finally, extracted 5000 strokes of continuous brushing Euler angles in 15 regions. It collected five data archives in each region.
This database was used to train and test the recurrent probabilistic neural network proposed in this project to evaluate the performance and efficiency of brush gesture recognition. Next, we defined the performance evaluation indicators for brushing posture recognition in machine learning (ML), information retrieval (IR), accuracy, precision, recall, and F1 (F1-Measure). It is widely used to evaluate the pros and cons of different algorithms and models. Before understanding the above evaluation methods, it was necessary to define true positive (TP), true negative (TN), false positive (FP), and false negative (FN) as the four classifications of the dichotomy, as shown in Table 2.
  • Accuracy:
It represents the ratio of the number of samples that the classifier model can correctly classify to the total number of samples for a given test dataset.
a c c u r a c y = T P + T N T P + F P + F N + T N
  • Precision:
It represents the proportion of information that is correctly retrieved as a percentage of the material that is actually retrieved.
p r e c i s i o n = T P T P + F P
  • Recall:
It represents the proportion of information that is correctly retrieved as a percentage of the material that should actually be retrieved.
r e c a l l = T P T P + F N
  • F1-Measure:
In some cases, the Precision and Recall values are contradictory. Therefore, in order to comprehensively evaluate the Precision and Recall, the most widely used method in the field of machine learning is the F-Measure. This method is a weighted average of Precision and Recall and is also known as the F-Score.
F = ( a 2 + 1 ) p r e c i s i o n r e c a l l a 2 ( p r e c i s i o n + r e c a l l )
The most common F1-Measure is when a value is 1, calculated as follows:
F 1 = 2 p r e c i s i o n r e c a l l p r e c i s i o n + r e c a l l
Many studies have pointed out that, due to insufficient brushing time, the chance of dental caries and oral diseases is higher than that between brushing teeth. Most people use insufficient brushing time compared to Beller’s brushing method, which requires at least two minutes.
The plan will also assess the integrity of the brushing of each subject. The formula for assessing the integrity of the brushing is as follows:
Brushing   completion = U s e r   b r u s h i n g   t i m e   i n   t h e   o r a l   a r e a D e n t i s t   s t i p u l a t e s   b r u s h i n g   t i m e   i n   t h e   o r a l   a r e a × 100 %
The experimental results will be compared to a typical CNN, recurrent neural networks, and LSTM identification results. We anticipate that the identification accuracy will go beyond the existing deep learning methods described above, while the computing time and hardware resource usage will be much lower than these existing methods.

6. Experiments

Experimental data were collected from five testers (user #1 to user #5). The brushing data were collected several times. Each brushing area collected 2000 training materials. The total training data totaled 150,000 pens, and there were 1000 more tests in each brushing area data.
We conducted comparative experiments of brushing posture recognition with a Convolutional Neural Network (CNN), Recurrent Neural Network (RNN), and posed recurrent probabilistic neural network (RPNN). In order to process the time series data, we used Multi-Scale Convolutional Neural Networks (MCNN) [50] as the CNN model, and LSTM [51], proposed by Yuwen Chen et al., as the recurrent neural network model. To compare the three deep learning models, all identification experiments were performed on the same PC (Intel Core i7-6770 CPU @ 3.40 GHz with 16.00 GB DDR3 2133 MHz).
The MCNN architecture consisted of four layers, encompassing two layers of convolutional layers and two layers of fully connected layers. The hyperparameters were set as shown in Table 3. The LSTM hyperparameter settings are shown in Table 4.
The DPNN model was optimized using PSO. We used PSO parameters with the inertia weight (W) initial value, social parameter (C1), cognitive coefficient (C2), number of particles, and iteration, such as in Table 5.
The experimental results are shown in Table 6 and Table 7.
The recognition accuracy of the RPNN model can reach 99.08%, and the average recognition rate is 16.2% higher than that of the CNN model. It is also 21.21% higher than the LSTM model, which greatly improves the recognition of the brushing area.
Typical deep neural network models such as the CNN and LSTM network architectures are complex, have many parameters, and are computationally intensive, requiring the use of better performing processors. The RPNN model is only one-thousandth and one-fiftieth the size of CNN and LSTM. The model parameters are also much lower than the CNN and LSTM, which can greatly save memory usage. Moreover, the real-time parameter is satisfied due to the lower computational load. Under the demand of brushing posture recognition, RPNN is more suitable for implementing edge devices with less hardware resources.

7. Conclusions

The existing smart toothbrush has insufficient accuracy and stability of posture recognition, and it is difficult to provide the user with correct information on the correctness and completeness of brushing. Therefore, this paper proposes a brushing attitude recognition model based on deep learning, which is applied to a smart toothbrush and can support the monitoring of the Bass Brushing Technique. Based on the brushing motion criterion of the Bass Brushing Technique, this study divided the teeth into 15 brushing areas and then defined the attitude angle of the toothbrush turning and brushing as the characteristics of the brushing area identification.
We used three deep neural networks for brushing gesture recognition in this study: Convolutional Neural Network (CNN) model MCNN for the time series data, the recurrent neural network with Long-Term and Short-Term Memory (RNN) Model LSTM, and the recurrent probabilistic neural network RPNN. The RPNN model used the probabilistic neural network as the core model of the neural network to identify the brush gesture. Referring to the long-term and short-term memory functions of the LSTM model, the probabilistic neural network has the ability to recognize continuous brushing motions and to recognize the conspicuous brushing areas.
Since each user’s brushing posture is closely related to the user’s own brushing inertia, employing a fixed identification model will make it difficult for users to deal with different brushing habits. In order to improve the user’s personalized brushing posture recognition, we propose a RPNN model migration method. The identification model of the source data domain (generalized model) is transformed into the identification model of the personalized domain, and the user model is trained by the PSO algorithm to make the model more suitable for personalized brushing habits, which can effectively improve identification accuracy. It also reduces the complexity of the model and computation time, enabling low-cost edge devices.
Compared with the deep learning neural network models CNN and LSTM, the results show that the recognition accuracy of the RPNN model can reach 99.08%, in which the average recognition rate is 16.2% higher than that of the CNN model and 21.21% higher than the LSTM model. The method proposed by this study can be adapted to implement edge devices with low hardware resources, such as smartphones lacking AI accelerators for instant brushing gesture recognition, while providing higher discrimination accuracy to ensure proper brushing during monitoring.

Author Contributions

Y.-Z.C. and C.-C.W. designed the framework of the model and carried out the experiment in this case study, which is a part of thesis work. C.-H.C. is the thesis director; he supervised the thesis work including the work in this paper and he also revised the paper on several occasions. C.-C.W. supervised the model development and the different data processing in this paper. Y.-Z.C. responsible for the data acquisition and C.-H.C. also revised the paper on several occasions. All authors have read and agreed to the published version of the manuscript.

Funding

This research comes from the Ministry of Science and Technology of Taiwan’s Industry-Academy Project (MOST), the project number is MOST105-2745-8-008-001-, to promote the development of Taiwan’s traditional industries, combined with the application of deep learning.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Written informed consent has been obtained from the patient(s) to publish this paper.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to confidentiality agreement with the company.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Eke, P.I.; Dye, B.; Wei, L.; Thornton-Evans, G.; Genco, R. Prevalence of periodontitis in adults in the United States: 2009 and 2010. J. Dent. Res. 2012, 91, 914–920. [Google Scholar] [CrossRef]
  2. American Dental Association (ADA) Division of Science. Tackling tooth decay. J. Am. Dent. Assoc. 2013, 144, 336. [Google Scholar] [CrossRef]
  3. Gibson, J.; Wade, A.B. Plaque removal by the Bass and Roll brushing techniques. J. Periodontol. 1977, 48, 456–459. [Google Scholar] [CrossRef]
  4. Poyato-Ferrera, M.; Segura-Egea, J.; Bullón-Fernández, P. Comparison of modified Bass technique with normal toothbrushing practices for efficacy in supragingival plaque removal. Int. J. Dent. Hyg. 2003, 1, 110–114. [Google Scholar] [CrossRef]
  5. der Weijden Van, G.; Timmerman, M.; Nijboer, A.; Lie, M.; der Velden Van, U. A comparative study of electric toothbrushes for the effectiveness of plaque removal in relation to toothbrushing duration. Timerstudy. J. Clin. Periodontol. 1993, 20, 476–481. [Google Scholar] [CrossRef]
  6. Van der Weijden, F.; Timmerman, M.; Snoek, I.; Reijerse, E.; Van der Velden, U. Toothbrushing duration and plaque removing efficacy of electric toothbrushes. Am. J. Dent. 1996, 9, S31–S36. [Google Scholar]
  7. Saxer, U.; Barbakow, J.; Yankell, S. New studies on estimated and actual toothbrushing times and dentifrice use. J. Clin. Dent. 1998, 9, 49–51. [Google Scholar]
  8. Beals, D.; Ngo, T.; Feng, Y.; Cook, D.; Grau, D.; Weber, D. Development and laboratory evaluation of a new toothbrush with a novel brush head design. Am. J. Dent. 2000, 13, 5A–14A. [Google Scholar]
  9. Creeth, J.E.; Gallagher, A.; Sowinski, J.; Bowman, J.; Barrett, K.; Lowe, S.; Patel, K.; Bosma, M.L. The effect of brushing time and dentifrice on dental plaque removal in vivo. J. Dent. Hyg. 2009, 83, 111–116. [Google Scholar]
  10. McCracken, G.; Janssen, J.; Swan, M.; Steen, N.; de Jager, M.; Heasman, P. Effect of brushing force and time on plaque removal using a powered toothbrush. J. Clin. Periodontol. 2003, 30, 409–413. [Google Scholar] [CrossRef]
  11. CCS Insight. Wearables Momentum Continues. 2016. Available online: http://www.ccsinsight.com/press/company-news/2516-wearables-momentum-continues (accessed on 17 February 2016).
  12. Avci, A.; Bosch, S.; Marin-Perianu, M.; Marin-Perianu, R.; Havinga, P. Activity recognition using inertial sensing for healthcare, wellbeing and sports applications: A survey. In Proceedings of the 23th International Conference on Architecture of Computing Systems 2010, Hannover, Germany, 22–23 February 2010; pp. 1–10. [Google Scholar]
  13. Bidargaddi, N.; Sarela, A.; Klingbeil, L.; Karunanithi, M. Detecting walking activity in cardiac rehabilitation by using accelerometer. In Proceedings of the 2007 3rd International Conference on Intelligent Sensors, Sensor Networks and Information, Melbourne, QLD, Australia, 3–6 December 2007; pp. 555–560. [Google Scholar]
  14. Chen, K.-H.; Chen, P.-C.; Liu, K.-C.; Chan, C.-T. Wearable sensor-based rehabilitation exercise assessment for knee osteoarthritis. Sensors 2015, 15, 4193–4211. [Google Scholar] [CrossRef] [Green Version]
  15. Tognetti, A.; Lorussi, F.; Bartalesi, R.; Quaglini, S.; Tesconi, M.; Zupone, G.; de Rossi, D. Wearable kinesthetic system for capturing and classifying upper limb gesture in post-stroke rehabilitation. J. Neuroeng. Rehabil. 2005, 2, 8. [Google Scholar] [CrossRef] [PubMed]
  16. Wu, W.H.; Bui, A.A.; Batalin, M.A.; Au, L.K.; Binney, J.D.; Kaiser, W.J. MEDIC: Medical embedded device for individualized care. Artif. Intell. Med. 2008, 42, 137–152. [Google Scholar] [CrossRef] [PubMed]
  17. Jiang, S.; Cao, Y.; Iyengar, S.; Kuryloski, P.; Jafari, R.; Xue, Y.; Bajcsy, R.; Wicker, S. CareNet: An integrated wireless sensor networking environment for remote healthcare. In Proceedings of the ICST 3rd International Conference on Body Area Networks, Tempe, AZ, USA, 13–15 March 2008. [Google Scholar]
  18. Hou, J.C.; Wang, Q.; AlShebli, B.K.; Ball, L.; Birge, S.; Caccamo, M.; Cheah, C.-F.; Gilbert, E.; Gunter, C.A.; Gunter, E.; et al. Pas: A wireless-enabled, sensor-integrated personal assistance system for independent and assisted living. In Proceedings of the IEEE Joint Workshop on High Confidence Medical Devices, Software, and Systems and Medical Device Plug-and-Play Interoperability, Boston, MA, USA, 25–27 June 2007; pp. 64–75. [Google Scholar]
  19. Osmani, V.; Balasubramaniam, S.; Botvich, D. Self-organising object networks using context zones for distributed activity recognition. In Proceedings of the ICST 2nd International Conference on Body Area Networks, Florence, Italy, 11–13 June 2007. [Google Scholar]
  20. Bosch, S.; Marin-Perianu, M.; Marin-Perianu, R.; Havinga, P.; Hermens, H. Keep on moving! Activity monitoring and stimulation using wireless sensor networks. In Proceedings of the 4th European Conference on Smart Sensing and Context, Guildford, UK, 16–18 September 2009. [Google Scholar]
  21. Marshall, A.; Medvedev, O.; Markarian, G. Self management of chronic disease using mobile devices and Bluetooth monitors. In Proceedings of the ICST 2nd International Conference on Body Area Networks, Florence, Italy, 11–13 June 2007. [Google Scholar]
  22. Ermes, M.; Pärkkä, J.; Mäntyjärvi, J.; Korhonen, I. Detection of daily activities and sports with wearable sensors in controlled and uncontrolled conditions. Ieee Trans. Inf. Technol. Biomed. 2008, 12, 20–26. [Google Scholar] [CrossRef] [PubMed]
  23. Long, X.; Yin, B.; Aarts, R.M. Single-accelerometer-based daily physical activity classification. In Proceedings of the 2009 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Minneapolis, MN, USA, 3–6 September 2009; pp. 6107–6110. [Google Scholar]
  24. Heinz, E.A.; Kunze, K.S.; Gruber, M.; Bannach, D.; Lukowicz, P. Using wearable sensors for real-time recognition tasks in games of martial arts-an initial experiment. In Proceedings of the 2006 IEEE Symposium on Computational Intelligence and Games, Reno, NV, USA, 22–24 May 2006; pp. 98–102. [Google Scholar]
  25. Markus, H.; Takafumi, H.; Sarah, N.; Sakol, T. Chi-ball, an interactive device assisting martial arts education for children. In Proceedings of the ACM Conference on Extended Abstracts on Human Factors in Computing Systems, Fort Lauderdale, FL, USA, 5–10 April 2003; pp. 962–963. [Google Scholar]
  26. Huang, H.; Lin, S. Toothbrushing monitoring using wrist watch. In Proceedings of the 14th ACM Conference on Embedded Network Sensor Systems CD-ROM, Stanford, CA, USA, 14–16 November 2016; pp. 202–215. [Google Scholar]
  27. Lee, J.-W.; Lee, K.-H.; Kim, K.-S.; Kim, D.-J.; Kim, K. Development of smart toothbrush monitoring system for ubiquitous healthcare. In Proceedings of the 2006 International Conference of the IEEE Engineering in Medicine and Biology Society, New York, NY, USA, 30 August–3 September 2006; pp. 6422–6425. [Google Scholar]
  28. Lee, Y.-J.; Lee, P.-J.; Kim, K.-S.; Park, W.; Kim, K.-D.; Hwang, D.; Lee, J.-W. Toothbrushing region detection using three-axis accelerometer and magnetic sensor. Ieee Trans. Biomed. Eng. 2012, 59, 872–881. [Google Scholar]
  29. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2012, 60, 84–90. [Google Scholar] [CrossRef]
  30. Bengio, Y.; Lamblin, P.; Popovici, D.; Larochelle, H. Greedy layer-wise training of deep networks. In Proceedings of the Twentieth Annual Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 4–7 December 2006; pp. 153–160. [Google Scholar]
  31. Hinton, G.E.; Osindero, S.; Teh, Y.-W. A fast learning algorithm for deep belief nets. Neural Comput. 2006, 18, 1527–1554. [Google Scholar] [CrossRef]
  32. Mohamed, A.-r.; Dahl, G.E.; Hinton, G. Acoustic modeling using deep belief networks. Ieee Trans. Audio Speech Lang. Process. 2012, 20, 14–22. [Google Scholar] [CrossRef]
  33. Dahl, G.E.; Yu, D.; Deng, L.; Acero, A. Context-dependent pre-trained deep neural networks for large-vocabulary speech recognition. IEEE Trans. Audio Speech Lang. Process. 2012, 20, 30–42. [Google Scholar] [CrossRef] [Green Version]
  34. Yu, Z.; Rennong, Y.; Guillaume, C.; Maoguo, G. Deep Residual Bidir-LSTM for Human Activity Recognition Using Wearable Sensors. arXiv 2017, arXiv:1708.08989. [Google Scholar]
  35. Li, K.; Zhao, X.; Bian, J.; Tan, M. Sequential learning for multimodal 3D human activity recognition with Long-Short Term Memory. In Proceedings of the 2017 IEEE International Conference on Mechatronics and Automation (ICMA), Takamatsu, Japan, 6–9 August 2017; pp. 1556–1561. [Google Scholar]
  36. Wang, J.; Chen, Y.; Hao, S.; Peng, X.; Hu, L. Deep learning for sensor-based activity recognition: A survey. arXiv 2017, arXiv:1707.03502. [Google Scholar] [CrossRef] [Green Version]
  37. Lara, O.D.; Labrador, M.A. A survey on human activity recognition using wearable sensors. IEEE Commun. Surv. Tutor. 2013, 15, 1192–1209. [Google Scholar] [CrossRef]
  38. Yang, J.; Nguyen, M.N.; San, P.P.; Li, X.; Krishnaswamy, S. Deep Convolutional Neural Networks on Multichannel Time Series for Human Activity Recognition. In Proceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence (IJCAI 2015), Buenos Aires, Argentina, 25–31 July 2015; pp. 3995–4001. [Google Scholar]
  39. Ordóñez, F.J.; Roggen, D. Deep convolutional and lstm recurrent neural networks for multimodal wearable activity recognition. Sensors 2016, 16, 115. [Google Scholar] [CrossRef] [Green Version]
  40. Zimmermann, T.; Taetz, B.; Bleser, G. IMU-to-Segment Assignment and Orientation Alignment for the Lower Body Using Deep Learning. Sensors 2018, 18, 302. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Hu, C.; Wu, Q.; Li, H.; Jian, S.; Li, N.; Lou, Z. Deep learning with a long short-term memory networks approach for rainfall-runoff simulation. Water 2018, 10, 1543. [Google Scholar] [CrossRef] [Green Version]
  42. Le, X.H.; Ho, H.V.; Lee, G.; Jung, S. Application of long short-term memory (LSTM) neural network for flood forecasting. Water 2019, 11, 1387. [Google Scholar] [CrossRef] [Green Version]
  43. Zhang, J.; Wang, P.; Yan, R.; Gao, R.X. Long short-term memory for machine remaining life prediction. J. Manuf. Syst. 2018, 48, 78–86. [Google Scholar] [CrossRef]
  44. Kratzert, F.; Klotz, D.; Brenner, C.; Schulz, K.; Herrnegger, M. Rainfall–runoff modelling using long short-term memory (LSTM) networks. Hydrol. Earth Syst. Sci. 2018, 22, 6005–6022. [Google Scholar] [CrossRef] [Green Version]
  45. Fischer, T.; Krauss, C. Deep learning with long short-term memory networks for financial market predictions. Eur. J. Oper. Res. 2018, 270, 654–669. [Google Scholar] [CrossRef] [Green Version]
  46. Kusy, M.; Kowalski, P.A. Weighted probabilistic neural network. Inf. Sci. 2018, 430, 65–76. [Google Scholar] [CrossRef]
  47. Shree, N.V.; Kumar, T.N.R. Identification and classification of brain tumor MRI images with feature extraction using DWT and probabilistic neural network. Brain Inform. 2018, 5, 23–30. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  48. Zhu, H.; Lu, L.; Yao, J.; Dai, S.; Hu, Y. Fault diagnosis approach for photovoltaic arrays based on unsupervised sample clustering and probabilistic neural network model. Sol. Energy 2018, 176, 395–405. [Google Scholar] [CrossRef]
  49. Jiang, Q.; Shen, Y.; Li, H.; Xu, F. New fault recognition method for rotary machinery based on information entropy and a probabilistic neural network. Sensors 2018, 18, 337. [Google Scholar] [CrossRef] [Green Version]
  50. Liu, Y.; Guan, Q.; Zhao, X.; Cao, Y. Image Forgery Localization based on Multi-Scale Convolutional Neural Networks. In Proceedings of the 6th ACM Workshop on Information Hiding and Multimedia Security—IH&MMSec ’18, Innsbruck, Austria, 20–22 June 2018; pp. 85–90. [Google Scholar]
  51. Kim, T.; Kim, H.Y. Forecasting stock prices with a feature fusion LSTM-CNN model using different representations of the same data. PLoS ONE 2019, 14, e0212320. [Google Scholar] [CrossRef]
Figure 1. Fifteen brush cleaning areas.
Figure 1. Fifteen brush cleaning areas.
Sensors 21 01238 g001
Figure 2. Brushing coordinates: (a) carrier coordinate system, (b) fixed coordinates and carrier coordinates, and (c) toothbrush steering.
Figure 2. Brushing coordinates: (a) carrier coordinate system, (b) fixed coordinates and carrier coordinates, and (c) toothbrush steering.
Sensors 21 01238 g002
Figure 3. Brushing area and brushing postures. According to the Bass Brushing Technique, the brushing area proceeds from (ao): outer incisors; upper right outer; lower right outer; upper left outer; lower left outer; upper right flank; upper incisor; upper left flank; lower right flank; lower incisor; lower left flank; upper right inner; lower right inner; upper left inner; lower left inner.
Figure 3. Brushing area and brushing postures. According to the Bass Brushing Technique, the brushing area proceeds from (ao): outer incisors; upper right outer; lower right outer; upper left outer; lower left outer; upper right flank; upper incisor; upper left flank; lower right flank; lower incisor; lower left flank; upper right inner; lower right inner; upper left inner; lower left inner.
Sensors 21 01238 g003
Figure 4. Recurrent neural network.
Figure 4. Recurrent neural network.
Sensors 21 01238 g004
Figure 5. Long- Short-Term Memory (LSTM) block.
Figure 5. Long- Short-Term Memory (LSTM) block.
Sensors 21 01238 g005
Figure 6. Probabilistic neural network architecture.
Figure 6. Probabilistic neural network architecture.
Sensors 21 01238 g006
Figure 7. Recurrent-probabilistic-neural-network-model.
Figure 7. Recurrent-probabilistic-neural-network-model.
Sensors 21 01238 g007
Figure 8. Brushing gesture recognition model migration.
Figure 8. Brushing gesture recognition model migration.
Sensors 21 01238 g008
Figure 9. Smart toothbrush prototype system.
Figure 9. Smart toothbrush prototype system.
Sensors 21 01238 g009
Table 1. Range of toothbrush steering characteristic values.
Table 1. Range of toothbrush steering characteristic values.
Brush face-up a z > 0.8   or   45 > r o l l > 45
Brush face-down a z < 0 . 8   or   180 > r o l l > 150
Brush faced to the left a x > 0.8   or 150 < p i t c h < 45
Brush faced to the right a x < 0.8   or   45 < p i t c h < 150
Table 2. Binary problem definition. TP: true positive, FP: false positive, FN: false negative, and TN: true negative.
Table 2. Binary problem definition. TP: true positive, FP: false positive, FN: false negative, and TN: true negative.
RelevantNonrelevant
RetrievedTPFP
Not RetrievedFNTN
Table 3. Multi-Scale Convolutional Neural Networks (MCNN) hyperparameter settings.
Table 3. Multi-Scale Convolutional Neural Networks (MCNN) hyperparameter settings.
DropoutLearning
Rate
IterationsBatch
Size
1.50.0010020,00064
Table 4. Long Short-Term Memory (LSTM) hyperparameter settings.
Table 4. Long Short-Term Memory (LSTM) hyperparameter settings.
Hidden
Layer
Learning
Rate
Lambda
Loss
IterationsBatch
Size
450.00150.001512,000,0008500
Table 5. Particle swarm optimization (PSO) parameter settings.
Table 5. Particle swarm optimization (PSO) parameter settings.
Parameter ItemWC1C2Particle DimensionNumber of ParticlesNumber of Iterations
Initial value0.302.02.0350100
Table 6. Performance evaluation of brushing posture recognition. RPNN: recurrent probabilistic neural network.
Table 6. Performance evaluation of brushing posture recognition. RPNN: recurrent probabilistic neural network.
UserModelTesting
Accuracy (%)
Precision (%)Recall (%)f1_Score (%)
#1CNN91.393394.519197.921496.19017
LSTM84.393378.776984.393381.5851
RPNN98.006798.112098.010098.0593
#2CNN80.900081.434980.981.1666
LSTM74.720076.277074.7275.4985
RPNN99.940099.940099.940099.9400
#3CNN79.540078.521879.540079.0276
LSTM75.240076.337575.240075.7888
RPNN99.306799.337699.306799.3221
#4CNN81.653384.661387.485786.0504
LSTM70.960073.414270.960072.1871
RPNN99.520099.534799.520099.5273
#5CNN80.213383.454985.942981.7891
LSTM84.053383.065684.053383.5565
RPNN98.620098.777098.620098.6984
Table 7. Comparison of the computing resources of the three deep learning models.
Table 7. Comparison of the computing resources of the three deep learning models.
Model SizeNumber of ParametersAverage Accuracy
CNN285 KB337182.88
LSTM560 KB33,63077.87
RPNN5 KB45399.08
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chen, C.-H.; Wang, C.-C.; Chen, Y.-Z. Intelligent Brushing Monitoring Using a Smart Toothbrush with Recurrent Probabilistic Neural Network. Sensors 2021, 21, 1238. https://doi.org/10.3390/s21041238

AMA Style

Chen C-H, Wang C-C, Chen Y-Z. Intelligent Brushing Monitoring Using a Smart Toothbrush with Recurrent Probabilistic Neural Network. Sensors. 2021; 21(4):1238. https://doi.org/10.3390/s21041238

Chicago/Turabian Style

Chen, Ching-Han, Chien-Chun Wang, and Yan-Zhen Chen. 2021. "Intelligent Brushing Monitoring Using a Smart Toothbrush with Recurrent Probabilistic Neural Network" Sensors 21, no. 4: 1238. https://doi.org/10.3390/s21041238

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop