Next Article in Journal
Dynamic Routings in Satellite Networks: An Overview
Previous Article in Journal
Objective Measurement of Walking Activity Using Wearable Technologies in People with Parkinson Disease: A Systematic Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Backhand-Approach-Based American Sign Language Words Recognition Using Spatial-Temporal Body Parts and Hand Relationship Patterns

by
Ponlawat Chophuk
1,
Kosin Chamnongthai
1,* and
Krisana Chinnasarn
2
1
Department of Electronic and Telecommunication Engineering, Faculty of Engineering, King Mongkut’s University of Technology Thonburi, Bangkok 10140, Thailand
2
Faculty of Informatics, Burapha University, Chon Buri 20131, Thailand
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(12), 4554; https://doi.org/10.3390/s22124554
Submission received: 15 May 2022 / Revised: 10 June 2022 / Accepted: 14 June 2022 / Published: 16 June 2022
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Most of the existing methods focus mainly on the extraction of shape-based, rotation-based, and motion-based features, usually neglecting the relationship between hands and body parts, which can provide significant information to address the problem of similar sign words based on the backhand approach. Therefore, this paper proposes four feature-based models. The spatial–temporal body parts and hand relationship patterns are the main feature. The second model consists of the spatial–temporal finger joint angle patterns. The third model consists of the spatial–temporal 3D hand motion trajectory patterns. The fourth model consists of the spatial–temporal double-hand relationship patterns. Then, a two-layer bidirectional long short-term memory method is used to deal with time-independent data as a classifier. The performance of the method was evaluated and compared with the existing works using 26 ASL letters, with an accuracy and F1-score of 97.34% and 97.36%, respectively. The method was further evaluated using 40 double-hand ASL words and achieved an accuracy and F1-score of 98.52% and 98.54%, respectively. The results demonstrated that the proposed method outperformed the existing works under consideration. However, in the analysis of 72 new ASL words, including single- and double-hand words from 10 participants, the accuracy and F1-score were approximately 96.99% and 97.00%, respectively.

1. Introduction

Sign language is a medium of communication for hearing-impaired people, a group which includes 466 million people worldwide [1], and it is expressed using the fingers and hands. American Sign Language is the most populous among its peers, and thus consists of over ten thousand word gestures [2]. Moreover, 65% of ASL gestures represent sign words during a full conversation [3]. In addition, communication between hearing people and hearing-impaired people is very important because they must work together; however, hearing people do not know much about sign language, so there exists a communication gap and this gap in communication occurs very often in society. For instance, in the case of medical diagnosis where the patient is hard-of-hearing, the information received from the patient can be inaccurate, which can affect critical healthcare decisions. Poor communication often has catastrophic effects. According to a World Health Organization report, the simple failure to follow doctor’s orders results in 125,000 deaths in the U.S. each year [4]. Moreover, unemployment statistics for persons with deafness, caused by a lack of communication, have reached 3.8% in the U.S. [5]. Therefore, an automatic sign language interpretation system is necessary to bridge a communication gap. Automatic sign language interpretation systems need to be portable and mobile; thus, a backhand approach using a portable sensing device was proposed in [6], with a sensor mounted on the user’s chest to make it portable and user-friendly.
Some authors, such as [7], have developed automatic sign language interpretation systems based on the relationship between hand positions and body part features to recognize dynamic word gestures. Ma et al. [7] used WiFi signals to detect the fingers, hands, arms, and head. Although this method showed good recognition performance, its use is restricted by the need for high-speed internet access and a good line of sight. However, dynamic sign language recognition using analysis of the hand and body relationship has demonstrated superb results, for example, in the visual recognition of traffic police gestures [8], bipolar disorder recognition [9] using upper body postures, and emotional stress state recognition [10]. However, using the relationship between hand positions and body part features may not be extendable to sign language used by the hearing-impaired group. This is because of the semantic nature of their sign language. In addition, these features are used with a forehand view, which leads to misclassification; therefore, it is difficult to apply this method with a backhand view. The authors in [2,3,6,11,12] extracted hand features using a backhand view approach, which led to outstanding performance; however, these studies may fail to recognize words with a similar shape, rotation, and movement (SRM words). To address this problem with the existing works of [2,3,6,11,12], in this paper we propose the spatial–temporal body parts and hand relationship patterns (ST-BHR patterns) as the main feature of analysis for 72 isolated signed words in the SRM group [13] based on the backhand approach. In this method, both single and double hands were applied, and their information was obtained using the 3D distance-based Cartesian product obtained from a 3D depth camera. In addition, three additional features, spatial–temporal finger joint angle patterns, spatial–temporal double-hand relationship patterns, and spatial–temporal 3D hand motion trajectory patterns, were built into the proposed model for greater efficiency. Then, a time-independent classification tool was used as a classifier. The main contributions of this study are as follows.
(a)
We propose a method for a portable sign language recognition system based on a backhand approach to allow mobility by attaching a 3D small depth sensor to the chest area to detect the skeletons of hands.
(b)
The proposed method includes a new feature-based recognition approach in terms of the relationship between the hands and key points of the essential body to identify sign words with a similar shape, rotation, and movement, but with different meanings.
(c)
The developed system can be used with both single and double hands.
The remainder of this article is structured as follows. In Section 1 we present the introduction; in Section 2, we present related works; in Section 3, we present a problem analysis; in Section 4, we provide a system overview; in Section 5, we describe the proposed method; in Section 6, we report on our experiments and results; in Section 7, we discuss the implemented approach, and Section 8 concludes the work.

2. Related Works

Many existing works have developed feature extraction techniques [14] to identify similar sign words and letters in a sign language recognition system. They can be divided into two groups. First, the sensor-based group [15,16,17,18,19,20] includes devices worn on the users’ hands, such as gloves, EMG sensors, cables, accelerometers, touch sensors, and flexion sensors. The degree of flexion of the fingers and the fingers’ motion are used as the extracted features. However, with these systems, the user feels uncomfortable due to the complicated equipment, and it is difficult to create a portable system. The second is the vision-based group, which is more popular, and these systems use cameras as the primary device to facilitate motion analysis and interpretation by capturing video sequences. The advantage of a camera is that it eliminates the need for sensors in sensory gloves and reduces system construction costs. The vision-based group consists of 2D and 3D approaches, as shown in Table 1. The 2D method [21,22,23,24,25,26,27,28,29,30,31,32,33] uses video and image data for sign language interpretation. For example, a frequency-based model [21,22] has been proposed which uses a histogram, discrete wavelet transform, and a Gabor filter for extracting features from 2D images. Another proposed approach is the deformable gabarit model [23,24], which uses an active contour and the hand shape as features. Subsequently, the motion-based model [25] was proposed, which uses the motion of the hand as a feature via hand trajectory tracking. Finally, the deep-based model [26,27,28,29,30,31,32,33,34,35] has been proposed, which uses a novel deep-learning-based architecture to learn spatial hierarchies of features automatically and adaptively for 2D images, such as the convolutional neural network model (CNN), the convolutional neural network–long short-term memory model (CNN-LSTM), and recurrent neural networks (RNN). However, the 2D approach has less information than the 3D approach, which makes it hard to solve the problem of similar gestures based on a backhand view. Hence, the 3D approach has been proposed, which can be classified into three groups. The first group is the single-hand group [36,37,38,39,40,41,42], in which a one-hand model-based method is proposed that uses a Leap Motion sensor to detect the finger joints in three dimensions. A feature-based model was used to identify sign letters or sign words, with features such as the angle between fingers; the distance between fingers; finger position; the pitch, roll, and yaw angles of hand; and the trajectory of the finger. Although this method is highly accurate and takes less time to process, and also involves few background problems, this model can only be used with one hand; however, sign language sometimes involves the use of both hands. The second group is the double-hand group [12]. A double-hand-based method was introduced using Leap Motion sensors to solve the problem of similar shapes but different movements and rotations. This method adds a special feature that can distinguish the palm’s rotation angle, such as the pitch, roll, and yaw angles. However, the hand’s rotation is similar in the case of SRM words because these words have similar hand shapes and movements. The third group is the group including both single- and double-hand signs [43,44,45,46,47,48,49], in which both single- and double-hands are used. This group proposed a method to extract features from the hand, such as joint angles, fingertip position, finger direction, and finger velocity. These methods used the 3D model feature extraction method, resulting in high accuracy. More importantly, the backhand approach has been applied [2,3,6,11,12], which is essential for the practical application of the system in daily life. However, these features are hard to use in the case of SRM signs because of the problems with their similar shape, rotation, and movement.

3. Problem Analysis

Many sign words have a similar shape, rotation, and movement, and these are called the SRM sign group [13]. Consequently, the features used in the existing work may not be sufficient to distinguish these types of sign words. In the previous research on the backhand view [2,12], a similar problem was solved using rotation-based analysis, such as pitch, roll, and yaw angles. Moreover, another previous work [3,6,11] used position-based feature extraction based on a backhand view. Nevertheless, it would be difficult to apply these features to the SRM sign group, as shown in Figure 1. In the first row of Figure 1, the images in the first column show a single-handed representation of the sign language words “fox” and “fruits”, and the images in the second column show the double-handed representation of the sign language words “brother” and “sister.” The outputs of the rotation-based (including pitch, yaw, and roll angles), motion-based, and shape-based (shape representation of thumb, pinky, and wrist joints in terms of time series) methods are demonstrated in the second, third, and fourth rows, respectively.
As a result, in both the single- and double-hand groups, there are signs with similar shapes, rotation, and movement. However, the position of the hands relative to a part of the body can be used to distinguish words in this group. Therefore, in this paper we present a method based on the spatial–temporal body parts and hand relationship patterns (ST-BHR) as the main feature, using the 3D distance-based Cartesian product, as displayed in Figure 2, which can be expressed as in Equation (5).
J i , ρ , t = x i , ρ , t , y i , ρ , t , z i , ρ , t
where i = {1,2,3,…,40}, ρ = {right hand ( R ), left hand ( L )}, and t = {1,…,T} stand for the total number of finger joints, ρ = R ; i = {1, 2, 3,…,20} and ρ = L ; i = {21, 22, 23,…,40}, and the total number of frames, respectively.
According to the solution presented above, F t and B t are the fingertip and palm position set and the key positions of the body part set, as displayed in Equations (2) and (3), respectively. The key position of the selected body part is created by pointing a fingertip of the index finger to the desired locations at the point of the set in Equation (3).
F t = { J 1 , L , t ,   J 5 , L , t , J 9 , L , t , J 13 , L , t , J 17 , L , t , J 20 , L , t ,   J 21 , R , t ,   J 25 , R , t , J 29 , R , t , J 33 , R , t , J 37 , R , t , J 40 , R , t }
Let F t , J 1 ,   L , t , L , R , t = {1,…,T} stand for the set of fingertip and palm positions, fingertip and palm positions, the left hand, the right hand, and the total number of frames, respectively.
B t = S 1 , S 2 , S 3 , S 4 , S 5 , S 6 , S 7 , S 8
Here, B t , S 1 , S 2 , S 3 , S 4 , S 5 , S 6 , S 7 ,   S 8 stand for the set of the key points of the body positions, the forehead   S 1 , the right ear S 2 , the left ear S 3 , the nose ( S 4 ), the chin ( S 5 ), the right shoulder S 6 , the left shoulder S 7 , and the chest   position S 8 , respectively.
The 3D distance between two points, such as F t i = x i , t , y i , t , z i , t and B t k = x k , t , y k , t , z k , t in xyz-space, is given by the following generalization of the distance formula in Equation (4).
φ t F t i , B t k = x i , t x k , t 2 + y i , t y k , t 2 + z i , t z k , t 2
where φ t , F t , B t , x, y, z, i = {1,2,3…,12}, k = {1,2,3…,8} stand for the 3D distance, the set of fingertip and palm positions, the set of the key positions of the body parts, the x-axis, the y-axis, the z-axis, the total of set F t , and the total of set B t , respectively.
The proposed method, 3D distance-based features based on Cartesian products, is used as the feature extraction method, as shown in Figure 3 and Equation (5), in which the output of the Cartesian product is normalized before feeding it into a classifier.
H 1 t = N o r m F t × B t = N o r m φ j k J i , h , t , S n J i , h , t F t S n B t
where H 1 t , F t ,   B t , φ j k , J i , h , t , and N o r m indicate the Cartesian of the set of F t and B t , the set of fingertip and palm positions, the set of the key positions of the body parts, the 3D distance, the finger joints, and min–max normalization, respectively.
This feature represents spatial–temporal body parts and hand relationship patterns based on the 3D distance, normalized via min–max normalization [50], as shown in Figure 4. As a result, the words “fox” and “fruit” show different patterns in terms of the time series of the video, represented by a red line between the circle and the rectangle, which stands for the normalized distance of the words “fox” and “fruit”, respectively.

4. System Overview

Hearing-impaired people usually communicate through sign language, whereas non-hearing-impaired people normally use speech based on a natural language. An approach that allows these two groups of people to communicate via a mobile device that converts sign language to natural language and vice versa in real-time is illustrated in Figure 5. When hearing-impaired people speak in front of hearing-impaired and non-hearing-impaired people, hearing-impaired people can understand the meaning based on the signs, while simultaneously, non-hearing-impaired people can listen to the speech. In this scenario, a mobile device is installed on a body part of a hearing-impaired signer, such as the chest, face, and head, and functionally converts the video of signs into speech for non-hearing-impaired people to understand and vice versa.

4.1. Hardware Unit

The Leap Motion sensor [51] or 3D sensor is set on the signer’s chest so that it automatically moves along the moving path of the signer and always detects hand signs in the same view or the backhand view, as illustrated in Figure 6. In this paper, we focus on translating sign words and showing them as text, presented in a red rectangle. In addition, the interaction zone of the sensor extends from 10 cm to 80 cm, and a 140° × 120° typical field of view. Therefore, if the sensor is in the normal plane with no tilt, an interaction zone is unable to detect the selected body point. The interaction zone can be designed as shown in Figure 6.
θ base   = θ 1 + δ ;   h 1
Therefore, θ b a s e ,   θ 3   = 120 ° , θ 1   , δ, and h 1 stand for the tilt angle of the sensor, the vertical angle of the 3D sensor, an angle, a bias angle, and the sensor’s position in terms of the chest area determined by the height from the shoulder to the chest, respectively.

4.2. Software Unit

The software unit is divided into three parts, as illustrated schematically in Figure 7. First, the preprocessing step is performed, containing the data mining and analysis processes, such as the receival of 3D skeleton hand data from the 3D depth sensor and the 3D body input, obtained from individual calibration, as described in Section 5.1. Second, the feature extraction process consists of four features: the spatial–temporal body parts and hand relationship patterns as the main feature, the spatial–temporal finger joint angle patterns, the spatial–temporal double-hand relationship patterns, and the spatial–temporal 3D hand motion trajectory patterns based on PCA, as reported in Section 5.2. Thirdly, the classification method, specifically, the two-layer BiLSTM neural network, can be used to deal with time-independent data, as described in Section 5.3. Otherwise, the system returns to the start of the loop, with new finger joint positions being obtained by the 3D sensor.

5. Proposed Method

In this section, we present the preprocessing, feature extraction, and classification technique, which is shown schematically in Figure 8.

5.1. Preprocessing Technique

The preprocessing technique consists of two parts: 3D skeleton joint data received from the data acquisition device describe the 3D skeleton hand. The calibration technique presents the 3D points of interest of the body area obtained via the calibration using the tip of the index finger as a pointer.

5.1.1. 3D Skeleton Joint Data

The 3D sensor is an optical hand tracking module that captures the movements of hands, such as the 3D position and direction. In the case of a single hand, the zero padding technique [2] is used to replace the absent of finger joints of the second hand; it is assumed to be zero. Then, the total number of finger joints is expressed in the t-frame, as demonstrated in Equation (1).

5.1.2. Calibration Technique

The implementation of a calibration technique is an important step before using the system, and this must be a calibration for an individual person. The result of this step is the three-dimensional positioning of key points of the body by placing the tip of the index finger on the desired locations at the points of set B t and obtaining the fingertip’s position using the Leap Motion sensor as a reference position point, as shown in Figure 2 and Equation (7).
P B t i ( x ,   y ,   z ) = T B t i ;   B t = { S 1 ,   S 2 ,   S 3 ,   S 4 ,   S 5 ,   S 6 ,   S 7 ,   S 8 }
where P B t i , T B t i ,   B t , i = 1,…,7 stand for the new 3D positions of key points of the body, the 3D position created by pointing a tip of the index finger to the desired locations at the point of set B t according to Equation (2), the set of the key points of the body positions, and the total number of selected body points, respectively.

5.2. Feature Extraction

In this section we use the spatial–temporal body parts and hand relationship patterns as the main feature to identify signed words with similar shapes, rotation, and movement. Moreover, in this study we have added more features to make the system more efficient because, in addition to solving the problems mentioned above, it can also be used with other samples [37], since in this work we have applied both single- and double-handed approaches. The case of double-handed signs involves the relationship of the left and the right hand. Therefore, we extract four features: the spatial–temporal body parts and hand relationship patterns, the spatial–temporal finger joint angle patterns, the spatial–temporal double-hand relationship patterns, and the spatial–temporal 3D hand motion trajectory patterns. These are proposed in Section 5.2.1, Section 5.2.2, Section 5.2.3 and Section 5.2.4, respectively.

5.2.1. Spatial–Temporal Body Parts and Hand Relationship Patterns

The spatial–temporal body parts and hand relationship patterns are the feature that determines the relationship between the position of the left and right hands with the selected points on the body to solve the problem of words which have similar shapes and movements, but different finger positions. However, information on the relationship of both hands and the selected 3D position points on the body can be used to solve this problem by calculating the distance between each 3D point, as shown in Figure 2. All distance is normalized to store a set of patterns in terms of spatial–temporal data, as shown in Equation (5).

5.2.2. Spatial–Temporal Finger Joint Angle Patterns

We have proposed the use of this feature due to the similar shapes used in sign words [2,12]. There are two kinds of features, as indicated in Figure 9. Firstly, finger joint angles are used to find the angles between the finger joints of the same finger and the angles between two adjacent fingertips, determined as follows in Equation (8). This feature can be used to characterize the shape of the hand. Secondly, pitch, yaw, and roll angles indicate the palm orientation, enabling us to obtain the pitch ( ρ ) (angle of the x-axis), yaw ( ϕ ) (angle of the y-axis), and roll ( ) (angle of the z-axis), as shown in Equations (9)–(11), respectively.
θ = arccos M · N M · N
where θ is the angle between M and N , which indicate the 3D finger joint positions.
ρ = arctan A X A Y 2 + A Z 2
ϕ = arctan A Y A X 2 + A Z 2
= arctan A X 2 + A Y 2 A Z
where ρ , ϕ , ,   A X     A Y , and A Z are the pitch angle, yaw angle, and roll angle of the palm orientation, the x-axis, the y-axis, and the z-axis, respectively.
In the final step, the pitch, yaw, and roll angles of the palm joint and the finger joint angles of the consecutive joint in the same finger must be collected in the set, in which all data are normalized to store a set of patterns in terms of the spatial–temporal data in H 2 t , as shown in Equation (12).
H 2 t = N o r m θ 1 t , θ 2 t , θ 3 t , , θ n t ,   ρ R t , ϕ R t , R t ,   ρ L t , ϕ L t , L t
where N o r m ,   t = 1 , , T , R, L, ρ t , ϕ t , and (t) are the min–max normalization, the total number of frames, the right hand, the left hand, the pitch angle, the yaw angle, and the roll angle, respectively.

5.2.3. Spatial–Temporal Double-Hand Relationship Patterns

Due to the similar shapes of sign words [6], the spatial–temporal double-hand relationship patterns are proposed based on 3D Euclidean distance, as demonstrated in Figure 10 and Equation (13). This feature describes the relationship of the left and right hands to solve the problem of words or letters signed with similar movements. Therefore, all 3D distances are collected in the set; then, the selected distance-based data are normalized to store a set of   H 3 t , which is the term of the spatial–temporal data.
H 3 t = N o r m U 1 t ,   U 2 t , U 3 t , , U n t
where U n t , N o r m ,   t = 1 , , T , and n= {1, 2, 3, …, 20} are the distance between two points in xyz-space, min–max normalization, the total number of frames, and the total number of points, respectively.

5.2.4. Spatial–Temporal 3D Hand Motion Trajectory Patterns

Due to the similar shapes used in different signs [3], the concept of the spatial–temporal 3D hand motion trajectory patterns are presented to extract the movement trajectory of the finger joints, in which the 3D positions of the joints are collected in the set, as shown in Figure 11. Then, principal component analysis (PCA) [52] is used to reduce the dimensions of the data set to one dimension. Finally, the selected data are normalized to store a set of H 4 t , the term of the spatial–temporal data, as shown in Equation (14).
H 4 t = N o r m [ PCA { ( T 1 t ) , ( T 2 t ) , ( T 3 t ) , , ( T n t ) } ]
where T n t , N o r m , PCA, and t = 1 , , T   are the tip and palm positions in the time series of xyz-space, min–max normalization, principal component analysis, and the total number of frames, respectively.
In the final stage of feature extraction, the four features consist of the spatial–temporal body parts and hand relationship patterns ( H 1 t ), the spatial–temporal finger joint angle patterns ( H 2 t ), the spatial–temporal double-hand relationship patterns ( H 3 t ), and the spatial–temporal 3D hand motion trajectory patterns ( H 4 t ). These features are concatenated into one-dimensional data by means of a concatenation technique, of which the equations are shown in Equation (15). The final feature extraction in terms of spatial temporal patterns ( X t ) is the input of a stacked BiLSTM in the classification process.
X t = H 1 t H 2 t H 3 t H 4 t
where H 1 t , H 2 t ,   H 3 t , H 4 t , ,   t = {1,…,T} stand for the spatial–temporal body parts and hand relationship patterns, the spatial–temporal finger joint angle patterns, the spatial–temporal double-hand relationship patterns, the spatial–temporal 3D hand motion trajectory patterns, the concatenation operator, and the total number of frames, respectively.

5.3. Classification

A deep learning algorithm, a recurrent neural network (RNN), is applied for the analysis of data in the form of a serial sequence, such as a video (a series of images) or text (a sequence of words). However, RNNs exhibit a vanishing and exploding gradient problem [53], which results in poor performance when dealing with long sequences. Therefore, bidirectional long short-term memory (BiLSTM) [54] is used for solving long sequences because it is possible to choose which data should be remembered or eliminated. Recent experimental works [3,12,37,55] have demonstrated that the BiLSTM network outperformed various models, such as the standard CNN, SVM, RNN, ARMA, and ARIMA. The use of a single LSTM network for sign word recognition led to low accuracy and overfitting, especially when learning complex sign sequences. To address this problem, stacking more than one BiLSTM unit, as in [2,43], enhances performance in the recognition of sign words. Therefore, inspired by these works, we designed our BiLSTM architecture using two BiLSTM units, as shown in Figure 12. This two-unit BiLSTM architecture allows us to achieve the high-level sequential modeling of the selected features.
The structure of the designed two-layer BiLSTM model consists of the input, BiLSTM hidden neurons, dropout, and classification layers, as shown in Figure 12. The input layer contains time series data, as demonstrated in Equation (15). Then, the BiLSTM hidden layer consists of four main gates: the forget gate, the input gate, the input modulation gate, and the output gate. The forget gate ( f t + 1 ) controls the flow of information to forget or keep the previous state ( C t ). The input data ( x n t + 1 ), previous hidden state ( h 1 t ), bias   ( b f ) , and the sigmoid function ( σ ) are used for making a decision, as shown in Equation (16). If the forget gate is set to 0, it decides to forget the previous state, but if the forget gate is set to 1, it keeps the previous state. However, the input gate ( i t + 1 ) is used to decide which information of the input ( x n t + 1 ) ,   sigmoid   function   σ , and   h 1 t should be passed to update the cell state, as demonstrated in Equation (17). Equation (18) expresses the cell states generated from the updated cell state in terms of the forget gate ( f t + 1 ), the input gate ( i t + 1 ), and the previous state ( C t ), respectively. Then, the output gate ( o t + 1 ) decides the value for the next sequence h 1 t + 1 , as shown in Equations (19) and (20). After that, a dropout layer is proposed to prevent overfitting by randomly turning off nodes in the network [56]. Based on previous works [57], the dropout layer was compared with each dropout level rate, resulting in dropout value of 0.2, which showed the best performance. Therefore, a dropout value of 0.2 was applied in this study. Lastly, the Softmax function is applied in the classification layer. The Softmax function is a form of logistic regression that normalizes an input value into vector values, following a probability distribution, in which the output values are in the range of [0,1].
  f t + 1 = σ x n t + 1 W f x + h 1 t W f h + b f
  i t + 1 = σ x n t + 1 W i x + h 1 t W i h + b i
C t + 1 = ( C t × f t + 1 ) + ( i t + 1 × t a n h x n t + 1 W Θ x + h 1 t W Θ h + b Θ )
  o t + 1 = σ x n t + 1 W o x + h 1 t W o h + b o
  h 1 t + 1 = o t + 1 × t a n
where   f t + 1 ,   i t + 1 , C t + 1 ,   o t + 1 ,   h 1 t + 1 , x n t + 1 , σ , h 1 t , C t , t a n h, W R U × V , and b R V stand for the forget gate vector, the input gate vector, the cell input vector, the output gate vector, the hidden state vector, the input data, the sigmoid function, the previously hidden state vector, the previous cell state vector, the hyperbolic tangent function, the weight matrices, in which the superscripts U   and V   refer to the number of input features and the number of hidden units, and the bias vector parameters, in which the superscripts V refer to the number of hidden units, respectively.

6. Experiments and Results

Figure 13 shows the experimental setup used for training and testing the proposed idea. It consisted of a sensor that was set up in the chest area and a computer system in which a laptop was used to collect the output from the sensor.

6.1. Experiments

This section consists of three parts: firstly, the dataset, which describes the dataset used in the experiment; the type of dataset used; and the dataset from the existing research that was used to compare the efficiency of the system. Secondly, the configuration parameters describe the parameter settings of the sign language recognition system and the classification model. Finally, we present an evaluation of the classification model and give details regarding the system’s validity.

6.1.1. Dataset

The American Sign Language dataset includes signed words and signed letters which are currently used in sign language recognition. Most are forehand methods, but in this study, we propose a backhand approach. Therefore, it is compulsory to create a new dataset due to insufficient information. The created datasets are divided into two types—single-handed and double-handed datasets of signed words, with 36 words, giving a total of 72 words. The datasets were collected by ten deaf and hard-of-hearing people, with each person collecting each word ten times so that both types would contain 7200 samples. This method uses the backhand approach; therefore, the sensor must be set on the chest and can be used in both day and night, as shown in Figure 2. In addition, this method has been tested on signed letters (the letters A–Z) that exhibited the problem of similar signed letters, using the datasets from [3], which presented a backhand dataset of 5200 samples. In the same way, this method was also tested with the dataset of signed words from [12], in which they collected 40 double-hand dynamic ASL words, giving a total of 4000 samples of similar signed words. Therefore, the total dataset contained 16,400 samples, as listed in Table 2. The protocol of k-fold cross validation was applied with these datasets.

6.1.2. Configuration Parameter

The configuration parameters can be categorized into two parts. The first part, the hardware specifications, describe the computer system and the Leap Motion sensor [51] details, as given in Table 3. In the second part, the classification parameter settings are used to configure the parameters of the BiLSTM model, as demonstrated in Table 4.

6.1.3. Evaluation of the Classification Model

Accuracy [12] is described as a measure of correct predictions. Accuracy is given by Equation (21). The standard deviation is used to measure the amount of variation of a set of values, as shown in Equation (22). Moreover, the error, precision, recall, and F1-score [12] are applied for this model.
  A c c u r a c y   ( A C C )   = Υ 1 + Υ 2 Υ 1 + Υ 2 + ϕ 1 + ϕ 2
where Υ 1 , Υ 2 , ϕ 1 , and ϕ 2 denote the true positive (correctly identified), true negative (correctly rejected), false positive (incorrectly identified), and false negative (incorrectly rejected) values, respectively.
S D = i = 1 n ( x i Υ 3 ) 2 n
where SD, Υ 3 , x i , and n stand for the standard deviation, the mean of all samples, each sample, and the total number of samples, respectively.

6.2. Results

The performance comparison in the task of signed-letter recognition (letters A–Z) using the experimental data set in [3] is demonstrated in Table 5. The overall evaluation results were obtained using measures of accuracy, error, precision, recall, F1-score, and standard deviation (SD), respectively. The recognition results of 26 signed letters using the proposed method were compared with the conventional method presented in [37] based on a feature-based model, specifically a finger-motion-based forehand view, and the method using a trajectory-based backhand view [3]. The results of the proposed method demonstrated that our proposed features improved the accuracy rate in signed-letter recognition (letters A–Z) by about 1.27%. The other metrics under consideration were not available in [3,37]; thus, these columns contain no data.
Moreover, the performance comparison of the experimental results obtained for 40 double-hand dynamic ASL words from [12] using our proposed method is shown in Table 6. Using the proposed method, the accuracy rate was improved by 0.54 % compared to the conventional method [12] through the use of shape, motion, position, and angle-based features. As shown in Table 6, the conventional method [12] achieved an accuracy of 97.98%, whereas our proposed method increased the accuracy by 0.54%, with an overall accuracy of 98.52%, an error of 1.48%, a precision of 98.56%, a recall of 98.52%, an F1-score of 98.54%, and an SD of 0.22%. The overall accuracy for 72 American Sign Language (ASL) words was 96.99%. The performance comparison of the experimental results in the recognition of signed-words (72 words, including single- and double-handed ASL words) is shown in Table 7. We performed an ablation test, to demonstrate the significance of the proposed features over those used in previous works.

Ablation Test

We conducted ablation tests to determine the significance of each feature in the proposed models. The feature combination used in this method consists of the spatial–temporal body parts and hand relationship patterns ( H 1 ), the spatial–temporal finger joint angle patterns ( H 2 ), the spatial–temporal double-hand relationship patterns ( H 3 ), and the spatial–temporal 3D hand motion trajectory patterns ( H 4 ). This method of feature extraction for signed-letter recognition evaluates different feature combinations among all the features, as shown in Table 8. We evaluated the 1 s t ( H 2 + H 3 ), 2 n d ( H 2 + H 3 + H 4 ), and 3 r d ( H 1 + H 2 + H 3 + H 4 ) combinations, and found that the 3 r d ( H 1 + H 2 + H 3 + H 4 ) combination was the best combination of features for signed-letter recognition (letters A–Z), achieving an accuracy of 97.34%, a precision of 97.39%, a recall of 97.34%, an F1-score of 97.36%, and an SD of 0.26%. However, the results showed that the use of the ( H 1 ,   H 4 ) features can improve the accuracy.
In our performance comparison, we conducted an ablation test related to signed-word recognition, and evaluated different feature combinations, as shown in Table 9. The feature combination consisted of the spatial–temporal body parts and hand relationship patterns ( H 1 ), the spatial–temporal finger joint angle patterns ( H 2 ), the spatial–temporal double-hand relationship patterns ( H 3 ), and the spatial–temporal 3D hand motion trajectory patterns ( H 4 ). There are three different feature combinations models: the 1 s t ( H 2 + H 3 ) , 2 n d ( H 2 + H 3 + H 4 ), and 3 r d ( H 1 + H 2 + H 3 + H 4 ) combinations. The results showed that the proposed model provided the best performance, with an accuracy of 98.52%, a precision of 98.56%, a recall of 98.52%, an F1-score of 98.54%, and an SD of 0.22%.
The experimental results for 72 sign words, including single- and double-handed words, based on a five-fold cross-validation [58], which is one of the most regularly used model evaluation methods, are shown in Table 10 and Table 11, and Figure 14. In Table 10, the group (G.) in the table indicates a pair of words with similar signs. The error column indicates other misspelled words. For example, in Table 11, group (G.) 13 of the word “father” displayed an error in relation to the word “mother” of 2.6%. As a result, this indicates that the system has high accuracy.

7. Discussion

In the case of using a single hand, there is some hand occlusion by the palm, which obstructs the view of the sensor, resulting in an incorrect predictive finger position and a missing position when the finger is occluded. For example, the word “spit” in the third row causes occlusion by the palm, as shown in Figure 15. Unfortunately, the predicted finger joint position in the second row is similar to the position of the word “grandmother” in the first row, causing the determined meaning to be incorrect. For future solutions, prior and post-information may be used to predict the lost position of the finger.
In the second case of single- and double-hand words, some hand occlusion problems are caused by fingers. For example, in Figure 16, the position of the word “are” in the third row, which is derived from the sensor, has the wrong position in the second row. Regrettably, this wrong position is similar to the word “true”, shown in the first row, thus causing a misclassification problem. Moreover, in terms of double-hand words, Figure 17 shows an instance in which the error position of the word “keep” in the second row is similar to the word “sister” in the first row.
In addition, since the Leap Motion sensor is mounted on the chest, which limits the scope of the area in which the hand can be detected, there is a problem with some signed words such as “introduce”, for which the hand position is sometimes difficult with the interaction zone. The solution for this problem is to increase an interaction zone by using two Leap Motion sensors.

8. Conclusions

When using a backhand approach, some signed words have a similar shape, rotation, and hand movement, but different hand positions; thus, detection systems suffer from misclassification errors, which results in lower accuracy. Therefore, in this study we propose the use of the spatial–temporal body parts and hand relationship patterns (ST-BHR) as the main feature, in which the set of the 3D positions of the finger joints and the set of the key positions of the body parts are applied, measuring the Euclidean distances based on a Cartesian product to derive a series of 3D distance-based features. Then, the bidirectional long short-term memory method is used as a classifier for time-independent data. The performance of the method was evaluated using 72 sign words from 10 participants, using both single- and double-handed words, and the accuracy was found to be approximately 96.99%. The method was further developed on 26 ASL sign letters and 40 double-hand dynamic ASL word datasets, improving upon the conventional method by 1.27% and 0.54%, respectively.
The evaluation results demonstrate that the use of ST-BHR features minimized the problems associated with the SRM sign group. Therefore, the proposed method can allow people who are hard-of-hearing to access the recognition system at their convenience. However, a major limitation of this work is the limited interactive space used for hand and face position detection. In future works, two Leap Motion sensors can be used to expand the interaction zone. Moreover, real social situations [59,60] will be examined in future work, introducing factors such as vibration [61] from driving and other means of transportation which would affect the position of the camera sensor. Therefore, this issue will be a challenging topic in future.

Author Contributions

Conceptualization, P.C., K.C. (Kosin Chamnongthai) and K.C. (Krisana Chinnasarn); formal analysis, P.C., K.C. (Kosin Chamnongthai) and K.C. (Krisana Chinnasarn); funding acquisition, K.C. (Kosin Chamnongthai) and K.C. (Krisana Chinnasarn); project administration, K.C. (Kosin Chamnongthai); supervision, K.C. (Kosin Chamnongthai) and K.C. (Krisana Chinnasarn); writing—original draft, P.C.; writing—review and editing, P.C., K.C. (Kosin Chamnongthai) and K.C. (Krisana Chinnasarn). All authors have read and agreed to the published version of the manuscript.

Funding

The research project is financially supported by Petchra Pra Jom Klao Ph.D. Research Scholarship (Grant no. 6/2560) from King Mongkut’s University of Technology Thonburi, Bangkok, and Faculty of Informatics, Burapha University, Chon Buri, Thailand.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. World Health Organization. Deafness and Hearing Loss. Available online: https://www.who.int/news-room/fact-sheets/detail/deafness-and-hearing-loss (accessed on 9 January 2022).
  2. Abdullahi, S.B.; Chamnongthai, K. American Sign Language Words Recognition using Spatio-Temporal Prosodic and Angle Features: A sequential learning approach. IEEE Access 2022, 10, 15911–15923. [Google Scholar] [CrossRef]
  3. Chophuk, P.; Chamnongthai, K. Backhand-view-based continuous-signed-letter recognition using a rewound video sequence and the previous signed-letter information. IEEE Access 2021, 9, 40187–40197. [Google Scholar] [CrossRef]
  4. In Medical Situations, Poor Communication Often Leads to Disaster. Deaf Services Unlimited. Available online: https://deafservicesunlimited.com/2016/05/in-medical-situations-poor-communication-often-leads-to-disaster/ (accessed on 11 January 2022).
  5. Garberoglio, C.L.; Palmer, J.L.; Cawthon, S.W.; Sales, A. Deaf People and Employment in the United States: 2019; National Deaf Center: Austin, TX, USA, 2019. [Google Scholar]
  6. Fang, B.; Co, J.; Zhang, M. DeepASL: Enabling Ubiquitous and Non-Intrusive Word and Sentence-Level Sign Language Translation. In Proceedings of the 15th ACM Conference on Embedded Network Sensor Systems, Shenzhen, China, 4–7 November 2018. [Google Scholar]
  7. Ma, Y.; Zhou, G.; Wang, S.; Zhao, H.; Jung, W. SignFi: Sign language recognition using WiFi. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2018, 2, 1–21. [Google Scholar] [CrossRef]
  8. He, J.; Zhang, C.; He, X.L.; Dong, R.H. Visual Recognition of traffic police gestures with convolutional pose machine and handcrafted features. Neurocomputing 2020, 390, 248–259. [Google Scholar] [CrossRef]
  9. Yang, L.; Li, Y.; Chen, H.; Jiang, D.; Oveneke, M.C.; Sahli, H. Bipolar disorder recognition with histogram features of arousal and body gestures. In Proceedings of the 2018 on Audio/Visual Emotion Challenge and Workshop, Seoul, Korea, 22 October 2018; pp. 15–21. [Google Scholar]
  10. Chen, H.; Liu, X.; Li, X.; Shi, H.; Zhao, G. Analyze spontaneous gestures for emotional stress state recognition: A micro-gesture dataset and analysis with deep learning. In Proceedings of the 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG), Lille, France, 14–18 May 2019; pp. 1–8. [Google Scholar]
  11. Pattanaworapan, K.; Chamnongthai, K.; Guo, J.-M. Signer independence finger alphabet recognition using discrete wavelet transform and area level run lengths. J. Vis. Commun. Image Represent. 2016, 38, 658–677. [Google Scholar] [CrossRef]
  12. Abdullahi, S.B.; Chamnongthai, K. American Sign Language Words Recognition of Skeletal Videos Using Processed Video Driven Multi-Stacked Deep LSTM. Sensors 2022, 22, 1406. [Google Scholar] [CrossRef]
  13. Tennant, R.A.; Gluszak, M.; Brown, M.G. The American Sign Language Handshape Dictionary; Gallaudet University Press: Washington, DC, USA, 1998. [Google Scholar]
  14. Starner, T.; Pentland, A. Real-time American Sign Language recognition from video using hidden Markov models. In Proceedings of the International Symposium on Computer Vision (ISCV 1995), Coral Gables, FL, USA, 21–23 November 1995; pp. 265–270. [Google Scholar]
  15. Lee, B.G.; Chong, T.W.; Chung, W.Y. Sensor fusion of motion-based sign language interpretation with deep learning. Sensors 2020, 20, 6256. [Google Scholar] [CrossRef]
  16. Li, W.; Luo, Z.; Xi, X. Movement Trajectory Recognition of Sign Language Based on Optimized Dynamic Time Warping. Electronics 2020, 9, 1400. [Google Scholar] [CrossRef]
  17. Gurbuz, S.Z.; Rahman, M.M.; Kurtoglu, E.; Malaia, E.; Gurbuz, A.C.; Griffin, D.J.; Crawford, C. Multi-frequency rf sensor fusion for word-level fluent asl recognition. IEEE Sens. J. 2021, 12, 11373–11381. [Google Scholar] [CrossRef]
  18. Dong, W.; Yang, L.; Gravina, R.; Fortino, G. Soft wrist-worn multi-functional sensor array for real-time hand gesture recognition. IEEE Sens. J. 2021. [Google Scholar] [CrossRef]
  19. Lee, S.; Jo, D.; Kim, K.B.; Jang, J.; Park, W. Wearable sign language translation system using strain sensors. Sens. Actuator A Phys. 2021, 331, 113010. [Google Scholar] [CrossRef]
  20. Zaccagnino, R.; Capo, C.; Guarino, A.; Lettieri, N.; Malandrino, D. Techno-regulation and intelligent safeguards. Multimed. Tools Appl. 2021, 80, 15803–15824. [Google Scholar] [CrossRef]
  21. Abid, M.; Petriu, E.; Amjadian, E. Dynamic sign language recognition for smart home interactive application using stochastic linear formal grammar. IEEE Trans. Instrum. Meas. 2015, 64, 596–605. [Google Scholar] [CrossRef]
  22. Ranga, V.; Yadav, N.; Garg, P. American sign language fingerspelling using hybrid discrete wavelet transform-gabor filter and convolutional neural network. J. Eng. Sci. Technol. 2018, 13, 2655–2669. [Google Scholar]
  23. Miguel, R.A.; Susana, O.-C.; Jorge, R.; Federico, S.-I. American sign language alphabet recognition using a neuromorphic sensor and an artificial neural network. Sensors 2017, 17, 2176. [Google Scholar]
  24. Liwicki, S.; Everingham, M. Automatic recognition of fingerspelled words in British Sign Language. In Proceedings of the 2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Miami, FL, USA, 20–25 June 2009. [Google Scholar]
  25. Aldelfy, H.A.; Al-Mufraji, M.H.; Saeed, T.R. An Efficient Feature Extraction of Isolated Word for Dynamic Sign Language Classification. In Proceedings of the 2018 Third Scientific Conference of Electrical Engineering (IEEE), Baghdad, Iraq, 19–20 December 2018; pp. 305–309. [Google Scholar]
  26. Mohammed, A.A.Q.; Lv, J.; Islam, M. A deep learning-based End-to-End composite system for hand detection and gesture recognition. Sensors 2019, 19, 5282. [Google Scholar] [CrossRef] [Green Version]
  27. Rahim, M.A.; Islam, M.R.; Shin, J. Non-Touch Sign Word Recognition Based on Dynamic Hand Gesture Using Hybrid Segmentation and CNN Feature Fusion. Appl. Sci. 2019, 9, 3790. [Google Scholar] [CrossRef] [Green Version]
  28. Sahoo, J.P.; Prakash, A.J.; Pławiak, P.; Samantray, S. Real-Time Hand Gesture Recognition Using Fine-Tuned Convolutional Neural Network. Sensors 2022, 22, 706. [Google Scholar] [CrossRef]
  29. Podder, K.K.; Chowdhury, M.E.; Tahir, A.M.; Mahbub, Z.B.; Khandakar, A.; Hossain, M.S.; Kadir, M.A. Bangla Sign Language (BdSL) Alphabets and Numerals Classification Using a Deep Learning Model. Sensors 2022, 22, 574. [Google Scholar] [CrossRef]
  30. Li, D.; Rodriguez, C.; Yu, X.; Li, H. Word-level deep sign language recognition from video: A new large-scale dataset and methods comparison. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Snowmass, CO, USA, 1–5 March 2020; pp. 1459–1469. [Google Scholar]
  31. Maruyama, M.; Ghose, S.; Inoue, K.; Roy, P.P.; Iwamura, M.; Yoshioka, M. Word-level sign language recognition with multi-stream neural networks focusing on local regions. arXiv 2021, arXiv:2106.15989. [Google Scholar]
  32. Aly, S.; Aly, W. DeepArSLR: A Novel Signer-Independent Deep Learning Framework for Isolated Arabic Sign Language Gestures Recognition. IEEE Access 2020, 8, 83199–83212. [Google Scholar] [CrossRef]
  33. Liao, Y.Q.; Xiong, P.W.; Min, W.D.; Min, W.Q.; Lu, J.H. Dynamic sign language recognition based on video sequence with BLSTM-3D residual networks. IEEE Access 2019, 7, 38044–38054. [Google Scholar] [CrossRef]
  34. Al-Hammadi, M.; Muhammad, G.; Abdul, W.; Alsulaiman, M.; Bencherif, M.A.; Alrayes, T.S.; Mekhtiche, M.A. Deep learningbased approach for sign language gesture recognition with efficient hand gesture representation. IEEE Access 2020, 8, 192527–192542. [Google Scholar] [CrossRef]
  35. Murhij, Y.; Serebrenny, V. Hand gestures recognition model for Augmented reality robotic applications. In Proceedings of the 15th International Conference on Electromechanics and Robotics “Zavalishin’s Readings”, Ufa, Russia, 15–18 April 2020; Springer: Singapore, 2021; pp. 187–196. [Google Scholar]
  36. Lee, C.K.; Ng, K.K.; Chen, C.H.; Lau, H.C.; Chung, S.; Tsoi, T. American sign language recognition and training method with recurrent neural network. Expert Syst. Appl. 2021, 167, 114403. [Google Scholar] [CrossRef]
  37. Chong, T.-W.; Lee, B.-G. American Sign Language Recognition Using Leap Motion Controller with Machine Learning Approach. Sensors 2018, 18, 3554. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  38. Vaitkevičius, A.; Taroza, M.; Blažauskas, T.; Damaševičius, R.; Maskeliūnas, R.; Woźniak, M. Recognition of american sign language gestures in a virtual reality using leap motion. Appl. Sci. 2019, 9, 445. [Google Scholar] [CrossRef] [Green Version]
  39. Ameur, S.; Khalifa, A.B.; Bouhlel, M.S. A novel hybrid bidirectional unidirectional LSTM network for dynamic hand gesture recognition with leap motion. Entertain. Comput. 2020, 35, 100373. [Google Scholar] [CrossRef]
  40. Naglot, D.; Kulkarni, M. Real time sign language recognition using the leap motion controller. In Proceedings of the 2016 IEEE International Conference on Inventive Computation Technologies (ICICT), Coimbatore, India, 26–27 August 2016; Volume 3, pp. 1–5. [Google Scholar]
  41. Shin, J.; Matsuoka, A.; Hasan, M.; Mehedi, A.; Srizon, A.Y. American Sign Language Alphabet Recognition by Extracting Feature from Hand Pose Estimation. Sensors 2021, 21, 5856. [Google Scholar] [CrossRef]
  42. Khelil, B.; Amiri, H.; Chen, T.; Kammüller, F.; Nemli, I.; Probst, C.W. Hand Gesture Recognition Using Leap Motion Controller for Recognition of Arabic Sign Language. In Proceedings of the 3rd International Conference on Automation, Control, Engineering and Computer Science ACECS, Hammamet, Tunisia, 22 March 2016; Volume 16, pp. 233–238. [Google Scholar]
  43. Avola, D.; Bernardi, M.; Cinque, L.; Foresti, G.L.; Massaroni, C. Exploiting recurrent neural networks and leap motion controller for the recognition of sign language and semaphoric hand gestures. IEEE Trans. Multimed. 2018, 21, 234–245. [Google Scholar] [CrossRef] [Green Version]
  44. Katoch, S.; Singh, V.; Tiwary, U.S. Indian Sign Language recognition system using SURF with SVM and CNN. Array 2022, 14, 100141. [Google Scholar] [CrossRef]
  45. Bird, J.J.; Ekárt, A.; Faria, D.R. British sign language recognition via late fusion of computer vision and leap motion with transfer learning to american sign language. Sensors 2020, 20, 5151. [Google Scholar] [CrossRef]
  46. Smedt, Q.D.; Wannous, H.; Vandeborre, J.-P. Heterogeneous Hand Gesture Recognition Using 3D Dynamic Skeletal Data. Comput. Vis. Image Underst. 2019, 181, 60–72. [Google Scholar] [CrossRef] [Green Version]
  47. Mapari, R.B.; Kharat, G. Analysis of multiple sign language recognition using leap motion sensor. Int. J. Res. Advent Technol. 2017. [Google Scholar]
  48. Kumar, P.; Gauba, H.; Roy, P.P.; Dogra, D.P. A multimodal framework for sensor based sign language recognition. Neurocomputing 2017, 259, 21–38. [Google Scholar] [CrossRef]
  49. Jamaludin, N.A.N.; Fang, O.H. Dynamic Hand Gesture to Text using Leap Motion. Int. J. Adv. Comput. Sci. Appl. 2019, 10, 199–204. [Google Scholar] [CrossRef] [Green Version]
  50. Moreira, L.L.; de Brito, M.M.; Kobiyama, M. Effects of different normalization, aggregation, and classification methods on the construction of flood vulnerability indexes. Water 2021, 13, 98. [Google Scholar] [CrossRef]
  51. Guzsvinecz, T.; Szucs, V.; Sik-Lanyi, C. Suitability of the Kinect Sensor and Leap Motion Controller—A Literature Review. Sensors 2019, 19, 1072. [Google Scholar] [CrossRef] [Green Version]
  52. Wen, Y.; Lin, P.; Nie, X. Research of stock price prediction based on PCA-LSTM model. IOP Conf. Ser. Mater. Sci. Eng. 2020, 790, 012109. [Google Scholar] [CrossRef]
  53. Kanai, S.; Fujiwara, Y.; Iwamura, S. Preventing gradient explosions in gated recurrent units. In Advances in Neural Information Processing Systems, Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Curran Associates, Inc.: Red Hook, NY, USA, 2017. [Google Scholar]
  54. Huang, R.; Wei, C.; Wang, B.; Yang, J.; Xu, X.; Wu, S.; Huang, S. Well performance prediction based on Long Short-Term Memory (LSTM) neural network. J. Pet. Sci. Eng. 2022, 208, 109686. [Google Scholar] [CrossRef]
  55. Song, W.; Gao, C.; Zhao, Y.; Zhao, Y. A Time Series Data Filling Method Based on LSTM—Taking the Stem Moisture as an Example. Sensors 2020, 20, 5045. [Google Scholar] [CrossRef]
  56. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  57. Zhou, Q.; Wu, H. NLP at IEST 2018: BiLSTM-attention and LSTM-attention via soft voting in emotion classification. In Proceedings of the 9th Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis, Brussels, Belgium, 31 October 2018; pp. 189–194. [Google Scholar]
  58. Umer, M.; Imtiaz, Z.; Ullah, S.; Mehmood, A.; Choi, G.S.; On, B.W. Fake news stance detection using deep learning architecture (cnn-lstm). IEEE Access 2020, 8, 156695–156706. [Google Scholar] [CrossRef]
  59. Kayhan, O.; Samur, E. A Wearable Haptic Guidance System Based on Skin Stretch around the Waist for Visually-Impaired Runners. In Proceedings of the IEEE Haptics Symposium (HAPTICS), IEEE, Santa Barbara, CA, USA, 21–24 March 2022; pp. 1–6. [Google Scholar]
  60. Ibrahim, N.B.; Zayed, H.H.; Selim, M.M. Advances, challenges and opportunities in continuous sign language recognition. J. Eng. Appl. Sci. 2020, 15, 1205. [Google Scholar] [CrossRef] [Green Version]
  61. Bhuiyan, M.H.; Fard, M.; Robinson, S.R. Effects of whole-body vibration on driver drowsiness: A review. J. Saf. Res. 2022, 81, 175–189. [Google Scholar] [CrossRef]
Figure 1. Problem analysis of different features used for similar sign words of (a) Rotation representation of hand of single hand; (b) Rotation representation of hand of double hands; (c) Motion representation of hand of single hand; (d) Motion representation of hand of double hands; (e) Shape representation of thumb, pinky, and wrist finger on time series of single hand; (f) Shape representation of thumb, pinky, and wrist finger on time series of double hands.
Figure 1. Problem analysis of different features used for similar sign words of (a) Rotation representation of hand of single hand; (b) Rotation representation of hand of double hands; (c) Motion representation of hand of single hand; (d) Motion representation of hand of double hands; (e) Shape representation of thumb, pinky, and wrist finger on time series of single hand; (f) Shape representation of thumb, pinky, and wrist finger on time series of double hands.
Sensors 22 04554 g001
Figure 2. Spatial–temporal body parts and hand relationship patterns (ST-BHR).
Figure 2. Spatial–temporal body parts and hand relationship patterns (ST-BHR).
Sensors 22 04554 g002
Figure 3. Distance-based features based on Cartesian products.
Figure 3. Distance-based features based on Cartesian products.
Sensors 22 04554 g003
Figure 4. Comparison of similarly signed words using H 1 t .
Figure 4. Comparison of similarly signed words using H 1 t .
Sensors 22 04554 g004
Figure 5. Scenario of the use of a sign-language interpretation device.
Figure 5. Scenario of the use of a sign-language interpretation device.
Sensors 22 04554 g005
Figure 6. Hardware system.
Figure 6. Hardware system.
Sensors 22 04554 g006
Figure 7. Flowchart of the proposed method.
Figure 7. Flowchart of the proposed method.
Sensors 22 04554 g007
Figure 8. Proposed method.
Figure 8. Proposed method.
Sensors 22 04554 g008
Figure 9. Spatial–temporal finger joint angle patterns.
Figure 9. Spatial–temporal finger joint angle patterns.
Sensors 22 04554 g009
Figure 10. Double-hand relationship patterns for single and double hands.
Figure 10. Double-hand relationship patterns for single and double hands.
Sensors 22 04554 g010
Figure 11. 3D hand motion trajectory patterns for single and double hands.
Figure 11. 3D hand motion trajectory patterns for single and double hands.
Sensors 22 04554 g011
Figure 12. Two-layer BiLSTM neural network.
Figure 12. Two-layer BiLSTM neural network.
Sensors 22 04554 g012
Figure 13. Photograph of the experimental setup.
Figure 13. Photograph of the experimental setup.
Sensors 22 04554 g013
Figure 14. Confusion matrix of the recognition performance of 72 American Sign Language (ASL) words.
Figure 14. Confusion matrix of the recognition performance of 72 American Sign Language (ASL) words.
Sensors 22 04554 g014
Figure 15. Example of an error caused by the palm.
Figure 15. Example of an error caused by the palm.
Sensors 22 04554 g015
Figure 16. Example of an error caused by a finger.
Figure 16. Example of an error caused by a finger.
Sensors 22 04554 g016
Figure 17. Example of a double-hand word error caused by a finger.
Figure 17. Example of a double-hand word error caused by a finger.
Sensors 22 04554 g017
Table 1. Vision-based methods of automatic sign language recognition.
Table 1. Vision-based methods of automatic sign language recognition.
ReferencesMethodologyAcquisition ModeResults
(%)
Forehand/
Backhand
Limitation
   1. 2D approach
[22]24 ASL letters + DWT Gabor filter + KNNImage96.70ForehandFails to track SRM sign group
[23]24 ASL letters + Contour-based + ANNImage79.58ForehandFails to track SRM sign group
[24]100 ASL words + HOG + HMMVideo98.90ForehandFails to track SRM sign group
[26]27 hand gestures + Lightweight CNN modelImage97.25ForehandFails to track SRM sign group
[27]20 ASL words + CNN model + SVMVideo97.28ForehandFails to track SRM sign group
[28]36 ASL letters + CNN modelImage90.26ForehandFails to track SRM sign group
   2. 3D approach
    2.1 Single-hand model
[36]26 ASL letters + 30 feature-based + LSTMVideo91.82ForehandFails to track SRM sign group
[37]26ASL letters + Position-based + DNNVideo93.81ForehandFails to track SRM sign group
[38]26 ASL letters + 56 feature-based + HMMVideo86.10ForehandFails to track SRM sign group
[39]12 ASL words + Position-based + D-LSTMVideo90.00ForehandFails to track SRM sign group
[40]26 ASL letters + Distance-based + ANNVideo96.15ForehandFails to track SRM sign group
[41]26 ASL letters + Distance-based + GBMImage87.60ForehandFails to track SRM sign group
[3]26 ASL letters + Trajectory-based + LSTMVideo96.07BackhandFails to track SRM sign group
    2.2 Double-hand model
[12]40 words + Motion and angle-based + BiLSTMVideo97.98BackhandFails to track SRM sign group
   3. Multi-single- and double-hand models
[43]30 words + Angle-based + LSTMVideo96.00ForehandFails to track SRM sign group
[45]18 words + CNN modelVideo82.55ForehandFails to track SRM sign group
[46]49 words + Hand skeletal + FV + SVMVideo86.86ForehandFails to track SRM sign group
[48]50 words + Position-based + BiLSTM-NNVideo94.55ForehandFails to track SRM sign group
[6]56 words + Trajectory-based + HB-RNNVideo94.50BackhandFails to track SRM sign group
[2]57 words + Angle-based + FFV-BiLSTMVideo98.60BackhandFails to track SRM sign group
Table 2. Dataset description.
Table 2. Dataset description.
DatasetsNo. of
Participants
Frequency
(Times/Word)
No. of Samples
36 single-hand ASL words (Created by author)10103600
36 double-hand ASL words (Created by author)10103600
26 signed letters (A–Z letters) by [3]10205200
40 double-hand ASL words by [12]10104000
Total samples16,400
Table 3. Hardware specifications.
Table 3. Hardware specifications.
SystemsSpecification
Computer systemDell G3 Gaming w56691425TH
CPU: Intel Core i7-8750H
GPU: NVidia GeForce GTX 1050Ti
Memory Size: 8 GB DDR4
Leap Motion sensorVideo: 120 frames per second
Infrared camera: 2 cameras
Pixel: 640 × 240
Interaction zone: 80 cm
FOV: 150 × 120 degrees
Accuracy: 0.01 mm
Table 4. Classification parameter settings for the two-layer BiLSTM neural network.
Table 4. Classification parameter settings for the two-layer BiLSTM neural network.
LayerParameter OptionsValue
Input layerSequence lengthLongest
Batch size27
Learning rate0.0001
Input per sequence170
Feature vector1 dimension
Hidden layerBiLSTM layerLongest
Hidden node(2/3) × (input size per series) [3]
Activation functionSoftMax
Dropout layerDropout0.2
Output layerLSTM modelMany to one
Output classModel 1 = 26 classes
Model 2 = 40 classes
Model 3 = 72 classes
Table 5. Performance comparison of signed-letter recognition (letters A–Z).
Table 5. Performance comparison of signed-letter recognition (letters A–Z).
ReferenceAccuracy (%)Error (%)Precision (%)Recall (%)F1-Score (%)SD (%)
[37]93.816.19----
[3]96.073.93----
Proposed method97.342.6697.3997.3497.360.26
Table 6. Performance comparison in the recognition of 40 double-hand dynamic ASL words.
Table 6. Performance comparison in the recognition of 40 double-hand dynamic ASL words.
ReferenceAccuracy (%)Error (%)Precision (%)Recall (%)F1-Score (%)SD (%)
[12]97.982.0296.7697.4996.97-
Proposed method98.521.4898.5698.5298.540.22
Table 7. Performance comparison in the recognition of signed-words (72 words, including single and double-hand ASL words).
Table 7. Performance comparison in the recognition of signed-words (72 words, including single and double-hand ASL words).
MethodAccuracy (%)Error (%)Precision (%)Recall (%)F1-Score (%)SD (%)
Proposed method96.993.0197.0196.9997.001.01
Table 8. Performance comparison via an ablation test of different feature combinations in signed-letter recognition (letters A–Z).
Table 8. Performance comparison via an ablation test of different feature combinations in signed-letter recognition (letters A–Z).
Feature ExtractionAccuracy (%)Precision (%)Recall (%)F1-Score (%)SD (%)
H 2   +   H 3 92.3892.6792.3892.520.51
H 2   +   H 3   +   H 4 96.4196.4896.4196.440.29
H 1   +   H 2   +   H 3   +   H 4 97.3497.3997.3497.360.26
Table 9. Performance comparison of ablation test of different feature combinations in 40 double-hand dynamic ASL words.
Table 9. Performance comparison of ablation test of different feature combinations in 40 double-hand dynamic ASL words.
Feature ExtractionAccuracy (%)Precision (%)Recall (%)F1-Score (%)SD (%)
H 2   +   H 3 86.9588.9586.9587.940.60
H 2   +   H 3   +   H 4 95.5195.8895.5195.690.34
H 1   +   H 2   +   H 3     +   H 4 98.5298.5698.5298.540.22
Table 10. Signed-word recognition based on single-hand data based on a backhand view and using 5-fold cross validation.
Table 10. Signed-word recognition based on single-hand data based on a backhand view and using 5-fold cross validation.
Single Hand Approach
G.WordsAcc.
(%)
Error (%)SD
(%)
G.WordsAcc.
(%)
Error (%)SD
(%)
1Mouse96.20Lonely (2.2), Nephew (1.6)1.1710Respect96.40Are (2.6), Nephew (1)1.20
Lonely96.60Mouse (2.2), Niece (1.2)1.20Are97.00Respect (1), True (1), Mouse (1)0.89
2Grandfather95.60Grandmother (4), spit (0.3), vehicle (0.1)1.3611Endorse97.20Latin (1.8), Am (1)0.40
Grandmother97.20Grandfather (2), spit (0.8)0.75Latin96.20Endorse (2.8), Nephew (1)1.83
3Tend96.20Delicious (3.2), Vehicle (0.4), Spit (0.2)1.9412Shave97.40Yesterday (1.6), Fruit (0.6), Earing (0.4)0.80
Delicious96.60Tend (2.8), Spit (0.4), Vehicle (0.2)1.36Yesterday96.60Shave (2.8), Onion (0.6)0.80
4Hear97.60Whisper (1.8), Earring (0.3), Hair (0.3)0.4913Apple96.40Onion (2.4), Niece (0.6), Eagle (0.6)1.20
Whisper96.20Hear (2.4), Fox (0.6), Grandmother (0.4), Fruit (0.4)2.04Onion97.60Apple (2.2), Yesterday (0.2)0.49
5Better97.00Forget (2.2), Saturdays (0.8)1.0914Deny97.40Drop (2.2), Spit (0.4)0.80
Forget96.20Better (2.8), Nephew (1)2.64Drop98.00Deny (1.2), Vehicle (0.8)0.89
6Fox96.60Fruit (2.4), Earring (0.4), Hair (0.3), Whisper (0.3)0.8015Niece96.00Nephew (2.4), Lonely (1.6)2.45
Fruit97.40Fox (1.6), Whisper (1)0.49Nephew97.60Niece (1.2), Mouse (1), Respect (0.2)1.20
7Past96.80Know (2.2), Onion (0.6), Apple (0.4)0.9816Eagle96.80Egypt (2), Onion (0.9), Apple (0.3)0.98
Know98.00Past (2)0.63Egypt97.20Eagle (1.6), Hair (0.7), Latin (0.5)0.40
8Spit95.60Grandmother (3.4), Vehicle (1)1.6217Am98.00True (1), Latin (1)0.63
Vehicle95.80Spit (4.2)1.60True98.20Am (1.2), Spit (0.6)0.75
9Earring96.20Hair (2), Fox (1.4), Hear (0.4)0.9718Saturdays97.00South (2.4), Lonely (0.6)1.26
Hair96.80Earring (1.8), Fox (1), Hear (0.4)0.75South96.60Saturdays (2.8), Am (0.2), Lonely (0.2), True (0.2)1.36
Table 11. Signed-word recognition based on double-hand data based on a backhand view and using 5-fold cross validation.
Table 11. Signed-word recognition based on double-hand data based on a backhand view and using 5-fold cross validation.
Double Hands Approach
G.WordsAcc.
(%)
Error (%)SD
(%)
G.WordsAcc.
(%)
Error (%)SD
(%)
1Holiday98.20Retired (1), Fire (0.5), Embarrass (0.3)0.7410Until96.80To (2.2), Keep (0.6), Pure (0.4)1.47
Retired97.00Holiday (2), Fire (1)1.09To97.60Until (1.4), Pure (0.5), Keep (0.5)1.36
2Bath96.40Drum (3.2), Act (0.4)0.4911Embarrass97.80Fire (1.2), Holiday (1)0.98
Drum96.40Bath (2.6), Act (0.6), Science (0.4)0.80Fire95.60Embarrass (2.4), Introduce (1.2), Convince (0.8)0.80
3Act96.60Science (2.4), Drum (0.7), Bath (0.3)1.4912Cool96.20Fear (2.8), Shock (0.7), Sweat (0.3)1.33
Science98.60Act (1.4).0.80Fear97.00Cool (2), Street (1)1.09
4Carve98.40Page (1), Bath (0.6)0.4913Father96.60Mother (2.6), Check (0.4), Fire (0.4)0.80
Page97.00Carve (2), Bath (1)0.63Mother97.20Father (1.8), Check (0.7), Fire (0.3)0.40
5Major97.00Street (2), Convince (1)1.6714Check97.20Pay (2.2), Convince (0.6)1.17
Street95.80Major (2.6), Convince (1), Fear (0.6)1.72Pay97.60Check (1.4), Clean (0.5), Laid off (0.5)1.20
6Convince96.20Introduce (2.4), Major (0.5), Pay (0.5), Fire (0.4)1.4715Apply97.80Plug (2), Keep (0.1), Pure (0.1)0.98
Introduce96.80Convince (3), Fire (0.2)0.75Plug98.20Apply (1.4), Keep (0.4)0.40
7Clean97.00Laid off (2.4), Pay (0.6)0.7416Shock98.00Sweat (1), Fear (0.5), Cool (0.5)0.63
Laid off98.20Clean (1.2), Pay (0.4), Page (0.2)0.40Sweat97.00Shock (1.8), Cool (0.7), Fear (0.5)0.33
8Brother98.60Sister (1), Check (0.2), Keep (0.2)0.4917Society96.20Team (2.8), Drum (1)0.75
Sister98.00Brother (1.4), Keep (0.4), Pure (0.2)0.63Team97.20Society (1.8), Science (1)0.98
9Awake96.20Surprise (2.8), Embarrass (1)0.4018Keep96.80Sister (2), Pure (1), Brother (0.2)1.17
Surprise96.80Awake (2.2), Embarrass (1)0.40Pure97.20Keep (2.4), Sister (0.2), Brother (0.2)1.17
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chophuk, P.; Chamnongthai, K.; Chinnasarn, K. Backhand-Approach-Based American Sign Language Words Recognition Using Spatial-Temporal Body Parts and Hand Relationship Patterns. Sensors 2022, 22, 4554. https://doi.org/10.3390/s22124554

AMA Style

Chophuk P, Chamnongthai K, Chinnasarn K. Backhand-Approach-Based American Sign Language Words Recognition Using Spatial-Temporal Body Parts and Hand Relationship Patterns. Sensors. 2022; 22(12):4554. https://doi.org/10.3390/s22124554

Chicago/Turabian Style

Chophuk, Ponlawat, Kosin Chamnongthai, and Krisana Chinnasarn. 2022. "Backhand-Approach-Based American Sign Language Words Recognition Using Spatial-Temporal Body Parts and Hand Relationship Patterns" Sensors 22, no. 12: 4554. https://doi.org/10.3390/s22124554

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop