Next Article in Journal
Simultaneous Detection of Ammonium and Nitrate in Environmental Samples Using on Ion-Selective Electrode and Comparison with Portable Colorimetric Assays
Previous Article in Journal
Two-Dimensional Angle Estimation of Two-Parallel Nested Arrays Based on Sparse Bayesian Estimation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

American Sign Language Recognition Using Leap Motion Controller with Machine Learning Approach

Department of Electronics Engineering, Keimyung University, Daegu 42601, Korea
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(10), 3554; https://doi.org/10.3390/s18103554
Submission received: 11 September 2018 / Revised: 16 October 2018 / Accepted: 17 October 2018 / Published: 19 October 2018
(This article belongs to the Section Physical Sensors)

Abstract

:
Sign language is intentionally designed to allow deaf and dumb communities to convey messages and to connect with society. Unfortunately, learning and practicing sign language is not common among society; hence, this study developed a sign language recognition prototype using the Leap Motion Controller (LMC). Many existing studies have proposed methods for incomplete sign language recognition, whereas this study aimed for full American Sign Language (ASL) recognition, which consists of 26 letters and 10 digits. Most of the ASL letters are static (no movement), but certain ASL letters are dynamic (they require certain movements). Thus, this study also aimed to extract features from finger and hand motions to differentiate between the static and dynamic gestures. The experimental results revealed that the sign language recognition rates for the 26 letters using a support vector machine (SVM) and a deep neural network (DNN) are 80.30% and 93.81%, respectively. Meanwhile, the recognition rates for a combination of 26 letters and 10 digits are slightly lower, approximately 72.79% for the SVM and 88.79% for the DNN. As a result, the sign language recognition system has great potential for reducing the gap between deaf and dumb communities and others. The proposed prototype could also serve as an interpreter for the deaf and dumb in everyday life in service sectors, such as at the bank or post office.

1. Introduction

Communication connects people by allowing them to convey messages to each other, to express their inner feelings, and to exchange thoughts, either verbally or non-verbally. However, deaf community incapable of communicating verbally. As such, the debut of sign language was designed to assist hearing-impaired communities to express their feelings to others. However, sign language is entirely different from spoken language: it has its own grammar and its own manner of expression. Sign language is a non-verbal language that expresses one’s meaning by involving the movement of fingers, hands, arms, head, and body, as well as through facial expressions [1]. Therefore, it can be very challenging for society to learn and practice sign language. This communication barrier has become one of the main reasons that these disability communities (deaf and dumb) are being isolated from society. Fortunately, with the advancement of human–computer interaction technology over the past decades, humans have been able to interact with computers and, in return, receive feedback from the computer. With such methods, many sign recognition systems have been proposed that capture human gestures and then analyze and provide the recognized sign language output, either by text or verbally.
Various research projects have been conducted in recent years regarding sign language recognition. In general, the methods for recognizing sign language can be divided into two general classes: sensor- and vision-based approaches. The sensor-based approach utilizes various sensors that mostly attach to the hands to capture hand gestures. In fact, flex sensors and Inertial Measurement Units (IMUs) are to the two most common sensors for tracking hand gestures [2]. These sensors are used to detect the flexion of fingers and orientation of the hand. Preetham et al. [3] proposed a low-cost data glove with five flex sensors attached to the fingers. The data glove mapped the collected data to a character set for gesture recognition by using a minimum mean square error (MMSE) algorithm. In general, the flex sensor is attached to the finger to measure its bending angle by converting the bendiness into digital output. The higher the degree of the bend in the flex sensor, the higher the value of the digital output. However, the flex sensor is unable to measure the orientation of the finger and hand. Thus, the IMU sensor was adopted to solve this issue. Wang et al. [4] incorporated five unidirectional flex sensors and an IMU into a data glove which classified hand gestures by a template matching algorithm, resulting in a recognition accuracy rate of trained gestures of over 91%. For further improvement, Lee et al. [2] added two pressure sensors to a smart wearable hand device for recognizing 26 American Sign Language (ASL) letters using a support vector machine (SVM). The smart wearable device adopted five flex sensors on the fingers, a three-axis inertial motion on the back of the palm, and two pressure sensors on the middle finger (on the top and on the left, assuming a right-handed glove). The fusion method showed a tremendous improvement in the accuracy rate—from 65.7% to 98.2%—with the introduction of the pressure sensors. Meanwhile, Wu et al. [5] introduced a sensor fusion wearable system for ASL recognition in real time using IMU and surface electromyography (sEMG). An information gain-based feature selection scheme was utilized for the best feature selection subset, and the experimental result showed an accuracy of 96.16% by the SVM classifier. The study revealed that the accuracy rate improved from 3.84% to 15.12% with the addition of the sEMG features using four different classifiers. With a similar approach, Cheng et al. [6] extracted features from an accelerometer and sEMG signals for 223 Chinese Sign Language (CSL) characters with a recognition accuracy of 96%. Both studies indicated that the muscle electrical activity tended to vary according to the different movements of the hand and fingers. In addition, Shukor et al. [7] developed a similar wearable glove with the integration of 10 tilt sensors to measure the fingers’ motion, and an accelerometer was placed on the back of the palm to capture the hand motion. The study reported an average accuracy rate of 89% in ASL recognition. Similarly, Mummadi et al. [8] presented an IMU-based glove equipped with five IMU sensors placed on the fingertips to obtain pitch, roll, and yaw data. The study revealed that the random forest (RF) classifier with 15 subtrees gave the highest accuracy of 92.95% for the recognition of 24 ASL letters. However, sensor-based approaches are often found to be intrusive to the users, as the sensors or data glove can be very bulky and uncomfortable to wear.
Hence, the vision-based approach is an alternative for sign language recognition. This approach utilizes a video camera or visual sensor to capture the movements and analyzes the gestures with certain motion algorithms. Roh et al. [9] found that a dynamic Bayesian network (DBN) model with a five-state hidden node achieved the highest accuracy of 94.6% for 48 ASL signs. Moreover, Bheda et al. [10] presented ASL alphabet and digit recognition separately using a deep convolutional network with an accuracy rate of 67% for the ASL letters and 70% for the digits. The study captured images for each sign and applied background subtraction techniques to remove the background. Meanwhile, instead of background subtraction, Pan et al. [11] presented hand segmentation by extracting skin color in the YCbCr color space determined by applying a Gaussian mixture model (GMM) to find the largest skin blob. YCbCr is a family of color spaces used in video and digital photography systems, where Y, Cb, and Cr represent the luma, blue, and red components, respectively. Overall, this study achieved an accuracy rate of 99.8% for Chinese Sign Language (CSL) recognition and 94% for ASL by an SVM classifier. Beside color spaces, Kang et al. [12] introduced depth sensors to recognize 31 ASL using convolutional neural network with accuracy rate from 83.58% to 85.49% for new signers.
On the other hand, the Microsoft and Leap Motion companies have developed distinct ways to detect and track a user’s hand and body motion by introducing Kinect and the Leap Motion Controller (LMC), respectively. Kinect recognizes the body skeleton and tracks the hands, whereas the LMC only detects and tracks hands with its built-in cameras and infrared sensors [13,14]. Using of the provided framework, Sykora et al. [15] utilized the Kinect system to capture the depth information of 10 hand gestures to classify them using a speeded up robust features (SURF) method that reached up to an 82.8% accuracy rate. Similarly, Huang et al. [16] proposed 10-word-based ASL recognition system using Kinect by 10-fold cross-validation with an SVM that achieved an accuracy rate of 97% using a set of frame-independent features. Meanwhile, Chai et al. [17] also utilized the Kinect system to capture the 3D trajectory of the motion of both hands and the body skeleton to recognize word- and sentence-based CSL with an accuracy rate of 96.32%. Yang et al. [18] proposed a system using Kinect to capture 3D information with a hierarchical conditional random field (CRF) that recognized a 24-word-based sign language. The system revealed an accuracy rate of 90.4%. Khelil et al. [19] developed a recognition system for 28 letters and digits of Arabic Sign Language (ArSL) using the LMC which presented a 91% accuracy rate with an SVM classifier. A similar study was also conducted by Du et al. [20], where 10-digit gestures were recognized using the LMC with an SVM classifier. The study reported that the performance improved significantly, from 83.36% to 98.08%, with the inclusion of histogram of oriented gradient (HOG) features extracted from the LMC sensor data. Additionally, Kumar et al. [21] proposed a recognition system for 20 ArSL signs integrated into one LMC and two digital cameras, resulting in an accuracy rate of 95%. The LMC in the system was used to capture finger movements and the digital cameras were used for capturing body movements and facial expressions. However, Funasaka et al. [22] and Mapari et al. [23] focused on the static (no movements involved) ASL alphabet with the LMC system. Funasaka et al. [22] presented an accuracy rate of 82.71% using an integration of a decision tree and a genetic algorithm, whereas Mapari et al. [23] introduced a 10% hold-out validation multiple layer perceptron (MLP) with a 90% accuracy rate. Recognition of ASL signs that involve movement for certain letters can be challenging, and several studies, such as those by [22,23,24], were only able to detect the ASL alphabet when the dynamic ASL signs for the letters ’J’ and ’Z’ were excluded.
However, sign language is not standardized globally, similar to the fact that each country has its own language. Some examples of sign languages are ASL, ArSL, Greek Sign Language (GSL), CSL, Korean Sign Language (KSL), etc. [1]. The aimed contribution of this study was to develop an ASL recognition system using the LMC system that involves complete sets of 26 letters (A–Z) and 10 digits (0–9) of ASL (depicted in Figure 1); such a system is not found in other existing studies. The extracted features served as inputs to two different classifiers—namely, a support vector machine (SVM) and deep neural network (DNN)—and the predicted results were compared for their accuracy rates. There are 36 classes of ASL fingerspelling recognition utilized in this study. In fact, ASL fingerspelling recognition is an important milestone in facilitating communication between two communities, as ASL is not a code for English [25]. ASL is an independent language which use signs to convey a message, and fingerspelling is used to spell out a name of anything lacking a specific corresponding sign [10].
This paper is organized into five sections. Background research and the most recent related studies are presented in the first section to give the readers clear contextual and background information of the topic. The next section is the proposed system overview and details of the LMC. Next, Section 3 presents four subsections: data collection, feature extraction, feature preprocessing, and classification. In Section 4, the classification results and discussion are presented. The discussion of each ASL letter and digit pair correlation are discussed as well, and, lastly, the paper ends with a summary of the proposed study.

2. System Overview

Nowadays, the LMC is widely used in different areas, such as gaming, device control, interactive art, virtual reality, and other fields. The LMC is a low-cost and palm-sized portable peripheral device which is specifically designed to track hand and finger motion with high precision in the 3D Cartesian coordinate system, as illustrated in the sensor module of Figure 2. The device consists of two monochromatic cameras and three infrared light-emitting diodes (LEDs) [13] and has an interaction field of approximately 8 cubic feet above the device. The LMC is provided with an application programming interface (API) which allows users to obtain data on the hand and fingers, e.g., fingertip position, hand palm position, etc. [13]. Unfortunately, not all the provided information is useful and suitable for sign language recognition. For instance, the absolute hand and finger positions are not quite relevant for gesture detection, but integration with other information can provide more meaningful information for sign language recognition [19].
The proposed system in this study consists of four modules, namely, the sensor, preprocessing, processing, and output modules, as shown in Figure 2. In the sensor module, the LMC is used for data acquisition, which includes a three-axis hand palm position, the hand sphere radius, and the position of the five fingertips. All collected data are subjected to the preprocessing module in order to extract 23 meaningful features. Lastly, the processing module provides the classification results in the range of 1–36, where the outputs 1–26 represent the letters from ‘A’ to ‘Z’, and the outputs 27–36 represent the digits from ‘0’ to ‘9’. Generally, this study utilized two classifiers, SVM and DNN, for comparison purposes. Detailed discussions are presented in Section 3, below.

3. Method

3.1. Data Collection

In order to collect the necessary data, 12 voluntary subjects were recruited from the university to participate in the experiments. It is clarified that the current study is the initial study; thus, the focus subject group did not include deaf or dumb individuals or experts in sign language, though they will definitely be included for future work. Before the experiment, the subjects were given approximately 1 h to become familiar with ASL as well as the LMC operating method before the real experiment was conducted. Once the subjects were confidently familiar with the prototype system and the sign language, experiments were conducted as illustrated in Figure 3. The LMC device was connected to a desktop PC and placed on the table to detect and track the subject’s hand and finger gestures. The monitor displayed the 3D model of the subject’s current gesture. The subject was requested to perform all ASL gestures, including the 26 letters and 10 digits. A laboratory member with experience in ASL sat beside the subject to assist and verify the correctness of his/her hand gestures throughout the experiments. All data collection was completed within 10 s for each sign at a 60 Hz sampling rate. The duration for performing each sign was slightly different depending on the complexity of the sign. However, during the recording phase, the subjects were requested to hold the gestures, except for the letters ‘J’ and ‘Z’, which require movement.

3.2. Feature Extraction

In this study, the adopted basic features included the hand palm sphere radius, the hand palm position, and fingertip position, as depicted in Figure 4a. The hand palm sphere radius R measures the radius of a sphere fit to the curvature of the hand [13], as illustrated in Figure 4b, where the hand is imagined as holding a virtual sphere, with the radius of the sphere defined by the degree of curliness of the fingers curled into a fist. The two red dots illustrated in Figure 4b are the top and bottom points of the sphere, and the line connecting the two red dots is the diameter of the sphere; thus, half of the diameter is the radius of the sphere, the so-called “hand palm sphere radius”. The palm position P represents the absolute 3D position of the palm center, whereas the fingertip positions F i are the positions of the fingertips, with i = 1, …, 5 representing the index of the fingers in order of thumb, index finger, middle finger, ring finger, and pinky finger, respectively. Subsequently, given the aforementioned data, features were extracted and organized into five distinct groups, known as S, R, D, A, and L, as shown in Table 1. Group S features represent the standard deviation of the absolute 3D (x-, y-, and z- coordinates) palm positions, computed as depicted in Equation (1),
S = 1 n 1 ( i = 1 n ( P i P ) 2 )
where n represents the size of the dataset. n is defined as 20, which allows the algorithm to perform three calculations per second. Meanwhile, group R features denote the palm sphere radius, whereas group D features are the values of the 3D Euclidean distances of the palm center to each of the fingertips, calculated as shown in Equation (2).
D F P = ( F i x P x ) 2 + ( F i y P y ) 2 + ( F i z P z ) 2
Group A features represent the angles between two adjacent fingers, i.e., the angles between the thumb fingertip and index fingertip, index fingertip and middle fingertip, middle fingertip and ring fingertip, and ring fingertip and pinky fingertip. These four features were computed as Equation (3). However, the angle between the thumb fingertip and pinky fingertip was excluded because this angle includes palm curliness, which is included in feature R.
A a F i = F i F i + 1 π
Lastly, group L features indicate the distance between one fingertip and the consecutive fingertips, e.g., the distance between thumb and index finger, thumb and middle finger, thumb and ring finger, thumb and pinky finger, index finger and middle finger, and so on and so forth. There are 10 features in this feature group, and all of these features were calculated similarly by the 3D Euclidean method, as in Equation (4),
L F F = ( F i x F j x ) 2 + ( F i y F j y ) 2 + ( F i z F j z ) 2
where i and j = 1, …, 5 represent all five fingertips, while ij means there is no distance calculated between a given fingertip and itself. Overall, a total of 23 features were extracted and served as input parameters to the classifiers. All 36 ASL signs have different gesture representations of hand and finger shape; that is, these 23 features were targeted to represent the hand movement, palm flexion, relation of palm and fingertips, and relation of fingertips to fingertips. For instance, feature group S is targeted to represent the hand movement, feature group R is targeted to describe the flexion of the palm, and feature group D represents the relation of the palm and fingertips. Lastly, feature group A and L are targeted to describe the relation of all fingertips for all 36 ASL signs.

3.3. Features Preprocessing

In order to reduce the processing time and complexity, all the features were normalized in the range [0, 1]. Due to the differences of each subject’s palm size and finger lengths, all the related features were further scaled on a subject-by-subject basis, where the features were scaled as shown in Equation (5),
x = x / max ( x s )
where s represents the maximum value of the features of each individual subject. To further study the implication of each feature group to ASL recognition, the features were divided into six groups, with the combinations shown below:
  • C1: D + A + R + L (S excluded)
  • C2: S + A + R + L (D excluded)
  • C3: S + D + R + L (A excluded)
  • C4: S + D + A + L (R excluded)
  • C5: S + D + A + R (L excluded)
  • C6: S + D + A + R + L (all included)
Categories 1–5 are a unique combination of four feature groups, and the last category includes all the features as inputs to the classifier.

3.4. Classification and Validation

In this study, SVM and DNN classifiers were compared for ASL recognition. The training dataset is denoted as Equation (6),
( x j i , y j ) , y j { 0 , , 35 }
where x and y represent the i-th feature and label for j-th training dataset, respectively. However, an SVM is a binary classifier, where the output is either −1 (negative) or 1 (positive). The decision boundary of the SVM is also known as a separating hyperplane. Specifically, an SVM with a linear kernel function was utilized in this study, and the separating hyperplane is defined in Equation (7),
f ( x ) = w x b
where the j-th data is classified as 1 (positive) if f ( x ) > 0 , and −1 (negative) if f ( x ) < 0 . The data points used to generate the separating hyperplane are known as support vectors, and the nearest distance between the hyperplane and the support vector is known as a margin. Thus, Equation (8) shows the formula of finding the smallest margin among all support vectors.
min w , b = w 2 2
However, given the multi-classes problem with a total of 36 classes (26 letters and 10 digits) in this study, a multi-class SVM was implemented. As such, by utilizing the Scikit-learn machine learning framework library [26], the decision function was modified by implementing the “one-vs-rest” (OVR) decision method. OVR is also known as “one-vs-all” (OVA) or “one-against-all”, and treats a class (class to observed) as a positive label and the rest of the classes as negative labels. The process is repeated with each class treated as a positive label exactly once, for a total of n times round, where n is the total of classes (36 in this case). Then, the results are integrated for a multi-class classification according to the highest score of each binary classifier x j , where x represents the confidence value (score) and j refers to the designated classifier. Hence, x is classified according to the class of the highest confidence value. On the other hand, DNN was implemented in the TensorFlow machine learning framework with the high-level Keras API [27]. The adopted neural network consisted of two hidden layers, with 128 neurons for each layer. All neurons are densely connected using the rectified linear unit (ReLU) activation function to process the data before proceeding to the following layer. The ReLU activation function is computed as Equation (9) below:
f ( x ) = m a x ( 0 , x )
where the output is 0 if the input is less than 0, while the output is the raw input if the input is greater than 0. Recently, ReLU has been widely used for the hidden layers in deep learning networks. Meanwhile, the last layer is computed as 36 neurons (26 in cases without digits) with the Softmax activation function. The Softmax activation function returns an array of 36 probability scores with the sum total of 1, as shown in Equation (10),
σ ( z ) j = e j z k = 1 K e k z
where z is the input vector from the previous layer; j indexes the output units from 1 to k; and k = 26/36 in this study. Each output nodes contain the probability of the input instance belonging to one of the 26/36 classes. Lastly, all DNN layers were subjected to compilation using the adaptive moment estimation (Adam) optimizer and sparse categorical cross-entropy loss metric into a classification model for training and testing. The Adam optimizer by TensorFlow was set to a learning rate of 0.001, exponential decay rate for first-moment estimates of 0.9, exponential decay rate for second-moment estimates of 0.999, and the very small number of 1 e 8 was used to prevent any division by zero in the implementation. The sparse categorical cross-entropy loss metric is used for the classification task to calculate the error, as the neural network aims to minimize loss.
The leave-one-subject-out (LOO) approach was utilized for dataset training and testing, where the k 1 subjects’ dataset formed the training dataset, and the leftover subjects’ dataset was treated as the testing dataset. The process was repeated, with each subject dataset used for testing exactly one time. In another word, this approach trained and tested from different combinations of subjects’ datasets accordingly, and there was no crossing between testing data from the same subject at any validation loop. This subject-oriented approach avoided subject bias successfully. Based on the study by Arlot et al. [28], LOO is the best approach for classifier model training and testing, where the testing dataset not participating in the training process can derive high-confidence classification results even with new data from other users that were not in the training dataset. Moreover, the LOO cross-validation approach has been used in several sign language recognition studies [24,29,30].

4. Result and Discussion

In order to determine the flexibility and credibility of the trained classifier, the accuracies ( A C C ) of the classifiers were computed, as shown in Equation (11):
A C C = T P + T N T P + T N + F P + F N
where T P , T N , F P , and F N denote true positive (correctly identified), true negative (correctly rejected), false positive (incorrectly identified), and false negative (incorrectly rejected), respectively. Table 2 illustrates the accuracy results of the trained model for the SVM and DNN classifiers by applying six distinct feature groups as the training dataset. The preliminary result showed that the DNN significantly outperformed the SVM in both the 26-class (90.58% vs. 75.5%) and 36-class (85.65% vs. 67.54%) ASL recognition. Overall, the feature group that, on average, performed best as the training dataset, as determined by accuracy rate, is group C6 with a mean accuracy rate of 83.78%, followed by C4 with a mean accuracy of 83.25%, C2 with a mean accuracy rate of 80.45%, C3 with a mean accuracy rate of 78.95%, C1 with a mean accuracy rate of 78.31%, and last, with the lowest accuracy, is C5 with a mean accuracy rate of 74.17%. This implies that group L makes the most significant contribution to the ASL recognition system among all other features. C5 had the lowest mean accuracy rate, resulting in group L’s exclusion as a training dataset. Moreover, the results also indicated that the DNN classifier had the best performance with the C6 feature group at 93.81% and 88.79% for the 26 classes (26 letters only) and 36 classes (26 letters and 10 digits), respectively. However, the SVM classifier presented distinct results: C4 actually performed better than C6 as the training dataset for the 26 classes (C4 accuracy of 80.30% vs. C6 accuracy of 79.73%). However, C6 was still the best-performing feature group for the 36 classes with the SVM at an accuracy rate of 72.79% compared to the other feature groups. Overall, the DNN had the best ASL recognition performance using the C6 feature group for both the 26 and 36 classes.
In addition, Figure 5 and Figure 6 present the confusion matrices for the 26 classes and 36 classes, respectively, for both classifiers. To further investigate the true classification rate of each letter and digit, the sensitivity and specificity of the classes were calculated and are presented in Table 3. Sensitivity (Se), also known as the true positive rate, represents the percentage ability of the classifier to correctly identify the actual class. On the other hand, specificity (Sp), also known as the true negative rate, shows the ability of the classifier to correctly reject an instance which is not an actual class. Equations (12) and (13) below show the computation of sensitivity and specificity, respectively.
S e = T P T P + F N
S p = T N T N + F P
In terms of the 26 classes (see Figure 5), the letters ‘B’, ‘C’, ‘F’, ‘I’, ‘W’, and ‘Y’ had the highest sensitivities, i.e., 100% (highest confidence level), which indicates that those classes were perfectly classified by both the classifiers. The letters ‘L’ and ‘M’ had 100% sensitivities with the SVM classifier but only 99.83% for ‘L’ and 95.17% for ‘M’ with DNN. On the contrary, the letters ‘H’, ‘S’, and ‘U’ were found to have the lowest confidence levels. Specifically, with the SVM classifier, the letter ‘U’ had the lowest sensitivity (8.5%), followed by the letters ‘S’ (33.33%), ‘H’ (41.67%), and ‘X’ (50%). The letter ‘U’ was incorrectly identified as the letter ‘V’ (33%), ‘R’ (29%), and ‘H’ (25%), according to Figure 5a. Conversely, with the DNN classifier, the letter ‘S’ had the lowest sensitivity percentage of 65.83%, followed by the letters ‘U’ (70%) and ‘H’ (72.17%). The letter ‘S’ was misclassified as the letter ‘A’ (14.67%), ‘H’ (1.17%), ‘M’ (2.83%), ‘N’ (3.5%), ‘R’ (3.83%), and ‘T’ (8.17%), as shown in Figure 5b. Overall, the letters ‘U’ and ‘H’ were reported as the lowest confidence level alphabets in both the SVM and DNN classifiers. Both letters were mutually misclassified as each other and also easily misclassified as the letter ‘R’. The main reason could be that the letters ‘H’ and ‘U’ are highly similarity in term of finger gestures; the only difference is the orientation of the hand.
On the other hand, Figure 6 depicts the confusion matrix for 36 classes. The results indicate that the letters ‘C’, ‘I’, ‘L’, and ‘Y’ and the digits ‘5’ and ‘7’ had a perfect classification rate (100%). Nevertheless, there are several letters and digits that had sensitivity levels below 50% with the SVM classifier: the letters ‘D’ (16.17%), ‘O’ (50%), ‘U’ (8.33%), and ‘X’ (41.67%), and the digit ‘0’ (17.5%). Meanwhile, in terms of the DNN classifier, the letters ‘D’ (62.33%), ‘H’ (69.67%), ‘O’ (64.5%), ‘S’ (64.33%), ‘U’ (71%), and ‘V’ (74.17%), and the digit ‘0’ (71.33%) are the letters and digit with sensitivity levels below 75%. By referring to Table 3, the letters ‘D’, ‘O’, and ‘U’ and the digit ‘0’ had the lowest sensitivity percentage in both classifiers. In general, the letters ‘U’, ‘H’, and ‘R’ were often found to be mutually misclassified with both classifiers. Also, the letter ‘D’ was found to be misclassified as the digit ‘1’ in both classifiers with percentages of 41.67% and 27.67% for the SVM and DNN, respectively. Moreover, the letter ‘O’ and the digit ‘0’ were significantly correlated, as both have similar gestures, and were mutually misclassified with mean accuracy rates of 45.62% and 30.92% for the SVM and DNN, respectively.
Similarly, for the 36 classes with the DNN, the letters ‘B’, ‘C’, ‘I’, ‘L’, and ‘Y’ and the digits ‘5’, ‘7’, and ‘8’ had the highest specificity (100%) and the letter ‘D’ and digit ‘0’had the lowest (99%). Meanwhile, of the 36 classes with the SVM classifier, none had 100% specificity, but the lowest specificity was obtained for the digits ‘2’ (98.22%) and the highest specificity was obtained for the letters ‘J’ (99.99%) and ‘Y’ (99.99%). In the case of the 26 classes, the letters ‘B’ and ‘C’ had perfect specificity with both classifiers for the 26 classes. Further, the letters ‘F’, ‘W’, and ‘Y’ had 100% specificity with the DNN and 99.99% in the SVM. The lowest true negative rate with both classifiers was attained with the letter ‘A’: 99.14% for the SVM and 99.41% for the DNN. Generally, this indicates that both classifiers generally performed well in rejecting the non-classes correctly.
Class-to-class comparison was performed in order to further investigate the similarities between gestures. Figure 7 depicts the correlation (similarity percentage) of each ASL letter pairwise for both the SVM and DNN classifiers. A high correlation percentage between two letters signifies that both letters showed high similarity in terms of gestures. To the contrary, a low correlation value between two letters depicts significant differences between the two letters. Similar to the previous results, the paired letters ‘H’ and ‘U’ had relatively higher correlation than the others, with values of 34% and 42% for the SVM and DNN classifiers, respectively. Also, the letters ‘H’ and ‘R’ had correlation percentages of 12% with the SVM classifier and 18% with the DNN classifier. These three letters have similar gestures, with the index and middle fingers fully open and the other fingers fully bent (see Figure 1). The only significant difference is the fingers’ pointing direction and the placement of the index finger under the middle finger for the letter ‘R’. In addition, there were also several significant similarities found in the letter-to-digit comparison for both classifiers: ‘O’-‘0’; ‘D’-‘1’; ‘W’-‘6’; ‘F’-‘9’. Among these pairs, ‘O’-‘0’ had the most significant similarity percentage: 55% and 44% with the SVM and DNN classifiers, respectively. This is due to the letter ‘O’ and digit ‘0’ having the exact same finger gestures, where the only difference is the hand orientation (finger pointing direction). The digit ‘1’ also exhibited high similarity with the letter ‘D’ (SVM: 42%; DNN: 33%) as well as with the letter ‘P’ (SVM: 30%; DNN: 11%). The only difference between the digit ‘1’ and the letter ‘D’ is that the thumb is placed under the index finger in the letter ‘D’, rather than overlapped on the index finger as it is in the digit ‘1’. Also, both gestures had exactly the same gesture, with a minor difference in the bendiness of the pinky finger (see Figure 1). Moreover, the digit ‘1’ and the letter ‘P’ differ in the index finger bendiness, but the hand orientation and finger depth information cannot be obtained with the LMC in its current state; thus, they look alike from the LMC data. The letter ‘F’ and digit ‘9’ had an average similarity percentage of 30% for the SVM and 25% for the DNN; these signs are only differentiated by the distance between the index and middle fingers as well as between the middle and ring fingers. The digit ‘2’ is also highly similar to the letter ‘V’: the index and middle fingers for the letter ‘V’ are not bent (slightly bent for the digit ‘2’), but the distances between these fingers are slightly different, with a larger distance for the digit ‘2’ compared to the letter ‘V’. Likewise, the letter ‘W’ and the digit ‘6’ had a similarity percentage of 47% with the SVM and 23% with the DNN classifiers. These two signs are only distinguishable by the small difference in the location of the pinky finger and the slight difference in the hand orientation, as letter ‘W’ is oriented to the right by several degrees.
Last, but not least, Table 4 illustrates the comparison of our proposed methods with the systems and methods conducted by other researchers. It was observed that the most common classifier applied is the SVM classifier [19,20,24,31]. Du et al. [20] reported the highest accuracy rate (99.42%) with the SVM classifier but only with the recognition system of 10 digits (no letters involved). Likewise, a similar method was also employed by Khelil et al. [19], with an accuracy rate of 91.3%. Funasaka et al. [22] and Marin et al. [24] proposed an incomplete ASL recognition system that only recognized 24 letters, with letters ‘J’ and ‘Z’ excluded from the system, and they reported an accuracy rate of 82.7% (using a decision tree) and 80.86% (using an SVM classifier), respectively. On the other hand, an ASL recognition system of 32 letters and digits which did not include ‘J’, ‘Z’, ‘2’, and ‘6’ was proposed by Mapari et al. [23] by utilizing an MLP classifier, which resulted in an accuracy rate of 90%. In fact, even though other existing studies had better-performing ASL recognition rates, those studies did not include all 26 letters and 10 digits (complete ASL set) in their studies. However, our proposed methods utilized all ASL letters and digits, and the systems resulted in mean accuracy rate, sensitivity, and specificity values of over 90%.

5. Conclusions

This paper presents an American Sign Language recognition system which involves 26 letters and 10 digits using the Leap Motion Controller. A total of 23 features were adopted in the study and further divided into six different groups of combinations. The results indicate that the distance between one fingertip and the adjacent fingertips (group L) is a significant feature for sign language recognition. Generally, the DNN classifier outperformed the SVM classifier for both the 26-class and 36-class ASL recognition systems. Also, the sign language recognition system with the digits included presented a lower mean accuracy rate due to the large similarity between certain letters and digits. Moreover, even though the accuracy rate obtained in this study is lower compared to existing studies, our study included all ASL letters and digits, whereas other studies did not. There are several issues to be considered in future work, such as the methods to measure the finger and hand depth information, computation of the finger and hand orientation, and detection of overlapping between fingers. Other future works should also consider expanding the sign language recognition system to word- and sentence-based recognition, as well as to other languages, instead of limiting recognition to only ASL.

Author Contributions

Both authors contributed extensively to the work presented in this paper. Overall research ideals and methodology were developed by T.-W.C., and B.-G.L. Meanwhile, T.-W.C. conducted the experiment and wrote the code for data analysis and classification model under supervision of B.-G.L. This manuscript is drafted by T.-W.C. and then further revised by B.-G.L., who provided valuable insights in this manuscript.

Funding

This work was supported by the Basic Science Research Program through the Ministry of Education of the Republic of Korea and National Research Foundation of Korea (NRF-2017-1678-2017R1D1A1A09000975). This work was supported by the Keimyung University Research Grant of 2017.

Acknowledgments

The authors would like to thank the volunteers from university for participating in data acquisition during the experiment.

Conflicts of Interest

The authors declare no conflict of interest.

Ethical Statements

All subjects gave their informed consent for inclusion before they participated in the study. The study was conducted in accordance with the Declaration of Helsinki, and the protocol was approved by the Ethics Committee of 40525-201809-HR-88.

References

  1. Cheok, M.J.; Omar, Z.; Jaward, M.H. A review of hand gestures and sign language recognition techniques. Int. J. Mach. Learn. Cybern. 2017, 8, 1–23. [Google Scholar] [CrossRef]
  2. Lee, B.G.; Lee, S.M. Smart Wearable Hand Device for Sign Language Interpretation system with Sensors Fusion. IEEE Sens. J. 2018, 18, 1224–1232. [Google Scholar] [CrossRef]
  3. Preetham, C.; Ramakrishnan, G.; Kumur, S.; Tamse, A.; Krishnapura, H. Hand Talk-Implemented of a Gesture Recognition Glove. In Proceedings of the 2013 Texas Instruments India Educators’ Conference, Bangalore, India, 4–6 April 2013; pp. 328–331. [Google Scholar]
  4. Wang, J.; Zhang, T. An ARM-Based Embedded Gesture Recognition System Using a Data Glove. In Proceedings of the 26th Chinese Control and Decision conference (CCDC), Changsha, China, 31 May–2 June 2014; pp. 1580–1584. [Google Scholar]
  5. Wu, J.; Sun, L.; Jafari, R. A Wearable System for Recognizing American Sign Language in Real-Time Using IMU and Surface EMG Sensors. IEEE J. Biomed. Health Inf. 2016, 20, 1281–1290. [Google Scholar] [CrossRef] [PubMed]
  6. Cheng, J.; Chen, X.; Liu, A.; Peng, H. A Novel Phonology- and Radical-Coded Chinese Sign Language Recognition Framework Using Accelerometer and Surface Electromyography Sensors. Sensors 2015, 15, 23303–23324. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Shukor, A.Z.; Miskon, M.F.; Jamaluddin, M.H.; Ali@Ibrahim, F.; Asyraf, M.F.; Bahar, M.B. A New Glove Approach for Malaysian Sign Language Detection. In Proceedings of the 2015 IEEE International Symposium on Robotics and Intelligent Sensors (IRIS 2015), Langkawi, Malaysia, 18–20 October 2015; pp. 60–67. [Google Scholar]
  8. Mummadi, C.K.; Leo, F.P.P.; Verma, K.D.; Kasireddy, S.; Scholl, P.M.; Kempfle, J.; Laerhoven, K.V. Real-Time and Embedded Detection of Hand Gesture with an IMU-Based Glove. Informatics 2018, 5, 28. [Google Scholar] [CrossRef]
  9. Roh, M.C.; Lee, S.W. Human gesture recogntion using a simplified dynamic Bayesian network. Multimed. Syst. 2015, 21, 557–568. [Google Scholar] [CrossRef]
  10. Bheda, V.; Radpour, D. Using Deep Convolutional Networks for Gesture Recognition in American Sign Language. arXiv. 2017. Available online: https://arxiv.org/abs/1710.06836v3 (accessed on 15 August 2018).
  11. Pan, T.; Lo, L.; Yeh, C.; Li, J.; Liu, H.; Hu, M. Real-Time Sign Language Recognition in Complex Background Scene Based on a Hierarchical Clustering Classification Method. In Proceedings of the 2016 IEEE Second International Conference on Multimedia Big Data (BigMM), Taipei, Taiwan, 20–22 April 2016; pp. 64–67. [Google Scholar] [CrossRef]
  12. Kang, B.; Tripathi, S.; Nguyen, T.Q. Real-time Sign Language Fingerspelling Recognition using Convolutional Neural Network from Depth map. arXiv. 2015. Available online: https://arxiv.org/abs/1509.03001v3 (accessed on 9 August 2018).
  13. Leap Motion. Available online: https://www.leapmotion.com/ (accessed on 5 July 2018).
  14. Kinect for Windows. Available online: https://developer.microsoft.com/en-us/windows/kinect (accessed on 5 July 2018).
  15. Sykora, P.; Kamencay, P.; Hudec, R. Comparison of SIFT and SURF Methods for Use on Hand Gesture Recognition Based on Depth Map. AASRI 2014, 9, 19–24. [Google Scholar] [CrossRef]
  16. Huang, F.; Huang, S. Interpreting American sign language with Kinect. J. Deaf Stud. Educ. 2011, 20, 1281–1290. [Google Scholar]
  17. Chai, X.; Li, G.; Lin, Y.; Xu, Z.; Tang, Y.; Chen, X.; Zhou, M. Sign Language Recognition and Translation with Kinect. In Proceedings of the IEEE International Conference on Automatic Face and Gesture Recognition, Shanghai, China, 22–26 April 2013. [Google Scholar]
  18. Yang, H.D. Sign Language Recognition with the Kinect Sensor Based on Conditional Random Fields. Sensors 2015, 15, 135–147. [Google Scholar] [CrossRef] [PubMed]
  19. Khelil, B.; Amiri, H. Hand Gesture Recognition Using Leap Motion Controller for Recognition of Arabic Sign Language. In Proceedings of the 3rd International Conference on Automation, Control, Engineering and Computer Science (ACECS’ 16), Hammamet, Tunisia, 20–22 March 2016. [Google Scholar]
  20. Du, Y.; Liu, S.; Feng, L.; Chen, M.; Wu, J. Hand Gesture Recognition with Leap Motion. arXiv. 2017. Available online: https://arxiv.org/abs/1711.04293 (accessed on 5 July 2018).
  21. Kumar, P.; Gauba, H.; Roy, P.P.; Dogr, D.P. A multimodal framework for sensor based sign language recognition. Neurocomputing 2017, 259, 21–38. [Google Scholar] [CrossRef]
  22. Funasaka, M.; Ishikawa, Y.; Takata, M.; Joe, K. Sign Language Recognition using Leap Motion. In Proceedings of the International Conference on Parallel and Distributed Processing Techniques and Applications (PDPTA), Las Vegas, NV, USA, 25–28 July 2016; pp. 263–269. [Google Scholar]
  23. Mapari, R.B.; Kharat, G. American Static Sign Recognition Using Leap Motion Sensor. In Proceedings of the Second International Conference on Information and Communication Technology for Competitive Strategies 2016 (ICTCS ’16), Udaipur, India, 4–5 March 2016. [Google Scholar]
  24. Marin, G.; Dominio, F.; Zanuttigh, P. Hand gesture recognition with Leap Motion and Kinect devices. In Proceedings of the 2014 IEEE International Conference on Image Processing (ICIP), Paris, France, 27–30 October 2014; pp. 1565–1569. [Google Scholar]
  25. Lane, H.L.; Grosjean, F. Recent Perspective on American Sign Language; Psychology Press Ltd: New York, NY, USA, 2017. [Google Scholar]
  26. Scikit Learn Machine Learning in Python. Available online: https://scikit-learn.org/stable/index.html (accessed on 31 August 2018).
  27. TenSorflow. An Open Source Machine Learning Framework for Everyone. Available online: https://www.tensorflow.org/ (accessed on 31 August 2018).
  28. Arlot, S.; Celisse, A. A survey of cross-validation procedures for model selection. Stat. Surv. 2010, 4, 40–79. [Google Scholar] [CrossRef] [Green Version]
  29. Marin, G.; Dominio, F.; Zanuttigh, P. Hand gesture recognition with jointlycalibrated Leap Motion and depth sensor. Multimed. Tools Appl. 2016, 22, 14991–15015. [Google Scholar] [CrossRef]
  30. Simos, M.; Nikolaidis, N. Greek sign language alphabet recognition using leap motion device. In Proceedings of the 9th Hellenic Conference on Artificial Intelligence, Thessaloniki, Greece, 18–20 May 2016; p. 34. [Google Scholar]
  31. Chuan, C.; Regina, E.; Guardino, C. American Sign Language Recognition Using Leap Motion Sensor. In Proceedings of the 13th International Conference on Machine Learning and Applications, Miami, FL, USA, 4–7 December 2013; pp. 541–544. [Google Scholar]
Figure 1. The 26 letters and 10 digits of American Sign Language (ASL).
Figure 1. The 26 letters and 10 digits of American Sign Language (ASL).
Sensors 18 03554 g001
Figure 2. Proposed system flow used in the study.
Figure 2. Proposed system flow used in the study.
Sensors 18 03554 g002
Figure 3. Experimental setup with a real-time hand gesture in 3D graphic display.
Figure 3. Experimental setup with a real-time hand gesture in 3D graphic display.
Sensors 18 03554 g003
Figure 4. The position of (a) palm center and fingertips illustrated in (b) sphere radius mode.
Figure 4. The position of (a) palm center and fingertips illustrated in (b) sphere radius mode.
Sensors 18 03554 g004
Figure 5. Confusion matrices of 26 classes. (a) Support vector machine. (b) Deep neural network.
Figure 5. Confusion matrices of 26 classes. (a) Support vector machine. (b) Deep neural network.
Sensors 18 03554 g005
Figure 6. Confusion matrices of 36 classes. (a) Support vector machine. (b) Deep neural network.
Figure 6. Confusion matrices of 36 classes. (a) Support vector machine. (b) Deep neural network.
Sensors 18 03554 g006
Figure 7. Class-to-class similarity rate matrix for both classifiers (%). (a) Support Vector Machine. (b) Deep Neural Network.
Figure 7. Class-to-class similarity rate matrix for both classifiers (%). (a) Support Vector Machine. (b) Deep Neural Network.
Sensors 18 03554 g007
Table 1. Feature group organization.
Table 1. Feature group organization.
GroupFeature# of Feature
SStandard deviation of palm position3
RHand palm curvature radius1
DDistance between palm center and each fingertips5
AAngle between two adjacent fingertips4
LDistance between one fingertip and the consecutive fingertip10
Table 2. Overall accuracy of 26 classes and 36 classes.
Table 2. Overall accuracy of 26 classes and 36 classes.
Accuracy (%)
Combination26 Classes36 Classes-
SVMDNNSVMDNNAverage
C174.2687.8467.8583.2974.31
C274.5792.7767.0887.3880.45
C375.3588.6167.9483.8978.95
C480.3093.2972.0487.3583.25
C568.8187.1557.5383.1874.17
C679.7393.8172.7988.7983.78
Average75.5090.5867.5485.65-
Table 3. Sensitivity (Se) and specificity (Sp) of the ASL letters for 26 classes and 36 classes for both  classifiers.
Table 3. Sensitivity (Se) and specificity (Sp) of the ASL letters for 26 classes and 36 classes for both  classifiers.
36 Classes26 Classes
LettersSVMDNNSVMDNN
Se (%)Sp (%)Se (%)Sp (%)Se (%)Sp (%)Se (%)Sp (%)
A99.5098.8186.3199.5994.1799.1490.8299.41
B83.3399.76100.00100.00100.00100.00100.00100.00
C100.0099.55100.00100.00100.00100.00100.00100.00
D16.1799.4862.3399.0074.3399.3289.6799.81
E63.5099.0398.6799.9383.8399.5899.0099.99
F75.0099.4979.3399.80100.0099.99100.00100.00
G91.6798.7189.1799.8591.6799.5890.0099.89
H52.6799.0069.6799.2241.6799.5972.1799.77
I100.0099.75100.00100.00100.0099.91100.0099.99
J94.8399.9999.6799.9094.3399.9999.6710.00
K75.0099.0593.1799.8491.6799.5599.5099.97
L100.0099.75100.00100.00100.0099.9099.8399.99
M91.6798.3396.6799.79100.0099.5095.1799.90
N58.3399.6496.3399.7166.6799.8093.6799.91
O50.0098.1264.5099.2083.3399.9099.3399.99
P70.5098.6681.1799.6258.3399.5886.6799.90
Q79.1799.5096.3399.6570.8399.7992.1799.91
R87.3398.9596.6799.8275.0099.37100.0099.85
S33.3399.2864.3399.5533.3399.6765.8399.83
T91.6799.5298.6799.7291.6799.6794.8399.90
U8.3399.6271.0099.048.5099.7270.0099.72
V58.3399.2874.1799.2480.6799.3496.3399.95
W58.3398.5789.5099.48100.0099.99100.00100.00
X41.6799.6193.0099.9550.0099.9998.1799.96
Y100.0099.99100.00100.00100.0099.99100.00100.00
Z97.8399.9796.1799.9997.8399.9999.6799.99
017.5099.0271.3399.00----
161.8398.0876.5099.05----
283.3398.2277.3399.39----
374.6799.2196.8399.80----
478.0099.6697.6799.99----
5100.0099.61100.00100.00----
657.1799.0581.8399.70----
7100.0099.92100.00100.00----
895.5099.73100.00100.00----
974.3399.1092.8399.41----
Table 4. Comparison of sign language recognition systems using the Leap Motion Controller.
Table 4. Comparison of sign language recognition systems using the Leap Motion Controller.
AuthorGestureDataset (M people × N repetitions per letter per person)Cross-ValidationFeaturesClassifierAccuracy (%)
Khelil et al. [19]10 ArSL digits gesture10 people × 10 sets-Angle between two fingertips, angle between fingertips and hand’s normal, distance between the hand center to each fingertipsSVM91.3
Du et al. [20]10 digits gesture13 people × 20 sets80% training set, and 20% testing set, experiment performed 50 timesFingertips angle (A), fingertips distance (D), fingertips elevation(E), fingertips tip distance (T)SVM83.36
A+D+E+T+ HOGSVM99.42
Funasaka et al. [22]24 ASL gestures (static)--Palm normal vector, fingertips position, arm direction and fingertips directionDecision Tree82.7
Marin et al. [24]10 ASL gestures14 people × 10 setsLeave-one-subject-out cross-validationFingertips angle fingertips distance, and fingertips elevationSVM80.86
Chuan et al. [31]26 ASL gestures2 people × 2 sets4 fold cross-validationPinch strength, grab strength, average distance, average spread, average tri-spread, extended distance, dip-tip projection, OrderX, and anglek-Nearest Neighbors
SVM
72.78
79.83
Mapari et al. [23]32 ASL gestures (J, Z, 2 and 6 are excluded)146 people × 1 set90% training set, and 10% testing set cross-validationFinger positions, palm position, distance between positions, angles between positionsMLP90
Proposed work36 ASL12 people × 1 setLeave-one-subject-out cross-validationStandard deviation of palm positions, hand palm curvature radius, distance between palm center and each fingertips, angle between two adjacent fingertips, distance between fingertips and each consecutive fingertipsDNN93.81 (26 classes)
88.79 (36 classes)

Share and Cite

MDPI and ACS Style

Chong, T.-W.; Lee, B.-G. American Sign Language Recognition Using Leap Motion Controller with Machine Learning Approach. Sensors 2018, 18, 3554. https://doi.org/10.3390/s18103554

AMA Style

Chong T-W, Lee B-G. American Sign Language Recognition Using Leap Motion Controller with Machine Learning Approach. Sensors. 2018; 18(10):3554. https://doi.org/10.3390/s18103554

Chicago/Turabian Style

Chong, Teak-Wei, and Boon-Giin Lee. 2018. "American Sign Language Recognition Using Leap Motion Controller with Machine Learning Approach" Sensors 18, no. 10: 3554. https://doi.org/10.3390/s18103554

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop