Next Article in Journal
Classification of Copper Minerals by Handheld Laser-Induced Breakdown Spectroscopy and Nonnegative Tensor Factorisation
Previous Article in Journal
RGB-D Object SLAM Using Quadrics for Indoor Environments
Previous Article in Special Issue
A Hierarchical Learning Approach for Human Action Recognition
Open AccessArticle

British Sign Language Recognition via Late Fusion of Computer Vision and Leap Motion with Transfer Learning to American Sign Language

1
ARVIS Lab—Aston Robotics Vision and Intelligent Systems, Aston University, Birmingham B4 7ET, UK
2
School of Engineering and Applied Science, Aston University, Birmingham B4 7ET, UK
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(18), 5151; https://doi.org/10.3390/s20185151
Received: 6 August 2020 / Revised: 2 September 2020 / Accepted: 4 September 2020 / Published: 9 September 2020
(This article belongs to the Special Issue Sensor Systems for Gesture Recognition)
In this work, we show that a late fusion approach to multimodality in sign language recognition improves the overall ability of the model in comparison to the singular approaches of image classification (88.14%) and Leap Motion data classification (72.73%). With a large synchronous dataset of 18 BSL gestures collected from multiple subjects, two deep neural networks are benchmarked and compared to derive a best topology for each. The Vision model is implemented by a Convolutional Neural Network and optimised Artificial Neural Network, and the Leap Motion model is implemented by an evolutionary search of Artificial Neural Network topology. Next, the two best networks are fused for synchronised processing, which results in a better overall result (94.44%) as complementary features are learnt in addition to the original task. The hypothesis is further supported by application of the three models to a set of completely unseen data where a multimodality approach achieves the best results relative to the single sensor method. When transfer learning with the weights trained via British Sign Language, all three models outperform standard random weight distribution when classifying American Sign Language (ASL), and the best model overall for ASL classification was the transfer learning multimodality approach, which scored 82.55% accuracy. View Full-Text
Keywords: sign language recognition; multimodality; late fusion sign language recognition; multimodality; late fusion
Show Figures

Graphical abstract

MDPI and ACS Style

Bird, J.J.; Ekárt, A.; Faria, D.R. British Sign Language Recognition via Late Fusion of Computer Vision and Leap Motion with Transfer Learning to American Sign Language. Sensors 2020, 20, 5151.

Show more citation formats Show less citations formats
Note that from the first issue of 2016, MDPI journals use article numbers instead of page numbers. See further details here.

Article Access Map by Country/Region

1
Search more from Scilit
 
Search
Back to TopTop