You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

28 February 2023

A Machine Learning Based Full Duplex System Supporting Multiple Sign Languages for the Deaf and Mute

,
,
,
and
1
Telecommunications Engineering School, University of Malaga, 29010 Málaga, Spain
2
Institute of Oceanic Engineering Research, University of Malaga, 29010 Málaga, Spain
3
Airbus Defence and Space, Hertfordshire SG1 2AS, UK
4
Department of Computer Engineering, Sir Syed University of Engineering & Technology, Karachi 75300, Pakistan

Abstract

This manuscript presents a full duplex communication system for the Deaf and Mute (D-M) based on Machine Learning (ML). These individuals, who generally communicate through sign language, are an integral part of our society, and their contribution is vital. They face communication difficulties mainly because others, who generally do not know sign language, are unable to communicate with them. The work presents a solution to this problem through a system enabling the non-deaf and mute (ND-M) to communicate with the D-M individuals without the need to learn sign language. The system is low-cost, reliable, easy to use, and based on a commercial-off-the-shelf (COTS) Leap Motion Device (LMD). The hand gesture data of D-M individuals is acquired using an LMD device and processed using a Convolutional Neural Network (CNN) algorithm. A supervised ML algorithm completes the processing and converts the hand gesture data into speech. A new dataset for the ML-based algorithm is created and presented in this manuscript. This dataset includes three sign language datasets, i.e., American Sign Language (ASL), Pakistani Sign Language (PSL), and Spanish Sign Language (SSL). The proposed system automatically detects the sign language and converts it into an audio message for the ND-M. Similarities between the three sign languages are also explored, and further research can be carried out in order to help create more datasets, which can be a combination of multiple sign languages. The ND-M can communicate by recording their speech, which is then converted into text and hand gesture images. The system can be upgraded in the future to support more sign language datasets. The system also provides a training mode that can help D-M individuals improve their hand gestures and also understand how accurately the system is detecting these gestures. The proposed system has been validated through a series of experiments resulting in hand gesture detection accuracy exceeding 95%.

1. Introduction

There are millions of deaf and mute (D-M) people who heavily rely on sign language to communicate with others. They are often faced with the challenge of establishing effective communication when others are unable to understand or use sign language. In general, people with no physical impairment do not make an effort to learn sign language. This creates a scenario where it is difficult to establish effective communication between the D-M and non-deaf and mute (ND-M). To overcome this problem, a system is needed that provides a platform where the D-M and the ND-M can communicate with each other.
This work focuses on the use of information and communication technology (ICT) to provide a platform for which a system can be developed to support this type of communication. The research covers the development of a software application that is interfaced with hardware and uses different algorithms to solve some of the above-discussed problems. This is expected to be a step towards improving the integration of the D-M in society. The development presented in this manuscript uses existing sign language datasets to establish automated communication. The proposed system is designed around commercial-off-the-shelf (COTS) hardware integrated with a software application using machine learning (ML) algorithms that provide full-duplex communication.
The use of ML is providing a growing number of solutions to complex problems in diverse areas related to human impairment. ML is already applied to establish effective communication systems so that the D-M and the ND-M can communicate. Researchers are using existing algorithms, as well as creating new algorithms, to enhance the existing systems. ML algorithms are not only being deployed to solve existing problems, but these algorithms are also employed to increase efficiency by reducing processing time, creating consistent datasets, and improving quality.
This research is the continuation of the previously published work [1]. In this manuscript, the capability of the previous system is significantly enhanced, and new features include the new combined sign language datasets, American Sign Language (ASL), Pakistani Sign Language (PSL) and Spanish Sign Language (SSL). The online versions of these datasets are enhanced by including more images of better quality and variations of individual gestures. The updated system now includes a language detection feature and an improved user profile which contains more information such as right/left hand, speed and accuracy of user’s hand gestures, default user’s sign language dataset, etc. The proposed system is further validated using new experiments. A comparison of the three-sign languages is also carried out. A literature review is extended to help improve the proposed system. The software application is updated and includes the features mentioned here. Some new features, i.e., detecting gestures when the user is holding or wearing other objects like a wristwatch, ring, etc., are also added. The database is updated with new images, and the algorithm is further improved to exploit the new features.
In this manuscript, limitations of the existing systems are studied, and a new two-way communication system for the D-M and the ND-M is presented based on a new dataset created, which is the combination of ASL, PSL, and SSL, with a focus on overcoming limitations in existing systems. Within this research, the three datasets which are individually available are further enhanced to improve the quality of these datasets. The details of the proposed system are presented in this manuscript and validated using a series of experiments. The acquired hand gesture data is processed using Convolutional Neural Network (CNN), and further processing, discussed later, is implemented using supervised ML. Figure 1 shows the scope of this work, which encompasses the four main building blocks, which include CNN, supervised ML algorithms, selection and processing of sign language datasets, and review of hardware tools for hand gesture acquisition.
Figure 1. Research scope.
A summary of the section-wise layout of the manuscript is outlined here. Section 2 provides a detailed review of existing systems as part of the related work to understand the current state of research, novel features, limitations of existing systems, and prototypes. Section 3 presents the research methodology through 4 steps. The next sections, i.e., 4 and 5, provide the details of the proposed communication system and its validation through experimental setups. The results of these experiments are presented in Section 5. A discussion section is presented in Section 6, followed by the last Section 7, which covers conclusions and future work.

3. Research Methodology

In this section, the research methodology is presented to explain the several steps which were carried out in this research. Figure 2 shows the block diagram with the steps and relevant details.
Figure 2. Research methodology.

3.1. Design Research

In this step, two tasks are performed. The first task is to list the activities that are within the research scope. The scope of this work is to implement a two-way communication system where the D-M and ND-M can communicate with each other without the need to learn sign language through sign language interpretation and conversion to audio. The following task is to select certain areas to review the existing work, which in this case includes deep learning algorithms, ML algorithms, and image-capturing devices. These details are provided in Section 2, literature review.

3.2. Conduct Research

In this second step, an in-depth review of the selected categories is carried out. The details are presented in Section 2. In this manuscript, some existing products and prototypes, which are developed to facilitate D-M communication, are reviewed. For this review, both the software and hardware aspects are considered, including the development platform, i.e., PC or Android, and hardware devices to capture hand gesture data. The literature review also covers how the existing systems are designed, whether based on COTS or custom-made. In the case of custom-made devices such as sensor-based gloves, the calibration and build process are also reviewed. The next step is to select and review sign language datasets, machine learning algorithms, and deep learning algorithms. Within the review, the software applications are also considered, and LabVIEW is selected for software development. The final activity is to define experimental setups so that the proposed system can be validated. A detailed review is conducted in order to define how the different requirements for the proposed system can be mapped to the experiments and then validated. It is important to have an optimum number of experiments for validation as a means of avoiding repetition and duplication. The conclusion of the literature review is necessary to determine which parameters to use for implementing the proposed system.

3.3. Design Implementation

In this step, the tasks related to the implementation of the proposed system are considered. The features of the system are selected after conclusions drawn from the literature review. The design is implemented through five tasks, as listed in Figure 2. In the first step, the sign language data is collected and enhanced, i.e., some images are updated, and new images are added. The sign language datasets are selected, and information is stored in the database. The software, including the algorithm implementation, is done in LabVIEW. After completing the software development, hardware deployment, and integration tasks, the user manual and other documentation work are completed.

3.4. Validation and Conclusions

This is the final step, where the complete system validation is carried out. The validation activities include testing the updated language dataset, system response, processing times, and accuracy. The validated system is then tested for performance through different experimental setups, and the results are presented.

4. Proposed Communication System for D-M

In this section, the details of the proposed system are presented, including some key features.

4.1. Communication System Novel Features

Figure 3 shows the novel features of the proposed system, which provides full duplex communication between the D-M and the ND-M. This low-cost system is user-friendly and easy to install, with no operational cost to the end user. There is a training option available, which allows the features to be customized for individual users. The ML-based algorithm provides a continuous improvement option where new and higher quality data, i.e., hand gesture images, can replace an existing image. The proposed system supports American Sign Language, Pakistani Sign Language, and Spanish Sign Language. More languages can be added. Hand gestures data is acquired using LMD, which is a COTS device and, unlike a glove, does not require maintenance or calibration. The image data is processed using the CNN algorithm.
Figure 3. Novel features of the full duplex communication system.

4.2. Communication System Block Details

Figure 4 is a full duplex communication system between the D-M and the ND-M. The figure shows the implementation of the proposed system. The D-M person is provided with an interface where the hand gestures are acquired using an LMD connected to the PC-based system. The LMD captures hand gestures as images and forwards them to the PC, where it is further processed. The details of the processing are discussed later. The system supports multiple sign language datasets, i.e., ASL, PSL, and SSL. The system is reconfigurable, and more datasets can be added in the future. The acquired image data is processed through multiple stages and is then converted into a voice message.
Figure 4. Block diagram of the communication system.
The ND-M person can listen to the voice message. The speech or voice data is generated through the PC sound card output. In response, the ND-M person records a voice message which is acquired by the PC using sound card input. The voice message is then processed by the software application and converted into text and hand gesture images. The D-M person is then able to read the text or see the images. Similarly, an ND-M person can also initiate a conversation.
The proposed system is low-cost, user-friendly, and does not require any special training. These are the main features that make it easy and attractive for people to use this system. The sign language datasets selected for this system are based on the number of people using these sign language datasets. The selection of ASL, PSL, and SSL is based on the availability and size of datasets and the number of people using them. A small dataset means the size of the training dataset will also be small; hence the trained system will be less effective or accurate, while a large dataset results in higher accuracy but a slower response. Considering this, medium-sized datasets are selected.
The system provides a training mode where the user can check the accuracy of the system while undergoing training. In this mode, the hand gestures of the D-M person are processed, and the detected results are displayed without going through the text to speech to text conversion. The D-M person can vary hand gestures to check for detection accuracy. In this mode, the system also updates the database by replacing existing images with higher-quality images. The ML-based process compares the newly acquired images with the stored images and then decides between replacing the existing image with the new one or storing the new image along with the existing image or not storing the new image at all. Having more image files means an improved dataset which will increase accuracy, but this also means the system will require more processing time. It is important to have a balance between these parameters. This decision is taken by the ML-based implementation.
The ML algorithm also reviews the user-stored data, which helps to increase processing speed and accuracy. For example, if a user profile shows that the user only understands SSL, then the process will bypass the language detection step for this particular user. Similarly, the user can update the profile by adding more sign language datasets and other parameters. The system also maintains a performance record, both for the individual user and the overall system, using the stored user data and new data acquired through the training and normal modes.
Figure 5 shows the ML-based implementation. The block diagram shows the input, output, and hidden layers displaying various activities. The data is fed through the input layer, as shown by ‘1’. The input data includes both the hand gesture image data and the user data, which is stored as part of the user profile. In ‘2’, the user profile data is processed. The user can update the profile to reflect any changes. The next step, ‘3’, is the training mode. This step is used when the training mode is selected. In normal mode, the ML algorithm can use some options from this step. The language detection is done in step ‘4’. Currently, there are three datasets, but more can be added. The user-generated gestures are acquired by the proposed system, comparing them to the existing dataset, which is a combination of ASL, PSL and SSL. The system checks up to five gestures and then selects the language based on the results of the gestures matched to individual datasets of ASL, PSL or SSL. The acquired image goes through initialization in step ‘5’. In step ‘6’, the image goes through initial processing, where some features are detected and extracted. CNN algorithm is applied to further process the image data in ‘7’. The next two steps, i.e., ‘8’ and ‘9’, are for data storage. The results are generated through the output layer as marked by step ‘10’.
Figure 5. Machine learning algorithm.
Figure 6 shows some graphical results of hand gesture image processing using the ML algorithm. The figure shows the original hand gesture and its extracted Red, Green and Blue (RGB) planes. The images in the third row show the feature and edge extraction of the original image for the particular gesture. The results of convolution using kernel size 3 are also presented here. The graphical results of detecting fingers using a region of interest and the pixel count of bright objects, i.e., the borders, are also shown.
Figure 6. Examples of processed images using the ML algorithm.

5. Experimental Setup and Results

In this section, the proposed system is validated using 8 experiments that are specifically defined for this research. It is important to validate the key features and understand the limitations of the system. This is achieved using the experiments presented in this section.

5.1. System Validation Criteria

The communication system is validated through two stages. The first stage is to select the sign language dataset, which in this case are ASL, PSL, and SSL datasets. The second stage is to select a subset of these datasets. It is important to use a known input dataset and a predicted output result. This will help in understanding the system accuracy and limitations of the proposed system. Once the system is validated using these criteria, the full dataset is applied to further determine the performance of the system.
It is important to use a known set of images for initial validation. Outside the scope of this work is a study of how sign language datasets are created and validated.
To validate the proposed communication system, the ASL dataset images are used [46]. These are colored images of hand gestures. The SSL dataset used for validation is available online [47]. The PSL dataset is created by recording images of letters and numbers. A reference to PSL is also available from [48].
Letters and numbers from the three sign language datasets are used for validating the communication system. For this research work, 70% of the hand gesture images from the datasets are used for training, while the remaining 30% are used for testing. The split of 70% and 30% is a good ratio for training and testing. However, a different ratio can also be used. As an example, approximately 1800 images of letters and 700 images of numbers from ASL are used, which are then split into the ratio mentioned before. Figure 7, Figure 8 and Figure 9 show subsets of ASL, PSL, and SSL dataset images. Figure 7 shows five numbers and letters from ASL, while in Figure 8, eight letters from PSL are shown. In Figure 9, six letters from SSL are shown. The dataset used by the proposed system includes approximately 6000 hand gesture images from the three individual datasets.
Figure 7. A subset of the American sign language dataset.
Figure 8. A subset of the Pakistani sign language dataset.
Figure 9. A subset of the Spanish sign language dataset.

5.2. Experiment 1—Proposed Communication System Accuracy

In this experiment, the accuracy of the detection of hand gestures is presented. The formula for accuracy is presented in Equation (1). The accuracy is presented in percentages in Table 3. The accuracy values are also used to determine the performance of the algorithm.
Accuracy = Correct   Detection Total   number   of   tries
Table 3. Accuracy of detection.
In this experiment, the repeatability of the algorithm is evaluated. The system is validated by processing the gestures of all letters and numbers. For validation, the most accurate gesture detected by the system is selected and processed. The results are presented in Table 3. The accuracy is listed for letters and numbers of the three individual sign language datasets. The table also lists the size of subsets used for the training and testing of the system. It is also observed that the accuracy of the proposed system depends on the accuracy of the hand gestures, i.e., how accurately the gesture is made.

5.3. Experiment 2—Processing Individual Hand Gestures

In this section, hand gesture detection results are presented. In Figure 10, three letters from the PSL dataset are randomly selected for this experiment and processed through the proposed system. The hand gesture detection algorithm acquires the hand gesture and compares it with the images from the dataset stored in the database. The results show that the detection can exceed 96%. Depending on the input hand gesture, more than one image from the database can be matched, and the one with the highest percentage is selected.
Figure 10. Processing hand gestures (PSL dataset).
Figure 11 shows the accuracy and loss graphs for the PSL dataset. The graphs include both actual and test results.
Figure 11. Hand gesture detection results for PSL dataset.
Figure 12 shows the hand gesture detection results for the SSL dataset. The results are encouraging and show high accuracy. In this figure, three random gestures are selected from SSL, and the detection accuracy is shown. It is observed that the accuracy of the captured hand gesture is high when it is unique among the stored hand gesture images. As an example, if two different hand gesture images look similar, then the detection accuracy of these will be slightly lower.
Figure 12. Processing hand gestures (SSL dataset).

5.4. Experiment 3—Processing Hand Gestures Image Quality

In this section, the experiment is carried out by changing the quality of the hand gesture image and then processed through the proposed system. Figure 13 shows the four variants of a letter from the PSL dataset. All the images are correctly detected, thus confirming that the proposed system is able to detect the correct gesture for varying image quality. In this experiment, the image quality of the acquired hand gestures varies, and the results show that the system is able to detect the letters correctly. This is an important result since the acquired image quality can vary due to background color, sharpness, background texture, etc.
Figure 13. Processing images of different quality for the same hand gesture.

5.5. Experiment 4—Processing Variations in Same Hand Gestures

In this section, the variation in hand gestures is processed through the proposed communication system. In Figure 14, Figure 15 and Figure 16, two variants of letters from the PSL dataset are processed, and the detection accuracy is calculated. In this experiment, three letters from PSL are selected, and the hand gestures of these letters are varied slightly and then processed through the system. The results show that the accuracy is between 80% and 100% for the first letter in Figure 14. This variation is due to the distance between the hand and the LMD. A part of the index finger of the image on the left is outside the camera vision.
Figure 14. Processing variation in same hand gesture scenario 1.
Figure 15. Processing variation in same hand gesture scenario 2.
Figure 16. Processing variation in same hand gesture scenario 3.
A similar experiment is repeated in Figure 15, where the accuracy is 100% due to minimal variation between the images. The results shown in Figure 16 are similar to that in Figure 14. Again, the low accuracy for the image on the right side in Figure 16 is due to the distance of the hand from the LMD as well as the variation in the gesture.
In Figure 17, a letter from the ASL dataset is shown with a detection accuracy exceeding 90%. The accuracy for the two images is high for the letter selected from the ASL dataset, mainly because the distance between the hand and camera for both images is similar, and the variation between the two gestures is minimal.
Figure 17. Processing variation in same hand gesture scenario 4.

5.6. Experiment 5—Processing Similar Hand Gestures

Some hand gestures are similar, although not identical, and consequently, there is a possibility of wrong detection. In this section, results of similar hand gesture detection are presented. Figure 18 and Figure 19 show the detection results of two sets of letters from the PSL dataset. In Figure 18, two hand gestures from the PSL dataset are compared. There is a significant difference between the two gestures, as there is only one finger used for the gesture on the right but two fingers for the gesture on the left. The system, in this instance, is able to correctly detect the two letters. The experiment is repeated for two more letters from PSL in Figure 19, which are different, and the system is able to detect these letters correctly.
Figure 18. Processing similar gestures (PSL dataset) scenario 1.
Figure 19. Processing similar gestures (PSL dataset) scenario 2.

5.7. Experiment 6—Detecting Other Objects

This experiment focuses on other objects that can be visible while detecting hand gestures. The proposed system has a unique feature where these objects are also detected, and their information is ignored in some cases so that the correct hand gesture can be recognized. Figure 20, Figure 21 and Figure 22 show images with and without other objects. A unique feature of this system is validated using this experiment. Here the user is wearing a watch in Figure 20 and Figure 21 and a ring in Figure 22. It is observed that the system detection accuracy can vary if these objects are also captured with a hand gesture. Therefore, the images of objects like watches, rings, etc., are also stored in the database, enabling the proposed system to detect these objects and ignore them.
Figure 20. Detecting other objects—wristwatch scenario 1.
Figure 21. Detecting other objects—wristwatch scenario 2.
Figure 22. Detecting other objects—ring.
Figure 23 shows another example of an object captured with a hand gesture. The proposed system is unable to correctly detect the hand gesture here. In this experiment, the object, which is a pen, is detected by the system, but the system is unable to detect the hand gesture correctly. This is due to the shape of the object and also how the user is holding it.
Figure 23. Detecting other objects—pen.

5.8. Experiment 7—Algorithm Performance

In this section, the performance of the PSL dataset is evaluated and presented using a confusion matrix. This matrix shows the results for each letter compared with all other letters. The matrix is used in evaluating the performance of the algorithm and highlighting any errors. The confusion matrix for letters within the PSL dataset is shown in Figure 24 and Figure 25. This is an important metric for validating this type of system.
Figure 24. Confusion matrix of letters within PSL dataset (set 1).
Figure 25. Confusion matrix of letters within PSL dataset (set 2).

5.9. Experiment 8—Creating a New Dataset

In this experiment, the similarities between the three sign language datasets are presented. In Figure 26, three letters, one each from ASL, SSL, and PSL, are shown. These letters have similar hand gestures. This approach can be used to create a new sign language dataset which can be a combination of multiple sign languages. This figure provides insight into the process of creating a new dataset. In the normal validation process, a confusion matrix is created, in which each letter within the dataset is compared with the rest of the letters, and accuracy is calculated. The figure shows a scenario where the gestures taken from ASL, PSL, and SSL are identical. In this case, the system is unable to detect the correct input sign language based on just one hand gesture.
Figure 26. Similarities between different sign languages.

6. Discussion

In this manuscript, a full duplex communication system for the D-M and the ND-M is presented. The scope of this work includes ML, CNN, the use of multiple Sign Language datasets, and COTS hardware devices. The system presented in this manuscript is the continuation of previously published work [1]. The system uses three sign language datasets combined into one. This is a unique feature where the system automatically detects sign language. An extensive review of the existing systems is carried out and discussed in the related work section, including a tabular summary providing a summary of the review and listing some of the features of the existing systems. This also includes an analysis of some available datasets of different sign languages, followed by an evaluation of some machine learning techniques.
The proposed system is developed, implemented, and then validated using a four-stage research methodology. The key features of the proposed system include the implementation of CNN, ML, text-to-speech-to-text, a training interface, and a database storing the sign language datasets and user profiles. The system acquires hand gesture data using a COTS LMD device, and then the data is processed using the CNN algorithm. Using a COTS device means the system has a stable hardware interface to capture hand gesture images and does not require extensive calibration, which is normally needed when custom-made devices like a glove with sensors are used.
The system is validated through a series of experiments, and the limitations of some features are also defined. The experiments include hand gesture detection accuracy, processing of individual hand gesture data, evaluation of hand gestures with varying image quality, detection of variations in the same gestures, and the identification and processing of different hand gestures that look similar, detection of hand gestures with visible objects such as a watch, ring, etc., and finally automatic identification of a sign language dataset. The performance of the system is also evaluated using a confusion matrix.

7. Conclusions and Future Work

The system presented in this manuscript provides a communication platform for the D-M and the ND-M to communicate with each other without the need to learn sign language. A unique feature of this work is the creation of a new dataset that combines three existing sign language datasets. The proposed system automatically detects the hand gestures of a D-M person and converts them into a voice for an ND-M person after going through various processing stages. The system is reliable, easy to use, and based on a COTS LMD device to acquire hand gesture data which is then processed using CNN. Apart from CNN, a supervised ML algorithm is also applied for processing and automatically detecting sign language.
The system provides an audio interface for the ND-M and a hand gesture capture interface for the D-M. The system currently recognizes ASL, PSL, and SSL datasets, and the total number of images is approximately 6000. The system includes a training mode that can help individuals review how accurately the system is detecting their hand gestures. The users can create their unique profiles, which the system can use when processing hand gestures. The proposed system is validated through a series of experiments using the combined dataset of ASL, PSL, and SSL, and results show an accuracy exceeding 95%. The results are encouraging and repeatable. Another key feature of this system is that the combined dataset can be used by millions of people who use English, Urdu, and Spanish sign languages. The proposed system is to be installed and can be upgraded. There is no annual licensing; hence there is a low initial expense and no ongoing reoccurring cost. The database is updated, which improves the system’s performance and accuracy.
In the future, more sign language datasets can be processed and added. Dataset size can also be increased based on the recommendations from ML-based algorithms. Datasets can also be improved by adding videos and other types of data, including word-level hand gestures. More work can be done in the undertaking of comparisons between different sign languages, understanding similarities between them, and afterward combining these to create larger datasets. The proposed communication system can be further customized for individual users by adding more training features.

Author Contributions

Conceptualization, M.I.S. and S.N.; methodology, formal analysis, programming, and validation, M.I.S.; data curation, S.N.; writing, M.I.S., A.S., S.N., M.-A.L.-N. and E.N.-B.; visualization and software architecture and development, review of machine learning algorithm, A.S.; supervision, M.-A.L.-N. and E.N.-B. All authors have read and agreed to the published version of the manuscript.

Funding

This research has been partially funded by Universidad de Málaga, Málaga, Spain.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

No data is available.

Acknowledgments

The authors would like to acknowledge the support provided by Universidad de Málaga, Málaga, Spain.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Imran Saleem, M.; Siddiqui, A.A.; Noor, S.; Luque-Nieto, M.Á. A Novel Machine Learning Based Two-Way Communication System for Deaf and Mute. Appl. Sci. 2022, 13, 453. [Google Scholar] [CrossRef]
  2. Saleem, M.I.; Otero, P.; Noor, S.; Aftab, R. Full Duplex Smart System for Deaf Dumb and Normal People. In Proceedings of the 2020 Global Conference on Wireless and Optical Technologies, GCWOT, Malaga, Spain, 6–8 October 2020. [Google Scholar] [CrossRef]
  3. Sood, A.; Mishra, A. AAWAAZ: A communication system for deaf and dumb’. In Proceedings of the 2016 5th International Conference on Reliability, Infocom Technologies and Optimization, ICRITO 2016: Trends and Future Directions, Noida, India, 7–9 September 2016; IEEE: Noida, India, 2016; pp. 620–624. [Google Scholar] [CrossRef]
  4. Boppana, L.; Ahamed, R.; Rane, H.; Kodali, R.K. Assistive sign language converter for deaf and dumb. In Proceedings of the 2019 IEEE International Congress on Cybermatics: 2019 International Conference on Internet of Things (iThings) and IEEE Green Computing and Communications (GreenCom) and IEEE Cyber, Physical and Social Computing (CPSCom) and IEEE Smart Data (SmartData), iThings/GreenCom/CPSCom/SmartData, Atlanta, GA, USA, 14–17 July 2019; pp. 302–307. [Google Scholar] [CrossRef]
  5. Ameur, S.; Ben Khalifa, A.; Bouhlel, M.S. A novel hybrid bidirectional unidirectional LSTM network for dynamic hand gesture recognition with Leap Motion’, Entertainment Computing. Entertain. Comput. 2020, 35, 100373. [Google Scholar] [CrossRef]
  6. Ameur, S.; Ben Khalifa, A.; Bouhlel, M.S. Chronological pattern indexing: An efficient feature extraction method for hand gesture recognition with Leap Motion. Journal of Visual Communication and Image Representation. J. Vis. Commun. Image Represent. 2020, 70, 102842. [Google Scholar] [CrossRef]
  7. Suharjito, R.A.; Wiryana, F.; Ariesta, M.C.; Kusuma, G.P. Sign Language Recognition Application Systems for Deaf-Mute People: A Review Based on Input-Process-Output. Procedia Comput. Sci. 2017, 116, 441–448. [Google Scholar] [CrossRef]
  8. Deb, S.; Suraksha Bhattacharya, P. Augmented Sign Language Modeling (ASLM) with interaction design on smartphone-An assistive learning and communication tool for inclusive classroom. Procedia Comput. Sci. 2018, 125, 492–500. [Google Scholar] [CrossRef]
  9. Bhadauria, R.S.; Nair, S.; Pal, D.K. A survey of deaf mutes’, Medical Journal Armed Forces India. Dir. Gen. Armed Med. Serv. 2007, 63, 29–32. [Google Scholar] [CrossRef]
  10. Patwary, A.S.; Zaohar, Z.; Sornaly, A.A.; Khan, R. Speaking System for Deaf and Mute People with Flex Sensors. In Proceedings of the 2022 6th International Conference on Trends in Electronics and Informatics, ICOEI 2022-Proceedings, (Icoei), Tirunelveli, India, 28–30 April 2022; pp. 168–173. [Google Scholar] [CrossRef]
  11. Asthana, O.; Bhakuni, P.; Srivastava, P.; Singh, S.; Jindal, K. Sign Language Recognition Based on Gesture Recognition/Holistic Features Recognition: A Review of Techniques. In Proceedings of the 2nd International Conference on Innovative Practices in Technology and Management, ICIPTM 2022, Tirunelveli, India, 23–25 February 2022; pp. 713–718. [Google Scholar] [CrossRef]
  12. Farhan, Y.; Madi, A.A.; Ryahi, A.; Derwich, F. American Sign Language: Detection and Automatic Text Generation. In Proceedings of the 2022 2nd International Conference on Innovative Research in Applied Science, Engineering and Technology, IRASET 2022, Meknes, Morocco, 3–4 March 2022; pp. 1–6. [Google Scholar] [CrossRef]
  13. Bisht, D.; Kojage, M.; Shukla, M.; Patil, Y.P.; Bagade, P. Smart Communication System Using Sign Language Interpretation. In Proceedings of the Conference of Open Innovation Association, FRUCT. FRUCT Oy, Moscow, Russia, 27–29 April 2022; pp. 12–20. [Google Scholar] [CrossRef]
  14. Billah, A.R.; Fahad, F.I.; Raaz, S.R.; Saha, A.; Eity, Q.N. Recognition of Bangla Sign Language Characters and Digits Using CNN. In Proceedings of the 2022 International Conference on Innovations in Science, Engineering and Technology, ICISET 2022, IDubai, United Arab Emirates, 16–18 February 2022; pp. 239–244. [Google Scholar] [CrossRef]
  15. Jamdar, V.; Garje, Y.; Khedekar, T.; Waghmare, S.; Dhore, M.L. Inner Voice-An Effortless Way of Communication for the Physically Challenged Deaf Mute People. In Proceedings of the 2021 1st IEEE International Conference on Artificial Intelligence and Machine Vision, AIMV 2021, Gandhinagar, India, 24–26 September 2021; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2021. [Google Scholar] [CrossRef]
  16. Ben Hamouda, S.; Gabsi, W. Arabic sign Language Recognition: Towards a Dual Way Communication System between Deaf and Non-Deaf People. In Proceedings of the 22nd IEEE/ACIS International Conference on Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed Computing, SNPD 2021-Fall, Gandhinagar, India, 24–26 September 2021; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2021; pp. 37–42. [Google Scholar] [CrossRef]
  17. Shareef, S.K.; Haritha IS, L.; Prasanna, Y.L.; Kumar, G.K. Deep Learning Based Hand Gesture Translation System. In Proceedings of the 5th International Conference on Trends in Electronics and Informatics, ICOEI 2021, Tirunelveli, India, 3–5 June 2021; pp. 1531–1534. [Google Scholar] [CrossRef]
  18. Illahi AA, C.; Betito MF, M.; Chen CC, F.; Navarro CV, A.; Or IV, L. Development of a Sign Language Glove Translator Using Microcontroller and Android Technology for Deaf-Mute. In Proceedings of the 2021 IEEE 13th International Conference on Humanoid, Nanotechnology, Information Technology, Communication and Control, Environment, and Management, HNICEM 2021, Manila, Philippines, 28–30 November 2021; pp. 1–5. [Google Scholar] [CrossRef]
  19. Janeera, D.A.; Raja, K.M.; Pravin UK, R.; Kumar, M.K. Neural Network based Real Time Sign Language Interpreter for Virtual Meet. In Proceedings of the 5th International Conference on Computing Methodologies and Communication, ICCMC 2021, (Iccmc), Erode, India, 8–10 April 2021; pp. 1593–1597. [Google Scholar] [CrossRef]
  20. Snehaa, A.; Suryaprakash, S.; Sandeep, A.; Monikapriya, S.; Mathankumar, M.; Thirumoorthi, P. Smart Audio Glove for Deaf and Dumb Impaired. In Proceedings of the 2021 International Conference on Advancements in Electrical, Electronics, Communication, Computing and Automation, ICAECA 2021, Coimbatore, India, 8–9 October 2021. [Google Scholar] [CrossRef]
  21. Vanaja, S.; Preetha, R.; Sudha, S. Hand Gesture Recognition for Deaf and Dumb Using CNN Technique. In Proceedings of the 6th International Conference on Communication and Electronics Systems, ICCES 202, Coimbatre, India, 8–10 July 2021. [Google Scholar] [CrossRef]
  22. Rishi, K.; Prarthana, A.; Pravena, K.S.; Sasikala, S.; Arunkumar, S. Two-Way Sign Language Conversion for Assisting Deaf-Mutes Using Neural Network. In Proceedings of the 8th International Conference on Advanced Computing and Communication Systems, ICACCS 2022, Coimbatore, India, 25–26 March 2022; pp. 642–646. [Google Scholar] [CrossRef]
  23. Rosero-Montalvo, P.D.; Godoy-Trujillo, P.; Flores-Bosmediano, E.; Carrascal-Garcia, J.; Otero-Potosi, S.; Benitez-Pereira, H.; Peluffo-Ordonez, D.H. Sign Language Recognition Based on Intelligent Glove Using Machine Learning Techniques. In Proceedings of the 2018 IEEE 3rd Ecuador Technical Chapters Meeting, ETCM 2018, Cuenca, Ecuador, 15–19 October 2018; pp. 5–9. [Google Scholar] [CrossRef]
  24. Soni, N.S.; Nagmode, M.S.; Komati, R.D. Online hand gesture recognition & classification for deaf & dumb. In Proceedings of the International Conference on Inventive Computation Technologies, ICICT 2016, Coimbatore, India, 26–27 August 2016. [Google Scholar] [CrossRef]
  25. Anupama, H.S.; Usha, B.A.; Madhushankar, S.; Vivek, V.; Kulkarni, Y. Automated sign language interpreter using data gloves. In Proceedings of the 2021 International Conference on Artificial Intelligence and Smart Systems (ICAIS), Coimbatore, India, 25–27 March 2021; pp. 472–476. [Google Scholar] [CrossRef]
  26. Gupta, A.M.; Koltharkar, S.S.; Patel, H.D.; Naik, S. DRISHYAM: An Interpreter for Deaf and Mute using Single Shot Detector Model. In Proceedings of the 8th International Conference on Advanced Computing and Communication Systems, ICACCS 2022, Coimbatore, India, 25–26 March 2022; pp. 365–371. [Google Scholar] [CrossRef]
  27. Samonte, M.J.C.; Gazmin, R.A.; Soriano, J.D.S.; Valencia, M.N.O. BridgeApp: An Assistive Mobile Communication Application for the Deaf and Mute. In Proceedings of the ICTC 2019-10th International Conference on ICT Convergence: ICT Convergence Leading the Autonomous Future, Jeju, Republic of Korea, 16–18 October 2019; pp. 1310–1315. [Google Scholar] [CrossRef]
  28. Lan, S.; Ye, L.; Zhang, K. Attention-Augmented Electromagnetic Representation of Sign Language for Human-Computer Interaction in Deaf-and-Mute Community. In Proceedings of the 2021 IEEE USNC-URSI Radio Science Meeting (Joint with AP-S Symposium), USNC-URSI 2021-Proceedings, Singapore, 4–10 December 2021; pp. 47–48. [Google Scholar] [CrossRef]
  29. Sobhan, M.; Chowdhury, M.Z.; Ahsan, I.; Mahmud, H.; Hasan, M.K. A Communication Aid System for Deaf and Mute using Vibrotactile and Visual Feedback. In Proceedings of the 2019 International Seminar on Application for Technology of Information and Communication: Industry 4.0: Retrospect, Prospect, and Challenges, iSemantic 2019, Semarang, Indonesia, 21–22 September 2019; pp. 184–190. [Google Scholar] [CrossRef]
  30. Chakrabarti, S. State of deaf children in West Bengal, India: What can be done to improve outcome. Int. J. Pediatr. Otorhinolaryngol. 2018, 110, 37–42. [Google Scholar] [CrossRef] [PubMed]
  31. Li, D.; Rodriguez, C.; Yu, X.; Li, H. Word-level deep sign language recognition from video: A new large-scale dataset and methods comparison. In Proceedings of the 2020 IEEE Winter Conference on Applications of Computer Vision, WACV 2020, Snowmass Village, CO, USA, 1–5 March 2020; pp. 1448–1458. [Google Scholar] [CrossRef]
  32. Ronchetti, F.; Quiroga, F.; Estrebou, C.A.; Lanzarini, L.C.; Rosete, A. LSA64: An Argentinian Sign Language Dataset; Congreso Argentino de Ciencias de La Computacion (CACIC): San Luis, Argentina, 2016; pp. 794–803. [Google Scholar]
  33. Sincan, O.M.; Keles, H.Y. AUTSL: A large scale multi-modal Turkish sign language dataset and baseline methods. IEEE Access 2020, 8, 181340–181355. [Google Scholar] [CrossRef]
  34. Huang, J.; Zhou, W.; Li, H.; Li, W. Attention-Based 3D-CNNs for Large-Vocabulary Sign Language Recognition. IEEE Trans. Circuits Syst. Video Technol. 2018, 29, 2822–2832. [Google Scholar] [CrossRef]
  35. Tavella, F.; Schlegel, V.; Romeo, M.; Galata, A.; Cangelosi, A. WLASL-LEX: A Dataset for Recognising Phonological Properties in American Sign Language. arXiv 2022, arXiv:2203.06096. [Google Scholar]
  36. Joze, H.R.V.; Koller, O. MS-ASL: A Large-Scale Data Set and Benchmark for Understanding American Sign Language. In Proceedings of the 30th British Machine Vision Conference 2019, BMVC 2019, Cardiff, UK, 9–12 September 2019. [Google Scholar]
  37. Kagirov, I.; Ivanko, D.; Ryumin, D.; Axyonov, A.; Karpov, A. TheRuSLan: Database of Russian Sign Language. In Proceedings of the LREC 2020—12th International Conference on Language Resources and Evaluation, Conference Proceedings, Marseille, France, 11–16 May 2020; pp. 6079–6085. [Google Scholar]
  38. Kumari, M.; Singh, V. Real-time glove and android application for visual and audible Arabic sign language translation. Procedia Comput. Sci. 2019, 163, 450–459. [Google Scholar] [CrossRef]
  39. Siddiqui, A.; Zia MY, I.; Otero, P. A universal machine-learning-based automated testing system for consumer electronic products. Electronics 2021, 10, 136. [Google Scholar] [CrossRef]
  40. Siddiqui, A.; Zia MY, I.; Otero, P. A Novel Process to Setup Electronic Products Test Sites Based on Figure of Merit and Machine Learning. IEEE Access 2021, 9, 80582–80602. [Google Scholar] [CrossRef]
  41. Engineer Ambitiously—NI. Available online: https://www.ni.com/en-gb.html (accessed on 2 January 2023).
  42. Martinek, P.; Krammer, O. Analysing machine learning techniques for predicting the hole-filling in pin-in-paste technology. Comput. Ind. Eng. 2019, 136, 187–194. [Google Scholar] [CrossRef]
  43. Sapkota, S.; Mehdy, A.N.; Reese, S.; Mehrpouyan, H. Falcon: Framework for anomaly detection in industrial control systems. Electronics 2020, 9, 1192. [Google Scholar] [CrossRef]
  44. Welcome to Python. Available online: https://www.python.org/ (accessed on 1 January 2023).
  45. Dorochowicz, A.; Kurowski, A.; Kostek, B. Employing Subjective Tests and Deep Learning for Discovering the Relationship between Personality Types and Preferred Music Genres. Electronics 2020, 9, 2016. [Google Scholar] [CrossRef]
  46. Kaggle Dataset. Available online: https://www.kaggle.com/datasets/alexalex1211/aslamerican-sign-language (accessed on 3 January 2022).
  47. Kaggle Dataset. Available online: https://www.kaggle.com/datasets/kirlelea/spanish-sign-language-alphabet-static (accessed on 3 January 2022).
  48. Kaggle Dataset. Available online: https://www.kaggle.com/datasets/hazrat/urdu-speech-dataset (accessed on 3 January 2022).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.