Previous Article in Journal
When Fear Backfires: How Emotionality Reduces the Online Sharing of Threatening Messages
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Moroccan Sign Language Recognition with a Sensory Glove Using Artificial Neural Networks

1
Faculty of Sciences Dhar-Mahraz, Sidi Mohamed Ben Abdellah University, Fez 30000, Morocco
2
DAKAI Laboratory, Nextronic by Aba Technology, Casablanca 20253, Morocco
*
Author to whom correspondence should be addressed.
Digital 2025, 5(4), 53; https://doi.org/10.3390/digital5040053
Submission received: 12 August 2025 / Revised: 3 October 2025 / Accepted: 7 October 2025 / Published: 8 October 2025

Abstract

Every day, countless individuals with hearing or speech disabilities struggle to communicate effectively, as their conditions limit conventional verbal interaction. For them, sign language becomes an essential and often sole tool for expressing thoughts and engaging with others. However, the general public’s limited understanding of sign language poses a major barrier, often resulting in social, educational, and professional exclusion. To bridge this communication gap, the present study proposes a smart wearable glove system designed to translate Arabic sign language (ArSL), especially Moroccan sign language (MSL), into a written alphabet in real time. The glove integrates five MPU6050 motion sensors, one on each finger, capable of capturing detailed motion data, including angular velocity and linear acceleration. These motion signals are processed using an Artificial Neural Network (ANN), implemented directly on a Raspberry Pi Pico through embedded machine learning techniques. A custom dataset comprising labeled gestures corresponding to the MSL alphabet was developed for training the model. Following the training phase, the neural network attained a gesture recognition accuracy of 98%, reflecting strong performance in terms of reliability and classification precision. We developed an affordable and portable glove system aimed at improving daily communication for individuals with hearing impairments in Morocco, contributing to greater inclusivity and improved accessibility.

1. Introduction

Spoken language is a fundamental medium of communication, but it excludes individuals with hearing impairments who rely on sign language. For deaf or hard of hearing people, sign languages are their main way to talk and understand each other. These languages use hand signs, face expressions, and body movement. In Morocco, many deaf people use Moroccan Sign Language (MSL) to communicate. But most hearing people do not know it, so it is hard for deaf people to receive things like education, healthcare, or take part in society [1].
To address these barriers, researchers are developing technologies that translate sign language gestures into written or spoken forms. These solutions aim to improve communication and promote social inclusion for deaf individuals [2,3]. Thanks to advances in machine learning and embedded systems, it is now feasible to create recognition tools that function efficiently on compact, low-power devices [4,5].
One effective method involves sensor gloves equipped with motion detectors such as accelerometers and gyroscopes. These gloves capture precise hand movements, which are then processed by artificial neural networks to accurately identify gestures. The recognized signs can be converted into text or voice outputs and delivered through user-friendly platforms like smartphone applications [6,7,8].
While many sign language recognition systems focus on languages like Arabic Sign Language or American Sign Language, the lessons learned from those efforts are valuable for developing MSL-specific tools. Moroccan Sign Language is unique and plays a crucial role in the cultural identity of the deaf community but remains largely unknown to the wider population, which sustains communication barriers and social isolation [1].
Accordingly, we propose a glove-based embedded system tailored for MSL gesture recognition. Tailoring solutions to the local language and cultural context ensure better adoption and effectiveness compared to adapting systems designed for other sign languages [3].
The solution proposed here translates MSL alphabet gestures into text using a glove fitted with MPU6050 sensors, which combine accelerometers and gyroscopes to measure hand movement dynamics. A Raspberry Pi Pico microcontroller processes the sensor data, selected for its compactness, low power needs, and sufficient computing power.
Through the acquisition of motion data (see Figure 1), a specialized dataset is constructed to train an artificial neural network capable of recognizing individual letters in Moroccan Sign Language (MSL) [5]. Upon detection of a gesture, the corresponding character is instantly identified and processed, enabling real-time interpretation. Furthermore, an integrated audio sensor captures ambient sounds, enhancing the interaction and paving the way for potential bidirectional communication in future developments.
The remainder of this paper is structured to follow logical progression. Section 2 presents the related works, highlighting previous research efforts on sign language recognition systems using different sensing technologies and machine learning models, and positioning the current contribution within this context. Section 3 explores various existing approaches and technologies in gesture recognition. Section 4 outlines the criteria applied in selecting the recognition strategy best suited for this application. Section 5 provides technical insight into the system’s hardware, focusing on the integration of the Raspberry Pi Pico with MPU6050 sensors. Section 6 delves into the hardware configuration, highlighting how the Raspberry Pi Pico is interfaced with the MPU6050 motion sensors to enable precise gesture data collection. Section 7 concludes the paper by summarizing the primary findings, identifying current system limitations, and proposing future improvements aimed at strengthening communication support for the deaf and hard-of-hearing community in Morocco.

2. Related Work

Recent studies have explored smart glove-based systems for sign language recognition using various sensing technologies and machine learning models. For instance, Wong et al. [9] developed a multi-feature capacitive sensor system achieving high gesture classification accuracy using traditional ML algorithms. Ravan et al. [10] proposed an inductive sensor-based glove integrated with neural networks for dynamic gesture recognition. Yudhana et al. [11] employed both flex and IMU sensors for real-time translation of sign language into text. However, few studies have focused specifically on Moroccan Sign Language (MSL), and most existing works rely on vision-based methods [4,5,6], which require high computational resources and are unsuitable for low-cost, portable devices. Our work distinguishes itself by developing an embedded, low-power glove-based system for MSL recognition, integrating IMU sensors, real-time processing on a Raspberry Pi Pico, and a custom dataset exceeding 7000 samples.

3. System Design and Methodology

3.1. Hardware Architecture

The proposed system includes a smart glove equipped with five MPU6050 inertial sensors, each carefully positioned on a finger to capture precise motion data such as acceleration and angular velocity. To handle the issue of identical I2C addresses among the sensors, a TCA9548A I2C multiplexer, is incorporated, allowing the central controller to communicate sequentially with each sensor without conflict [12].
At the heart of the system is the Raspberry Pi Pico microcontroller [13], which manages data collection, filtering, and real-time classification of gestures via an embedded Artificial Neural Network (ANN). This ANN, trained in advance on a specialized dataset of Moroccan Sign Language (MSL) gestures, interprets the sensor data to identify the corresponding Arabic alphabet character.
When a gesture is recognized, the detected letter is immediately transmitted to a voice output module. This module utilizes either a digital-to-analog converter (DAC) or pre-recorded audio clips to produce the spoken equivalent of the recognized letter in Arabic, offering auditory feedback for hearing users.
An ambient microphone sensor is also included to capture environmental sounds, enhancing the system’s interactive features and enabling potential two-way communication in future updates.
The Raspberry Pi Pico, featuring a dual-core RP2040 microcontroller based on the ARM Cortex-M0+ architecture and designed by ARM Ltd., (Cambridge, UK), supports real-time data acquisition and preprocessing. Its multiple I2C buses and general-purpose input/output (GPIO) pins enable seamless interfacing with the multiplexer and various sensors. This optimized architecture ensures efficient conversion of physical gestures into digital signals with minimal latency, making it highly suitable for embedded intelligent systems.
The integration of these components results in a responsive system capable of capturing and interpreting gestures with high precision, forming the foundation for translating sign language into understandable visual and auditory formats, thereby improving communication accessibility for deaf users in Moroccan society.

3.2. Sensor Placement and Motion Data Acquisition

The effectiveness of the gesture recognition process largely depends on the precise placement of the motion sensors on the glove. In this system, each MPU6050 sensor is carefully positioned on a finger to ensure optimal capture of both linear acceleration and angular velocity across three axes. This arrangement enables the system to detect even subtle variations in finger movement, which is crucial for distinguishing between signs with similar trajectories or shapes. The sensor layout was designed to align with the natural articulation of the hand, providing reliable and consistent data acquisition during sign execution.
Although Figure 2 illustrates the theoretical placement of the sensors, it does not convey the physical assembly and integration of components. To complement this schematic, Figure 3 displays real photographs of the glove prototype from various angles, highlighting the wiring structure, sensor alignment, and overall ergonomic configuration of the wearable system.
The glove integrates five MPU6050 motion sensors—one per finger—connected via individual colored wires. The various angles show the placement of the sensors, the wiring layout, and the overall flexibility of the glove during gesture performance.
The MPU6050, manufactured by TDK InvenSense (San Jose, CA, USA), is an integrated 6-axis motion tracking device combining a 3-axis gyroscope and a 3-axis accelerometer. It allows precise measurement of angular velocity and linear acceleration, which is essential for real-time gesture recognition in embedded systems.
An additional OLED display (SSD1306, 0.96-inch) was integrated into the glove, enabling real-time visualization of recognized Moroccan Sign Language (MSL) letters. This ensures immediate textual feedback in addition to audio output.
The SSD1306 OLED display is a 0.96-inch monochrome screen commonly used in embedded systems. It communicates via the I2C interface and is sourced from manufacturers in China, such as Jiangxi Wisevision Optronics Co., Ltd., or OpenELAB Technology Ltd. The display provides clear real-time feedback, which is essential for interactive gesture recognition.
To improve the glove’s durability and user safety, all sensor wires were protected with flexible silicone sleeves to prevent accidental disconnections. A detachable connector interface was implemented between the MPU6050 sensors and the Raspberry Pi Pico to facilitate maintenance. In future iterations, the glove design will transition to a flexible PCB-based layout to enhance robustness, comfort, and long-term usability.

3.3. Embedded Machine Learning Implementation

The implementation of Embedded Machine Learning (EML) for sign language recognition was achieved by designing and training an Artificial Neural Network (ANN) model using Python 3.11 and TensorFlow 2.14 on Google Colab. The trained model was then quantized and deployed to a Raspberry Pi Pico microcontroller. The ANN architecture is composed of an input layer of 30 neurons, corresponding to six features (3-axis accelerometer and 3-axis gyroscope) from each of the five MPU6050 sensors. This is followed by a hidden layer with 20 neurons, and an output layer with 8 neurons, representing the recognized gesture classes.
The hidden layer utilizes the ReLU (Rectified Linear Unit) activation function [14]:
R e L U x = x     ,   x > 0   0     ,   x 0
The ReLU function is widely used in embedded machine learning due to its efficiency and its contribution to stable training without vanishing gradients [15].
The output layer applies the Softmax function [16]:
S o f t m a x ( x ) = e Z i j = 1 n e Z j
This function converts the raw output scores into a probability distribution over the 28 gesture classes, ensuring that the highest-probability output corresponds to the predicted letter in Moroccan Sign Language.
The ANN was trained using the Adam optimizer and categorical cross-entropy loss for multi-class classification, achieving over 97% accuracy on the test set. After training, the model was converted to TensorFlow Lite Micro (TFLM) format using post-training quantization, reducing both size and inference latency. [17]. This lightweight model runs directly on the Raspberry Pi Pico, with an inference time of less than 30 ms, making it suitable for real-time gesture recognition.
Our gesture recognition system is divided into three essential stages: data collection, training, and recognition (shown in Figure 4). In the first phase, five MPU6050 sensors—each placed on a different finger—record motion and orientation data across three axes, producing 30 numeric features per gesture. A group of 50 participants performed 28 unique hand gestures, each representing a letter of the Arabic alphabet, with multiple repetitions per gesture, leading to a dataset of over 7000 labeled samples. For each gesture, an audio file (.mp3) is also linked, enabling auditory feedback. Once gathered, the sensor data is cleaned and normalized, then stored in a structured database along with audio references. In the training phase, these processed features are used to train an Artificial Neural Network (ANN) capable of accurately distinguishing between gestures. Finally, in the recognition phase, the system captures a new gesture in real time, extracts its features, identifies the corresponding Arabic letter using the trained ANN, and outputs both the letter and its audio.

3.4. Dataset Construction

To ensure the development of a robust and realistic gesture recognition system, we constructed a dedicated dataset specifically tailored to Moroccan Sign Language (MSL), covering the 28 letters of the Arabic alphabet (ا to ي).
The data collection involved 50 voluntary participants, most of whom were non-signers, each wearing the smart glove equipped with five MPU6050 motion sensors—one per finger. Each participant performed multiple repetitions for each of the 28 signs, leading to a wide diversity of gesture executions in terms of hand orientation, gesture dynamics, speed, and personal variability.
In total, over 7000 labeled gesture samples were recorded. Each sample consists of:
A 30-dimensional feature vector (3-axis accelerometer + 3-axis gyroscope × 5 sensors);
A corresponding Arabic letter label; and
An associated audio file (.mp3) used for potential voice feedback.
The dataset was divided into training (70%), validation (15%), and test (15%) subsets to guarantee robust evaluation and reproducibility.
To ensure gesture quality and repeatability, the data collection phase was conducted under controlled conditions and extended over a period of more than three months. The recorded data were systematically filtered, normalized, and organized in a Raw Gesture & Audio Database, which will serve as the foundation for training and evaluating machine learning models in subsequent stages. An overview of the dataset used in this study is presented in Table 1.
Table 1. Dataset Overview.
Table 1. Dataset Overview.
Sign GestureRepetitions per ParticipantNumber of ParticipantsTotal Samples
Arabic Alphabet [5]Multiple50>7000
Figure 4. Block diagram of the system.
Figure 4. Block diagram of the system.
Digital 05 00053 g004

4. Training and Evaluation

4.1. Model Training

The artificial neural network (ANN) developed for gesture classification was designed to balance performance and computational efficiency, ensuring suitability for deployment on a resource-constrained embedded device such as the Raspberry Pi Pico.
Training was conducted using Python and TensorFlow on Google Colab, allowing access to GPU acceleration and efficient experimentation. The input consists of 30 motion features derived from five MPU6050 sensors, each delivering six-axis data (3-axis accelerometer and 3-axis gyroscope). This implementation was guided by official deployment examples provided by the Raspberry Pi Foundation for TensorFlow Lite on the RP2040 platform [11]
The proposed ANN architecture is composed of the following layers:
Input Layer: 30 neurons, corresponding to the real-time motion features from all fingers.
Dense Layer 1: 64 neurons with ReLU activation, enabling the network to learn complex nonlinear relationships efficiently.
Dense Layer 2: 32 neurons with ReLU activation, refining feature extraction and intermediate representation.
Dense Layer 3: 16 neurons with ReLU activation, adding depth while remaining lightweight enough for microcontroller inference.
Output Layer: 28 neurons with Softmax activation, each neuron representing one Arabic letter from Moroccan Sign Language (MSL). The Softmax function ensures that the output forms a valid probability distribution over the gesture classes.
Figure 5 illustrates a simplified schematic of the proposed artificial neural network (ANN) architecture used for Moroccan Sign Language gesture classification. While the number of neurons per layer is reduced for clarity, the structure reflects the actual sequence and function of the model’s layers as implemented on the Raspberry Pi Pico.
The network consists of an input layer (30 features), three hidden layers (64, 32, and 16 neurons, respectively, with ReLU activation), and an output layer with 28 neurons using Softmax activation.
The model was compiled with the following settings:
Optimizer: Adam, chosen for its adaptive learning rate and fast convergence
Loss function: Categorical cross-entropy, suitable for multi-class classification
Metrics: Accuracy, tracked during training and validation
After training on the constructed dataset (see Section 3.4), the model achieved strong classification performance and was subsequently quantized using TensorFlow Lite Micro (TFLM). The final quantized model had a memory footprint under 150 KB, and inference latency remained below 30 ms on the Raspberry Pi Pico, confirming its ability to support real-time gesture recognition with minimal delay.

4.2. Evaluation Metrics

To assess the performance and generalization capability of the trained artificial neural network (ANN), a range of evaluation metrics were monitored throughout the training process. These include:
Training accuracy: Measures how well the model fits the training data.
Validation accuracy: Evaluates performance on unseen samples during training.
Categorical cross-entropy loss: Indicates the model’s prediction error for multi-class classification.
Test accuracy: Computed on a held-out dataset to estimate real-world performance.
To better visualize the learning dynamics of the ANN, the evolution of training and validation accuracy, as well as the loss function, was plotted over 15 epochs. As shown in Figure 6 and Figure 7, the model demonstrates consistent improvement and convergence. The training and validation accuracy curves increase steadily and remain closely aligned, indicating low overfitting. Likewise, the training and validation loss curves exhibit a smooth and monotonic decrease, further confirming the effectiveness and stability of the training process.
Figure 6 and Figure 7 show a consistent improvement in both accuracy and loss over the training epochs. The training and validation accuracy curves remain closely aligned, reaching 98% and 96%, respectively, indicating strong generalization and stable learning. Similarly, the training and validation losses decrease steadily and converge below 0.1, confirming effective optimization and the absence of significant overfitting.
These results highlight the suitability of the proposed ANN architecture and dataset for accurate and efficient gesture classification.

4.3. Experimental Setup

To evaluate the gesture recognition system in real-world conditions, the final trained and quantized ANN model was deployed onto a Raspberry Pi Pico microcontroller. The experimental setup includes the following components:
A wearable glove equipped with five MPU6050 inertial sensors, each attached to a finger to capture 6-axis motion data (accelerometer and gyroscope).
A TCA9548A I2C multiplexer to resolve address conflicts and enable sequential sensor data acquisition.
The OLED display enhances usability by providing silent, visual feedback, which is particularly beneficial in noisy environments or when audio playback is not desired.
An OLED display module for visual output of recognized letters.
A voice output module (based on pre-recorded Arabic audio clips) connected to the Pico for auditory feedback.
A microphone sensor for potential environmental sound capture and future interaction extensions.
Sensor data were sampled at 50 Hz and passed through low-pass filtering to reduce noise. Real-time inference was performed directly on the Raspberry Pi Pico using the quantized TFLite Micro version of the trained model. Upon recognizing a gesture, the corresponding Arabic letter was displayed on the OLED screen and vocalized via the speaker module.
This hardware configuration proved to be compact, responsive, and suitable for embedded gesture recognition, with an average inference time of less than 30 ms per gesture.

4.4. Performance Evaluation

To validate these findings, the system was tested by multiple users performing each gesture in a variety of conditions, including different lighting environments and hand positions.
The deployed system was tested across various real-world scenarios to assess its responsiveness, reliability, and robustness under practical constraints. The evaluation focused on key performance indicators such as inference latency, accuracy consistency, and environmental robustness.
The model achieved an average inference time below 30 ms per gesture on the Raspberry Pi Pico, ensuring real-time interaction without noticeable delay. During testing, the system consistently maintained high recognition accuracy across different users, lighting conditions, and gesture execution styles.
In addition to accuracy metrics, the system’s stability was evaluated by running extended sessions in diverse indoor environments. No significant performance degradation or sensor failure was observed, confirming the viability of the hardware-software integration for portable and embedded use.
Overall, the glove-based system demonstrated reliable real-time performance, strong classification precision, and operational stability, making it a suitable assistive technology for daily use by deaf and hard-of-hearing individuals.
The 98% training accuracy refers to the best-performing model on the training set, while the 96% test accuracy represents performance on unseen data. Misclassification of certain letters is due to visual similarity, as discussed in Section 5.2.

5. Results and Discussion

5.1. Recognition Accuracy

The proposed ANN-based system achieved a peak test accuracy of 96%, with training accuracy reaching 98%, demonstrating the model’s ability to generalize well to unseen data. This high recognition rate is attributed to the well-structured dataset, precise sensor placement, and optimized network architecture. The real-time inference on the Raspberry Pi Pico further validated the system’s responsiveness and deployment readiness.

5.2. Limitations and Challenges

Despite strong overall performance, several limitations were observed:
Sensor drift over prolonged use can affect calibration and recognition reliability.
Variations in glove fit across users may alter motion dynamics and introduce noise.
Some signs involving subtle finger differences (e.g., س vs. ش, ح vs. خ) were more prone to misclassification due to gesture similarity.
Despite these isolated misclassifications, their overall impact on system accuracy was minor. The model maintained an average test accuracy of 96%, as the confused gestures represented less than 2% of the total dataset. This confirms the robustness of the proposed ANN model and the consistency of its generalization performance.
Addressing these issues may require more advanced preprocessing, adaptive calibration, or incorporating additional sensor modalities (e.g., flex or pressure sensors).

5.3. Real-Time Application

The system is compact, cost-effective (estimated under $25), and provides instantaneous visual and auditory feedback. This makes it particularly suitable for use in educational settings, healthcare, or public services to assist deaf individuals in daily communication scenarios such as transportation, shopping, or classroom interactions.

5.4. Comparative Analysis

To further contextualize the performance of the proposed system, a comparison was conducted with other relevant sign language recognition solutions. Table 2 summarizes the performance of our system compared to existing sensor-based and vision-based approaches for sign language recognition.
According to the updated Table 2, the proposed glove demonstrates high recognition accuracy while remaining low-cost, portable, and entirely embedded. Other approaches, such as inductive, capacitive, or flex sensor gloves, also show good performance but often require more complex design, higher cost, or limited durability.
Vision-based solutions generally achieve strong accuracy as well, but they depend on heavy computation and image preprocessing, which restricts their use in real-time wearable systems. These comparisons underline the strength of our design, which balances accuracy with simplicity, affordability, and suitability for embedded deployment.

6. Future Work and Improvements

Future work will explore advanced sequence models such as Long Short-Term Memory (LSTM) networks and lightweight Transformer encoders. These architectures are better suited for sequential sign language data and will be trained on our custom MSL dataset to assess performance improvements over the current ANN-based approach.
Extending recognition from isolated letters to continuous words and sentences will be a key priority in future research. Sequence modeling techniques will be applied to enable dynamic recognition for more natural and practical communication.
While the proposed system demonstrates strong performance in recognizing isolated letters from Moroccan Sign Language (MSL), several directions remain for future enhancement:
Alphabet-to-Word and Sentence Translation: Extending the system to recognize complete words or sentences will require sequence modeling techniques, such as Recurrent Neural Networks (RNNs), Long Short-Term Memory (LSTM) networks, or Transformer-based architectures.
Gesture Intensity and Duration Analysis: Incorporating analog signal processing or pressure/flex sensors could help distinguish gestures with subtle temporal or force-related variations, enhancing recognition of expressive content or emotion.
User Adaptation and Personalization: Implementing adaptive calibration routines would allow the glove to automatically adjust to different hand sizes and user movement profiles, improving long-term robustness.
Multimodal Fusion and Context Awareness: Integrating additional sensors (e.g., microphones, cameras, EMG) could support multimodal learning, enabling the system to better handle ambiguous or context-dependent gestures.
Analog Artificial Neural Networks (AANNs): Investigating neuromorphic computing and transistor-based analog circuits may enable ultra-low-power inference suitable for continuous use in wearable applications.
These future developments aim to move beyond isolated gesture recognition toward a fully autonomous, context-aware interface that better supports deaf and hard-of-hearing individuals.

7. Conclusions

This study presented a low-cost, portable, and real-time system for recognizing Moroccan Sign Language (MSL) alphabet gestures using a wearable glove equipped with motion sensors and an embedded artificial neural network (ANN). The system leverages five MPU6050 inertial sensors and a Raspberry Pi Pico microcontroller to perform gesture recognition directly on-device, eliminating the need for external computing resources. Through a carefully constructed dataset of over 7000 labeled samples collected from 50 participants, the ANN model was trained and optimized to achieve high accuracy and fast inference times. The system reached 98% training accuracy and 96% validation accuracy, with inference latency below 30 ms, validating its suitability for real-world deployment. Beyond its technical achievements, this system represents a meaningful step toward more inclusive communication tools for individuals who rely on sign language for the deaf and hard-of-hearing population in Morocco. Future work will focus on extending the system to full-word and sentence recognition, improving sensor integration, and exploring neuromorphic alternatives for ultra-low-power operation. This research provides a promising foundation for the development of intelligent, context-aware sign language recognition solutions that are both accessible and impactful.

Author Contributions

H.E.K. conceived and led the research work presented in this article, including the overall system design, data collection, hardware integration, experimentation, analysis, and manuscript writing. M.A.S. contributed significantly to the neural network development and the implementation of algorithms. A.B., I.E.M., Y.A. and A.A. provided technical support and administrative or academic guidance. All authors have read and agreed to the published version of the manuscript.

Funding

This research did not receive any specific grant from funding agencies in the public, commercial, or not-for-profit sectors.

Institutional Review Board Statement

In compliance with Moroccan Law No. 28-13 (2015) on the Protection of Participants in Biomedical Research, formal ethical approval was not required for this study, as it involved no medical interventions, collection of biological samples, or handling of personal or sensitive data. The procedures were entirely non-invasive: participants only wore a sensory glove to capture numerical finger movement data via dedicated software, with no identifying information recorded. All participants were verbally informed about the purpose of the study and voluntarily agreed to participate, in accordance with the principles of the Declaration of Helsinki (1975, revised in 2013).

Informed Consent Statement

Verbal informed consent was obtained from all participants involved in the study. Verbal consent was obtained rather than written because no personal or identifiable information was collected, and the procedure involved only non-invasive motion data recording.

Data Availability Statement

The dataset generated and analyzed during the current study is available from the corresponding author upon reasonable request. To enhance reproducibility, the complete source code and dataset will be made publicly available on GitHub upon acceptance of the manuscript.

Acknowledgments

The authors wish to sincerely acknowledge their debt to the anonymous reviewers for their highly accurate comments and their constructive help in improving this paper. The authors would also like to thank all the volunteers who participated in the data collection process, and the Sidi Mohamed Ben Abdellah University for logistical support. During the preparation of this manuscript, the authors used ChatGPT (OpenAI, version August 2024) for the purposes of improving the clarity and structure of the English language. The authors have reviewed and edited the output and take full responsibility for the content of this publication.

Conflicts of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Boukdir, A.; Benaddy, M.; Ellahyani, A.; El Meslouhi, O.; Kardouchi, M. Isolated Video-Based Arabic Sign Language Recognition Using Convolutional and Recursive Neural Networks. Arab. J. Sci. Eng. 2022, 47, 2187–2199. [Google Scholar] [CrossRef]
  2. Noor, T.H.; Alharbi, A.F.; Faisal, A.; Alrashidi, R.; Alsaedi, A.S.; Alharbi, G.; Alsanoosy, T.; Alsaeedi, A. Real-Time Arabic Sign Language Recognition Using a Hybrid Deep Learning Model. Sensors 2024, 24, 3683. [Google Scholar] [CrossRef] [PubMed]
  3. Alayed, A. Machine Learning and Deep Learning Approaches for Arabic Sign Language Recognition: A Decade Systematic Literature Review. Sensors 2024, 24, 7798. [Google Scholar] [CrossRef] [PubMed]
  4. Oulad-Naoui, S.; Ben-Abderrahmane, H.; Chagha, A.; Cherif, A. An LSTM-based System for Dynamic Arabic Sign Language Recognition. In Proceedings of the International Conference on Emerging Intelligent Systems for Sustainable Development (ICEIS 2024), Aflou, Algeria, 26–27 June 2024. [Google Scholar]
  5. El-Bendary, N.; Zawbaa, H.; Daoud, M.; Hassanien, A.E.; Nakamatsu, K. Sign Language Recognition for Arabic Alphabets Using Transfer Learning Technique. Int. J. Comput. Inf. Syst. Ind. Manag. Appl. 2010, 2010, 590–595. [Google Scholar]
  6. Al-Jarrah, A.; Al-Haj, M.; Al-Jarrah, M. Image Based Arabic Sign Language Recognition System. Int. J. Adv. Comput. Sci. Appl. 2018, 9, 185–194. [Google Scholar] [CrossRef]
  7. Rahman, M.A.; Rahman, M.M.; Islam, M.S. Real-Time Bangla Sign Language Translator. arXiv 2024, arXiv:2412.16497. [Google Scholar] [CrossRef]
  8. Wang, Y. A Chinese Sign Language Recognition System Using Leap Motion. Int. J. Robot. Autom. 2018, 33, 1–8. [Google Scholar]
  9. Wong, W.K.; Juwono, F.H.; Khoo, B.T.T. Multi-Features Capacitive Hand Gesture Recognition Sensor: A Machine Learning Approach. IEEE Sens. J. 2021, 21, 8441–8450. [Google Scholar] [CrossRef]
  10. Ravan, M.; Abbasnia, A.; Davarzani, S.; Amineh, R.K. A Smart Glove Based on Inductive Sensors for Hand Gesture Recognition. IEEE Trans. Hum.-Mach. Syst. 2025, 55, 549–558. [Google Scholar] [CrossRef]
  11. Yudhana, A.; Rahmawan, J.; Negara, C.U.P. Flex sensors and MPU6050 sensors responses on smart glove for sign language translation. IOP Conf. Ser. Mater. Sci. Eng. 2018, 403, 012032. [Google Scholar] [CrossRef]
  12. McDermott, S. Multiplexing Multiple I2C Sensors: TCA/PCA9548A; UTA Senior Design Knowledge Base: Arlington, TX, USA, 2024. [Google Scholar]
  13. Raspberry Pi Foundation. TensorFlow Lite on RP2040; GitHub: San Francisco, CA, USA, 2023. [Google Scholar]
  14. Nair, V.; Hinton, G.E. Rectified linear units improve restricted boltzmann machines. In Proceedings of the 27th International Conference on Machine Learning, Haifa, Israel, 21–24 June 2010. [Google Scholar]
  15. Glorot, X.; Bordes, A.; Bengio, Y. Deep Sparse Rectifier Neural Networks. In Proceedings of the 14th International Conference on Artificial Intelligence and Statistics (AISTATS) 2011, Fort Lauderdale, FL, USA, 11–13 April 2011. [Google Scholar]
  16. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  17. Saha, S.S.; Sandha, S.S.; Srivastava, M. Machine Learning for Microcontroller-Class Hardware: A Review. IEEE Sens J. 2022, 22, 21362–21390. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Arabic sign language alphabet [5].
Figure 1. Arabic sign language alphabet [5].
Digital 05 00053 g001
Figure 2. Sensor placement on the glove.
Figure 2. Sensor placement on the glove.
Digital 05 00053 g002
Figure 3. Multiple views of the sensor glove prototype designed for Moroccan Sign Language recognition.
Figure 3. Multiple views of the sensor glove prototype designed for Moroccan Sign Language recognition.
Digital 05 00053 g003
Figure 5. Simplified representation of the ANN architecture designed for embedded gesture recognition.
Figure 5. Simplified representation of the ANN architecture designed for embedded gesture recognition.
Digital 05 00053 g005
Figure 6. Training and validation accuracy over epochs.
Figure 6. Training and validation accuracy over epochs.
Digital 05 00053 g006
Figure 7. Training and validation loss over epochs.
Figure 7. Training and validation loss over epochs.
Digital 05 00053 g007
Table 2. Comparative Performance of Sign Language Recognition Systems.
Table 2. Comparative Performance of Sign Language Recognition Systems.
SystemRecognition MethodAccuracyPlatformSensor TypeDataset SizeRemarksCost
Our Work (MSL Glove)ANN on 5 × MPU6050 IMUs96%Raspberry Pi PicoIMU (Inertial)7000+ (Custom)Real-time, low power, fully embedded, portableLow
Ravan et al. [10]ANN on inductive sensors~95%FPGAInductive2000 (Custom)High accuracy, robust to noise, but higher fabrication complexity and costMedium–High
Al-Jarrah et al. [6]CNN on image sequences94%PC + CameraVision-based3000 (Public)High accuracy, requires image preprocessingHigh
Yudhana et al. [11]LSTM on flex sensors95%Arduino + BluetoothFlex + IMU2500 (Custom)Wireless, limited on-device processingMedium
Wong et al. [9]ML on capacitive sensors93%ARM Cortex-MCapacitive1800 (Custom)Lightweight, suitable for embedded devicesMedium
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

El Khoukhi, H.; Belatik, A.; El Manaa, I.; Sabri, M.A.; Abouch, Y.; Aarab, A. Moroccan Sign Language Recognition with a Sensory Glove Using Artificial Neural Networks. Digital 2025, 5, 53. https://doi.org/10.3390/digital5040053

AMA Style

El Khoukhi H, Belatik A, El Manaa I, Sabri MA, Abouch Y, Aarab A. Moroccan Sign Language Recognition with a Sensory Glove Using Artificial Neural Networks. Digital. 2025; 5(4):53. https://doi.org/10.3390/digital5040053

Chicago/Turabian Style

El Khoukhi, Hasnae, Assia Belatik, Imane El Manaa, My Abdelouahed Sabri, Yassine Abouch, and Abdellah Aarab. 2025. "Moroccan Sign Language Recognition with a Sensory Glove Using Artificial Neural Networks" Digital 5, no. 4: 53. https://doi.org/10.3390/digital5040053

APA Style

El Khoukhi, H., Belatik, A., El Manaa, I., Sabri, M. A., Abouch, Y., & Aarab, A. (2025). Moroccan Sign Language Recognition with a Sensory Glove Using Artificial Neural Networks. Digital, 5(4), 53. https://doi.org/10.3390/digital5040053

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop