Next Article in Journal
DP-YOLO: A Lightweight Real-Time Detection Algorithm for Rail Fastener Defects
Previous Article in Journal
Fast Spot Locating for Low-Density DNA Microarray
Previous Article in Special Issue
A Gesture Elicitation Study of Nose-Based Gestures
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Real-Time American Sign Language Interpretation Using Deep Learning and Keypoint Tracking

by
Bader Alsharif
1,2,*,
Easa Alalwany
3,
Ali Ibrahim
1,
Imad Mahgoub
1 and
Mohammad Ilyas
1
1
Department of Electrical Engineering and Computer Science, Florida Atlantic University, 777 Glades Road, Boca Raton, FL 33431, USA
2
Department of Computer Science and Engineering, College of Telecommunication and Information, Technical and Vocational Training Corporation, Riyadh 12464, Saudi Arabia
3
College of Computer Science and Engineering, Taibah University, Yanbu 46421, Saudi Arabia
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(7), 2138; https://doi.org/10.3390/s25072138
Submission received: 3 March 2025 / Revised: 25 March 2025 / Accepted: 26 March 2025 / Published: 28 March 2025
(This article belongs to the Special Issue Sensor Systems for Gesture Recognition (3rd Edition))

Abstract

:
Communication barriers pose significant challenges for the Deaf and Hard-of-Hearing (DHH) community, limiting their access to essential services, social interactions, and professional opportunities. To bridge this gap, assistive technologies leveraging artificial intelligence (AI) and deep learning have gained prominence. This study presents a real-time American Sign Language (ASL) interpretation system that integrates deep learning with keypoint tracking to enhance accessibility and foster inclusivity. By combining the YOLOv11 model for gesture recognition with MediaPipe for precise hand tracking, the system achieves high accuracy in identifying ASL alphabet letters in real time. The proposed approach addresses challenges such as gesture ambiguity, environmental variations, and computational efficiency. Additionally, this system enables users to spell out names and locations, further improving its practical applications. Experimental results demonstrate that the model attains a mean Average Precision (mAP@0.5) of 98.2%, with an inference speed optimized for real-world deployment. This research underscores the critical role of AI-driven assistive technologies in empowering the DHH community by enabling seamless communication and interaction.

1. Introduction

For millions of Deaf and Hard-of-Hearing individuals worldwide, communication barriers hinder their ability to engage effectively in everyday interactions. Traditional methods, such as sign language interpretation services, are often limited in availability, costly, and reliant on human interpreters. In an increasingly digital world, there is a growing need for intelligent assistive technologies that provide real-time, accurate, and accessible solutions to bridge the communication gap [1].
American Sign Language (ASL) is one of the most widely used sign languages, consisting of distinct hand gestures that represent letters, words, and phrases. However, existing ASL recognition systems often struggle with real-time performance, accuracy, and robustness across diverse environments. To address these limitations, this study introduces a real-time ASL interpretation system that integrates deep learning with keypoint tracking. By leveraging YOLOv11 for rapid hand gesture recognition and MediaPipe for precise landmark detection, the proposed approach enhances the efficiency and reliability of ASL interpretation [2,3].
At the core of our system, a built-in webcam functions as a non-contact optical sensor that captures real-time visual data. This vision-based sensor converts light into digital image frames, which serve as the primary input for gesture analysis. MediaPipe extracts 21 keypoints per hand from each frame to generate a skeletal representation, while YOLOv11 detects and classifies specific ASL alphabet letters based on these visual inputs. The webcam’s role as a sensing device enables the system to acquire gesture-based data continuously and reliably. It also ensures that the entire recognition pipeline—from sensing to classification—can operate in real time, even under different lighting conditions and backgrounds, using only standard hardware. This highlights the system’s practical viability as an accessible and scalable assistive technology.
In addition to enhancing communication accessibility, AI-driven ASL recognition systems have the potential to revolutionize various industries. In education, they can facilitate more effective learning environments for Deaf students, allowing them to interact seamlessly with teachers and peers. In healthcare, real-time ASL interpretation can bridge the gap between Deaf patients and medical professionals, ensuring better access to quality care [4]. Similarly, workplaces can become more inclusive by integrating ASL recognition tools, enabling Deaf employees to participate fully in discussions and decision-making processes. The integration of AI and deep learning in assistive technologies not only fosters inclusivity but also paves the way for more advancements in human–computer interaction, making communication more intuitive and accessible for individuals with hearing impairments [5,6].
The significance of this research lies in its potential to empower the Deaf community by providing an AI-driven tool that translates ASL gestures into text, facilitating smoother interactions in education, workplaces, healthcare, and social settings. By developing a robust and accessible ASL interpretation system, this study contributes to the advancement of assistive technologies aimed at promoting inclusivity and bridging communication barriers for the DHH population [7].
The key contributions of our research paper can be summarized as follows:
  • Developed a real-time AI-driven ASL interpretation system capable of accurately spelling out names and locations through hand gesture recognition.
  • Integrated YOLOv11 for gesture detection and MediaPipe for hand tracking, ensuring both high accuracy and fast inference speeds.
  • Implemented keypoint-based annotation techniques to enhance the model’s ability to capture subtle variations in finger positioning, improving classification precision.
  • Utilized a standard built-in webcam as a vision-based optical sensor for gesture data acquisition, demonstrating the system’s practical application in real-time human–computer interaction and AI-driven assistive technology.
  • Achieved an impressive 98.2% mean Average Precision (mAP@0.5), demonstrating the system’s reliability in recognizing ASL letters.
  • Curated and processed a large-scale dataset of 130,000 ASL hand gesture images, each annotated with 21 keypoints to enhance model learning and generalization.
The research paper is structured as follows: Section 2 and Section 3 discuss relevant research and background concepts, including Transfer Learning, YOLO, and MediaPipe, which are fundamental to real-time ASL alphabet recognition. Section 4 elaborates on the methodology used in this study, detailing dataset preparation, preprocessing, augmentation, annotation, and data balancing techniques. Section 5 presents and discusses the results obtained, evaluating model performance through precision–recall analysis, confidence-based evaluations, and confusion matrix analysis, along with the real-time system’s robustness and accuracy. Finally, a general conclusion is provided in Section 6.

2. Background

2.1. Transfer Learning

Transfer learning is a deep learning technique that adapts a model pre-trained on a large dataset to a new, related task with a smaller dataset. Instead of training from scratch, this approach leverages the pre-trained model’s knowledge, improving generalization and reducing training time. It is particularly advantageous in object detection, where pre-trained models learn fundamental visual features—such as edges, textures, and shapes—that can be fine-tuned for specific applications [8,9].
For this task, a pre-trained YOLO11 model, originally trained on the COCO dataset, serves as the base model. This model has already learned essential visual patterns, making it well suited for hand detection. By utilizing its embedded knowledge, transfer learning enables adaptation for ASL-specific detection. During this process, the backbone of YOLO11—responsible for extracting core visual features—remains frozen, preserving its learned representations. Since the lower convolutional layers are already optimized for detecting edges, textures, and object structures, retraining them is redundant. Freezing these layers retains essential knowledge, prevents unwarranted modifications, and reduces computational overhead, allowing the model to focus on refining task-specific features [10].
While the backbone remains unchanged, the Neck and Head layers are fine-tuned to specialize in detecting ASL alphabet signs. These layers play a critical role in refining feature representations, predicting bounding boxes, and accurately classifying hand gestures. Fine-tuning involves several optimization strategies, including using a lower learning rate for stable weight updates, restricting gradient updates to trainable layers to preserve pre-trained features, and applying data augmentation techniques (e.g., flipping, rotation, and lighting variations) to enhance generalization [11].
Throughout this phase, the model progressively learns to distinguish between ASL letters, refine bounding box predictions, and improve classification accuracy, ensuring the reliable recognition of hand gestures. See Figure 1 for a visual representation of the workflow of the transfer learning methodology using the YOLO model.

2.2. YOLO

YOLO (You Only Look Once) is a state-of-the-art real-time object detection model designed for high-speed and accurate detection. YOLOv11 is an advanced version that builds upon previous iterations with improved performance in localization, detection accuracy, and efficiency. It is particularly optimized for detecting small and complex objects, making it ideal for applications such as ASL hand gesture recognition [12].

2.2.1. Detection Pipeline

YOLOv11 follows a single-stage detection pipeline, meaning it processes the entire image in one pass instead of multiple steps like region-based models (e.g., Faster R-CNN) [13].
The input image is divided into grid cells (e.g., a 640 × 640 resolution image is divided into 32 × 32 cells). Each cell is responsible for detecting objects within its region. YOLOv11 utilizes a deep convolutional neural network (CNN) to extract meaningful spatial features from the image, ensuring efficient object detection [14].

2.2.2. Bounding Box Prediction

Each grid cell predicts multiple bounding boxes [15], including
  • (x, y): Center coordinates of the object.
  • (w, h): Width and height relative to the image size.
  • Confidence Score: Probability of an object being present.
Since objects may overlap, multiple bounding boxes are generated to handle these cases effectively.

2.2.3. Class Prediction

Each bounding box is assigned a class probability score using softmax classification [16]. For ASL hand gesture detection, the class labels range from 0 to 25, corresponding to the A–Z alphabet.

2.2.4. Non-Maximum Suppression (NMS)

YOLO follows a technique called Non-Maximum Suppression (NMS), where multiple boxes may predict the same object. NMS removes redundant detections by keeping only the highest-confidence detection while suppressing lower-confidence overlapping boxes [17].

2.2.5. Final Output

The final output consists of localized bounding boxes and class labels. In ASL detection, YOLOv11 detects hands and assigns them to ASL letter categories based on keypoints, ensuring accurate real-time recognition.
Figure 2 illustrates the step-by-step process of YOLO, including grid cell division, bounding box regression, classification with confidence scores, and final object detection.

2.3. MediaPipe

MediaPipe is an open-source framework developed by Google for building cross-platform, high-performance machine learning (ML) pipelines for real-time applications. It provides pre-built solutions for various computer vision tasks, including hand tracking, face detection, object tracking, pose estimation, and gesture recognition [18].
The MediaPipe Hand Tracking system employs a two-stage ML pipeline for precise and efficient hand detection and tracking.

2.3.1. BlazePalm Detector

BlazePalm is an optimized deep learning model for real-time palm detection. As a single-shot detector, it processes an entire image in one pass, enabling the fast and efficient localization of hands. It generates a robust bounding box around the detected hand, even in cases of partial occlusion, ensuring stable tracking for applications like gesture recognition and sign language interpretation [19,20]. Its speed and accuracy make it ideal for mobile and embedded systems.

2.3.2. Hand Landmark Model

The Hand Landmark Model is a lightweight convolutional neural network (CNN) trained on diverse datasets to detect 21 keypoints across the hand, including the fingers and wrist. It outputs (x, y, z) coordinates for each keypoint, where z provides depth estimation. This enables 3D gesture analysis [21].

2.3.3. Keypoint Representation

The 21 detected keypoints are stored in a normalized form [22] using the following representation:
K i = ( x i , y i ) , for i [ 0 , 20 ]
where
  • K i represents the i-th keypoint.
  • x i , y i are the normalized coordinates within the range [ 0 , 1 ] .
This normalization ensures consistency across different image resolutions and scales, making the keypoints adaptable for various machine learning applications [22,23].
Table 1 shows the keypoint IDs and the hand part associated with each ID.
Figure 3 illustrates the 21 keypoints tracked by MediaPipe’s hand tracking model, capturing the articulation of the fingers and wrist for gesture recognition and 3D hand analysis.
By integrating YOLOv11 with MediaPipe Hands, this study enhances real-time ASL alphabet recognition, improving both its accuracy and its computational efficiency.

3. Related Work

Research on real-time American Sign Language (ASL) interpretation has advanced significantly with deep learning, particularly through object detection frameworks such as YOLO. Recent studies demonstrate that high accuracy can be achieved for static hand gesture recognition (e.g., fingerspelling alphabets) in real time. However, challenges persist in continuous sign language recognition. Below, we summarize key studies, their methodologies, and performance metrics, followed by a comparison table.
Early deep learning approaches applied object detection models for classifying hand gestures. A study by [24] trained a YOLOv5 model on the MU Hand Images ASL dataset (2425 images) and achieved a recall of 97% and a precision of 95%, resulting in an F1-score of approximately 96%. The model also attained a mean Average Precision (mAP) of 98% at IoU 0.5 (mAP@0.5). However, this dataset featured only one signer, limiting the model’s generalizability and reducing its practical applicability in diverse user scenarios.
Expanding upon this, [25] utilized a larger custom dataset of 8000 images (covering 40 classes, including letters and numbers) to develop a YOLOv4-based recognizer. Their model achieved an mAP@0.5 of 98.0%, with a recall of approximately 96% and an F1-score of 96%. They reported real-time performance at 28.9 FPS. However, their evaluation was conducted using pre-recorded video rather than live webcam integration, limiting the assessment of its robustness in dynamic, real-world environments.
More recent advancements in YOLO variants have further improved recognition accuracy. The study by [26] introduced a hybrid approach that combined YOLOv6 for static signs and an LSTM for dynamic signs. Their YOLOv6 model, trained on a custom-collected dataset, achieved a test accuracy of about 95%, with a precision of 96% and a recall of 92% for ASL alphabet gestures (mAP 96%). This outperformed their sequential-sign model, where an LSTM applied to continuous sign sequences achieved an accuracy of 88–92%, underscoring that static gesture recognition is currently more advanced than continuous sign recognition.
The study by [27] explored the application of YOLOv9 in real-time ASL interpretation. Their study utilized a dataset of 1334 images for training, validation, and testing. Despite achieving high accuracy across precision and recall metrics, the limited dataset size presents challenges for building a reliable real-time interpretation system.
In a more recent study, ref. [28] further refined sign language detection with YOLOv8, focusing on the Indonesian Sign Language System (SIBI). Their study trained YOLOv8 on video and image datasets, employing extensive preprocessing and data mining techniques to improve model efficiency. The model, optimized for 168 layers with over 11 million parameters, achieved an average inference speed of 4.6 ms per image. Detection results demonstrated exceptionally high success rates, particularly for the letters D, F, N, O, and Q, which achieved up to 96% accuracy. Overall, 22 out of 26 letters displayed “excellent” detection results (above 90%), while four letters (H, M, T, Z) showed “good” detection results (86–89%). The study highlights the potential of YOLOv8 in enhancing real-time sign language recognition.
Despite these advancements, several research gaps remain. Many studies focus solely on isolated signs (individual letters or digits) using relatively uniform image datasets, often featuring a single signer in a controlled background. While achieving near-perfect accuracy in such settings is impressive, it does not necessarily translate to real-world applications involving multiple users, varying lighting conditions, and continuous signing. Performance significantly declines when extending to word-level or sentence-level ASL, which involves gesture transitions and motion. For instance, the LSTM-based continuous sign recognition model in [26] attained only 88% accuracy, highlighting the difficulty of fluid ASL recognition. Additionally, the lack of a standardized large-scale benchmark dataset for continuous ASL gestures impedes direct comparisons among studies, as most researchers curate their own datasets. Some models also require on-screen hands or specific camera setups, making it challenging to achieve comparable accuracy with arbitrary video feeds in diverse settings.
To address these limitations, we propose a novel YOLOv11-based real-time ASL recognition system integrated with MediaPipe keypoint tracking. This approach enhances precision and robustness, overcoming challenges faced in previous studies. Unlike prior works that relied on small or single-signer datasets, our model is trained on a large-scale dataset of 130,000 images, each annotated with 21 keypoints. This ensures improved generalization across diverse signers, hand shapes, and lighting conditions. Moving beyond static gesture recognition, our system facilitates real-time ASL translation, allowing users to interactively spell names, locations, and simple phrases with high accuracy. The proposed model achieves an mAP@0.5 of 98.2% with an inference speed of 1.3 ms per image, making it highly suitable for real-world applications. By integrating deep learning with keypoint tracking, our system advances AI-driven assistive technologies, bridging the gap between research and practical ASL interpretation tools. This work sets the stage for more robust, real-time sign language translation in diverse environments.
Table 2 below provides a comparative summary of representative studies, detailing real-time capability, dataset size, and key performance metrics. All listed works focus on the deep-learning-based recognition of ASL alphabet gestures and report performance on static gesture classification

4. Methodology

This study aims to develop a real-time American Sign Language (ASL) alphabet recognition system that accurately identifies and translates ASL hand gestures into text, enabling users to spell names and locations interactively. The proposed approach integrates YOLOv11 for gesture detection and MediaPipe for hand tracking to enhance precision and real-time responsiveness.
A major challenge in ASL recognition systems lies in distinguishing visually similar gestures, such as “A” and “T” or “M” and “N”, which often leads to misclassifications. Additionally, the dataset quality presents significant obstacles, including poor image resolution, motion blur, inconsistent lighting, and variations in hand sizes, skin tones, and backgrounds. These factors introduce bias and reduce the model’s ability to generalize across different users and environments [29].
To address these challenges, a large-scale dataset was curated, incorporating custom keypoint-based annotations to enhance the model’s ability to detect subtle variations in finger positioning. This ensures improved accuracy, robustness against variations in hand structure, and better adaptability to real-world conditions.
Figure 4 presents an overview of the data processing pipeline for the proposed real-time ASL recognition system. It illustrates the sequential stages, including data preprocessing, augmentation, annotation, and model training, leading to the final real-time recognition system.

4.1. Dataset Description

The ASL Alphabet Hand Gesture Dataset is a comprehensive collection of hand gesture images designed to train deep learning models for real-time ASL recognition. This dataset enhances computer vision models by enabling the precise classification of ASL alphabet gestures, thereby supporting real-time sign language interpretation systems.
Comprising 130,000 images, the dataset captures diverse hand gestures under varying conditions to improve model generalization. These conditions include
  • Different lighting environments: bright, dim, and shadowed.
  • Various backgrounds: natural outdoor settings and indoor scenes.
  • Multiple angles and hand orientations for robustness.
Each image is annotated with 21 keypoints, accurately marking critical hand structures such as fingertips, knuckles, and the wrist (see Figure 5). These annotations provide a skeletal representation of the hand, enabling the model to distinguish between similar gestures with greater precision.
For optimal training and evaluation, the dataset is systematically partitioned as follows:
  • Training set: 80%.
  • Validation set: 10%.
  • Testing set: 10%.
This structured division ensures that the trained models achieve high performance in real-world ASL recognition applications, making the dataset a valuable resource for AI-driven sign language accessibility solutions.

4.2. Data Preprocessing

To standardize the dataset and enhance model performance, several preprocessing techniques were applied:
  • Auto-orientation: Ensuring the correct alignment of images regardless of the original orientation.
  • Resizing: Standardizing all images to 640 × 640 pixels for consistency.
  • Contrast adjustment: Enhancing visual clarity using contrast stretching.
  • Grayscale conversion: Applied to 10% of images to improve robustness to color variations.

4.3. Data Augmentation

To enhance generalization and mitigate overfitting, 30% of the dataset underwent augmentation with the following transformations:
  • Flipping: Horizontal and vertical flips to introduce diverse orientations.
  • Rotation: Random variations between −10° and +10° to simulate natural hand positioning.
  • Brightness adjustment: Random modifications within ±15% to accommodate lighting variations.
  • Grayscale conversion: Selectively applied to certain images to enhance the model’s focus on gesture structure.
These augmentations significantly improve the model’s ability to recognize ASL gestures across different environmental conditions, ensuring higher accuracy and reliability in real-time applications.

4.4. Data Annotation

To optimize YOLO for ASL hand gesture recognition, images were annotated using 21 keypoints extracted from hand landmarks. Unlike traditional bounding boxes that enclose the entire hand without precisely identifying its central structure, this approach prioritizes the 21 keypoints detected by MediaPipe. From these keypoints, a bounding box is computed based on their spatial distribution, ensuring it accurately encapsulates the gesture while preserving crucial articulation details. By leveraging this method, the annotation process enhances precision, as it focuses on the actual structure of the hand rather than just its outer boundaries, leading to more reliable recognition and classification.
The annotation process was automated and underwent the following steps:
  • Hand Landmark Detection:
    • MediaPipe extracted 21 keypoints corresponding to joints and finger positions of the hand.
    • The detected keypoints were scaled and normalized to maintain consistency across varying image resolutions.
  • Bounding Box Calculation:
    A bounding box was generated to enclose only the 21 keypoints that were previously captured on the hand by determining the minimum and maximum coordinates from the extracted keypoints.

4.4.1. Bounding Box Computation

Since YOLO requires object detection labels in the format (x_center, y_center, width, height), the bounding box was calculated using the following mathematical expressions:

Finding the Bounding Box Extent

The bounding box must enclose all 21 detected keypoints. To determine its position, we compute the top-left corner ( x min , y min ) and bottom-right corner ( x max , y max ) :
x min = min ( x 0 , x 1 , , x 20 ) , y min = min ( y 0 , y 1 , , y 20 )
x max = max ( x 0 , x 1 , , x 20 ) , y max = max ( y 0 , y 1 , , y 20 )
Here, x min and y min represent the smallest x and y values among all detected keypoints, defining the top-left corner of the bounding box. Similarly, x max and y max represent the largest x and y values, defining the bottom-right corner.

Computing Bounding Box Dimensions

The width and height of the bounding box are calculated as
width = x max x min , height = y max y min
These values represent the total span of the hand in the x and y directions.

Determining the Bounding Box Center

Since YOLO requires bounding box coordinates in a center-based format, the center of the bounding box is computed as
x center = x min + x max 2 , y center = y min + y max 2
This ensures the bounding box is properly centered around the detected hand.

Normalization for YOLO Format

Given that keypoints are already normalized between 0 and 1 relative to image dimensions, the bounding box coordinates must also be constrained within the same range:
x center , y center , width , height [ 0 , 1 ]
This prevents values from exceeding the image boundaries, ensuring consistency across different image resolutions.

4.4.2. Annotation Format for YOLO

Each detected hand was assigned a class label corresponding to the ASL letter and formatted according to YOLO’s labeling convention:
  • Class Label: An integer (0–25) representing the ASL letter.
  • 21 Keypoint Coordinates: Normalized values stored as
    K i = ( x i , y i ) , i [ 0 , 20 ]
    where K i represents the i-th keypoint, and x i , y i are in the range [0, 1].
  • Bounding Box: Stored in the YOLO format as
    ( x center , y center , width , height )
Figure 6 illustrates the effectiveness of the annotation process during batch processing. It shows bounding boxes that focus solely on the 21 keypoints, ensuring precise localization.
This structured annotation approach enables YOLO to detect and classify ASL gestures with greater precision, leading to a fine-tuned, real-time ASL recognition system optimized for robust gesture classification.

4.4.3. Data Balancing

Balancing the dataset is essential for training YOLOv11 to detect and interpret the American Sign Language (ASL) alphabet in real time. Without balance, the model favors more frequent classes, leading to bias where certain letters are detected with high accuracy while others are misclassified or ignored. This results in unreliable performance, making the system unsuitable for users relying on ASL [30].
An imbalanced dataset also affects YOLOv11’s detection confidence. If certain letters appear more often, the model assigns higher confidence to them while lowering confidence for less frequent ones, leading to poor recall and inconsistent real-time detection. Balancing the dataset ensures uniform exposure, resulting in stable confidence scores and improved accuracy across all classes.
Overfitting is another critical issue. If the model encounters significantly more samples of some letters, it becomes biased toward those while failing to learn the characteristics of rare ones. This weakens its ability to recognize less frequent letters, reducing overall effectiveness [31].
To address these challenges, the upsampling technique has been utilized through data augmentation to balance the dataset to 5000 images per class. This ensures equal learning across all ASL letters, prevents bias, enhances generalization, avoids overfitting, and improves real-world performance. A well-balanced dataset is crucial for building a reliable, inclusive, and effective real-time ASL interpretation system.
Figure 7 illustrates the dataset’s distribution and the spatial characteristics of the bounding boxes, offering insight into the overall structure and organization of the data.
Top-left (Bar Chart): This plot confirms that the dataset is balanced, with each of the 26 ASL letters having an equal number of instances (5000 images per class). The uniform distribution ensures that no letter is overrepresented or underrepresented, preventing model bias.
Top-right (Bounding Box Overlays): This visualization displays multiple bounding boxes superimposed on each other. The tight clustering of bounding boxes suggests that the hand positions across different images are well aligned and consistently annotated, reducing variations that could impact the model’s learning.
Bottom-left (Heatmap of x-y Coordinates): This heatmap shows the spatial distribution of the hand’s center positions in normalized coordinates. The density around the center suggests that most hand gestures are captured within a specific region of the frame, making it easier for YOLOv11 to detect and recognize them accurately.
Bottom-right (Heatmap of Width–Height Distribution): This plot highlights the distribution of bounding box sizes. The concentration of values in a specific region indicates that the hands in the dataset mostly follow a uniform scale, ensuring that the model generalizes well without being biased toward specific sizes.
This visualization confirms that the dataset is structured optimally for YOLOv11 training, ensuring consistency across annotations and reducing variability in hand positioning and bounding box dimensions.

5. Results and Discussion

5.1. Model Performance and Evaluation Metrics

The performance evaluation of the proposed American Sign Language (ASL) alphabet recognition system was conducted using standard multiclass classification metrics, including precision, recall, F1-score, and mean Average Precision (mAP). The system is powered by the trained YOLO11 model, a deep learning architecture with 181 layers and 2,594,910 parameters, optimized for efficient real-time inference. With a computational complexity of 6.5 GFLOPs (Giga Floating Point Operations per Second) per inference, the model was rigorously tested on a balanced dataset of ASL alphabet hand gesture images. The training was conducted with 300 epochs, a batch size of 16, and an input image size of 640 × 640. The AdamW optimizer was employed to enhance training stability.
The chosen configuration ensures the model’s reliability, accuracy, and scalability for real-world ASL translation applications.
The model’s performance is summarized in Table 3.
These results demonstrate the model’s strong ability to accurately detect and classify ASL hand gestures with a high mean Average Precision (mAP@0.5) of 98.2%, confirming its effectiveness in a real-time sign language interpretation system.

5.2. Precision–Recall Analysis in Multiclass Classification

The precision–recall (PR) curve provides insight into how the model balances precision and recall across different confidence thresholds for a multiclass classification setting [32].
In a multiclass classification setting, the precision and recall for each class c are computed as
Precision c = T P c T P c + F P c
Recall c = T P c T P c + F N c
where
  • T P c (True Positives): Correctly classified instances of class c.
  • F P c (False Positives): Instances incorrectly classified as class c.
  • F N c (False Negatives): Instances of class c that were misclassified.
From Figure 8, we observe that precision remains above 0.98 across most recall values, confirming minimal false positive detections. The slight degradation in precision at extreme recall values suggests that the model maintains a conservative detection strategy.

5.3. Confidence-Based Evaluations for Multiclass Classification

To further analyze model performance, we examine the recall–confidence, precision–confidence, and F1–confidence curves [33].

5.3.1. Recall–Confidence Curve

In a multiclass classification scenario, the recall–confidence curve illustrates how recall changes as the confidence threshold τ increases [34]. The global recall at a given confidence level τ is computed as
Recall ( τ ) = c = 1 N T P c ( τ ) c = 1 N T P c ( τ ) + F N c ( τ )
As seen in Figure 9, recall remains at its highest value (close to 0.99) for lower confidence thresholds, ensuring that most detections are retained. However, as the confidence threshold surpasses τ = 0.8 , recall starts to decline sharply. This behavior indicates a trade-off where the model prioritizes precision over recall, filtering out lower-confidence predictions to improve accuracy while reducing the number of retrieved positive instances. The sharp drop in recall near τ = 0.99 suggests that only the highest-confidence predictions are retained, which can be useful for applications requiring high precision such ASL interpretation systems.

5.3.2. Precision–Confidence Curve

The global precision at confidence τ is computed as
Precision ( τ ) = c = 1 N T P c ( τ ) c = 1 N T P c ( τ ) + F P c ( τ )
From Figure 10, precision remains above 0.90 across nearly all confidence levels. The model achieves perfect precision at confidence τ = 0.985 , indicating that high-confidence detections are nearly always correct.

5.3.3. F1–Confidence Curve

The F1-score at confidence τ balances precision and recall [35]:
F 1 ( τ ) = 2 × Precision ( τ ) × Recall ( τ ) Precision ( τ ) + Recall ( τ )
Figure 11 shows the optimal F1-score occurs at τ = 0.99 , suggesting the best trade-off between detection accuracy and reliability.

5.4. Confusion Matrix Analysis

The confusion matrix provides a class-wise breakdown of model performance [36].
For a multiclass, the confusion matrix element at ( i , j ) represents
C M i j = P ( y pred = j y true = i )
where y pred and y true are the predicted and actual labels, respectively. Refer to Figure 12 for a visualization.
Key findings:
  • The model demonstrates high classification accuracy with minimal misclassifications.
  • Class-wise performance remains consistent, ensuring robustness across all ASL alphabet classes.
  • Confusion is minimal, confirming the effectiveness of the YOLO11n architecture in distinguishing ASL gestures.
These findings confirm the model’s strong classification ability, making it a promising solution for real-time ASL translation applications.

5.5. Evaluation of a Real-Time ASL Interpretation System

Evaluating a real-time American Sign Language (ASL) interpretation system for users to spell names and locations requires a comprehensive assessment of key performance aspects, including accuracy, latency, robustness, and usability. To conduct this evaluation, we utilized a laptop’s built-in webcam for real-time hand tracking and ASL recognition.
In our system, the webcam serves as a non-contact vision-based sensor that captures gesture data in the form of image frames. These frames are processed using MediaPipe to extract 21 hand keypoints and passed to YOLOv11 for ASL gesture classification. This sensor-driven pipeline enables real-time processing under various lighting and background conditions using only standard hardware.

5.5.1. System Specifications

The system was tested on a laptop with the following specifications:
  • Operating Environment: Python 3.10.11 and PyTorch 2.2.1.
  • Hardware: 12th Gen Intel Core i7-1260P CPU.

5.5.2. Performance Analysis

Our analysis of image processing per frame revealed the following time distribution:
  • Preprocessing: 1.1 ms.
  • Inference (Model Prediction): 70.9 ms.
  • Loss Calculation: 0.0 ms.
  • Postprocessing: 0.5 ms.

5.5.3. Robustness Across Different Conditions

To further evaluate the system’s effectiveness, we tested it under various conditions, including different hand shapes, skin tones, and background complexities. The system demonstrated strong generalization capabilities, accurately recognizing these variations with minimal error.
This robustness can be attributed to the integration of MediaPipe, which focuses solely on detecting 21 keypoints of the hand. Once the hand shape is determined, this information is passed to YOLOv11 for the final recognition process. Despite the two-step approach—first detecting the hand shape and keypoints and then processing the recognition—our model remains fast and highly efficient in detecting hand gestures and translating them into text.

5.5.4. ASL Translation to Names and Locations

Figure 13 and Figure 14 illustrate the system’s ability to interpret ASL gestures into meaningful text.
These results highlight the system’s efficiency in real-time ASL interpretation, ensuring minimal latency while maintaining high accuracy across diverse conditions.

6. Conclusions

The development of assistive technologies for the Deaf and Hard-of-Hearing community is crucial in fostering inclusivity and eliminating communication barriers. This study presents a real-time ASL interpretation system that effectively integrates deep learning with keypoint tracking, demonstrating high accuracy and efficiency in recognizing ASL alphabet gestures. By leveraging YOLOv11 for gesture detection and MediaPipe for precise hand tracking, the proposed system ensures robust performance across diverse conditions.
Experimental results validate the system’s effectiveness, achieving a mean Average Precision (mAP@0.5) of 98.2% with minimal latency, making it highly suitable for real-time applications. The ability to translate ASL gestures into text in real time significantly enhances accessibility in various domains, including education, healthcare, and professional settings. Future work will focus on extending the system’s capabilities beyond individual letter recognition to full ASL sentence interpretation, enabling more natural and fluid communication. This will require advancements in sequence modeling techniques, such as recurrent neural networks (RNNs), transformers, or temporal convolutional networks (TCNs), to effectively capture contextual dependencies between consecutive gestures.
In conclusion, this research underscores the transformative potential of AI-driven assistive technologies in empowering the Deaf community. By bridging the communication gap through real-time ASL recognition, this system contributes to a more inclusive society where individuals with hearing impairments can interact seamlessly with the world around them, whether introducing themselves or indicating locations.

Author Contributions

Conceptualization, B.A.; methodology, B.A.; software, B.A.; validation, B.A., E.A. and A.I.; formal analysis, B.A.; investigation, E.A. and A.I.; resources, B.A. and E.A.; data curation, B.A.; writing—original draft preparation, B.A.; writing—review and editing, E.A. and A.I.; visualization, B.A.; supervision, I.M. and M.I. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are available upon request.

Conflicts of Interest

Author Bader Alsharif is employed by the Technical and Vocational Training Corporation. The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Papatsimouli, M.; Sarigiannidis, P.; Fragulis, G.F. A survey of advancements in real-time sign language translators: Integration with IoT technology. Technologies 2023, 11, 83. [Google Scholar] [CrossRef]
  2. Alsharif, B.; Ilyas, M. Internet of things technologies in healthcare for people with hearing impairments. In IoT and Big Data Technologies for Health Care; Springer: Cham, Switzerland, 2022; pp. 299–308. [Google Scholar]
  3. Cui, R.; Liu, H.; Zhang, C. Recurrent Convolutional Neural Networks for Continuous Sign Language Recognition by Staged Optimization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 7361–7369. [Google Scholar]
  4. Rautaray, S.S.; Agrawal, A. Vision based hand gesture recognition for human computer interaction: A survey. Artif. Intell. Rev. 2015, 43, 1–54. [Google Scholar] [CrossRef]
  5. ZainEldin, H.; Gamel, S.A.; Talaat, F.M.; Aljohani, M.; Baghdadi, N.A.; Malki, A.; Badawy, M.; Elhosseini, M.A. Silent no more: A comprehensive review of artificial intelligence, deep learning, and machine learning in facilitating Deaf and mute communication. Artif. Intell. Rev. 2024, 57, 188. [Google Scholar] [CrossRef]
  6. Cheok, M.J.; Omar, Z.; Jaward, M.H. A review of hand gesture and sign language recognition techniques. Int. J. Mach. Learn. Cybern. 2019, 10, 131–153. [Google Scholar] [CrossRef]
  7. Zdravkova, K.; Krasniqi, V.; Dalipi, F.; Ferati, M. Cutting-edge communication and learning assistive technologies for disabled children: An artificial intelligence perspective. Front. Artif. Intell. 2022, 5, 970430. [Google Scholar] [CrossRef]
  8. Guo, J.; Liu, Y.; Lin, L.; Li, S.; Cai, J.; Chen, J.; Huang, W.; Lin, Y.; Xu, J. Chromatic plasmonic polarizer-based synapse for all-optical convolutional neural network. Nano Lett. 2023, 23, 9651–9656. [Google Scholar] [CrossRef]
  9. Alsharif, B.; Alanazi, M.; Altaher, A.S.; Altaher, A.; Ilyas, M. Deep Learning Technology to Recognize American Sign Language Alphabet Using Mulit-Focus Image Fusion Technique. In Proceedings of the 2023 IEEE 20th International Conference on Smart Communities: Improving Quality of Life Using AI, Robotics and IoT (HONET), Boca Raton, FL, USA, 4–6 December 2023; pp. 1–6. [Google Scholar]
  10. Ali, M.L.; Zhang, Z. The YOLO framework: A comprehensive review of evolution, applications, and benchmarks in object detection. Computers 2024, 13, 336. [Google Scholar] [CrossRef]
  11. Alaftekin, M.; Pacal, I.; Cicek, K. Real-time sign language recognition based on YOLO algorithm. Neural Comput. Appl. 2024, 36, 7609–7624. [Google Scholar] [CrossRef]
  12. Saxena, S.; Paygude, A.; Jain, P.; Memon, A.; Naik, V. Hand Gesture Recognition using YOLO Models for Hearing and Speech Impaired People. In Proceedings of the 2022 IEEE Students Conference on Engineering and Systems (SCES), Prayagraj, India, 1–3 July 2022; pp. 1–6. [Google Scholar] [CrossRef]
  13. Khanam, R.; Hussain, M. Yolov11: An overview of the key architectural enhancements. arXiv 2024, arXiv:2410.17725. [Google Scholar]
  14. Ye, T.; Huang, S.; Qin, W.; Tu, H.; Zhang, P.; Wang, Y.; Gao, C.; Gong, Y. YOLO-FIX: Improved YOLOv11 with Attention and Multi-Scale Feature Fusion for Detecting Glue Line Defects on Mobile Phone Frames. Electronics 2025, 14, 927. [Google Scholar] [CrossRef]
  15. Hoermann, S.; Henzler, P.; Bach, M.; Dietmayer, K. Object detection on dynamic occupancy grid maps using deep learning and automatic label generation. In Proceedings of the 2018 IEEE Intelligent Vehicles Symposium (IV), Suzhou, China, 26–30 June 2018; pp. 826–833. [Google Scholar]
  16. Li, Y.; Wang, T.; Kang, B.; Tang, S.; Wang, C.; Li, J.; Feng, J. Overcoming classifier imbalance for long-tail object detection with balanced group softmax. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10991–11000. [Google Scholar]
  17. Dobranskỳ, M. Object Detection for Video Surveillance Using the SSD Approach. Master’s Thesis, Technical University of Košice, Košice, Slovakia, 2019. [Google Scholar]
  18. Rajendran, A.K.; Sethuraman, S.C. A survey on yogic posture recognition. IEEE Access 2023, 11, 11183–11223. [Google Scholar] [CrossRef]
  19. Pinochet, D. Computational Gestural Making: A Framework for Exploring the Creative Potential of Gestures, Materials, and Computational Tools; Massachusetts Institute of Technology: Cambridge, MA, USA, 2023. [Google Scholar]
  20. Noroozi, F.; Corneanu, C.A.; Kamińska, D.; Sapiński, T.; Escalera, S.; Anbarjafari, G. Survey on emotional body gesture recognition. IEEE Trans. Affect. Comput. 2018, 12, 505–523. [Google Scholar]
  21. Zhang, H.; Tian, Y.; Zhang, Y.; Li, M.; An, L.; Sun, Z.; Liu, Y. Pymaf-x: Towards well-aligned full-body model regression from monocular images. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 12287–12303. [Google Scholar]
  22. Zhao, L.; Li, X.; Zhuang, Y.; Wang, J. Deeply-learned part-aligned representations for person re-identification. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 3219–3228. [Google Scholar]
  23. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  24. Dima, T.F.; Ahmed, M.E. Using YOLOv5 algorithm to detect and recognize American sign language. In Proceedings of the 2021 International Conference on Information Technology (ICIT), Amman, Jordan, 14–15 July 2021; pp. 603–607. [Google Scholar]
  25. Al-Shaheen, A.; Çevik, M.; Alqaraghulı, A. American sign language recognition using yolov4 method. Int. J. Multidiscip. Stud. Innov. Technol. 2022, 6, 61–65. [Google Scholar]
  26. Buttar, A.M.; Ahmad, U.; Gumaei, A.H.; Assiri, A.; Akbar, M.A.; Alkhamees, B.F. Deep learning in sign language recognition: A hybrid approach for the recognition of static and dynamic signs. Mathematics 2023, 11, 3729. [Google Scholar] [CrossRef]
  27. Imran, A.; Hulikal, M.S.; Gardi, H.A. Real Time American Sign Language Detection Using Yolo-v9. arXiv 2024, arXiv:2407.17950. [Google Scholar]
  28. Kurniawan, R.; Wijaya, Y.A. YOLOv8 Algorithm to Improve the Sign Language Letter Detection System Model. J. Artif. Intell. Eng. Appl. (JAIEA) 2025, 4, 1379–1385. [Google Scholar]
  29. Alsharif, B.; Altaher, A.S.; Altaher, A.; Ilyas, M.; Alalwany, E. Deep learning technology to recognize american sign language alphabet. Sensors 2023, 23, 7970. [Google Scholar] [CrossRef]
  30. Alsharif, B.; Alanazi, M.; Ilyas, M. Machine Learning Technology to Recognize American Sign Language Alphabet. In Proceedings of the 2023 IEEE 20th International Conference on Smart Communities: Improving Quality of Life Using AI, Robotics and IoT (HONET), Boca Raton, FL, USA, 4–6 December 2023; pp. 173–178. [Google Scholar]
  31. Raschka, S. Model evaluation, model selection, and algorithm selection in machine learning. arXiv 2018, arXiv:1811.12808. [Google Scholar]
  32. Brownlee, J. Imbalanced Classification with Python: Better Metrics, Balance Skewed Classes, Cost-Sensitive Learning; Machine Learning Mastery: San Francisco, CA, USA, 2020. [Google Scholar]
  33. Sundaresan Geetha, A.; Alif, M.A.R.; Hussain, M.; Allen, P. Comparative Analysis of YOLOv8 and YOLOv10 in Vehicle Detection: Performance Metrics and Model Efficacy. Vehicles 2024, 6, 1364–1382. [Google Scholar] [CrossRef]
  34. Alalwany, E.; Alsharif, B.; Alotaibi, Y.; Alfahaid, A.; Mahgoub, I.; Ilyas, M. Stacking Ensemble Deep Learning for Real-Time Intrusion Detection in IoMT Environments. Sensors 2025, 25, 624. [Google Scholar] [CrossRef]
  35. Aldahoul, N.; Karim, H.A.; Sabri, A.Q.M.; Tan, M.J.T.; Momo, M.A.; Fermin, J.L. A comparison between various human detectors and CNN-based feature extractors for human activity recognition via aerial captured video sequences. IEEE Access 2022, 10, 63532–63553. [Google Scholar] [CrossRef]
  36. Naing, K.M.; Boonsang, S.; Chuwongin, S.; Kittichai, V.; Tongloy, T.; Prommongkol, S.; Dekumyoy, P.; Watthanakulpanich, D. Automatic recognition of parasitic products in stool examination using object detection approach. PeerJ Comput. Sci. 2022, 8, e1065. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Workflow of transfer learning methodology for ASL hand gesture detection using YOLO.
Figure 1. Workflow of transfer learning methodology for ASL hand gesture detection using YOLO.
Sensors 25 02138 g001
Figure 2. YOLO detection workflow for ASL hand gesture recognition.
Figure 2. YOLO detection workflow for ASL hand gesture recognition.
Sensors 25 02138 g002
Figure 3. Illustration of the 21 keypoints tracked by MediaPipe’s hand tracking model.
Figure 3. Illustration of the 21 keypoints tracked by MediaPipe’s hand tracking model.
Sensors 25 02138 g003
Figure 4. Workflow of the proposed ASL recognition model.
Figure 4. Workflow of the proposed ASL recognition model.
Sensors 25 02138 g004
Figure 5. Annotated ASL Alphabet Hand Gestures with 21 keypoints for model training.
Figure 5. Annotated ASL Alphabet Hand Gestures with 21 keypoints for model training.
Sensors 25 02138 g005
Figure 6. Effectiveness of annotation in batch processing with precise keypoint localization.
Figure 6. Effectiveness of annotation in batch processing with precise keypoint localization.
Sensors 25 02138 g006
Figure 7. Overview of the dataset distribution and spatial properties of the bounding boxes.
Figure 7. Overview of the dataset distribution and spatial properties of the bounding boxes.
Sensors 25 02138 g007
Figure 8. Precision–recall curve for ASL recognition model.
Figure 8. Precision–recall curve for ASL recognition model.
Sensors 25 02138 g008
Figure 9. Recall–confidence curve.
Figure 9. Recall–confidence curve.
Sensors 25 02138 g009
Figure 10. Precision–confidence curve.
Figure 10. Precision–confidence curve.
Sensors 25 02138 g010
Figure 11. F1–confidence curve.
Figure 11. F1–confidence curve.
Sensors 25 02138 g011
Figure 12. Normalized confusion matrix for ASL alphabet recognition.
Figure 12. Normalized confusion matrix for ASL alphabet recognition.
Sensors 25 02138 g012
Figure 13. The model translating ASL into a name.
Figure 13. The model translating ASL into a name.
Sensors 25 02138 g013
Figure 14. The model translating ASL into a location.
Figure 14. The model translating ASL into a location.
Sensors 25 02138 g014
Table 1. MediaPipe hand keypoints and corresponding hand parts.
Table 1. MediaPipe hand keypoints and corresponding hand parts.
Keypoint ID(s)Hand Part
0Wrist
1, 2, 3, 4Thumb
5, 6, 7, 8Index Finger
9, 10, 11, 12Middle Finger
13, 14, 15, 16Ring Finger
17, 18, 19, 20Pinky Finger
Table 2. Comparison of our proposed approach with recent advancements in YOLO-based ASL recognition models, highlighting dataset size, performance metrics, and real-time capability.
Table 2. Comparison of our proposed approach with recent advancements in YOLO-based ASL recognition models, highlighting dataset size, performance metrics, and real-time capability.
RefYearModelDataset (Images)PrecisionRecallF1-ScoremAP@0.5Real-Time
[24]2021YOLOv5242595%97%96%98%No
[25]2022YOLOv4800096.00%96.00%96.00%98.01%Yes
[26]2023YOLOv6800096.00%96.00%96.00%96.22%Yes
[27]2024YOLOv9133496.83%92.96%-97.84%Yes
[28]2025YOLOv8Video-based dataset100%99%89%96%Yes
Our work2025YOLOv11130,00098.50%98.10%99.10%98.20%Yes
Table 3. Performance metrics of the ASL recognition model.
Table 3. Performance metrics of the ASL recognition model.
MetricValue
Precision (P)0.985
Recall (R)0.981
F1-Sore (F)0.991
mAP@0.50.982
mAP@0.5–0.950.933
Inference Speed1.3 ms per image
Post-processing Speed0.3 ms per image
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alsharif, B.; Alalwany, E.; Ibrahim, A.; Mahgoub, I.; Ilyas, M. Real-Time American Sign Language Interpretation Using Deep Learning and Keypoint Tracking. Sensors 2025, 25, 2138. https://doi.org/10.3390/s25072138

AMA Style

Alsharif B, Alalwany E, Ibrahim A, Mahgoub I, Ilyas M. Real-Time American Sign Language Interpretation Using Deep Learning and Keypoint Tracking. Sensors. 2025; 25(7):2138. https://doi.org/10.3390/s25072138

Chicago/Turabian Style

Alsharif, Bader, Easa Alalwany, Ali Ibrahim, Imad Mahgoub, and Mohammad Ilyas. 2025. "Real-Time American Sign Language Interpretation Using Deep Learning and Keypoint Tracking" Sensors 25, no. 7: 2138. https://doi.org/10.3390/s25072138

APA Style

Alsharif, B., Alalwany, E., Ibrahim, A., Mahgoub, I., & Ilyas, M. (2025). Real-Time American Sign Language Interpretation Using Deep Learning and Keypoint Tracking. Sensors, 25(7), 2138. https://doi.org/10.3390/s25072138

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop