Enhancing Hand Gesture Image Recognition by Integrating Various Feature Groups
Abstract
:1. Introduction
- It presents an effective technique for hand gesture recognition that combines the reliable ResNet50 model for feature extraction with Tamura features, which record global texture characteristics, to improve the recognition of gestures with the same shapes but different textures.
- The combination can improve recognition resilience and accuracy, especially in situations with varying gesture texture, image quality, or lighting.
- By enabling the model to distinguish motions based on both shape and small textural changes, the proposed method reduces misclassification errors, particularly when gestures have the same shapes but variable surface textures.
- A median filter is applied to the images to perform low-pass filtering, smoothing the images and reducing noise, thereby enhancing the reliability of the analysis.
- The Tamura–ResNet50–Optimized GAM framework improves overall accuracy and resilience in recognizing a variety of hand motions while reducing the time required for processing.
2. Related Works
3. Proposed Method
3.1. Image Preparation and Preprocessing
Algorithm 1: Image preparation and preprocessing for hand gesture recognition. |
Input: Unprocessed hand gesture images. |
Output: Preprocessed images that are prepared for feature extraction and classification stage. |
Begin |
For |
1. Import every image through the dataset in RGB or greyscale format. |
2. Transform the Unprocessed images to grayscale for simplified processing and concentrate on important details. |
3. Resize all images to a consistent dimension (for example, 224 × 224 pixels) to make sure that every image is compatible for the ResNet50 model. |
4. Normalize pixel values to an interval (usually [0, 1] or [−1, 1]) to enhance training convergence. |
5. Gaussian or median filters can be used to eliminate noise in images, removing unnecessary noise while maintaining important information. |
6. Store the preprocessed images that they can be used later for the feature extraction and classification stage. |
End for |
End |
3.1.1. Image Conversion
3.1.2. Resizing and Normalization
3.1.3. Noise Removal and Image Smoothing
3.2. Feature Extraction
Algorithm 2: Feature extraction for hand gesture recognition. |
Input: Preprocessed grayscale image I of size M × N. |
Output: 1 × 1 × 2048 Feature vector dimension |
Begin |
For |
1. Apply Tamura texture descriptor: |
1.1 For every Preprocessed grayscale image, Calculate the Tamura features (coarseness, contrast, and directionality). |
1.2 Store the generated Tamura feature vector with 1 × 3 dimension. |
2: Combine Tamura Features with Image Data using the following: |
2.1 Create a 1D vector by flattening the Tamura features. |
2.2 To create a more comprehensive feature collection, immediately combine the Tamura features with the image data that has been flattened. |
2.3 Save the combined texture-based and pixel-based data that has been generated. |
3. Apply ResNet50 as a Feature Extractor: |
3.1 Load the pre-trained ResNet50 model |
3.1.1 Import the pre-trained ResNet50 model from a library. |
3.1.2 To concentrate on feature extraction, remove the last classification layer. |
3.2 Freeze the first several layers of ResNet50 to maintain the pre-trained features. |
3.3 Preprocessed the combined data by resizing it to 224 × 224 pixels and normalizing it (typically within 0 and 1). |
3.4 Feed the preprocessed data over the ResNet50 model to extract deep features. |
3.5 The feature can be extracted from the final pooling layer (average pooling 5). |
3.6 The features map is flattened to an array of one dimensions. |
4. Store the extracted feature vector in 1 × 1 × 2048 dimensions. |
End for |
End |
3.2.1. Tamura Feature Extraction
3.2.2. Combine Tamura Features with Image Data
3.2.3. ResNet50 Model as a Feature Extractor
3.3. Optimizing Classification Using Feature Selection
Algorithm 3: GAM optimization based on Sequential Feature Selection (SFS). |
Input: 1 × 1 × 2048 Feature vector dimension |
Output: Selected feature subset (Fselected) and Performance measurements. |
Begin |
For |
1. Load the ResNet50 extracted 2048-dimensional features vector F = {f1, f2, …, f2048}. |
2. For all feature vectors, Apply Sequential Feature Selection (SFS) method. |
2.1 Begin with an empty set of features Fselected = ∅. |
2.2 Implement feature selection iteratively: |
2.2.1 Add or remove features from Fselected. |
2.2.2 Use cross-validation accuracy as a measure to assess performance. |
2.3 Repeat rounds until performance is improved by no more feature additions or deletions. |
2.4 Save the final selected features vector (Fselected). |
3. Train a Generalized Additive Model (GAM) classifier |
3.1 Train a Generalized Additive Model (GAM) classifier on the Fselected. |
3.2 Record the nonlinear connections between specific features and the classes of hand gestures that correspond to them. |
4. Evaluate the GAM Model |
4.1 Use the test dataset to test the trained GAM classifier. |
4.2 Calculate the F1-score, recall, accuracy, and precision as indicators of performance. |
End for |
End |
4. Results and Discussion
- HP laptop;
- Intel(R) Core(TM) i7-6500U CPU @ 2.50 GHz 2.60 GHz;
- 8 GB Memory;
- 2 GB GPU Memory;
- MATLAB version R2020a;
- Windows 10 Pro 64Bit Edition;
- Intel® HD Graphics 520 (NVIDIA GTX 950M).
4.1. Dataset Description
4.2. Performance Evaluation Metric
4.3. Evaluation Results
4.4. Multi-Class Hand Gesture Classification Results
4.5. Results of Comparing Existing Methods
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Wu, C.-H.; Chen, W.-L.; Lin, C.H. Depth-based hand gesture recognition. Multimed. Tools Appl. 2016, 75, 7065–7086. [Google Scholar] [CrossRef]
- Zhang, T.; Ding, Y.; Hu, C.; Zhang, M.; Zhu, W.; Bowen, C.R.; Han, Y.; Yang, Y. Self-Powered Stretchable Sensor Arrays Exhibiting Magnetoelasticity for Real-Time Human–Machine Interaction. Adv. Mater. 2023, 35, 2203786. [Google Scholar] [CrossRef]
- Al Farid, F.; Hashim, N.; Abdullah, J.; Bhuiyan, M.R.; Shahida Mohd Isa, W.N.; Uddin, J.; Haque, M.A.; Husen, M.N. A structured and methodological review on vision-based hand gesture recognition system. J. Imaging 2022, 8, 153. [Google Scholar] [CrossRef]
- Adeyanju, I.A.; Bello, O.O.; Adegboye, M.A. Machine learning methods for sign language recognition: A critical review and analysis. Intell. Syst. Appl. 2021, 12, 200056. [Google Scholar] [CrossRef]
- Qi, J.; Xu, K.; Ding, X. Approach to hand posture recognition based on hand shape features for human–robot interaction. Complex Intell. Syst. 2022, 8, 2825–2842. [Google Scholar] [CrossRef]
- Chakraborty, B.K.; Sarma, D.; Bhuyan, M.K.; MacDorman, K.F. Review of constraints on vision-based gesture recognition for human–computer interaction. IET Comput. Vis. 2018, 12, 3–15. [Google Scholar] [CrossRef]
- Guo, L.; Lu, Z.; Yao, L. Human-machine interaction sensing technology based on hand gesture recognition: A review. IEEE Trans. Hum.-Mach. Syst. 2021, 51, 300–309. [Google Scholar] [CrossRef]
- Hammad, B.T.; Jamil, N.; Rusli, M.E.; Z’Aba, M.R.; Ahmed, I.T. Implementation of Lightweight Cryptographic Primitives. J. Theor. Appl. Inf. Technol. 2017, 95, 5571–5586. [Google Scholar]
- Zahra, R.; Shehzadi, A.; Sharif, M.I.; Karim, A.; Azam, S.; De Boer, F.; Jonkman, M.; Mehmood, M. Camera-based interactive wall display using hand gesture recognition. Intell. Syst. Appl. 2023, 19, 200262. [Google Scholar] [CrossRef]
- Oudah, M.; Al-Naji, A.; Chahl, J. Hand gesture recognition based on computer vision: A review of techniques. J. Imaging 2020, 6, 73. [Google Scholar] [CrossRef]
- López, L.I.B.; Ferri, F.M.; Zea, J.; Caraguay, Á.L.V.; Benalcázar, M.E. CNN-LSTM and post-processing for EMG-based hand gesture recognition. Intell. Syst. Appl. 2024, 22, 200352. [Google Scholar]
- Gupta, B.; Shukla, P.; Mittal, A. K-nearest correlated neighbor classification for Indian sign language gesture recognition using feature fusion. In Proceedings of the 2016 International Conference on Computer Communication and Informatics (ICCCI), Coimbatore, India, 7–9 January 2016; pp. 1–5. [Google Scholar]
- Sadeddine, K.; Djeradi, R.; Chelali, F.Z.; Djeradi, A. Recognition of static hand gesture. In Proceedings of the 2018 6th International Conference on Multimedia Computing and Systems (ICMCS), Rabat, Morocco, 10–12 May 2018; pp. 1–6. [Google Scholar]
- Zhang, F.; Liu, Y.; Zou, C.; Wang, Y. Hand gesture recognition based on HOG-LBP feature. In Proceedings of the 2018 IEEE International Instrumentation and Measurement Technology Conference (I2MTC), Houston, TX, USA, 14–17 May 2018; pp. 1–6. [Google Scholar]
- Sahoo, J.P.; Ari, S.; Ghosh, D.K. Hand gesture recognition using DWT and F-ratio based feature descriptor. IET Image Process. 2018, 12, 1780–1787. [Google Scholar] [CrossRef]
- Ahmed, I.T.; Der, C.S.; Hammad, B.T. Recent Approaches on No-Reference Image Quality Assessment for Contrast Distortion Images with Multiscale Geometric Analysis Transforms: A Survey. J. Theor. Appl. Inf. Technol. 2017, 95, 561–569. [Google Scholar]
- Gajalakshmi, P.; Sharmila, T.S. Hand gesture recognition by histogram based kernel using density measure. In Proceedings of the 2019 2nd International Conference on Power and Embedded Drive Control (ICPEDC), Chennai, India, 21–23 August 2019; pp. 294–298. [Google Scholar]
- Li, J.; Li, C.; Han, J.; Shi, Y.; Bian, G.; Zhou, S. Robust hand gesture recognition using HOG-9ULBP features and SVM model. Electronics 2022, 11, 988. [Google Scholar] [CrossRef]
- Sayed, U.; Bakheet, S.; Mofaddel, M.A.; El-Zohry, Z. Robust Hand Gesture Recognition Using HOG Features and machine learning. Sohag J. Sci. 2024, 9, 226–233. [Google Scholar] [CrossRef]
- Nyirarugira, C.; Choi, H.-R.; Kim, J.; Hayes, M.; Kim, T. Modified levenshtein distance for real-time gesture recognition. In Proceedings of the 2013 6th International Congress on Image and Signal Processing (CISP), Hangzhou, China, 16–18 December 2013; Volume 2, pp. 974–979. [Google Scholar]
- Aurangzeb, K.; Javeed, K.; Alhussein, M.; Rida, I.; Haider, S.I.; Parashar, A. Deep Learning Approach for Hand Gesture Recognition: Applications in Deaf Communication and Healthcare. Comput. Mater. Contin. 2024, 78, 127. [Google Scholar] [CrossRef]
- Ozcan, T.; Basturk, A. Transfer learning-based convolutional neural networks with heuristic optimization for hand gesture recognition. Neural Comput. Appl. 2019, 31, 8955–8970. [Google Scholar] [CrossRef]
- Sahoo, J.P.; Prakash, A.J.; Pławiak, P.; Samantray, S. Real-time hand gesture recognition using fine-tuned convolutional neural network. Sensors 2022, 22, 706. [Google Scholar] [CrossRef]
- Mujahid, A.; Awan, M.J.; Yasin, A.; Mohammed, M.A.; Damaševičius, R.; Maskeliūnas, R.; Abdulkareem, K.H. Real-time hand gesture recognition based on deep learning YOLOv3 model. Appl. Sci. 2021, 11, 4164. [Google Scholar] [CrossRef]
- Ewe, E.L.R.; Lee, C.P.; Kwek, L.C.; Lim, K.M. Hand gesture recognition via lightweight VGG16 and ensemble classifier. Appl. Sci. 2022, 12, 7643. [Google Scholar] [CrossRef]
- Wang, F.; Hu, R.; Jin, Y. Research on gesture image recognition method based on transfer learning. Procedia Comput. Sci. 2021, 187, 140–145. [Google Scholar] [CrossRef]
- Kika, A.; Koni, A. Hand Gesture Recognition Using Convolutional Neural Network and Histogram of Oriented Gradients Features. In Proceedings of the 3rd International Conference on Recent Trends and Applications in Computer Science and Information Technology (RTA-CSIT 2018), Tirana, Albania, 23–24 November 2018; pp. 75–79. [Google Scholar]
- Rastgoo, R.; Kiani, K.; Escalera, S. Multi-modal deep hand sign language recognition in still images using restricted Boltzmann machine. Entropy 2018, 20, 809. [Google Scholar] [CrossRef]
- Damaneh, M.M.; Mohanna, F.; Jafari, P. Static hand gesture recognition in sign language based on convolutional neural network with feature extraction method using ORB descriptor and Gabor filter. Expert Syst. Appl. 2023, 211, 118559. [Google Scholar] [CrossRef]
- Ahmed, I.T.; Der, C.S.; Hammad, B.T.; Jamil, N. Contrast-distorted image quality assessment based on curvelet domain features. Int. J. Electr. Comput. Eng. 2021, 11, 2595. [Google Scholar] [CrossRef]
- El-Shafai, W.; Almomani, I.; AlKhayer, A. Visualized malware multi-classification framework using fine-tuned CNN-based transfer learning models. Appl. Sci. 2021, 11, 6446. [Google Scholar] [CrossRef]
- Tamura, H.; Mori, S.; Yamawaki, T. Textural features corresponding to visual perception. IEEE Trans. Syst. Man. Cybern. 1978, 8, 460–473. [Google Scholar] [CrossRef]
- Theckedath, D.; Sedamkar, R.R. Detecting affect states using VGG16, ResNet50 and SE-ResNet50 networks. SN Comput. Sci. 2020, 1, 79. [Google Scholar] [CrossRef]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Davis, J.; Shah, M. Recognizing hand gestures. In Proceedings of the Computer Vision—ECCV’94: Third European Conference on Computer Vision, Stockholm, Sweden, 2–6 May 1994; Springer: Berlin/Heidelberg, Germany, 1994; Volume I 3, pp. 331–340. [Google Scholar]
- Fregoso, J.; Gonzalez, C.I.; Martinez, G.E. Optimization of convolutional neural networks architectures using PSO for sign language recognition. Axioms 2021, 10, 139. [Google Scholar] [CrossRef]
- Hammad, B.T.; Ahmed, I.T.; Jamil, N. An secure and effective copy move detection based on pretrained model. In Proceedings of the 2022 IEEE 13th Control and System Graduate Research Colloquium (ICSGRC), Shah Alam, Malaysia, 23 July 2022; pp. 66–70. [Google Scholar]
- Ahmed, I.T.; Jamil, N.; Din, M.M.; Hammad, B.T. Binary and Multi-Class Malware Threads Classification. Appl. Sci. 2022, 12, 12528. [Google Scholar] [CrossRef]
- Ahmed, I.T.; Der, C.S.; Jamil, N.; Hammad, B.T. Analysis of Probability Density Functions in Existing No-Reference Image Quality Assessment Algorithm for Contrast-Distorted Images. In Proceedings of the 2019 IEEE 10th Control and System Graduate Research Colloquium (ICSGRC), Shah Alam, Malaysia, 2–3 August 2019; pp. 133–137. [Google Scholar]
- Neiva, D.H.; Zanchettin, C. A dynamic gesture recognition system to translate between sign languages in complex backgrounds. In Proceedings of the 2016 5th Brazilian Conference on Intelligent Systems (BRACIS), Recife, Brazil, 9–12 October 2016; pp. 421–426. [Google Scholar]
- Srinath, S.; Sharma, G.K. Classification approach for sign language recognition. In Proceedings of the International Conference on Signal, Image Processing, Communication & Automation, Bengaluru, India, 6–7 July 2017; pp. 141–148. [Google Scholar]
- Sharma, A.; Mittal, A.; Singh, S.; Awatramani, V. Hand gesture recognition using image processing and feature extraction techniques. Procedia Comput. Sci. 2020, 173, 181–190. [Google Scholar] [CrossRef]
- Das, A.; Gawde, S.; Suratwala, K.; Kalbande, D. Sign language recognition using deep learning on custom processed static gesture images. In Proceedings of the 2018 International Conference on Smart City and Emerging Technology (ICSCET), Mumbai, India, 5 January 2018; pp. 1–6. [Google Scholar]
- Gao, Q.; Liu, J.; Ju, Z.; Li, Y.; Zhang, T.; Zhang, L. Static hand gesture recognition with parallel CNNs for space human-robot interaction. In Proceedings of the Intelligent Robotics and Applications: 10th International Conference, ICIRA 2017, Wuhan, China, 16–18 August 2017; Part I 10. pp. 462–473. [Google Scholar]
Authors and Publication Year | Feature Extraction Type | No. of Gestures | Feature Extraction Method | Classification Method | Recognition Rate (%) | Limitations |
---|---|---|---|---|---|---|
Gupta et al. (2016) [12] | Hand-Crafted Features | 10 | HOG and SFIT | KNN | 90 | - Sensitivity to background noise and occlusions. - Sensitivity to variations in lighting. |
Sadeddine et al. (2018) [13] | Hand-Crafted Features | 10 30 26 | Hu’s MD), ZMD, GFD and LBPD. | PNN | NUS: 96.67 ArSL: 92.64 ASL: 94.13 | - Sensitivity to background noise and occlusions. - Complexity of computation. - Restricted set of gestures. |
Zhang et al. (2018) [14] | Hand-Crafted Features | 10 | HOG and LBP | SVM | 96.7 | - Sensitivity to variations in lighting. - Complexity of computation. |
Sahoo et al. (2018) [15] | Hand-Crafted Features | 5 | DWT and F-ratio | SVM | 94.79 | - Sensitivity to background noise. - Sensitivity to changes in gesture shape and orientation. |
Gajalakshmi and Sharmila (2019) [17] | Hand-Crafted Features | 10 | CCH | SVM | 91 | -Sensitivity to changes in gesture shape and orientation. - CCH sensitivity to background noise and occlusions. |
Li et al. (2022) [18] | Hand-Crafted Features | 12 | HOG and 9ULBP | SVM | 94.1 | - Complexity of computation. - Sensitivity to extreme variations in scale, orientation, and partial occlusions. |
Sayed et al. (2024) [19] | Hand-Crafted Features | 15 | HOG | SVM, KNN, and DT | 97.64 | - Sensitivity to background noise and occlusions. - Sensitivity to variations in lighting. |
Nyirarugira et al. (2013) [20] | Deep Features | 10 | CNN | RF, boosting algorithms, and DT | 66.2 | - Poor generalization. - Sensitivity to variations in lighting. |
Aurangzeb et al. (2024) [21] | Deep Features | MUD: 12. ASLAD: 26. | VGG16 | Optimized CNN | MUD: 96.75 ASLAD: 94.36 | - Complexity of computation. - Sensitivity to variations in lighting. |
Ozcan and Basturk (2019) [22] | Deep Features | ASL Digits: 10. ASL Dataset: 26. | AlexNet | AlexNet | ASL Digits: 98.09 ASL: 94.33. | - Higher computational cost. - Full dependence on quality and amount of training datasets. |
Sahoo et al. (2022) [23] | Deep Features | MU: 12. HUST-ASL: 28. | AlexNet and VGG16 | Ensemble | MU: 90.26 HUST-ASL: 56.18 | - Poor generalization. - Poor performance in variable gesture recognition tasks. |
Mujahid et al. (2021) [24] | Deep Features | - | YOLO v3 and DarkNet-53 | YOLO v3 and DarkNet-53 | 97.10 | - Poor generalization. - Weak efficiency in practical applications. |
Edmond et al. (2020) [25] | Deep Features | ASL: 26. ASL Digits: 10 NUS: 10. | Lightweight VGG16 | Ensemble classifier | 96.7 | - Complexity of computation. - Poor generalization. |
Wang et al. (2021) [26] | Deep Features | 12 | MobileNet | RF | 94.5 | -Higher computational cost. - Sensitivity to variations in lighting. -Poor generalization. |
Kika et al. (2018) [27] | Combined | 15 | HOG and CNN | SVM | 96 | - Sensitivity to background noise and occlusions. - Decreased performance in real-world applications. |
Damaneh et al. (2022) [29] | Combined | 12 26 30 | ORB descriptor and Gabor filter | CNN | Massey: 99.92 ASL Alphabet: 99.80 ASL: 99.80 | -Potential inability to handle noise and changes in lighting. - Poor generalization to invisible gestures. |
Rastgoo et al., 2018) [28] | Combined | ASL: 30. Massey: 12. NYU:10. ASL: 26. | Gabor filters and Haar-like features | Restricted Boltzmann machines (RBMs) | Massey:91.3 ASL Alphabet: 90.5. ASL: 93.3. | - Sensitivity to background noise. - Complicated feature extraction process. |
Layer/Component | Description/Configuration |
---|---|
Input Layer | (224, 224, 3) |
Removed Top Layer | The final fully connected classification layer is removed (include_top = False) |
Feature Map Output Size | (7, 7, 2048) |
Global Average Pooling (GAP) | Reduces feature map dimensions from (7, 7, 2048) to (1, 1, 2048) |
Optimizer | Adam optimizer (learning_rate = 0.001, beta_1 = 0.9, beta_2 = 0.999, epsilon = 1 × 10−7) |
Activation Function | ReLU (Rectified Linear Unit) |
Batch Size | 32 |
Epochs | 10–20 |
Feature Extraction | Features are extracted from the Global Average Pooling (GAP) layer, which provides a 2048-dimensional feature vector |
Attribute | Description |
---|---|
Image Type | Static images of hand gestures |
Number of Classes | 29 |
Total Images | 87,000 |
Samples per Class | 3000 |
Class Names | A, B, C……Z, (as well as possibly additional symbols) |
Image Size | 200 × 200 |
Image Format | PNG or JPEG |
Images for training | 69,600 |
Images for testing | 8700 |
Task type | Multi-class classification problem |
Class ID | Name | Evaluation Measures (%) | ||||
---|---|---|---|---|---|---|
Accuracy | Error | F1-Score | Precision | Recall | ||
#1 | A | 96.56 | 3.44 | 98.25 | 96.61 | 99.95 |
#2 | B | 96.74 | 3.26 | 98.34 | 96.73 | 100 |
#3 | C | 96.28 | 3.72 | 98.10 | 96.28 | 100 |
#4 | D | 96.64 | 3.36 | 98.29 | 96.64 | 100 |
#5 | Del | 96.70 | 3.3 | 98.32 | 96.70 | 100 |
#6 | E | 96.45 | 3.55 | 98.19 | 96.46 | 99.98 |
#7 | F | 96.67 | 3.33 | 98.30 | 96.67 | 100 |
#8 | G | 96.66 | 3.34 | 98.30 | 96.66 | 100 |
#9 | H | 96.56 | 3.44 | 98.25 | 96.56 | 100 |
#10 | I | 96.59 | 3.41 | 98.82 | 96.59 | 100 |
#11 | J | 96.78 | 3.22 | 98.36 | 96.78 | 100 |
#12 | K | 96.68 | 3.32 | 98.31 | 96.68 | 100 |
#13 | L | 96.62 | 3.38 | 98.28 | 96.75 | 99.86 |
#14 | M | 96.70 | 3.3 | 98.32 | 96.70 | 100 |
#15 | N | 96.69 | 3.31 | 98.32 | 96.69 | 100 |
#16 | Nothing | 96.69 | 3.31 | 98.83 | 96.77 | 99.90 |
#17 | O | 96.32 | 3.68 | 98.13 | 96.33 | 99.99 |
#18 | P | 96.70 | 3.3 | 98.32 | 96.70 | 100 |
#19 | Q | 96.69 | 3.31 | 98.32 | 96.69 | 100 |
#20 | R | 96.76 | 3.24 | 98.35 | 96.77 | 99.99 |
#21 | S | 96.41 | 3.59 | 98.17 | 96.41 | 100 |
#22 | Space | 96.67 | 3.33 | 98.31 | 96.67 | 100 |
#23 | T | 96.49 | 3.51 | 98.22 | 96.49 | 100 |
#24 | U | 96.47 | 3.53 | 98.20 | 96.48 | 99.99 |
#25 | V | 96.46 | 3.54 | 98.20 | 96.46 | 100 |
#26 | W | 96.44 | 3.56 | 98.19 | 96.44 | 100 |
#27 | X | 96.40 | 3.6 | 98.17 | 96.40 | 100 |
#28 | Y | 99.40 | 0.6 | 99.69 | 99.70 | 99.68 |
#29 | Z | 96.72 | 3.28 | 98.33 | 96.72 | 100 |
Average | 96.687 | 3.312 | 98.351 | 96.707 | 99.977 |
Methods | Data Analysis | Feature Extraction Type | Classifier Kind | Dataset | Accuracy (%) |
---|---|---|---|---|---|
[41] | Static | LBP AND GLCM | SVM | ASL | 86 |
[43] | Static | CNN | CNN | ASL | 90 |
[42] | Static | HOG AND PCA | KNN | ASL | 95.1 |
[13] | Static | Hu’s Moment + LBPD + Zernike moments + GFD | PNN | ASL | 93.33 |
[17] | Static | CCH | SVM | ASL | 90.00 |
[44] | Static | RGB-CNN and Depth-CNN | SoftMax | ASL | 93.30 |
[24] | Static | YOLOv3 and DarkNet-53 | YOLOv3 and DarkNet-53 | Pascal VOC and YOLO format, | 97 |
[22] | Static | AlexNet | AlexNet | ASL | 85 |
Proposed | Static | Tamura and ResNet50 | GAM | ASL | 96.68 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Ahmed, I.T.; Gwad, W.H.; Hammad, B.T.; Alkayal, E. Enhancing Hand Gesture Image Recognition by Integrating Various Feature Groups. Technologies 2025, 13, 164. https://doi.org/10.3390/technologies13040164
Ahmed IT, Gwad WH, Hammad BT, Alkayal E. Enhancing Hand Gesture Image Recognition by Integrating Various Feature Groups. Technologies. 2025; 13(4):164. https://doi.org/10.3390/technologies13040164
Chicago/Turabian StyleAhmed, Ismail Taha, Wisam Hazim Gwad, Baraa Tareq Hammad, and Entisar Alkayal. 2025. "Enhancing Hand Gesture Image Recognition by Integrating Various Feature Groups" Technologies 13, no. 4: 164. https://doi.org/10.3390/technologies13040164
APA StyleAhmed, I. T., Gwad, W. H., Hammad, B. T., & Alkayal, E. (2025). Enhancing Hand Gesture Image Recognition by Integrating Various Feature Groups. Technologies, 13(4), 164. https://doi.org/10.3390/technologies13040164