Real-Time Hand Posture Recognition for Human-Robot Interaction Tasks
AbstractIn this work, we present a multiclass hand posture classifier useful for human-robot interaction tasks. The proposed system is based exclusively on visual sensors, and it achieves a real-time performance, whilst detecting and recognizing an alphabet of four hand postures. The proposed approach is based on the real-time deformable detector, a boosting trained classifier. We describe a methodology to design the ensemble of real-time deformable detectors (one for each hand posture that can be classified). Given the lack of standard procedures for performance evaluation, we also propose the use of full image evaluation for this purpose. Such an evaluation methodology provides us with a more realistic estimation of the performance of the method. We have measured the performance of the proposed system and compared it to the one obtained by using only the sampled window approach. We present detailed results of such tests using a benchmark dataset. Our results show that the system can operate in real time at about a 10-fps frame rate. View Full-Text
Share & Cite This Article
Hernandez-Belmonte, U.H.; Ayala-Ramirez, V. Real-Time Hand Posture Recognition for Human-Robot Interaction Tasks. Sensors 2016, 16, 36.
Hernandez-Belmonte UH, Ayala-Ramirez V. Real-Time Hand Posture Recognition for Human-Robot Interaction Tasks. Sensors. 2016; 16(1):36.Chicago/Turabian Style
Hernandez-Belmonte, Uriel H.; Ayala-Ramirez, Victor. 2016. "Real-Time Hand Posture Recognition for Human-Robot Interaction Tasks." Sensors 16, no. 1: 36.
Note that from the first issue of 2016, MDPI journals use article numbers instead of page numbers. See further details here.