Next Article in Journal
Energy-Efficient Control with Harvesting Predictions for Solar-Powered Wireless Sensor Networks
Next Article in Special Issue
Design and Analysis of a Sensor System for Cutting Force Measurement in Machining Processes
Previous Article in Journal
Self-Powered WSN for Distributed Data Center Monitoring
Previous Article in Special Issue
Sensor Fusion Based Model for Collision Free Mobile Robot Navigation
Article Menu

Export Article

Open AccessArticle
Sensors 2016, 16(1), 36; doi:10.3390/s16010036

Real-Time Hand Posture Recognition for Human-Robot Interaction Tasks

Universidad de Guanajuato DICIS, Carr. Salamanca-Valle Km. 3.5 + 1.8, Palo Blanco, Salamanca, C.P. 36885, Mexico
*
Author to whom correspondence should be addressed.
Academic Editor: Lianqing Liu
Received: 30 October 2015 / Revised: 9 December 2015 / Accepted: 18 December 2015 / Published: 4 January 2016
(This article belongs to the Special Issue Sensors for Robots)
View Full-Text   |   Download PDF [3299 KB, uploaded 5 January 2016]   |  

Abstract

In this work, we present a multiclass hand posture classifier useful for human-robot interaction tasks. The proposed system is based exclusively on visual sensors, and it achieves a real-time performance, whilst detecting and recognizing an alphabet of four hand postures. The proposed approach is based on the real-time deformable detector, a boosting trained classifier. We describe a methodology to design the ensemble of real-time deformable detectors (one for each hand posture that can be classified). Given the lack of standard procedures for performance evaluation, we also propose the use of full image evaluation for this purpose. Such an evaluation methodology provides us with a more realistic estimation of the performance of the method. We have measured the performance of the proposed system and compared it to the one obtained by using only the sampled window approach. We present detailed results of such tests using a benchmark dataset. Our results show that the system can operate in real time at about a 10-fps frame rate. View Full-Text
Keywords: human-robot interaction; vision-based hand posture recognition; full image evaluation; AdaBoost; bootstrapping human-robot interaction; vision-based hand posture recognition; full image evaluation; AdaBoost; bootstrapping
This is an open access article distributed under the Creative Commons Attribution License which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. (CC BY 4.0).

Scifeed alert for new publications

Never miss any articles matching your research from any publisher
  • Get alerts for new papers matching your research
  • Find out the new papers from selected authors
  • Updated daily for 49'000+ journals and 6000+ publishers
  • Define your Scifeed now

SciFeed Share & Cite This Article

MDPI and ACS Style

Hernandez-Belmonte, U.H.; Ayala-Ramirez, V. Real-Time Hand Posture Recognition for Human-Robot Interaction Tasks. Sensors 2016, 16, 36.

Show more citation formats Show less citations formats

Note that from the first issue of 2016, MDPI journals use article numbers instead of page numbers. See further details here.

Related Articles

Article Metrics

Article Access Statistics

1

Comments

[Return to top]
Sensors EISSN 1424-8220 Published by MDPI AG, Basel, Switzerland RSS E-Mail Table of Contents Alert
Back to Top