Next Article in Journal
An Unsupervised Deep Hyperspectral Anomaly Detector
Next Article in Special Issue
Pressure Distribution during Negative Pressure Wound Therapy of Experimental Abdominal Compartment Syndrome in a Porcine Model
Previous Article in Journal
Label-Free Quantification of Anti-TNF-α in Patients Treated with Adalimumab Using an Optical Biosensor
Previous Article in Special Issue
Compact Hip-Force Sensor for a Gait-Assistance Exoskeleton System
Open AccessArticle

Enhancing Perception with Tactile Object Recognition in Adaptive Grippers for Human–Robot Interaction

System Engineering and Automation Department, University of Málaga, 29071 Málaga, Spain
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(3), 692; https://doi.org/10.3390/s18030692
Received: 28 December 2017 / Revised: 17 February 2018 / Accepted: 19 February 2018 / Published: 26 February 2018
(This article belongs to the Special Issue Force and Pressure Based Sensing Medical Application)
The use of tactile perception can help first response robotic teams in disaster scenarios, where visibility conditions are often reduced due to the presence of dust, mud, or smoke, distinguishing human limbs from other objects with similar shapes. Here, the integration of the tactile sensor in adaptive grippers is evaluated, measuring the performance of an object recognition task based on deep convolutional neural networks (DCNNs) using a flexible sensor mounted in adaptive grippers. A total of 15 classes with 50 tactile images each were trained, including human body parts and common environment objects, in semi-rigid and flexible adaptive grippers based on the fin ray effect. The classifier was compared against the rigid configuration and a support vector machine classifier (SVM). Finally, a two-level output network has been proposed to provide both object-type recognition and human/non-human classification. Sensors in adaptive grippers have a higher number of non-null tactels (up to 37% more), with a lower mean of pressure values (up to 72% less) than when using a rigid sensor, with a softer grip, which is needed in physical human–robot interaction (pHRI). A semi-rigid implementation with 95.13% object recognition rate was chosen, even though the human/non-human classification had better results (98.78%) with a rigid sensor. View Full-Text
Keywords: adaptive gripper; tactile sensors; human–robot interaction; object recognition adaptive gripper; tactile sensors; human–robot interaction; object recognition
Show Figures

Graphical abstract

MDPI and ACS Style

Gandarias, J.M.; Gómez-de-Gabriel, J.M.; García-Cerezo, A.J. Enhancing Perception with Tactile Object Recognition in Adaptive Grippers for Human–Robot Interaction. Sensors 2018, 18, 692.

Show more citation formats Show less citations formats
Note that from the first issue of 2016, MDPI journals use article numbers instead of page numbers. See further details here.

Article Access Map by Country/Region

1
Search more from Scilit
 
Search
Back to TopTop