Next Article in Journal
Single-Trial Electroencephalography Discrimination of Real, Regulated, Isometric Wrist Extension and Wrist Flexion
Next Article in Special Issue
Technology Acceptance and Usability of a Therapy System with a Humanoid Robot Serving as Therapeutic Assistant for Post-Stroke Arm and Neurovisual Rehabilitation—An Evaluation Based on Stroke Survivors’ Experience
Previous Article in Journal
FedDyH: A Multi-Policy with GA Optimization Framework for Dynamic Heterogeneous Federated Learning
Previous Article in Special Issue
Neuromusculoskeletal Control for Simulated Precision Task versus Experimental Data in Trajectory Deviation Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Human-like Dexterous Grasping Through Reinforcement Learning and Multimodal Perception

by
Wen Qi
1,†,
Haoyu Fan
1,†,
Cankun Zheng
1,
Hang Su
2,* and
Samer Alfayad
2
1
School of Future Technology, South China University of Technology, Guangzhou 511442, China
2
The IBISC Laboratory, UEVE, University of Paris-Saclay, 91000 Evry, France
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Biomimetics 2025, 10(3), 186; https://doi.org/10.3390/biomimetics10030186
Submission received: 17 January 2025 / Revised: 3 March 2025 / Accepted: 13 March 2025 / Published: 18 March 2025
(This article belongs to the Special Issue Biomimetic Innovations for Human–Machine Interaction)

Abstract

:
Dexterous robotic grasping with multifingered hands remains a critical challenge in non-visual environments, where diverse object geometries and material properties demand adaptive force modulation and tactile-aware manipulation. To address this, we propose the Reinforcement Learning-Based Multimodal Perception (RLMP) framework, which integrates human-like grasping intuition through operator-worn gloves with tactile-guided reinforcement learning. The framework’s key innovation lies in its Tactile-Driven DCNN architecture—a lightweight convolutional network achieving 98.5% object recognition accuracy using spatiotemporal pressure patterns—coupled with an RL policy refinement mechanism that dynamically correlates finger kinematics with real-time tactile feedback. Experimental results demonstrate reliable grasping performance across deformable and rigid objects while maintaining force precision critical for fragile targets. By bridging human teleoperation with autonomous tactile adaptation, RLMP eliminates dependency on visual input and predefined object models, establishing a new paradigm for robotic dexterity in occlusion-rich scenarios.

1. Introduction

The replication of human dexterity in robotic manipulation remains an open challenge, particularly in non-visual environments where objects with diverse geometries and material properties demand adaptive grasping strategies. While recent advances in robotic hand design have achieved remarkable morphological similarity to human hands [1,2], existing control paradigms struggle to handle the infinite physical permutations encountered in real-world manipulation tasks. Traditional model-based approaches [3,4] rely on precise mathematical formulations of hand-object interactions, yet their effectiveness diminishes when confronted with unstructured scenarios requiring real-time tactile adaptation. This fundamental limitation becomes particularly acute when manipulating objects that demand continuous force modulation without visual feedback—a capability intrinsic to human manipulation but elusive in robotic systems. Emerging data-driven solutions [5,6,7] show promise in bridging this gap, yet often neglect the biological intelligence encoded in human sensorimotor coordination.
Recent advances in reinforcement learning (RL) have propelled human-robot interaction (HRI) toward resolving two interdependent challenges: dynamic adaptation to physiological variability and safety-guaranteed coordination in unstructured environments [8]. Central to this progress are human-in-the-loop RL frameworks that empirically optimize collaborative parameters through iterative physical interactions [9], effectively bridging deterministic control policies with stochastic human behaviors. Building on these foundations, hierarchical multi-agent architectures now integrate adversarial training and causal reasoning to disentangle intent ambiguities during complex manipulation tasks [10], while inverse RL methodologies enable robots to infer latent reward functions from expert demonstrations, replicating human-like force modulation without tactile feedback [11]. In safety-critical domains, hybrid paradigms combining augmented reality with deep RL achieve real-time collision prediction through multimodal sensor fusion [12], complemented by intrinsic reward mechanisms that autonomously balance task efficiency and risk mitigation [13]. Despite these innovations, current RL-driven HRI systems predominantly prioritize task-level metrics over the embodied cognitive principles governing human dexterity [14], limiting their ability to generalize across non-visual manipulation scenarios requiring symbiotic sensorimotor intelligence.
With increasing task complexity, the integration of multifinger dexterous robotic hands for grasping various objects has driven advances in data set availability and algorithmic sophistication, with recent studies showcasing the efficacy of large-scale reinforcement learning (RL) in addressing these challenges [15,16,17]. Unlike traditional deep learning methods constrained by task-specific limitations, RL thrives in dynamic, unstructured environments, providing remarkable adaptability and performance, especially in navigating the complexities of volatile grasping scenarios, where unsupervised strategies like RL offer significant innovative potential [18]. Model-free RL approaches, though promising, involve challenges such as precise reward configuration, prolonged iterative training, and substantial computational resources, often necessitating industrial-scale systems, thus complicating the application of dexterous manipulation to real-world problems [4]. These constraints raise a pivotal question: can dexterous behaviors be learned with greater sample efficiency, reducing computational and time demands while preserving or enhancing performance, thus unlocking new avenues for efficient and versatile solutions in complex robotic manipulation tasks?
Previous research has explored a wide range of gesture recognition models to facilitate human-robot teleoperation, utilizing various data sources such as skeletal motion, EMG, and joint angle information [19]. A notable approach integrated sensor fusion of hand kinematics for efficient control of underactuated bionic hands with minimal input data, while another innovative strategy employed a multilayer LSTM-RNN to classify hand gestures from depth vision data, highlighting the potential for advanced gesture recognition [20]. Although these models are effective, they typically rely on extensive data collection and preprocessing, without being tested in conjunction with real robotic systems to address potential issues that may arise in practical implementation. Reinforcement learning-based approaches offer a compelling alternative, enabling the system to learn from predefined rules, reducing dependence on large datasets, and providing a flexible framework for adapting to complex teleoperation tasks.
While recent advances in robotic manipulation have improved grasping reliability, existing systems still struggle with adaptive force control when handling unknown objects in unstructured environments—particularly in integrating human demonstration with autonomous tactile adaptation. This paper introduces the Reinforcement Learning-Based Multimodal Perception (RLMP) framework, illustrated in Figure 1, which aims to enhance the effectiveness of dexterous grasping through multimodal hand teleoperation and reinforcement learning-based modeling. (a) Human-Robot Interaction: The robotic hand’s functionality is evaluated in scenarios such as stabilizing a paper cup or performing clinking gestures, showcasing its interactive dexterity. (b) Teleoperation through Multimodal Inputs: A custom glove transmits precise data by modeling and processing the Euler angles of finger joints, enabling accurate guidance of the robotic hand’s movements. (c) Reinforcement Learning for Grasp Refinement: Adaptive reinforcement learning techniques are employed, leveraging diverse objects to iteratively enhance grasping efficiency and versatility. By integrating multiple sensory modalities, the RLMP framework facilitates more precise and adaptive grasping in complex and dynamic environments. The key contributions of this work include:
  • A human-robot co-learning framework (RLMP) that synergizes novice-friendly teleoperation with reinforcement learning, enabling robots to acquire adaptive grasping strategies for objects with diverse geometries and material properties, without requiring pre-defined object models or visual feedback.
  • A five fingers tactile-centric RL architecture that seamlessly correlates finger kinematics with tactile sensory input, eliminating the necessity for prior domain knowledge.
  • An innovative tactile recognition method that utilizes deep convolutional networks to extract material specific features from high-dimensional pressure matrices, achieving visual object recognition without supervision or manual labeling, which is crucial for adaptive grasping in visually blurry environments.
This study introduces the RLMP framework to address challenges in dexterous robotic grasping. Section 2 review previous advancements in robotic hand grasping, emphasizing model-based and RL approaches, highlight the integration of multimodal perception in teleoperation systems, and examine RL applications in human-robot interaction (HRI). Section 3 details the framework, including mapping glove movements to robotic hand actions for intuitive teleoperation, RL-based modeling to associate finger kinematics with tactile feedback, and the deployment of a DCNN for robust object recognition. In Section 4, comprehensive real-world tests demonstrate the effectiveness of the RLMP framework in achieving precise, contact-intensive grasps and accurate object identification in diverse scenarios. Finally, Section 5 highlights the significant contributions of the framework to improving dexterous robotic manipulation through multimodal perception and RL, offering a promising foundation to advance HRI applications in complex environments.

2. Related Works

The functional domains of robotic hands have been extensively explored in contemporary research, with a primary focus on the execution of intricate movements and tasks, as well as the precise mechanisms underlying object grasping and manipulation. The foundation of the proposed system draws on pivotal advances in the fields of dexterous grasping, teleoperation, and reinforcement learning. This section provides a concise synthesis of prior studies that form the contextual basis for the present investigation.

2.1. Dexterous Grasping via Robotic Hand

The pursuit of anthropomorphic dexterity in multi-fingered robotic grasping has driven innovations across domains ranging from precision manufacturing to teleoperated surgical systems. Recent advances in semantic-aware grasp synthesis leverage point cloud representations to encode object geometry and material properties, enabling robots to approximate human-like grasping adaptability under partial visual observations [21,22]. Particularly noteworthy are frameworks such as pregrasp-informed manipulation (PGDM), which circumvent the need for task-specific configurations by dynamically optimizing contact points through self-supervised learning, thereby enhancing generalization across unseen object categories [23]. Despite these methodological advancements, fundamental limitations persist across three critical fronts: the curse of dimensionality in reinforcement learning policy optimization, mechanical compromises between actuation speed and finger flexibility, and perceptual fragility during multi-directional target engagement. For instance, while generalized modeling approaches have improved rotational motion precision through kinematic constraints [4], they struggle to synchronize force modulation with trajectory planning in unstructured environments [24]. These unresolved challenges highlight the pressing need for control paradigms that holistically integrate biological principles with data-driven adaptability.

2.2. Multimodal Perception-Driven Teleoperation

The evolution of dexterous teleoperation systems increasingly relies on synergistic integration of heterogeneous sensor modalities to decode human manipulation intent [25]. Central to these efforts is the dual sensing paradigm: intrinsic measurements through position and bend sensors capture kinematic hand configurations, while extrinsic feedback from joint torque and tension sensors quantifies interaction dynamics, enabling comprehensive reconstruction of both positional and force-related manipulation states [26]. Complementing these approaches, haptic interfaces combined with electromyography (EMG) sensors establish bidirectional human-robot communication channels—translating muscle activation patterns into robotic force modulation while providing tactile feedback to operators [27]. However, prevailing methodologies exhibit two systemic limitations: (1) the persistent segregation of unimodal data streams (e.g., visual vs. proprioceptive) during perception fusion [28], and (2) the architectural decoupling between gripper hardware design and adaptive learning algorithms. This fragmentation fundamentally constrains the development of truly embodied teleoperation systems capable of simultaneous multimodality perception and dynamic tool-environment adaptation.

2.3. RL-Based HRI

The ability of a robotic system to securely grasp objects is vital for the successful execution of complex manipulation tasks. Over the years, a multitude of approaches grounded in deep learning have been proposed to address this challenge [29]. For instance, it framed grasping tasks as regression problems and leveraged the capabilities of Convolutional Neural Networks (CNNs) to tackle them. Although these vision-based models exhibit robust generalization in various grasping scenarios, their susceptibility to collisions often hampers their efficacy [30]. This limitation becomes particularly evident in tasks that require the manipulation of objects in close proximity, where the likelihood of collisions increases significantly [31].
However, RL aims to derive optimal policies that maximize the cumulative expected rewards. Recent advances in this domain have introduced RL-based frameworks specifically designed for dexterous grasping. These frameworks utilize multimodal grippers to streamline actions, enhancing efficiency in handling intricate manipulation tasks [28]. By integrating visual and tactile feedback, RL-based systems enable agents to continually gather new sensory data, facilitating dynamic refinement of their operational strategies in real time.
Unlike conventional deep learning approaches, which often depend on task-specific constraints and predefined conditions, RL methods exhibit significant potential for achieving superior performance in dexterous grasping. This advantage is particularly evident in dynamic and unstructured environments, where adaptability is crucial. The following sections of this paper provide an in-depth exploration of RL frameworks, highlighting their applicability and advantages in the context of advanced grasping tasks.

3. Method of RL-Based Multimodal Perception

The RLMP framework encapsulates a human-robot interaction process structured into three pivotal phases (see Figure 2): Initially, multimodal teleoperation is achieved through a sensorized glove that captures human hand kinematics via inertial measurement units and flexion sensors, with the acquired data being processed by a pose detection algorithm to generate robotic control signals through kinematic retargeting. Subsequently, reinforcement learning optimizes dexterous grasping by integrating tactile feedback matrices from 25-element pressure sensors with finger joint proprioception, where a Q-learning policy iteratively refines grasp stability through reward maximization based on object displacement minimization and contact force regulation. Finally, tactile patterns from historical grasp trials train a deep convolutional neural network classifier that enables real time object recognition by identifying material-specific pressure signatures, thereby closing the perception-action loop. This tripartite architecture establishes a bidirectional mapping between human demonstration and robotic execution while enabling adaptive grasping through continuous tactile learning.

3.1. Robotic Hand Mapping

For the robotic transposition of human hand movements, we utilize a glove sensor system that captures both curvature and Euler angles. As illustrated in Figure 3, the glove is equipped with six inertial measurement units (IMUs) and five flex sensors, which collectively record acceleration data A i = [ a x i , a y i , a z i ] , gyroscopic measurements G i = [ ω x i , ω y i , ω z i ] , and flex values F i for i { 1 , , 6 } , corresponding to the back of the palm and each of the fingers. To accommodate the inherent variability in human joint postures, it is crucial to perform a localized mapping of the IMU data for each distal interphalangeal joint (DIP). This process involves transforming the data from a global coordinate system to a palm-centered reference frame, ensuring a more accurate representation of the movements of the hands [32,33].
The Kalman filter operates directly on raw gyroscopic ( A i ) and accelerometric ( G i ) measurements, producing denoised A i ^ and G i ^ sequences [34].
State vector : X k = ω x i , ω y i , ω z i , a x i , a y i , a z i k | k 1 State transition : X k = F X k 1 + w k , F = I 3 Δ t · I 3 0 3 I 3 Measurement : Z k = H X k + v k , H = I 6
where Δ t denotes the sampling interval, with w k N ( 0 , Q ) and v k N ( 0 , R ) representing process and measurement noise respectively. The filter recursively computes:
X ^ k = X ^ k | k 1 + K k ( Z k H X ^ k | k 1 )
where the Kalman gain K k minimizes the estimation error covariance.
To mitigate orientation-induced signal artifacts caused by arbitrary hand positions during operation, we first establish a unified coordinate framework through sensor alignment. This critical preprocessing step eliminates waveform distortions and signal level discrepancies that would otherwise propagate through subsequent calculations when the hand assumes non-canonical orientations. Initially, the orientation of the IMU in the dorsal hand is consistent with the global system, implying that the rotation matrix R defaults to the identity matrix I.
R = 1 0 0 0 1 0 0 0 1
For each sampling time step Δ t , rotation matrix R with:
R i = R i · exp ( G i · Δ t ) = R i · [ e ω x i Δ t , e ω y i Δ t , e ω z i Δ t ]
Subsequently, determine the relative rotation matrix R c t i from the center of the palm to the five fingers. Transform the acceleration, angular velocity, and gravitational acceleration g from the global to the local coordinate system.
R c i = R 0 T · R i A c i = R 0 T · A i G c i = R 0 T · G i g = R 0 T · g
Therefore, the DIP angles could be calculated as:
a l i n e a r i = a x i 2 + a y i 2 + a z i 2 g 2 θ i = a r c t a n ( a l i n e a r i r )
where r is the radius of the glove’s bending axis.
To address the limitations of single-sensor mapping, we consider a Multimodal mapping strategy to processes IMU orientations θ i and flex values F i through coordinate transformation and confidence-aware fusion [20,35]:
θ fused = w imu · θ imu + w flex · F i
where the adaptive weights are computed from sensor variances:
w imu = σ flex 2 σ imu 2 + σ flex 2 , w flex = 1 w imu
with σ imu 2 = 0.8 and σ flex 2 = 0.3 obtained through 200 static calibration trials.
Kinematic Mapping:
q robot = K · θ fused + b
where the scaling matrix K = diag ( [ 1.05 , 0.97 , 1.02 , 1.1 , 0.95 ] ) and offset b are derived through least-squares fitting on calibration trajectories.

3.2. Reinforcement Learning Framework

Our reinforcement learning framework employs a multimodal state representation that synergistically integrates proprioceptive, tactile, and object-aware information. As depicted in Figure 2, the learning process operates through continuous interaction with the environment, where the agent’s policy π : S A maps observed states to joint torque adjustments. Departing from conventional single-modality approaches, we formulate the Markov Decision Process (MDP) with three complementary state components and a constrained action space, specifically designed for dexterous grasping scenarios. The key innovations include: (1) hierarchical abstraction of high-dimensional tactile data, (2) safety-aware action constraints to prevent mechanical overload, and (3) an adaptive reward mechanism balancing grasp stability and force minimization.
The Markov Decision Process (MDP) is formally defined as follows: State Space S :
  • Joint state:
    s j = [ θ 1 , ω 1 , , θ 5 , ω 5 ] R 10
    where θ i and ω i represent the angular position (rad) and velocity (rad/s) of the i-th finger joint.
  • Tactile state:
    s t = flatten ( M 5 × 5 ) R 25
    converting the pressure matrix M into a vector through row-wise flattening. Tactile perception of dexterous hands and data states under different grazing states as shown in Figure 4).
  • Object properties:
    s o = [ m , k ] R 2
    where m denotes mass (kg) and k represents stiffness coefficient (N/m).
Action Space A :
a = [ τ 1 , , τ 5 ] R 5 , Δ τ i [ 0.2 , 0.2 ] N · m
with the safety constraint:
i = 1 5 | τ i | 1.5 N · m
The composite reward function combines three essential objectives:
r ( s , a ) = w 1 · e β | Δ p | w 2 · | f i | + w 3 · I stable
  • Δ p : Position displacement error (mm) between current and target grasp
  • f i : Pressure deviation at i-th tactile element from desired range
  • I stable : Time-dependent stability bonus (0.1 per second)
The weighting coefficients are empirically set as:
w 1 = 2.0 , w 2 = 0.5 , w 3 = 0.3 , β = 0.1
Q-learning operates as an offline policy optimization method, whereas SARSA serves as an online policy optimization technique, inherently managing the trade-off between exploration and exploitation throughout the learning phase and adhering to the prevailing policy. Within the system, three distinct grasping states are delineated: standard grasp, overly loose grasp, and overly tight grasp. During teleoperation and object perception, the tactile feedback thresholds are fine-tuned to achieve standard grasping for various objects [36].
For Q-Learning, the update procedure is given as:
Q n e w ( s t , a t ) = ( 1 α ) Q ( s t , a t ) + α ( r t + γ m a x a Q ( s t + 1 , a ) )
For SARSA:
Q n e w ( s t , a t ) = ( 1 α ) Q ( s t , a t ) + α ( r t + γ Q ( s t + 1 , a t + 1 ) )
The subsequent state and action, s t + 1 and a t + 1 , are derived via the ϵ -greedy method. Before running the update, it is essential to define the optimal hyperparameters: learning rate α , ϵ in ϵ -greedy strategy and reward matrix R. This initialization promotes the integration of recent data with obsolete insights, improving dynamic learning. Hyperparameter selection plays a pivotal role in balancing immediate versus future returns in the reinforcement learning paradigm. In this context [37,38], decisions are temporally based on the agent’s discrete choices of states s and actions a at specific moments, namely t and t + 1 .

3.3. Object Recognition

To address the challenge of tactile feedback-driven object recognition in dexterous robotic manipulation, we propose a Tactile-Driven DCNN architecture for processing pressure matrices (as shown in Figure 5). The network’s input is a 25 × 5 tactile pressure matrix, representing the concatenated result of the tactile pressure measurement matrices from five fingers.
The architecture begins with a feature extraction module comprising four sequential convolutional layers. Each layer employs 3 × 3 convolutional kernels with ReLU activation and batch normalization, progressively increasing the number of filters ( 32 , 64 , 128 , 256 ) to hierarchically encode multi-scale spatial patterns. A 2 × 2 max-pooling operation follows every convolutional layer to reduce feature map dimensions while preserving dominant activation patterns. To ensure robust generalization, batch normalization is applied after each convolution to stabilize training dynamics, and a dropout layer (rate 0.5 ) is introduced after the final pooling stage to mitigate overfitting.
The classification module consists of a fully connected layer with 512 ReLU-activated units, which processes the flattened output from the convolutional base, followed by an output layer with softmax activation for multi-class probability estimation [39]. The model is trained end-to-end using the Adam optimizer with an initial learning rate of 1 × 10 2 , mini-batches of 50 samples, and a dropout rate of 0.001 applied during optimization. Training spans 300 epochs, with the mean squared error (MSE) serving as the loss function to minimize divergence between predicted and ground-truth class vectors in the hierarchical feature space. This architecture balances spatial feature extraction and regularization strategies to ensure adaptability in dynamic tactile perception tasks.

4. Experimental Evaluation and Discussion

4.1. HRI Platform

The experimental platform, as depicted in Figure 6, represents a comprehensive humanoid arm robot system, which is composed of three key components:
  • Humanoid Dual-arm Robot: The core of the experimental setup is a dual-arm humanoid robot, consisting of a fixed base, anthropomorphic mechanical arms, a tactile five-finger dexterous hand, a depth vision unit, and a main control unit. The anthropomorphic mechanical arms feature shoulder joints with three rotational degrees of freedom (DoFs), enabling forward-backward swinging, inward-outward expansion, and rotational movements; the elbow has one pitch DoF for flexion-extension control, while the wrist integrates three rotational DoFs. The tactile five-finger dexterous hand comprises six DoFs in total, with each finger equipped with a single-axis linkage bending joint for grasping actions, and the thumb additionally provided with a flexion-extension DoF to enable bidirectional motion capability. The fingertip area integrates a high-density 5 × 5 tactile sensor array, capable of detecting force interactions within a range of 0–5 N. Additionally, the depth vision module, located at the top of the robot, can acquire RGBD view data within its range, while the main control unit board, positioned at the center of the robot’s body, handles signal transmission and reception, performs related edge computing, and is equipped with connectivity options, including a standard serial port and a modern Wi-Fi interface.
  • Custom Data Glove:Complementing the robotic system is a custom-designed data glove, aimed at providing real-time kinesthetic feedback of human hand movements. The glove embeds five inertial measurement units (IMUs) at the fingertips and one IMU at the back of the palm to track motion information, while five flexible bending sensors are installed in the finger sleeve areas to capture the bending state of each finger during movement. The glove is equipped with a state-of-the-art transceiver system, utilizing a router to synchronously and robustly receive signals from both gloves, enabling synchronized data collection at multiple frequencies (25 Hz, 50 Hz, and 100 Hz).
  • Computational Apparatus: The system’s data analysis, real-time processing capabilities, and algorithmic workload are powered by a high-performance experimental PC equipped with an Intel i7 processor clocked at 2.80 GHz, complemented by 16 GB of RAM, enabling efficient multitasking and high data throughput. For GPU-intensive tasks and advanced simulations, the PC is also integrated with a state-of-the-art NVIDIA RTX 3070 Ti graphics processing unit.

4.2. Multimodal Hand-Robot Mapping Estimation

To enable teleoperation of the robotic hand that closely emulates human gestures, the system integrates a sophisticated multisensor data fusion methodology. Central to this approach is a comparative study that critically evaluates the performance of the fusion technique against two alternative approaches: one based on controlling the DIP joint angles and the other focused on the flexural bending of the joints. The experimental results, visually presented in Figure 7, reveal an important observation: reliance on the angle control of the DIP joint alone leads to significant mapping errors, particularly when there are small variations in angle. This issue can be attributed to the fact that natural hand movements require the simultaneous adjustment of multiple joints, and subtle variations in DIP angles are often compounded by changes in adjacent joints, which are not considered in this approach. In contrast, the flexural bending method introduces challenges of its own, primarily due to the complex interdependencies among the various degrees of freedom in the robotic hand. The motion of a single joint can affect adjacent joints, creating difficulties in achieving effective alignment.
Table 1 provides a numerical analysis of the results. Compared to directly using DIP mapping or Flex sensor mapping, the multimodal fusion method significantly outperforms other techniques. Except for the more complex variations in the Thumb Rotation angle, the multimodal mapping achieves the highest accuracy in the bending error comparison across the five fingers. This approach mitigates mapping errors caused by sensor inaccuracies or mechanical interference, which are common in alternative methods. Moreover, it demonstrates exceptional accuracy in teleoperation of the dexterous hand, enabling precise replication of intricate manipulation tasks and complex gestures. By integrating multiple sensors, the system effectively transmits the operator’s intent to the robotic hand, thereby minimizing operational deviations and errors. These findings strongly support the effectiveness of the multimodal fusion method, highlighting its immense potential in advancing humanoid robotics and enhancing their practical applications [40,41].

4.3. RL Performance

Our system integrates an RL-driven modeling strategy to facilitate the autonomous acquisition of object-specific grasping techniques by the dexterous robotic hand. In this system, the mechanical arms are only controlled through a fixed program to assist the interaction between the five finger tactile dexterous hand and humans. In this setup, human operators remotely control the hand to interact with four distinct objects: the rubber doll, the bounce ball, the plastic cup, and the paper cup. The selection of these four objects is grounded in their inherent characteristics, which lead to distinct variations in the tactile pressure matrices observed during grasping. Both the doll and the ball are elastic objects that deform when subjected to pressure, but the irregular shape of the doll results in a markedly different deformation pattern and pressure distribution compared to the spherical ball. In contrast, plastic and paper cups are rigid objects that are not expected to undergo deformation under normal conditions, with any deformation signaling potential damage. These two materials, differing in composition, exhibit distinct hardness levels and surface smoothness, factors that directly influence grasping performance. In sum, these four objects, chosen for their diversity in material properties and shapes, serve as a robust test bed to evaluate the grasping capabilities of the robotic hand in various scenarios. By analyzing tactile pressure matrices during interactions with these objects, the system continuously refines its grasping methodologies, adapting to the distinct curvature characteristics of each object.
The experimental procedure follows a three-stage architecture of “demonstration imitation-online learning-tactile optimization” (under the framework of Figure 2), implemented as follows: First, eight volunteer operators wearing data gloves randomly performed multiple grasping demonstrations (over 10 trials) for four types of objects: rubber doll, elastic ball, plastic cup, and corrugated paper cup. During grasping, the multimodal mapping results from glove data were fed back to drive the five-fingered dexterous hand in synchronous motion to replicate identical postures, while sensing the 5 × 5 pressure distribution matrix from fingertip tactile sensors. The reinforcement learning module constructs an initial Q-table by integrating demonstration data through an ϵ -greedy strategy (exploration rate ϵ = 0.4). Thereafter, the online learning phase commences: employing a dynamic learning rate α (baseline value 0.2, adaptively adjusted with tactile pressure gradients) and multiple discount factors γ = 0.1, 0.3, and 0.5 to update Q-values based on real-time tactile feedback.
The empirical results depicted in Figure 8 provide a comparative analysis of Q-learning and SARSA under varying discount factors ( γ = 0.1 , 0.3 , 0.5 ). The Q matrices are designed to encode distinct tactile-driven grasping states derived from the 5 × 5 pressure distribution matrices, where states such as ‘loose grasp’ and ‘stabilized grasp’ are mapped to reward structures. In particular, the matrices
Q 1 = 2 1 3 3 2 1 1 3 2
and
Q 2 = 1 1 3 3 1 1 1 3 1
capture distinct exploration-exploitation tradeoffs tied to γ values. The entries of the Q 1 and Q 2 were determined through a task-driven reward design combined with empirical calibration. Specifically, the values encode utility assessments for distinct tactile grasping states: high entries prioritize stable grasps, while penalties deter unsafe states. For γ = 0.1 , Q 1 prioritizes immediate rewards through rapid state transitions, aligning with the short-term focus required for initial contact phases. In contrast, Q 2 ’s penalty terms better support long-term stability goals when γ = 0.5 , albeit with delayed convergence. The interplay between γ and tactile-driven learning rates α reveals critical dynamics: Lower γ values complement adaptive α scaling (via pressure gradients), enabling fast adaptation to abrupt force changes (elastic objects). Higher γ ( 0.5 ) enhances coordination of multi-step force modulation (contact→stabilize→hold), as evidenced by reduced object slip events in post-optimization trials. SARSA’s inherent conservatism provides more stable policies for rigid objects, while Q-learning’s optimistic updates excel in exploratory grasping of deformable targets. These findings validate the framework’s capability to balance real-time responsiveness and strategic force planning through tactile-aware RL design.

4.4. Tactile-Driven Object Identification

Figure 9 provides a meticulous evaluation of the robotic hand’s performance in a diverse set of object grasping scenarios through visual imagery and tactile matrices, while distributed matrices with moderate intensity typically signify successful grasping. Conversely, non-uniform distribution with pronounced concentration often suggests either overly firm or insufficiently secure grasps. The tactile sensors embedded in the hand play a pivotal role, serving as the essential instruments to capture critical metrics that are fundamental to assessing the effectiveness of the grasping process. Through a careful analysis of the tactile data, particularly examining the subtle variations in sensor responses that correspond to the distinct material properties of the objects, our study uncovers valuable insights into the various grasping strategies. This chapter underscores the significant role of grasping techniques in object recognition, with particular attention to the variance in grasp quality. Grasps that exhibit excessive force often lead to object deformation, which not only harms the object, but also hinders the ability of the robotic hand to maintain stable hold. In contrast, grasps that lack sufficient force risk instability, which can cause the object to slip from the grasp of the hand. In contrast, the ideal grasp demonstrates a harmonious balance, where force is applied with precision, ensuring both stability and effective manipulation without any detrimental effects on the object.
The tactile-driven object recognition module was rigorously evaluated through specialized trials. For classifying the four target objects (rubber doll, bouncy ball, plastic cup, paper cup), we implemented a streamlined DCNN architecture (detailed in Section 3.3) to balance accuracy and computational efficiency—a critical requirement for real-time deployment on the robotic mainboard. The training utilized various types of data successfully grasped in RL experiments. The protocol involved: (1) Data preprocessing: Normalizing pressure values to [0, 1] based on sensor range; (2) Optimization: Adam solver (lr = 1 × 10 3 , batch size = 32) with early stopping (10-epoch patience). The model is based on PyTorch Mobile 2.5.0 for lightweight direct loading and inference. The training graph of the model and the error matrix graph of the test results are shown in the Figure 10.
Tactile-based identification achieves 98.5% classification accuracy on the test set, demonstrating the efficacy of the proposed lightweight DCNN architecture. Notably, rigid objects (plastic/paper cups) exhibit near-perfect precision due to consistent pressure signatures during initial contact. Deformable targets show marginally lower recall, attributable to transient pressure variations induced by elastic deformation.
A fascinating pattern emerges when the grasping of elastic objects is analyzed. The optimal grasp is characterized by a uniformly distributed force matrix, reflecting an equilibrium in pressure application that ensures a firm yet non-deforming grip on the object. However, the force distribution is not entirely independent of the shape of the object. In the case of an irregularly shaped doll, the pressure matrix remains evenly distributed, although its center does not align with the center of the matrix, as the fingers interact with the nonuniform surface of the object. On the other hand, for regular, spherical objects, the pressure distribution aligns closely with the center of the matrix. As the force deviates from the optimal range, significant consequences ensue; excessive force can lead to irreversible deformation of the object, which is evident in the uneven distribution of the pressure matrix. However, insufficient force does not secure a firm grip, jeopardizing the stability of the grip and risking the escape of the object from the hand.
The handling of rigid objects, exemplified by everyday objects such as cups, introduces a new layer of complexity, where the art of grasping lies in the precise modulation of force based on the properties of the object’s material. Different types of cups, such as plastic and paper cups, present varying challenges due to their differing hardness and surface textures. For smoother and more fragile plastic cups, the robotic hand must apply a moderate and evenly distributed force to ensure stability while avoiding deformation. As illustrated in the figure, when excessive force is exerted, the plastic cup, which cannot withstand pressure, undergoes deformation. In contrast, paper cups, being relatively stronger and rougher in texture, require a greater applied force to maintain a stable grip. However, beyond a critical threshold, excessive force can lead to deformation of the paper cup, with the pressure matrix showing markedly reduced values. Ensuring a uniform force distribution across all digits of the robotic hand is paramount; any deviations in force application can cause the object to warp due to excessive pressure or result in an unstable grip from insufficient force, both compromising the success of the grasp.

5. Conclusions

This study tackles the fundamental challenge of enabling robots to grasp geometrically and materially diverse objects in vision-denied environments, where traditional model-based approaches struggle due to their reliance on predefined object models and visual feedback. The proposed RLMP framework overcomes these limitations by embedding human-like adaptive grasping capabilities through a synergistic integration of tactile-guided reinforcement learning and novice-accessible teleoperation. Experiments have shown that RLMP achieves very small errors in human-computer interaction mapping, successfully learning to grasp different objects while achieving 98.5% classification accuracy in lightweight classifiers based solely on tactile data.
The RLMP framework addresses the fundamental challenge of enabling robots to grasp geometrically and materially diverse objects in non-visual environments, where conventional methods fail to adapt to high-precision force and contact-state requirements. By learning from human demonstrations via teleoperation, RLMP translates human grasping intuition—refined through years of biological evolution—into tactile-guided reinforcement learning policies. This allows robots to dynamically adjust finger kinematics and force profiles based on real-time pressure feedback, eliminating dependency on visual input or predefined object models. By closing the gap between human dexterity and robotic autonomy, RLMP achieves a new approach to manipulation in environments where vision is unreliable or absent.
Future work will focus on expanding RLMP’s robustness in dynamic environments: (1) addressing data drift through online incremental learning to maintain tactile recognition accuracy under sensor degradation or environmental variations (e.g., temperature/humidity changes), and (2) mitigating concept drift via adaptive reinforcement learning that dynamically updates reward functions when task requirements evolve (e.g., shifting from deformation prevention to rapid grasping). Additionally, we will explore hardware-software co-design to enhance real-time adaptability while extending the framework to collaborative tasks requiring human-robot force negotiation.

Author Contributions

W.Q.: Conceptualization, Methodology, Writing—Original draft preparation. H.F.: Software, Data curation. C.Z.: Visualization, Investigation. H.S.: Supervision. S.A.: Writing—Reviewing and Editing. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by National Nature Science Foundation of China under Grant 62303187, and in part by Guangdong Provincial Key Laboratory of Human Digital Twin under Grant 2022B1212010004.

Institutional Review Board Statement

The study utilized non-invasive wearable devices to collect human motion data and did not involve medical interventions, biological sampling, or personal privacy risks. Ethical approval for this research was reviewed and exempted by the Ethics Committee of South China University of Technology (SCUT), in accordance with institutional guidelines and China’s Regulations on Ethical Review of Biomedical Research Involving Human Subjects. This exemption applies to non-interventional observational studies that solely record kinematic parameters (e.g., limb movement trajectories, acceleration data) through sensor-based measurements without posing physical or psychological risks. All participants were fully informed about the research purpose, assured of data anonymity, and provided written consent regarding data collection, storage, and future academic use. No conflicts of interest or participant compensation were involved in the study.

Data Availability Statement

The raw data supporting the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhu, T.; Wu, R.; Lin, X.; Sun, Y. Toward human-like grasp: Dexterous grasping via semantic representation of object-hand. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 15741–15751. [Google Scholar]
  2. Hampali, S.; Rad, M.; Oberweger, M.; Lepetit, V. Honnotate: A method for 3d annotation of hand and object poses. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 3196–3206. [Google Scholar]
  3. Dollar, A.M.; Howe, R.D. The SDM hand as a prosthetic terminal device: A feasibility study. In Proceedings of the 2007 IEEE 10th International Conference on Rehabilitation Robotics, Noordwijk, The Netherlands, 13–15 June 2007; pp. 978–983. [Google Scholar]
  4. Arunachalam, S.P.; Silwal, S.; Evans, B.; Pinto, L. Dexterous imitation made easy: A learning-based framework for efficient dexterous manipulation. In Proceedings of the 2023 Ieee International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; pp. 5954–5961. [Google Scholar]
  5. Puhlmann, S.; Harris, J.; Brock, O. RBO Hand 3: A platform for soft dexterous manipulation. IEEE Trans. Robot. 2022, 38, 3434–3449. [Google Scholar] [CrossRef]
  6. Beattie, A.; Mulink, P.; Sahoo, S.; Christou, I.T.; Kalalas, C.; Gutierrez-Rojas, D.; Nardelli, P.H. A Robust and Explainable Data-Driven Anomaly Detection Approach For Power Electronics. In Proceedings of the 2022 IEEE International Conference on Communications, Control, and Computing Technologies for Smart Grids (SmartGridComm), Singapore, 25–28 October 2022; pp. 296–301. [Google Scholar]
  7. Liu, L.; Zhang, Y.; Liu, G.; Xu, W. Variable motion mapping to enhance stiffness discrimination and identification in robot hand teleoperation. Robot. Comput.-Integr. Manuf. 2018, 51, 202–208. [Google Scholar] [CrossRef]
  8. Newbury, R.; Gu, M.; Chumbley, L.; Mousavian, A.; Eppner, C.; Leitner, J.; Bohg, J.; Morales, A.; Asfour, T.; Kragic, D.; et al. Deep learning approaches to grasp synthesis: A review. IEEE Trans. Robot. 2023, 39, 3994–4015. [Google Scholar] [CrossRef]
  9. Oliff, H.; Liu, Y.; Kumar, M.; Williams, M.; Ryan, M. Reinforcement learning for facilitating human-robot-interaction in manufacturing. J. Manuf. Syst. 2020, 56, 326–340. [Google Scholar] [CrossRef]
  10. Zhang, R.; Lv, Q.; Li, J.; Bao, J.; Liu, T.; Liu, S. A reinforcement learning method for human-robot collaboration in assembly tasks. Robot. Comput.-Integr. Manuf. 2022, 73, 102227. [Google Scholar] [CrossRef]
  11. Modares, H.; Ranatunga, I.; Lewis, F.L.; Popa, D.O. Optimized assistive human–robot interaction using reinforcement learning. IEEE Trans. Cybern. 2015, 46, 655–667. [Google Scholar] [CrossRef]
  12. Semeraro, F.; Griffiths, A.; Cangelosi, A. Human–robot collaboration and machine learning: A systematic review of recent research. Robot. Comput.-Integr. Manuf. 2023, 79, 102432. [Google Scholar] [CrossRef]
  13. Liu, Q.; Liu, Z.; Xiong, B.; Xu, W.; Liu, Y. Deep reinforcement learning-based safe interaction for industrial human-robot collaboration using intrinsic reward function. Adv. Eng. Inform. 2021, 49, 101360. [Google Scholar] [CrossRef]
  14. Modares, H.; Ranatunga, I.; AlQaudi, B.; Lewis, F.L.; Popa, D.O. Intelligent human–robot interaction systems using reinforcement learning and neural networks. In Trends in Control and Decision-Making for Human–Robot Collaboration Systems; Springer: Cham, Switzerland, 2017; pp. 153–176. [Google Scholar]
  15. Veiga, F.; Edin, B.; Peters, J. Grip stabilization through independent finger tactile feedback control. Sensors 2020, 20, 1748. [Google Scholar] [CrossRef]
  16. Deng, Z.; Jonetzko, Y.; Zhang, L.; Zhang, J. Grasping force control of multi-fingered robotic hands through tactile sensing for object stabilization. Sensors 2020, 20, 1050. [Google Scholar] [CrossRef]
  17. James, J.W.; Lepora, N.F. Slip detection for grasp stabilization with a multifingered tactile robot hand. IEEE Trans. Robot. 2020, 37, 506–519. [Google Scholar] [CrossRef]
  18. Veiga, F.; Akrour, R.; Peters, J. Hierarchical tactile-based control decomposition of dexterous in-hand manipulation tasks. Front. Robot. AI 2020, 7, 521448. [Google Scholar] [CrossRef] [PubMed]
  19. Su, H.; Zhang, J.; Fu, J.; Ovur, S.E.; Qi, W.; Li, G.; Hu, Y.; Li, Z. Sensor fusion-based anthropomorphic control of under-actuated bionic hand in dynamic environment. In Proceedings of the 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic, 27 September–1 October 2021; pp. 2722–2727. [Google Scholar]
  20. Qi, W.; Ovur, S.E.; Li, Z.; Marzullo, A.; Song, R. Multi-sensor guided hand gesture recognition for a teleoperated robot using a recurrent neural network. IEEE Robot. Autom. Lett. 2021, 6, 6039–6045. [Google Scholar] [CrossRef]
  21. Yu, S.; Zhai, D.H.; Xia, Y.; Wu, H.; Liao, J. SE-ResUNet: A novel robotic grasp detection method. IEEE Robot. Autom. Lett. 2022, 7, 5238–5245. [Google Scholar] [CrossRef]
  22. Cheng, H.; Wang, Y.; Meng, M.Q.-H. A robot grasping system with single-stage anchor-free deep grasp detector. IEEE Trans. Instrum. Meas. 2022, 71, 1–12. [Google Scholar] [CrossRef]
  23. Dasari, S.; Gupta, A.; Kumar, V. Learning dexterous manipulation from exemplar object trajectories and pre-grasps. In Proceedings of the 2023 IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; pp. 3889–3896. [Google Scholar]
  24. Xu, K.; Hu, Z.; Doshi, R.; Rovinsky, A.; Kumar, V.; Gupta, A.; Levine, S. Dexterous manipulation from images: Autonomous real-world rl via substep guidance. In Proceedings of the 2023 IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; pp. 5938–5945. [Google Scholar]
  25. de Carvalho, K.B.; Villa, D.K.D.; Sarcinelli-Filho, M.; Brandao, A.S. Gestures-teleoperation of a heterogeneous multi-robot system. Int. J. Adv. Manuf. Technol. 2022, 118, 1999–2015. [Google Scholar] [CrossRef]
  26. Xia, Z.; Deng, Z.; Fang, B.; Yang, Y.; Sun, F. A review on sensory perception for dexterous robotic manipulation. Int. J. Adv. Robot. Syst. 2022, 19, 17298806221095974. [Google Scholar] [CrossRef]
  27. Luo, J.; Lin, Z.; Li, Y.; Yang, C. A teleoperation framework for mobile robots based on shared control. IEEE Robot. Autom. Lett. 2019, 5, 377–384. [Google Scholar] [CrossRef]
  28. Liu, F.; Sun, F.; Fang, B.; Li, X.; Sun, S.; Liu, H. Hybrid robotic grasping with a soft multimodal gripper and a deep multistage learning scheme. IEEE Trans. Robot. 2023, 39, 2379–2399. [Google Scholar] [CrossRef]
  29. Yang, Y.; Liang, H.; Choi, C. A deep learning approach to grasping the invisible. IEEE Robot. Autom. Lett. 2020, 5, 2232–2239. [Google Scholar] [CrossRef]
  30. Guan, H.; Li, J.; Yan, R. An efficient robotic grasping pipeline base on fully convolutional neural network. In Proceedings of the 2019 5th International Conference on Control, Automation and Robotics (ICCAR), Beijing, China, 19–22 April 2019; pp. 172–176. [Google Scholar]
  31. Mandikal, P.; Grauman, K. Learning dexterous grasping with object-centric visual affordances. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xian, China, 30 May–5 June 2021; pp. 6169–6176. [Google Scholar]
  32. Zheng, X.Z.; Nakashima, R.; Yoshikawa, T. On dynamic control of finger sliding and object motion in manipulation with multifingered hands. IEEE Trans. Robot. Autom. 2000, 16, 469–481. [Google Scholar] [CrossRef]
  33. Gustus, A.; Stillfried, G.; Visser, J.; Jörntell, H.; van der Smagt, P. Human hand modelling: Kinematics, dynamics, applications. Biol. Cybern. 2012, 106, 741–755. [Google Scholar] [CrossRef] [PubMed]
  34. Bousbaine, A.; Fareha, A.; Josaph, A.K.; Fekik, A.; Azar, A.T.; Moualek, R.; Benyahia, N.; Benamrouche, N.; Kamal, N.A.; Al Mhdawi, A.K.; et al. Design and Implementation of a Robust 6-DOF Quadrotor Controller Based on Kalman Filter for Position Control. In Mobile Robot: Motion Control and Path Planning; Springer: Cham, Switzerland, 2023; pp. 331–363. [Google Scholar]
  35. Su, H.; Qi, W.; Yang, C.; Sandoval, J.; Ferrigno, G.; De Momi, E. Deep neural network approach in robot tool dynamics identification for bilateral teleoperation. IEEE Robot. Autom. Lett. 2020, 5, 2943–2949. [Google Scholar] [CrossRef]
  36. Kumar Shastha, T.; Kyrarini, M.; Gräser, A. Application of reinforcement learning to a robotic drinking assistant. Robotics 2019, 9, 1. [Google Scholar] [CrossRef]
  37. Watkins, C.J.; Dayan, P. Q-learning. Mach. Learn. 1992, 8, 279–292. [Google Scholar] [CrossRef]
  38. Zou, S.; Xu, T.; Liang, Y. Finite-sample analysis for sarsa with linear function approximation. Adv. Neural Inf. Process. Syst. 2019, 32, 8668–8678. [Google Scholar]
  39. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  40. Jleilaty, S.; Ammounah, A.; Abdulmalek, G.; Nouveliere, L.; Su, H.; Alfayad, S. Distributed real-time control architecture for electrohydraulic humanoid robots. Robot. Intell. Autom. 2024, 44, 607–620. [Google Scholar] [CrossRef]
  41. Ovur, S.E.; Su, H.; Qi, W.; De Momi, E.; Ferrigno, G. Novel adaptive sensor fusion methodology for hand pose estimation with multileap motion. IEEE Trans. Instrum. Meas. 2021, 70, 1–8. [Google Scholar] [CrossRef]
Figure 1. The presented graph illustrates the three-layered framework of our anthropomorphic dexterous grasping system.
Figure 1. The presented graph illustrates the three-layered framework of our anthropomorphic dexterous grasping system.
Biomimetics 10 00186 g001
Figure 2. RLMP framework overall structure diagram, human-like dexterous grasping is realized through a series of interconnected processes.
Figure 2. RLMP framework overall structure diagram, human-like dexterous grasping is realized through a series of interconnected processes.
Biomimetics 10 00186 g002
Figure 3. Introduction of palm structure and data glove perception information.
Figure 3. Introduction of palm structure and data glove perception information.
Biomimetics 10 00186 g003
Figure 4. Introduction of tactile perception of dexterous hands and data states under different grasping states.
Figure 4. Introduction of tactile perception of dexterous hands and data states under different grasping states.
Biomimetics 10 00186 g004
Figure 5. Structure diagram of classifier based on tactile data.
Figure 5. Structure diagram of classifier based on tactile data.
Biomimetics 10 00186 g005
Figure 6. Overview of the designed Human Robot Teleoperation system.
Figure 6. Overview of the designed Human Robot Teleoperation system.
Biomimetics 10 00186 g006
Figure 7. Comparison of angles in different ways of mapping from hand to manipulator joint angles.
Figure 7. Comparison of angles in different ways of mapping from hand to manipulator joint angles.
Biomimetics 10 00186 g007
Figure 8. The quantitative analysis of changing Q values for Q-Learning (Left: Q = Q1, Right: Q = Q2) and SARSA (Left: Q = Q1, Right: Q = Q2) when γ = 0.1 , 0.3 , 0.5 . Epoch refers to the total number of complete traversal attempts.
Figure 8. The quantitative analysis of changing Q values for Q-Learning (Left: Q = Q1, Right: Q = Q2) and SARSA (Left: Q = Q1, Right: Q = Q2) when γ = 0.1 , 0.3 , 0.5 . Epoch refers to the total number of complete traversal attempts.
Biomimetics 10 00186 g008
Figure 9. The efficacy of robotic hands in grasping diverse objects.
Figure 9. The efficacy of robotic hands in grasping diverse objects.
Biomimetics 10 00186 g009
Figure 10. Training process and recognition test results of Tactile-Driven DCNN.
Figure 10. Training process and recognition test results of Tactile-Driven DCNN.
Biomimetics 10 00186 g010
Table 1. Comparison of absolute differences between joint angles of robotic arms and different mapping methods.
Table 1. Comparison of absolute differences between joint angles of robotic arms and different mapping methods.
JointAbsolute of Angle DifferenceMean ± Standard Deviation
Thumb Rotation angle | θ R o b o t H a n d θ M u l t i m o d a l | 3.60 ± 2.16
| θ R o b o t H a n d θ D I P | 11.48 ± 5.60
| θ R o b o t H a n d θ F l e x | 3.09 ± 1.54
Thumb bend angle | θ R o b o t H a n d θ M u l t i m o d a l | 0.43 ± 0.45
| θ R o b o t H a n d θ D I P | 2.93 ± 2.15
| θ R o b o t H a n d θ F l e x | 2.81 ± 0.85
Index bend angle | θ R o b o t H a n d θ M u l t i m o d a l | 1.34 ± 0.71
| θ R o b o t H a n d θ D I P | 12.54 ± 7.59
| θ R o b o t H a n d θ F l e x | 3.20 ± 3.54
Middle bend angle | θ R o b o t H a n d θ M u l t i m o d a l | 1.57 ± 1.14
| θ R o b o t H a n d θ D I P | 15.47 ± 7.13
| θ R o b o t H a n d θ F l e x | 6.08 ± 7.15
Thumb bend angle | θ R o b o t H a n d θ M u l t i m o d a l | 1.04 ± 0.72
| θ R o b o t H a n d θ D I P | 8.47 ± 5.90
| θ R o b o t H a n d θ F l e x | 2.75 ± 1.99
Pinky bend angle | θ R o b o t H a n d θ M u l t i m o d a l | 1.14 ± 0.99
| θ R o b o t H a n d θ D I P | 8.36 ± 5.14
| θ R o b o t H a n d θ F l e x | 3.00 ± 2.64
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qi, W.; Fan, H.; Zheng, C.; Su, H.; Alfayad, S. Human-like Dexterous Grasping Through Reinforcement Learning and Multimodal Perception. Biomimetics 2025, 10, 186. https://doi.org/10.3390/biomimetics10030186

AMA Style

Qi W, Fan H, Zheng C, Su H, Alfayad S. Human-like Dexterous Grasping Through Reinforcement Learning and Multimodal Perception. Biomimetics. 2025; 10(3):186. https://doi.org/10.3390/biomimetics10030186

Chicago/Turabian Style

Qi, Wen, Haoyu Fan, Cankun Zheng, Hang Su, and Samer Alfayad. 2025. "Human-like Dexterous Grasping Through Reinforcement Learning and Multimodal Perception" Biomimetics 10, no. 3: 186. https://doi.org/10.3390/biomimetics10030186

APA Style

Qi, W., Fan, H., Zheng, C., Su, H., & Alfayad, S. (2025). Human-like Dexterous Grasping Through Reinforcement Learning and Multimodal Perception. Biomimetics, 10(3), 186. https://doi.org/10.3390/biomimetics10030186

Article Metrics

Back to TopTop