Next Article in Journal
The Innovative Design and Performance Testing of a Mobile Robot for the Automated Installation of Spacers on Six-Split Transmission Lines
Next Article in Special Issue
An Autonomous Fluoroscopic Imaging System for Catheter Insertions by Bilateral Control Scheme: A Numerical Simulation Study
Previous Article in Journal
Comparative Analysis of Dry, Minimum Quantity Lubrication, and Nano-Reinforced Minimum Quantity Lubrication Environments on the Machining Performance of AZ91D Magnesium Alloy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Methodology for Human–Robot Collaborative Assembly Based on Human Skill Imitation and Learning

1
Shenzhen Research Institute of Nankai University, Nankai University, Shenzhen 518083, China
2
College of Design and Innovation, Zhejiang Normal University, Jinhua 321004, China
3
School of Information and Software Engineering, University of Electronic Science and Technology of China, Chengdu 610054, China
4
College of Mathematical Sciences, Sichuan Normal University, Chengdu 610066, China
*
Author to whom correspondence should be addressed.
Machines 2025, 13(5), 431; https://doi.org/10.3390/machines13050431
Submission received: 11 April 2025 / Revised: 14 May 2025 / Accepted: 16 May 2025 / Published: 19 May 2025

Abstract

With the growing demand for personalized and flexible production, human–robot collaboration technology receives increasing attention. However, enabling robots to accurately perceive and align with human motion intentions remains a significant challenge. To address this, a novel human–robot collaborative control framework is proposed, which utilizes electromyography (EMG) signals as an interaction interface and integrates human skill imitation with reinforcement learning. Specifically, to manage the dynamic variation in muscle coordination patterns induced by joint angle changes, a temporal graph neural network enhanced with an Angle-Guided Attention mechanism is developed. This method adaptively models the topological relationships among muscle groups, enabling high-precision three-dimensional dynamic arm force estimation. Furthermore, an expert reward function and a fuzzy experience replay mechanism are introduced in the reinforcement learning model to guide the human skill learning process, thereby enhancing collaborative comfort and smoothness. The proposed approach is validated through a collaborative assembly task. Experimental results show that the proposed arm force estimation model reduces estimation errors by 10.38%, 8.33%, and 11.20% across three spatial directions compared to a conventional Deep Long Short-Term Memory (Deep-LSTM). Moreover, it significantly outperforms state-of-the-art methods, including traditional imitation learning and adaptive admittance control, in terms of collaborative comfort, smoothness, and assembly accuracy.

1. Introduction

In recent years, robotic technologies have been increasingly applied in industrial production, particularly in domains such as welding and assembly [1,2]. However, the level of automation remains constrained when facing a wide range of unstructured environments and flexible manufacturing tasks [3]. To address this issue, the concept of human–robot collaboration (HRC) has been proposed, aiming to integrate human decision-making capabilities with the high payload capacity of robots, thereby demonstrating substantial application potential [4].
A key challenge in the development of HRC technologies lies in the lack of efficient human–robot interaction interfaces, which hampers accurate transmission of human motion intent. Currently, physical interaction, primarily based on force sensors, remains the dominant mode of HRC [5]. However, since force sensors acquire signals based on contact-induced strain, it is often difficult to distinguish between forces originating from the human operator and those resulting from interactions with the external environment [6,7]. This significantly limits the accuracy of capturing actual human-applied forces.
With the continuous integration of interdisciplinary research, scholars have increasingly recognized the strong correlation between EMG signals, commonly used in the medical field to assess human motion, and muscle force output [8]. Early arm force estimation methods primarily relied on muscle dynamics models, such as the Hill-type model and its variants [9]. However, constructing these models requires precise acquisition of individual-specific anatomical parameters, including muscle and skeletal structures. As the dimensionality of force estimation increases, the complexity and computational burden of these models rise sharply, presenting significant challenges for practical deployment [10].
With the advancement of computational technologies, researchers have increasingly turned to machine learning algorithms to directly model EMG signals, enabling force estimation without depending on musculoskeletal structural parameters. Several EMG-to-force mapping approaches have been proposed, including Artificial Neural Networks (ANNs) [11], Long Short-Term Memory networks (LSTMs) [12], Principal Component Interaction (PCI) modeling [13], and Convolutional Neural Networks (CNNs) [14].
However, current machine learning models often emphasize unidirectional force estimation, which limits their effectiveness in complex industrial applications [15,16]. To enhance model generalizability, Su et al. proposed decomposing arm force output into three orthogonal directions in Cartesian space and modeling each direction independently using a Deep Convolutional Neural Network (DCNN) [17]. Building on this idea, Zhang et al. introduced a Bayesian approach that first identifies the dominant force direction and then applies parallel LSTM networks for direction-specific force estimation [18]. While these methods improve multidimensional force prediction, they overlook a key physiological reality: in real-world tasks, muscle groups typically operate in a coordinated, multi-directional fashion, resulting in complex interdependencies and mutual interference. As a result, modeling each direction separately may oversimplify the intricate dynamics of muscle coordination and limit the model’s ability to capture the underlying biomechanical interactions.
As demand grows for multidimensional dynamic arm force estimation, researchers have increasingly recognized the importance of arm posture in modeling. During human movement, joint angles reflect physiological factors such as muscle length variations and contraction types. These not only affect the amplitude dynamics of EMG signals but also modulate inter-muscle coordination, thereby significantly influencing the generalization ability of force estimation models [19,20]. To address this, Mobasser et al. proposed a Fast Orthogonal Search (FOS)-based elbow joint force estimation model that incorporates both EMG and joint angle information. The method demonstrated its effectiveness under isometric, isobaric, and low-load conditions [21].
Although incorporating joint angles improves model accuracy, increasing movement dimensionality and complexity introduces greater challenges for training and fitting. To address this, Xie et al. proposed a method that takes both joint angles and EMG signals as inputs and constructs a Deep Long Short-Term Memory (Deep-LSTM) network for dynamic arm force estimation [22]. This approach is considered one of the most effective techniques for improving dynamic force estimation accuracy. However, current methods lack explicit modeling of how posture affects muscle synergy. Instead, these approaches rely on the model to implicitly learn the complex nonlinear coupling between EMG features and joint configurations from data, which undoubtedly increases the learning burden and may limit generalization under multi-task conditions. Therefore, achieving accurate multidimensional dynamic arm force estimation remains a substantial challenge.
Beyond perception of human intent, another core issue in HRC lies in the design of effective collaboration control strategies [23]. To enhance the responsiveness and compliance of robotic systems during collaboration, mainstream approaches commonly utilize impedance or admittance control frameworks, which dynamically adjust control parameters to adapt to interaction forces [24]. For instance, Bednarczyk et al. proposed a variable impedance control method that integrates human-applied forces into the control model to optimize robot interaction performance [25]. Yao introduced an adaptive admittance control method using an admittance filter, which adjusts parameters based on estimated joint torques, thereby improving the applicability of rehabilitation robots in assisted motion [26].
Meanwhile, data-driven strategies have gradually been integrated into variable impedance frameworks due to advancements in machine learning, significantly enhancing the adaptability of control models. For example, impedance parameter optimization via iterative learning has been successfully applied in industrial tasks such as polishing and assembly [27,28]. Yang and Anand introduced reinforcement learning (RL) into controller design, enabling adaptive adjustment of impedance parameters through policy optimization, thus substantially improving compliant control performance [29,30]. However, despite notable success in improving flexibility and precision, these methods still struggle to meet real-world demands in highly dynamic and uncertain HRC environments, due to limitations of static models and single-strategy optimization [31].
In recent years, imitation learning has emerged as a promising approach for real-time HRC. By learning from human demonstrations, robots can participate in collaborative tasks more naturally [32,33]. Liao et al. restructured the Dynamic Movement Primitives (DMP) formulation using Riemannian metrics, incorporating position, orientation, and stiffness information to enable the transfer and reproduction of humanoid multi-space skills [34]. Additionally, traditional machine learning techniques such as Gaussian Mixture Models, Random Forests, and Hidden Markov Models have laid solid foundations for imitation learning [35,36]. However, since traditional imitation learning typically relies on predefined trajectories and is more suited for offline behavior cloning, it faces limitations in processing high-dimensional data and handling dynamic real-time interactions, which restricts its effectiveness in collaborative scenarios [37].
Notably, reinforcement learning has demonstrated outstanding performance even in the absence of precise dynamic models, making it well-suited for complex and frequently changing collaborative environments [38]. For example, Zhang proposed using RL to directly extract force-related information from EMG signals, enabling robot motion control without relying on traditional control models, and successfully validated this approach in a human–robot collaborative sawing task [39]. The organic integration of imitation learning and reinforcement learning is expected to become a promising direction for novel skill acquisition and the exploitation of advanced collaborative control models. Currently, imitation reinforcement learning has been predominantly applied in domains such as autonomous driving and robotic manipulation [32,40]. However, research specifically targeting human skill acquisition and the enhancement of human–robot collaborative control through this paradigm remains relatively limited.
To address the aforementioned challenges, an EMG-driven dynamic arm force estimation model and an imitation reinforcement learning-based collaborative control framework are proposed to enhance performance in human–robot collaborative assembly. The overall research framework is illustrated in Figure 1.
First, joint angle information and EMG signals collected from the human body are processed to extract features for arm force estimation. To account for the influence of joint angle variations on force estimation, a Temporal Graph Neural Network incorporating an Angle-Guided Attention mechanism (AGA-LSTM) is introduced to capture changes in inter-muscle synergies and estimate three-dimensional (3D) arm force. The estimated human arm force is then used to extract key indicators of human skill from demonstration data. Specifically, this skill refers to the ability to adjust speed in response to changes in external collaborative forces. Additionally, the estimated arm force also serves as an interface for human–robot interaction.
Based on the skill indicators extracted from human demonstrations, an expert reward function is constructed. A performance-optimized control model is then trained using an imitation reinforcement learning algorithm, enhanced with a fuzzy rule-based experience replay partitioning strategy. The proposed control framework adopts a human-in-the-loop design, converting human interaction forces directly into robot motion commands. By leveraging human perceptual and decision-making abilities, the robot is guided to collaborate effectively with the operator. In cases of trajectory deviations, the operator can intuitively correct the robot’s motion by adjusting the applied force. This approach avoids rigid trajectory replication, enabling more flexible and autonomous human–robot interaction.

2. Construction of an EMG Signals-Based Dynamic Arm Force Estimation Model

To meet the requirements for interaction force smoothness and accuracy in human–robot collaborative control, this section focuses on the acquisition and processing of arm motion signals and the construction of arm force estimation model.

2.1. Acquisition and Processing of Arm Motion Signals

2.1.1. Methods for Acquiring Arm Motion Signals

Figure 2 presents the human arm motion signal acquisition platform, which supplies training data for the EMG-based dynamic force estimation model. A goniometer is positioned on the outer side of the elbow to record one-dimensional elbow joint angle data. An IMU, attached to the inner side of the upper arm, captures the shoulder’s three-dimensional Euler angles. Six EMG electrodes are placed on key muscle groups, including the biceps/triceps brachii, anterior/posterior deltoids, pectoralis major, and infraspinatus. Arm-generated forces are simultaneously recorded using a six-axis force sensor.
During data collection, a human operator manually guides the robot by applying forces with varying intensities and directions, enabling free movement. At the same time, EMG intensity, joint angles, and force measurements are recorded. The operator is oriented toward the YZ-plane of the robot’s base frame, and the applied forces are decomposed into three orthogonal components (X, Y, and Z) relative to the robot’s coordinate system.

2.1.2. Feature Extraction from EMG Signals

It is necessary to extract features that correlate with the arm force from these time-varying signals. Given that the amplitude of the EMG signal is strongly correlated with arm force, a moving root mean square (RMS) filter is applied to the raw EMG signal to extract amplitude features. The processing is described by the following equation:
E RMS i = 1 N m = i N + 1 i E raw m 2
where, Eraw[m] denotes the m-th denoised EMG signal and ERMS[i] represents the i-th signal after processing by the RMS filter with a sliding window of size N.
As shown in Figure 3, even after RMS filtering, the feature signal still exhibits considerable fluctuations. Therefore, a simple discrete low-pass filter is applied to smooth the extracted features. The smoothing process can be described as follows:
E LPF n = ξ E RMS n + ( 1 ξ ) E LPF n 1
where ELPF[n] denotes the n-th output of the low-pass filter and ξ is the filter’s difference weight.
According to Ref. [41], the sliding window size N and the filter’s difference weight ξ are set to 64 and 0.00628, respectively, demonstrating a good balance between signal smoothness and responsiveness.

2.2. A 3D Arm Force Estimation Model Based on an Angle-Guided Attention Temporal Graph Neural Network (AGA-TGNN)

In EMG-based arm force estimation model, muscle activation patterns are significantly influenced by joint angles. The contribution of individual muscles varies with different joint angles; for instance, the biceps brachii exerts more force at smaller flexion angles, whereas its contribution diminishes during extension, with the triceps brachii taking a dominant role during extension. Moreover, muscle coordination patterns change with joint angle variations—alterations in shoulder angles, for example, can modify the synergistic behavior between the pectoralis major and the anterior deltoid. These angle-dependent dynamics make it challenging to accurately estimate arm force using traditional methods.
Existing approaches such as CNNs are effective at extracting local spatial features from EMG signals but fall short in capturing the global inter-muscle cooperative relationships. Although LSTM networks are well-suited for temporal modeling, they lack the capability to capture the spatial dependencies among muscles, making them less adaptable to the variations in muscle coordination under different joint angles. Therefore, a model that can simultaneously capture the topological structure of muscle groups and their temporal variations is needed to more accurately model muscle coordination patterns under varying joint angles.
Graph Neural Networks (GNNs) provide an effective approach to modeling the topological structure of EMG signals by employing an adjacency matrix to represent inter-muscle relationships. However, conventional GNNs typically rely on a fixed adjacency matrix, which fails to capture the dynamic changes in muscle coordination induced by variations in joint angles. To overcome this limitation, an Angle-Guided Attention mechanism is introduced, enabling the adaptive adjustment of the adjacency matrix based on joint angle information. This enhancement significantly improves the model’s generalization capability across diverse motion conditions.
To further strengthen the temporal dependency modeling of EMG signals, an LSTM is integrated following the GNN, forming a Temporal Graph Neural Network (TGNN). In this architecture, the GNN extracts spatial features related to muscle topology, while the LSTM models the time dependencies of the EMG signals. The combination of GNN and LSTM enables the model to capture both the spatial cooperation among muscles and the temporal patterns, thereby improving the accuracy of arm force estimation. Based on these considerations, the AGA-TGNN is proposed.
(a) Motivation of EMG Signals Modulated by Joint Angles
Human arm output force arises from the coordinated activation of multiple muscle groups, governed by both muscular dynamics and skeletal biomechanics. According to physiological coordination dynamics, muscle-generated tension is first transformed into joint torques through a moment matrix R(θ) and then mapped to the arm output force via the Jacobian matrix J(θ). The dynamic formulation of the estimated arm force Festm is given by the following:
F e s t m = J T θ R T θ F msc
where, Fmsc is the muscle force, and both R(θ) and J(θ) are functions of joint angles θ.
This formulation illustrates that the resulting arm output force is influenced not only by muscle activity (as measured by EMG) but also by the arm’s current posture. In this paper, the muscle force Fmsc is modeled using a neural network, leading to an updated force estimation expression:
F e s t m = J T θ R T θ f NN E LPF
Given the complexity of directly measuring muscle forces and modeling the analytical forms of R(θ) and J(θ), joint angle information is incorporated as a prior into the neural network to guide the learning process. This allows the model to learn the topological coordination pattern among muscle groups. Accordingly, the dynamic force estimation model consists of two submodules as follows:
F e s t m = f NN 1 E a d j = f NN 1 E LPF f NN 2 θ
where, Eadj denotes the EMG amplitudes modulated by joint angles; fNN1 represents a Temporal Graph Neural Network (TGNN) that estimates 3D arm force based on EMG signals; fNN2 is the Angle-Guided Attention (AGA) module that learns a posture-dependent adjacency matrix to guide the TGNN. By integrating joint angle information as a prior, this mechanism enables the network to dynamically capture inter-muscle coordination patterns under varying postures.
(b) Angle-Guided Attention Module
Traditional attention mechanisms, such as the self-attention mechanism in Transformers, adjust weights by computing the “importance” or “relevance” between elements. Inspired by this principle, the Angle-Guided Attention mechanism is developed to dynamically capture changing muscle coordination patterns. Specifically, a three-layer multilayer perceptron (MLP) is used to transform the joint angle information θ[t] into an adjacency matrix A′[t]:
A t = ReLU W 3 ReLU W 2 ReLU W 1 θ t + b 1 + b 2 + b 3
where, θ[t] is a four-dimensional vector comprising the joint angles of the elbow and the Euler angles of the shoulder; W1, W2, and W3 are the weight matrices of the MLP; b1, b2, and b3 are the bias vectors of the MLP.
Since the adjacency matrix must satisfy a row-normalization constraint (i.e., the sum of the connection probabilities for each node equals 1), A′[t] is normalized using the softmax function to obtain the final adjacency matrix A[t]:
A i , j t = exp A i , j ' t k = 1 N A exp A i , k ' t
where, Ai,j[t] represents the connection weight between EMG channel i and j at time t. In this study, six muscles were utilized for arm force estimation. Accordingly, to represent the interactions between all pairs of muscle groups, the adjacency matrix was designed with a dimension of 6 × 6.
(c) Temporal Graph Neural Network Model
The GNN employs a graph convolution operation based on the Angle-Guided Attention-derived adjacency matrix to extract spatial dependencies among EMG signals. The resulting temporal features are then fed into an LSTM network for sequential modeling. The LSTM’s hidden state is passed through a fully connected (FC) layer to obtain the final force estimation. This process is illustrated in the following network structure:
G t = σ A t H t W g h t = LSTM G t , h t 1 F pred t = W f h t + b f
where, H[t] represents the GNN input consisting of d1 time-series ELPF from six muscle groups (with d1 = 6 in this parper); Wg is the weight matrix of the GNN; G[t] is the GNN’s feature matrix at time t; h(t) denotes the LSTM’s hidden state at time t; Fpred[t] represents the estimated arm force at time t.
(d) Loss Function and Regularization
To minimize the estimation error, the mean squared error (MSE) between the true arm force Ftrue[t] and the estimated arm force Fpred[t] is used as the loss function:
L 1 = 1 N L t = 1 N L F true t F pred t 2
where, NL is the number of samples.
Additionally, to prevent overfitting and to ensure that the adjacency matrix A(t) does not undergo extreme variations, a regularization term based on the Frobenius norm is added:
L = 1 N L t = 1 N L F true t F pred t 2 + λ A t A t 1 F 2
where, λ is the regularization coefficient. The regularization parameter λ is selected from a commonly used range between 1 × 10−5 and 1 × 10−3. In this paper, preliminary experimental results show that λ = 1 × 10−4 achieve the best balance between training accuracy and generalization performance, yielding the lowest validation error and minimal overfitting.
The overall loss function thus combines the MSE and the regularization term, guiding the model to learn a more accurate and stable arm force estimation.

3. Mathematical Modeling of Human Skill Imitation and Learning Algorithm

3.1. Skill Criteria from Human Demonstrations

Most current imitation learning control models rely on fitting speed and force curves from human demonstrations to replicate movements. However, the application of this skill imitation approach is heavily limited by the demonstration tasks. In human–robot collaboration tasks, the uncertainty of movement increases the demand for skill generalization. To address this, this paper aims to extract key skill criteria from human demonstrations to guide the online learning and optimization of the control model, enabling it to more closely mimic human behavior.
(1)
Collaborative comfort
During collaboration, the human’s output force primarily guides the robot’s movement, while the robot handles most of the physical labor. For a given task, a smaller control force from the operator indicates stronger robot assistance, and greater comfort. Therefore, the concept of the equivalent output force is introduced as a measure of collaborative comfort.
In human demonstrations, the equivalent output force Fimt is defined as the ratio of the energy exerted by the leader to the total displacement. It is calculated as follows:
F imt = 1 3 d = 1 3 t = 1 N F d [ t ] Δ x d [ t ] t = 1 N Δ x d [ t ]
where, F d [ t ] and Δ x d [ t ] are the leader output force and the follower movement displacement at time t, respectively; the subscript d represents the direction of the follower movement.
In the human–robot collaboration scenario, the equivalent output force Fcnt is defined as the ratio of the energy exerted by the operator to the robot’s effective motion distance. Since the robot may occasionally move against the operator’s intended direction during the training of collaborative control models, such reverse motion should be excluded from the calculation of effective motion distance. Accordingly, Fcnt is expressed as follows:
F cnt = 1 3 d = 1 3 t = 1 T F d [ t ] Δ x d [ t ] t = 1 T sign ( F d [ t ] Δ x d [ t ] ) Δ x d [ t ]
where, T is the number of samples.
To evaluate actual behavior using the skill criteria from human demonstrations, according to the Ref. [42], the collaborative comfort RF in the expert reward function is defined as follows:
R F = 1 F cnt F imt F cnt + F imt
(2)
Collaborative smoothness
Collaborative smoothness refers to the continuity of a robot’s movements during task execution. It plays a critical role in human–robot interaction, particularly in scenarios that demand precision and coordinated actions. Smooth robotic motion ensures stable and sustained force assistance, enhances predictability, and improves both safety and controllability. For the robot, smoothness is typically characterized by speed, while for the human operator, it is reflected in the consistency of the applied control force.
In human demonstrations, the smoothness of both motion speed Vimt and control force Pimt can be quantified as follows:
V imt = d = 1 3 t = 1 N V d [ t ] V d [ t 1 ] 3 N P imt = d = 1 3 t = 1 N F d [ t ] F d [ t 1 ] 3 N
where, Vd[t] is the robot speed at time t.
In the human–robot collaboration scenario, the smoothness of both motion speed Vcnt and control force Pcnt can be quantified as follows:
V cnt = d = 1 3 t = 1 T V d [ t ] V d [ t 1 ] 3 T P cnt = d = 1 3 t = 1 T F d [ t ] F d [ t 1 ] 3 T
Thus, the collaborative smoothness RV in the expert reward function is defined as follows:
R V = 1 1 2 V cnt V imt V cnt + V imt 1 2 P cnt P imt P imt + P cnt

3.2. Expert Reward Function Construction

Based on the skill criteria defined in the previous section, the optimization problem for the human–robot collaborative control model is formulated as learning a decision-making policy that maps human arm forces to robot motion velocities, with the objective of maximizing both comfort and smoothness during collaboration.
P 1 : max v , F cnt R F , R V     s . t . C 1 :   - 0.02 V d t 0.02   C 2 : V d t F d [ t ] > 0
Here, C1 represents the constraint on the robot’s speed; C2 denotes the constraint on the control signal, ensuring that the robot’s motion aligns with the direction of the force applied by the human, thereby enhancing the naturalness of the control.
In the collaboration performance evaluation, collaborative comfort and smoothness are considered equally important, requiring the combined rewards to be maximized. An auxiliary penalty strategy is introduced to help the control system adapt more effectively to the operator’s movement habits. An auxiliary penalty value Ra is introduced when the direction of the guiding force is misaligned with the robot’s motion, which has been shown to facilitate the robot’s rapid adaptation to human motion tendencies. When the learned behavior aligns closely with human demonstrations, the combined reward for collaborative comfort and motion smoothness approaches a total value of two. Experimental evaluations demonstrated that values greater than two are effective in accelerating training; however, excessively large penalties are found to negatively impact long-term reward optimization. In this study, a value of Ra = two is selected to ensure that the reward function retains a clear distinction between positive and negative values after applying the penalty mechanism. The expert reward function of human skill imitation learning algorithm r is as follows:
r = R F + R V V d t F d [ t ] > 0 r = R F + R V R a V d t F d [ t ] 0

3.3. Human–Robot Collaborative Control Model Based on Imitation Reinforcement Learning Algorithm

3.3.1. Construction of Imitation Reinforcement Learning Model Based on Actor–Critic Algorithm

In imitation learning, particularly in scenarios with limited data or insufficient human demonstrations, traditional behavior cloning methods are prone to overfitting. This overfitting leads to learned strategies that become excessively reliant on the expert’s examples, resulting in poor generalization to new situations. The integration of reinforcement learning algorithms addresses this issue by allowing the model to not only learn from expert demonstrations but also improve the strategy through continuous interaction with the environment, thereby enhancing the generalization capability of the control strategy.
As shown in Figure 4, the Actor–Critic (AC) reinforcement learning algorithm is applied to human–robot collaborative control. The AC algorithm combines both policy optimization and value evaluation methods by assessing the current value function to optimize the policy function. The control model consists of AC networks and an experience replay pool.
During the human–robot collaboration, the human output force serves as a crucial interaction medium and plays an important role in determining the robot’s interaction speed. Therefore, the state space of the agent S[t] includes two key features: the current force Fd[t] and the force change ΔFd[t], i.e., S[t] = [Fd[t], ΔFd[t]]. The action space of the agent corresponds to the robot’s speed v[t]. Using the policy function π(v[t]|S[t]) learned through the Actor network, the agent determines Vd[t] based on the current force state S[t] and sends v[t] to the robot. The operators adjust their arm force according to the intended movement, enabling human-in-the-loop control.
The reward value r[t] is used to evaluate how well the operator’s motion intention is achieved, and the generated samples {S[t], v[t], r[t], and S[t − 1]} are stored in the experience replay pool for model training and updating. The specific structural framework of the imitation learning model based on the AC algorithm is as follows:
(1)
Actor network architecture
In the control model, both the Actor and Critic networks use artificial neural network structures. To ensure the exploration ability of the Actor network, the output of the Actor network consists of the mean μ and variance σ of the robot speed. The network structure is as follows:
h 1 a = f ReLU W 1 a F d t , Δ F d t + b 1 a h 2 a = f ReLU W 2 a h 1 a + b 2 a μ F d t , Δ F d t = W μ h 2 a + b μ σ F d t , Δ F d t = f softplus W σ h 2 a + b σ
where, h1a and h2a represent the outputs of the first and second hidden layers, respectively; b 1 a , b 2 a , b μ , and b σ are the bias vectors for each layer of the network; W 1 a , W 2 a , W μ , and W σ are the weight matrices for each layer of the network; f ReLU represents ReLU activation function; f softplus represents Softplus activation function.
The actual robot speed v[t] is determined by sampling from this distribution:
v t ~ N μ F d t , Δ F d t , σ F d t , Δ F d t
(2)
Critic network architecture
The Critic network’s output is the value function V F d t , Δ F d t , which evaluates the value of the robot speed v[t] under the current human output force state F d t , Δ F d t . The Critic network structure is as follows:
h 1 c = f ReLU W 1 a F d t , Δ F d t , v t + b 1 c h 2 c = f ReLU W 2 c h 1 c + b 2 c V F d t , Δ F d t = W 3 c h 2 c + b 3 c
where, h1c and h2c represent the outputs of the first and second hidden layers, respectively; b1c, b2c, and b3c are the bias vectors for each layer of the network; W1c, W2c, and W3c are the weight matrices or each layer of the network.

3.3.2. Fuzzy Rule-Based Experience Replay Partitioning Mechanism

Figure 5 illustrates a leader–follower collaborative assembly demonstration. During the human demonstration data collection, the follower completes peg-in-hole by responding to the guiding forces applied by the leader. To eliminate the influence of visual feedback and ensure the purity of physical interaction, the follower is blindfolded throughout the entire procedure. Moreover, the demonstration speed is limited to match the robot’s motion constraints, which in this study is set to no more than 0.02 m/s.
As shown in Figure 6, the curves of the arm force and speed in the Y-direction during human demonstrations are presented. It can be observed that the estimated arm-applied force and motion velocity exhibit a strong correlation during the collaborative assembly process. Additionally, the assembly task naturally decomposes into four distinct sub-motions: Acceleration (Accel), Rapid Transfer (RT), Deceleration (Decel), and Fine Adjustment (FA). In fact, the vast majority of collaborative tasks are composed of these four sub-motions.
Considering that key indicators, such as arm force stability and equivalent output force, differ significantly across these sub-motion phases, directly mixing all demonstration data into a single training stream would introduce substantial variance in reward estimation, thereby impairing convergence and overall learning stability. To mitigate this, a fuzzy rule-based experience replay partitioning strategy is proposed. This strategy classifies each sample into a corresponding sub-motion phase based on the magnitude of the arm force and its rate of change. Each motion phase maintains its own dedicated experience buffer and stores its phase-specific reward function values accordingly.
Although sub-motion types can be roughly identified by analyzing the magnitude and temporal trends of interaction forces, the inherent physiological uncertainty and the fuzzy nature of human force output make rigid classification unreliable. To address this challenge, the proposed fuzzy rule-based experience replay partitioning strategy incorporates a fuzzy-rule-based sub-motion inference mechanism to enhance adaptability and robustness under dynamic conditions.
Table 1 presents the fuzzy rules adopted in the experience replay partitioning method. Arm force and its rate of change are categorized into six levels based on direction (positive or negative) and magnitude (high, medium, or small), denoted as Positive High (PH), Positive Medium (PM), Positive Small (PS), Negative Small (NS), Negative Medium (NM), and Negative High (NH), respectively.
Both the arm force and its rate of change are fuzzified using triangular membership functions, as illustrated in Figure 7. Taking the positive-direction arm force (or the rate of change of arm force) as an example, its membership function is defined as follows:
λ A F d t = max min 1 , a F d t a , 0 0 F d t a F d t PS   max min F d t a , b F d t b a , 0 0 F d t b F d t PM   max min 1 , F d t a b a , 0 b F d t F d t PH
where, λ A denotes the membership function for force, with parameters a and b defining the boundaries of the membership levels. In the triangular membership function for positive arm force, a represents half of the preset maximum interaction force (or maximum rate of change), while b corresponds to the preset maximum interaction force (or maximum rate of change).
This configuration provides practical benefits: In real-world human–robot interaction scenarios, a considerable portion of movements falls within medium-force levels. A narrow classification margin may lead to misclassification of such transitional actions. By setting a to half of the maximum expected value, the model not only aligns better with realistic force distributions but also broadens the central membership region. This enhances the system’s robustness to boundary fuzziness and improves its ability to discriminate intermediate-force motions under dynamic conditions.
The arm force and its changes are fuzzified according to the fuzzy rule table using the maximum membership method to determine the current sub-motion type. The process is expressed as follows:
z p i = A & B p i λ A F d [ t ] λ B Δ F d [ t ] p i * = arg max p i z p 1 , z p 2 , z p 3 , z p 4
where, z p i represents the membership value of each sub-motion type; λ B represents the membership function of the force change; p i * represents the current sub-motion type corresponds to the state space.
This fuzzy-rule-based sub-motion inference mechanism enables a soft partitioning of motion types. Based on the identified sub-motion types, actions, behaviors, and reward values in imitation reinforcement learning are stored in separate partitions. By incorporating partitioned experience replay, the method ensures balanced sampling and training across the four sub-motion types, thereby significantly improving learning stability and adaptability to complex motion dynamics.

3.3.3. Imitation Reinforcement Learning Model Update

The value function V(S[t]) represents the expected return at state S[t], but it alone cannot fully assess policy quality. To address this, the advantage function is introduced as a metric to quantify the relative the excellence of an action in a given state. For computational simplicity, the Temporal Difference (TD) error from the Critic is used as an approximation of the advantage function. The advantage function A(S[t], v[t]) is defined as follows [43]:
A S t , v t = Q S t , v t V S t
where, Q(S[t], v[t]) is the state-action value function, representing the expected return after taking action v[t] at state S[t], which includes both the current action and all future rewards.
According to the Bellman equation, the state-action value function Q(S[t], v[t]) and the action reward value V(S[t]) are related as follows:
Q S t , v t = r t + γ V S t + 1
where, γ is the discount factor, used to balance the relative importance of current rewards and future rewards. Experimental results show that lower discount factors (e.g., γ = 0.96) lead to faster convergence but result in suboptimal task performance, suggesting an overemphasis on immediate rewards. In contrast, higher discount factors (e.g., γ > 0.98) slow down the learning process without yielding substantial improvements in final performance. A value of γ = 0.97 is found to offer the best trade-off, enabling efficient learning while preserving strong generalization to dynamic task variations.
Thus, combining the Equations (24) and (25), A(S[t], v[t]) can be expressed as follows:
A S t , v t = r t + γ V S t + 1 V S t
To balance the optimization of various sub-motion types and prevent dominance by any single sub-motion mode, the experience replay pools from different regions are sampled evenly. The sample size for each pool is T. In this case, the gradient update for the Actor-Critic network can be expressed as follows:
θ c L θ c = 1 4 T d 1 2 A S t , v t θ c V S t d θ a J θ a = 1 4 T d θ a log π S t | v t A S t , v t d
where, θ c L θ c represents the gradient of the loss function of the Critic network with respect to the parameters θ c ; θ a J θ a represents the gradient of the objective function of the Actor network with respect to the parameters θ a .

4. Experimental Analysis and Discussion

4.1. Experimental Setup

Figure 8 illustrates the schematic of the human–robot collaborative assembly experimental platform. The platform is built using a GSK 8 kg industrial robot Guangzhou GSK Technology Co., Ltd. Guangzhou, China). EMG signals and elbow joint angles are collected using sensors and goniometers from Biometrics Ltd. (Cwmfelinfach, UK), while the shoulder joint Euler angles are acquired through an inertial measurement unit (IMU) from Hipnuc Electronics Technology Co., Ltd. (Beijing, China). A six-axis force/torque sensor from Me-Systeme GmbH (Stuttgart, Germany) is mounted at the robot’s end-effector; however, it is not involved in the control loop. Instead, it is used to collect ground-truth force data for evaluating the performance of the force estimation model.
During the experiment, the robot adjusts its position by following the human arm force, enabling collaborative assembly between the human and the robot. The initial offset between the peg and the hole is set to Lx = 300, Ly = 200 mm, and Lz = 40 mm along the X, Y, and Z axes, respectively. The assembly peg has a diameter of 9.98 mm, while the hole diameter is 11.03 mm.

4.2. Ablation Study of the Proposed Arm Force Estimation Model

To validate the effectiveness of the AGA module, two ablation models are constructed for comparison:
Fixed-Adjacency Matrix TGNN: This model constructs inter-muscle connections using a static adjacency matrix. The model uses the same EMG and joint angle inputs as the full AGA-TGNN, enabling a fair comparison and highlighting the limitations of Fixed-Adjacency Matrix structures under dynamic tasks.
LSTM-only (TGNN without GNN): In this version, the GNN is removed entirely, and only a LSTM is used to process EMG sequences. Joint angle inputs and spatial topology information are excluded, allowing assessment of the importance of spatial topology and joint angle information in modeling muscle coordination.
The results are summarized in Figure 9 and Table 2. The AGA-TGNN model achieves superior performance across all key metrics, including median, mean, and maximum estimation errors. The performance gains are primarily attributed to the following factors:
The Fixed-Adjacency Matrix TGNN lacks adaptability to joint angles (arm posture) variations, as its connectivity structure remains constant regardless of joint configuration. This limits its ability to capture joint angles-dependent muscle coordination patterns.
The LSTM-only model fails to incorporate spatial topology and joint angles information, resulting in poor representation of the nonlinear coupling between muscle activations and arm output force.
In conclusion, the Angle-Guided Attention module effectively integrates joint angle information to dynamically adjust muscle interaction modeling, significantly improving the accuracy of force estimation in dynamic scenarios.

4.3. Validation of Arm Force Estimation Model

This study selects several representative models that have demonstrated strong performance in force estimation tasks. The specific configurations are as follows:
Comparison Model 1 (FOS Model): The FOS model constructs an arm force fitting model based on joint angles and the amplitude of EMG signals using an orthogonal search strategy. It represents a traditional analytical approach to force estimation.
Comparison Model 2 (LSTM Model): This model employs LSTM networks to model the temporal sequences of EMG signals. As a widely adopted method for force estimation, it exhibits strong capabilities in learning temporal dependencies.
Comparison Model 3 (Deep-LSTM Model): Based on Deep-LSTM, this model performs feature extraction and modeling of EMG signals. With its powerful spatial pattern recognition and nonlinear modeling capabilities, it is one of the most competitive end-to-end force estimation approaches to date.
The experimental results of the three comparison models are shown in Figure 10, with the corresponding error distribution box plots presented in Figure 11. The statistical performance metrics are listed in Table 3.
As observed from Table 3, Figure 10 and Figure 11, the proposed AGA-TGNN model outperforms existing methods in terms of median error, maximum error, and mean error across all three force components. Compared with the Deep-LSTM model, the proposed model achieves reductions in mean estimation error of 10.38%, 8.33%, and 11.20% in the X, Y, and Z directions, respectively, demonstrating a significant performance advantage.
The performance differences can be attributed to the following factors:
First, FOS is a traditional model based on analytical expressions. While it can rapidly construct predictive functions in low-dimensional spaces, it lacks sufficient expressive power to capture the highly nonlinear relationship between EMG signals and arm forces. Furthermore, its fitting process is sensitive to signal fluctuations and lacks the ability to model dynamic features, resulting in unstable estimation performance. As the number of input muscles increases, the computational complexity of FOS grows exponentially without a corresponding improvement in fitting accuracy, limiting its performance in multi-dimensional dynamic scenarios.
Second, while LSTM and Deep-LSTM are both deep learning approaches with strong temporal modeling and feature extraction capabilities, respectively, neither considers the dynamic variation in inter-muscle coordination arising from changing joint states. Specifically, both models directly couple EMG and angle signals without explicitly accounting for the underlying physiological and physical interaction mechanisms. By completely relying on neural networks to decouple the relationship between EMG signals and joint angles, they fail to capture the complex, dynamic interactions, leading to insufficient generalization ability.
In contrast, the proposed method introduces an Angle-Guided Attention mechanism, which dynamically adjusts the adjacency matrix of the GNN based on joint angles. This approach effectively decouples the dynamic relationships between muscle groups and joint angles, enhancing the physiological plausibility of the input signal structure. By improving the model’s sensitivity to key muscle coordination patterns, this mechanism significantly enhances the robustness and generalization ability of the network, enabling stable and accurate force estimation across varying motion conditions.

4.4. Training of the Imitation Reinforcement Learning Model

As illustrated in Figure 4, the imitation reinforcement learning framework employed in this work comprises an Actor–Critic (AC) network and an experience replay pool. Based on the reward mechanism defined in Equation (17), the model is trained online with the goal of directly translating human arm forces into robot control commands. This approach aims to overcome the parameter tuning challenges inherent in conventional impedance control strategies.
However, as the initial imitation reinforcement learning model is untrained, it cannot immediately perform collaborative assembly tasks. To ensure the safety of both the experimental equipment and human participants, the training process is divided into two phases: pre-training and task training.
In the pre-training phase, a kinesthetic teaching setup is adopted, where the human operator physically guides the robot without engaging in actual assembly. It is assumed that human motor adaptation skills in response to external collaborative forces are generally similar across different directions. Therefore, to reduce training complexity, motion coordination is learned along a single axis, the X-direction, as a representative case. The human operator repetitively moves the robot along this axis to help the model acquire basic motion alignment and compliance characteristics.
Following pre-training, the imitation reinforcement learning model gains an initial ability to coordinate with human motion—namely, aligning the robot’s movement direction with the direction of human-applied forces while maintaining a degree of compliance. The model then enters the task training phase, which takes place in the collaborative assembly scenario shown in Figure 8. To prevent excessive contact forces during training, a safety mechanism is implemented: if the force between the peg and assembly hole exceeds 50 N, the current cycle is terminated, and the robot returns to its initial position before restarting. This task training stage aims to further improve motion smoothness and collaborative comfort by optimizing the policy with respect to the expert-defined reward.
During training, the robot operates on a 50 ms control cycle. In each cycle, the system performs arm force estimation, sub-motion classification, policy update via imitation reinforcement learning, and robot control command generation.
Figure 12a shows the reward curve during the pre-training stage. It can be observed that, under the guidance of the penalty mechanism, the robot achieves alignment between its movement direction and the human’s applied force direction after approximately 2000 training iterations, demonstrating good following behavior. However, due to the limited diversity of movement in long-term reciprocating training, this phase tends to result in high robot velocities even in response to small human forces. Therefore, once the robot has acquired basic following ability, further optimization through task training becomes necessary.
In Figure 12b, the blue curve represents the total reward for each task cycle. Significant fluctuations in the early training stages suggest that the control model at that time did not fully conform to the expert demonstration trajectories. The red curve shows a fitted trend for the task rewards, indicating an overall upward trend with fluctuations, eventually converging to a stable value. After multiple rounds of training, the model’s performance surpasses that of the demonstration (the reward value of the expert demonstration is 2). This outcome may be attributed to the challenges of collecting human data, which is often limited by the degree of coordination and synergy between the leader and follower.

4.5. Comparative Analysis of Control Models

To evaluate the collaboration performance of the proposed imitation reinforcement learning-based control strategy, a comparative study is conducted against two state-of-the-art control approaches: the Adaptive Admittance Control (ADC) model [18] and the Gaussian Process Regression-based Imitation Learning (GPRIL) model [26].
To ensure a fair comparison within the experimental context of this study, necessary adaptations are made to each of the comparison models.
For the ADC model, the algorithm proposed in Ref. [21] is considered, which employs joint stiffness estimation to adaptively tune PD parameters. This approach is particularly effective in addressing the trade-off between precise trajectory tracking and compliant assistance. However, since this study does not involve joint stiffness estimation, it is substituted with the estimated arm output force. Under this modification, the PD parameters are adjusted as follows:
k d = 0.2 k p k p = k p max k p min F d F min F max F min + k p min
where, kd is the derivative term parameter, kp is the proportional term parameter, k p max and k p min indicates the maximum and minimum value of the parameter kp, and F max and F min are the maximum and minimum arm output forces from the operator during the experiment.
Regarding the GPRIL model, the multi-model Gaussian Process Regression technique, as proposed in the Ref. [26], is employed to model the relationship between the leader’s force and the follower’s speed during demonstrations.
To further evaluate the effectiveness of the proposed fuzzy experience replay mechanism, an ablation model, used as an additional comparison model, is constructed by removing this component from the original imitation reinforcement learning framework.
In this study, each control model is evaluated through 30 repeated experiments to assess the collaborative comfort, smoothness, task completion speed, and accuracy under different control methods. The performance statistics are summarized in Table 4, while Figure 13 illustrates the operator’s output force and the robot’s movement speed across the three control models. The methods used to calculate collaborative comfort and smoothness are detailed in Equations (13) and (16). The evaluation criteria for accuracy and task completion speed are outlined as follows:
(a) Assembly Accuracy
Assembly error is a commonly used metric to evaluate collaboration precision. After the robot reaches the target position, the deviation between the robot’s final position and the desired position in the XY plane is used to assess the accuracy of the collaborative assembly. The assembly deviation can be expressed as follows:
e = e x 2 + e y 2
where, e x and e y denote the assembly errors in the X and Y directions, respectively.
(b) Success Rate
The success rate of assembly serves as a direct indicator of collaboration effectiveness. To avoid collisions during the process, a safety threshold is set: if the contact force between the peg and hole surpasses 50 N, the robot stops instantly, and the attempt is deemed unsuccessful. Likewise, if the assembly duration exceeds 30 s, the task is also classified as a failure.
(c) Efficiency
The time from task initiation to the moment the assembly peg is inserted more than 10 mm into the assembly hole is defined as the assembly time. In collaborative scenarios, human operators typically prefer shorter assembly times to improve task efficiency.
As shown in Table 4, the GPRIL model achieves collaborative comfort and smoothness values close to one. However, due to its limited learning capacity, it struggles to surpass expert-level performance.
In the ADC model, PD parameters are adjusted proportionally to the output force: higher forces lead to increased PD gains, while lower forces reduce them. This adjustment enables the robot to perform precise control when the operator exerts low force and executes rapid movements under high force. However, auxiliary motions during collaborative tasks often require coordination across three dimensions, and not all directions are suitable for high-speed execution. As a result, frequent acceleration and deceleration occur, as illustrated by the orange and blue curves in Figure 13f.
As shown in Figure 13, the robot’s motion under the proposed imitation reinforcement learning-based control model is noticeably smoother than that of the comparison models. As shown in Table 4, the proposed method achieves comprehensive improvements over the ADC model across all evaluation metrics. Specifically, collaborative comfort increases by 2.75%, and collaborative smoothness improves by 8.42%, reflecting enhanced physical interaction quality. In terms of task efficiency, the assembly time is reduced by 1.63%, and the assembly error decreases by 17.65%, indicating higher precision. Moreover, the assembly success rate improves by 7.2%.
Moreover, as shown in Table 4, the removal of the fuzzy experience replay mechanism resulted in significant performance degradation. Without sub-motion-based categorization, experiences from different motion phases interfered with each other during training, leading to decreased learning efficiency, reduced task execution accuracy, and visibly diminished coordination between human and robot during collaboration. These findings underscore the critical role of fuzzy experience replay in mitigating motion interference, promoting balanced learning across diverse sub-motion patterns, and ultimately enhancing the stability and responsiveness of human–robot collaboration under dynamic conditions.
To further validate the robustness of the proposed method across different individuals, two healthy adult participants are recruited to perform both the calibration of the arm force estimation model and the human–robot collaborative assembly tasks. The corresponding performance metrics are summarized in Table 5. The experimental results indicate that both participants successfully complete the assembly tasks with high accuracy, which further confirms the effectiveness of the proposed method.
Notably, inter-subject variability in EMG signals remains a significant challenge. The proposed dynamic force estimation model is designed as a learning framework rather than a universal solution. Different individuals require personalized calibration following the outlined training procedure. Applying a model trained on one subject directly to another leads to a marked decline in estimation accuracy and overall collaborative performance. Future work will explore transfer learning techniques with the aim of reducing the training time required to adapt force estimation models across different users.

5. Conclusions

Inspired by the remarkable human ability to coordinate force and speed in collaborative tasks, this study proposes an innovative framework addressing two key challenges: three-dimensional arm force estimation and human-robot collaborative control.
For force estimation, an AGA-TGNN algorithm is designed, which significantly enhances the model’s capacity to capture dynamic interactions between muscle coordination patterns and joint states. Experimental results demonstrate that the proposed model outperforms representative comparison methods (FOS, LSTM, and Deep-LSTM) across multiple metrics, including median error, mean error, and maximum error. Specifically, compared to the Deep-LSTM model, estimation errors in the three directions are reduced by 10.38%, 8.33%, and 11.20%, respectively.
In the domain of collaborative control, to overcome the limitations of traditional methods in compliance and generalization, an expert reward mechanism is introduced alongside a fuzzy rule-based experience replay partitioning strategy. A novel imitation reinforcement learning algorithm is developed, which requires no explicit control model and exhibits strong adaptability. This approach enables the robot to respond flexibly to operator motion intentions by perceiving arm forces in real time, thereby enhancing the comfort and smoothness of human–robot collaboration.
The proposed method is validated in a human–robot collaborative assembly scenario. Experimental results demonstrate that, compared with the representative GPRIL and ADC models, the proposed approach achieves substantial improvements across key performance metrics, including collaborative comfort, motion stability, task efficiency, and assembly accuracy. Specifically, compared to the ADC model, collaborative comfort increases by 2.75%, collaborative smoothness improves by 8.42%, assembly time is reduced by 1.63%, assembly error decreases by 17.65%, and the assembly success rate improves by 7.2%. These findings support the feasibility of the proposed approach in practical human–robot collaboration.
The current arm force estimation model only predicts three-dimensional Cartesian forces and does not account for torque components. As a result, the human–robot collaborative assembly system lacks the capability to perform posture adjustments, which limits its effectiveness in tasks requiring precise rotational control. To address this, future work will explore the correlation between EMG signals and arm torques, with the goal of enabling posture-aware collaboration.

Author Contributions

Conceptualization, H.S. and Y.Z.; methodology, H.S. and N.T.; software, H.S. and N.T.; validation, Y.Z. and Z.L.; formal analysis, H.S. and Z.L.; investigation, Y.Z.; writing—original draft preparation, H.S. and Y.Z.; writing—review and editing, H.S.; visualization, N.T. and Z.L.; supervision, H.S.; funding acquisition, H.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Postdoctoral Fellowship Program of CPSF under Grant Number GZC20240752.

Data Availability Statement

Data used in this work can be requested by contacting the correspondence author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wu, X.Y.; Tian, R.Y.; Lei, Y.C.; Gao, H.L.; Fang, Y.J. Real-Time Space Trajectory Judgment for Industrial Robots in Welding Tasks. Machines 2024, 12, 360. [Google Scholar] [CrossRef]
  2. Cao, Y.; Zhou, Q.; Yuan, W.; Ye, Q.; Popa, D.; Zhang, Y.M. Human-robot collaborative assembly and welding: A review and analysis of the state of the art. J. Manuf. Process. 2024, 131, 1388–1403. [Google Scholar] [CrossRef]
  3. Stancic, I.; Music, J.; Grujic, T. Gesture recognition system for real-time mobile robot control based on inertial sensors and motion strings. Eng. Appl. Artif. Intell. 2017, 66, 33–48. [Google Scholar] [CrossRef]
  4. Subramanian, K.; Thomas, L.; Sahin, M.; Sahin, F. Supporting Human-Robot Interaction in Manufacturing with Augmented Reality and Effective Human-Computer Interaction: A Review and Framework. Machines 2024, 12, 706. [Google Scholar] [CrossRef]
  5. Hu, L.Y.; Zhai, D.H.; Yu, D.D.; Xia, Y.Q. A Hybrid Framework Based on Bio-Signal and Built-in Force Sensor for Human-Robot Active Co-Carrying. IEEE Trans. Autom. Sci. Eng. 2025, 22, 3553–3566. [Google Scholar] [CrossRef]
  6. Ding, R.Q.; Mu, X.S.; Cheng, M.; Xu, B.; Li, G. Terminal force soft sensing of hydraulic manipulator based on the parameter identification. Measurement 2022, 200, 111551. [Google Scholar] [CrossRef]
  7. Khoshdel, V.; Akbarzadeh, A. An optimized artificial neural network for human-force estimation: Consequences for rehabilitation robotics. Ind. Robot. 2018, 45, 416–423. [Google Scholar] [CrossRef]
  8. Xiong, D.Z.; Zhang, D.H.; Zhao, X.G.; Zhao, Y.W. Deep Learning for EMG-based Human-Machine Interaction: A Review. IEEE CAA J. Autom. Sin. 2021, 8, 512–533. [Google Scholar] [CrossRef]
  9. Fleischer, C.; Hommel, G. A human-exoskeleton interface utilizing electromyography. IEEE Trans. Robot. 2008, 24, 872–882. [Google Scholar] [CrossRef]
  10. Artemiadis, P.K.; Kyriakopoulos, K.J. An EMG-Based Robot Control Scheme Robust to Time-Varying EMG Signal Features. IEEE Trans. Inf. Technol. Biomed. 2010, 14, 582–588. [Google Scholar] [CrossRef]
  11. Jali, M.H.; Izzuddin, T.A.; Bohari, Z.H.; Sarkawi, H.; Sulaima, M.F.; Baharom, M.F.; Bukhari, W.M. Joint Torque Estimation Model of sEMG Signal for Arm Rehabilitation Device Using Artificial Neural Network Techniques. In Proceedings of the 1st International Conference on Communication and Computer Engineering (ICOCOE), Malacca, Malaysia, 20–21 May 2014; pp. 671–682. [Google Scholar]
  12. Zhang, L.B.; Soselia, D.; Wang, R.L.; Gutierrez-Farewik, E.M. Estimation of Joint Torque by EMG-Driven Neuromusculoskeletal Models and LSTM Networks. IEEE Trans. Neural Syst. Rehabil. Eng. 2023, 31, 3722–3731. [Google Scholar] [CrossRef] [PubMed]
  13. Hashemi, J.; Morin, E.; Mousavi, P.; Hashtrudi-Zaad, K. Enhanced Dynamic EMG-Force Estimation through Calibration and PCI Modeling. IEEE Trans. Neural Syst. Rehabil. Eng. 2015, 23, 41–50. [Google Scholar] [CrossRef] [PubMed]
  14. Wahid, A.; Ullah, K.; Ullah, S.I.; Amin, M.; Almutairi, S.; Abohashrh, M. sEMG-Based Upper Limb Elbow Force Estimation Using CNN, CNN-LSTM, and CNN-GRU Models. IEEE Access 2024, 12, 128979–128991. [Google Scholar] [CrossRef]
  15. Zhang, T.; Chu, H.B.; Zou, Y.B. An Online Human Dynamic Arm Strength Perception Method Based on Surface Electromyography Signals for Human-Robot Collaboration. IEEE Trans. Instrum. Meas. 2023, 72, 7507714. [Google Scholar] [CrossRef]
  16. DelPreto, J.; Rus, D. Sharing the Load: Human-Robot Team Lifting Using Muscle Activity. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 7906–7912. [Google Scholar]
  17. Su, H.; Qi, W.; Li, Z.; Chen, Z.; Ferrigno, G.; De Momi, E. Deep Neural Network Approach in EMG-Based Force Estimation for Human–Robot Interaction. IEEE Trans. Artif. Intell. 2021, 2, 404–412. [Google Scholar] [CrossRef]
  18. Zhang, T.; Sun, H.L.; Zou, Y.B.; Chu, H.B. An electromyography signals-based human-robot collaboration method for human skill learning and imitation. J. Manuf. Syst. 2022, 64, 330–343. [Google Scholar] [CrossRef]
  19. Wang, J.Q.; Cao, D.G.; Liang, G.J.; Wu, Y.Q. A Surface Electromyography-Driven State-Space Model for Joint Angle Estimation. IEEE Sens. J. 2025, 25, 3048–3060. [Google Scholar] [CrossRef]
  20. Wang, Z.; Chen, C.; Chen, H.; Zhou, Y.; Wang, X.; Wu, X. Dual Transformer Network for Predicting Joint Angles and Torques from Multi-Channel EMG Signals in the Lower Limbs. IEEE J. Biomed. Health Inform. 2025, in press. [Google Scholar] [CrossRef]
  21. Mountjoy, K.; Morin, E.; Hashtrudi-Zaad, K. Use of the Fast Orthogonal Search Method to Estimate Optimal Joint Angle For Upper Limb Hill-Muscle Models. IEEE Trans. Biomed. Eng. 2010, 57, 790–798. [Google Scholar] [CrossRef]
  22. Xie, C.L.; Li, Z.J.; Song, R. Deep-LSTM-Based HumanRobot Collaboration Control with Static-Dynamic Calibration. IEEE Trans. Ind. Electron. 2024, 71, 16165–16175. [Google Scholar] [CrossRef]
  23. Xie, C.L.; Yang, Q.Q.; Huang, Y.; Su, S.V.; Xu, T.; Song, R. A Hybrid Arm-Hand Rehabilitation Robot with EMG-Based Admittance Controller. IEEE Trans. Biomed. Circuits Syst. 2021, 15, 1332–1342. [Google Scholar] [CrossRef] [PubMed]
  24. Zahedi, F.; Lee, H. Biomechanics-Based User-Adaptive Variable Impedance Control for Enhanced Physical Human-Robot Interaction Using Bayesian Optimization. Adv. Intell. Syst. 2024, 7, 2400760. [Google Scholar] [CrossRef]
  25. Bednarczyk, M.; Omran, H.; Bayle, B. EMG-Based Variable Impedance Control with Passivity Guarantees for Collaborative Robotics. IEEE Robot. Autom. Lett. 2022, 7, 4307–4312. [Google Scholar] [CrossRef]
  26. Yao, S.W.; Zhuang, Y.; Li, Z.J.; Song, R. Adaptive Admittance Control for an Ankle Exoskeleton Using an EMG-Driven Musculoskeletal Model. Front. Neurorobotics 2018, 12, 16. [Google Scholar] [CrossRef]
  27. Chen, C.J.; Zhang, C.X.; Pan, Y. Active compliance control of robot peg-in-hole assembly based on combined reinforcement learning. Appl. Intell. 2023, 53, 30677–30690. [Google Scholar] [CrossRef]
  28. Ding, Y.F.; Zhao, J.C.; Min, X.P. Impedance control and parameter optimization of surface polishing robot based on reinforcement learning. Proc. Inst. Mech. Eng. Part B J. Eng. Manuf. 2023, 237, 216–228. [Google Scholar] [CrossRef]
  29. Yang, Y.L.; Ding, Z.H.; Wang, R.; Modares, H.; Wunsch, D.C. Data-Driven Human-Robot Interaction Without Velocity Measurement Using Off-Policy Reinforcement Learning. IEEE CAA J. Autom. Sin. 2022, 9, 47–63. [Google Scholar] [CrossRef]
  30. Anand, A.S.; Gravdahl, J.T.; Abu-Dakka, F.J. Model-based variable impedance learning control for robotic manipulation. Robot. Auton. Syst. 2023, 170, 104531. [Google Scholar] [CrossRef]
  31. Al-Yacoub, A.; Zhao, Y.C.; Eaton, W.; Goh, Y.M.; Lohse, N. Improving human robot collaboration through Force/Torque based learning for object manipulation. Robot. Comput. Integr. Manuf. 2021, 69, 102111. [Google Scholar] [CrossRef]
  32. Hua, J.; Zeng, L.C.; Li, G.F.; Ju, Z.J. Learning for a Robot: Deep Reinforcement Learning, Imitation Learning, Transfer Learning. Sensors 2021, 21, 1278. [Google Scholar] [CrossRef]
  33. Yang, L.; Ariffin, M.Z.; Lou, B.C.; Lv, C.; Campolo, D. A Planning Framework for Robotic Insertion Tasks via Hydroelastic Contact Model. Machines 2023, 11, 714. [Google Scholar] [CrossRef]
  34. Liao, Z.W.; Jiang, G.D.; Zhao, F.; Wu, Y.Q.; Yue, Y.; Mei, X.S. Dynamic Skill Learning From Human Demonstration Based on the Human Arm Stiffness Estimation Model and Riemannian DMP. IEEE ASME Trans. Mechatron. 2023, 28, 1149–1160. [Google Scholar] [CrossRef]
  35. Duque, D.A.; Prieto, F.A.; Hoyos, J.G. Trajectory generation for robotic assembly operations using learning by demonstration. Robot. Comput. Integr. Manuf. 2019, 57, 292–302. [Google Scholar] [CrossRef]
  36. Kim, S.; Haschke, R.; Ritter, H. Gaussian Mixture Model for 3-DoF orientations. Robot. Auton. Syst. 2017, 87, 28–37. [Google Scholar] [CrossRef]
  37. Xing, X.Y.; Maqsood, K.; Zeng, C.; Yang, C.G.; Yuan, S.; Li, Y.N. Dynamic Motion Primitives-based Trajectory Learning for Physical Human-Robot Interaction Force Control. IEEE Trans. Ind. Inform. 2024, 20, 1675–1686. [Google Scholar] [CrossRef]
  38. Wang, W.T.; Chen, Y.; Li, R.; Jia, Y.Y. Learning and Comfort in Human-Robot Interaction: A Review. Appl. Sci. 2019, 9, 5152. [Google Scholar] [CrossRef]
  39. Zhang, T.; Sun, H.L.; Zou, Y.B. An electromyography signals-based human-robot collaboration system for human motion intention recognition and realization. Comput. Integr. Manuf. 2022, 77, 102359. [Google Scholar] [CrossRef]
  40. Zou, Q.J.; Xiong, K.; Fang, Q.; Jiang, B.H. Deep imitation reinforcement learning for self-driving by vision. CAAI Trans. Intell. Technol. 2021, 6, 493–503. [Google Scholar] [CrossRef]
  41. Yu, H.J.; Lee, A.Y.; Choi, Y. Human elbow joint angle estimation using electromyogram signal processing. IET Signal Process. 2011, 5, 767–775. [Google Scholar] [CrossRef]
  42. Shahbazi, M.; Atashzar, S.F.; Tavakoli, M.; Patel, R. V Robotics-Assisted Mirror Rehabilitation Therapy: A Therapist-in-the-Loop Assist-as-Needed Architecture. IEEE ASME Trans. Mechatron. 2016, 21, 1954–1965. [Google Scholar] [CrossRef]
  43. Lin, Y.; Xie, A.T.; Liu, X. Autonomous Vehicle Decision and Control through Reinforcement Learning with Traffic Flow Randomization. Machines 2024, 12, 264. [Google Scholar] [CrossRef]
Figure 1. Flowchart of human–robot collaborative assembly based on human skill imitation and learning.
Figure 1. Flowchart of human–robot collaborative assembly based on human skill imitation and learning.
Machines 13 00431 g001
Figure 2. The schematic diagram of the data acquisition platform for human arm motion signals.
Figure 2. The schematic diagram of the data acquisition platform for human arm motion signals.
Machines 13 00431 g002
Figure 3. Amplitude curve of the extracted features from the EMG signals.
Figure 3. Amplitude curve of the extracted features from the EMG signals.
Machines 13 00431 g003
Figure 4. Structure of imitation reinforcement learning model based on Actor–Critic algorithm.
Figure 4. Structure of imitation reinforcement learning model based on Actor–Critic algorithm.
Machines 13 00431 g004
Figure 5. Schematic of leader–follower collaborative assembly demonstration.
Figure 5. Schematic of leader–follower collaborative assembly demonstration.
Machines 13 00431 g005
Figure 6. The curves of the arm force and speed in the Y-direction during leader–follower demonstration. (a) The curves of the arm force; (b) The curves of the speed. In the figure, T1 represents the acceleration phase, T2 represents the rapid transfer phase, T3 represents the deceleration phase, and T4 represents the fine adjustment phase.
Figure 6. The curves of the arm force and speed in the Y-direction during leader–follower demonstration. (a) The curves of the arm force; (b) The curves of the speed. In the figure, T1 represents the acceleration phase, T2 represents the rapid transfer phase, T3 represents the deceleration phase, and T4 represents the fine adjustment phase.
Machines 13 00431 g006
Figure 7. Illustration of triangular membership functions.
Figure 7. Illustration of triangular membership functions.
Machines 13 00431 g007
Figure 8. Schematic of the human–robot collaborative assembly setup.
Figure 8. Schematic of the human–robot collaborative assembly setup.
Machines 13 00431 g008
Figure 9. Ablation results of the proposed arm force estimation model. (a) AGA-LSTM model; (b) Fixed-Adjacency Matrix TGNN model; (c) LSTM-only model; (d) Boxplot of estimation errors.
Figure 9. Ablation results of the proposed arm force estimation model. (a) AGA-LSTM model; (b) Fixed-Adjacency Matrix TGNN model; (c) LSTM-only model; (d) Boxplot of estimation errors.
Machines 13 00431 g009
Figure 10. Estimated arm force curves generated by various models. (a) AGA-LSTM model; (b) FOS model; (c) LSTM model; (d) Deep-LSTM model.
Figure 10. Estimated arm force curves generated by various models. (a) AGA-LSTM model; (b) FOS model; (c) LSTM model; (d) Deep-LSTM model.
Machines 13 00431 g010
Figure 11. The error distribution box plots for different arm force estimation models.
Figure 11. The error distribution box plots for different arm force estimation models.
Machines 13 00431 g011
Figure 12. Reward value curves during the training process of the human–robot collaborative control model. (a) Reward value curve during the pre-training stage; (b) Reward value curve during the task training stage.
Figure 12. Reward value curves during the training process of the human–robot collaborative control model. (a) Reward value curve during the pre-training stage; (b) Reward value curve during the task training stage.
Machines 13 00431 g012
Figure 13. Force and speed curves of different collaborative control models: (a,b) Force and speed curves of the proposed control model; (c,d) Force and speed curves of the GPRIL model; (e,f) Force and speed curves of the ADC model; (g,h) Force and speed curves of the ablation model.
Figure 13. Force and speed curves of different collaborative control models: (a,b) Force and speed curves of the proposed control model; (c,d) Force and speed curves of the GPRIL model; (e,f) Force and speed curves of the ADC model; (g,h) Force and speed curves of the ablation model.
Machines 13 00431 g013
Table 1. The fuzzy rules in the experience replay partitioning mechanism.
Table 1. The fuzzy rules in the experience replay partitioning mechanism.
Arm Force Level
PHPMPSNSNMNH
Arm
Force Change Level
PHAccelAccelAccelDecelDecelDecel
PMRTAccelAccelDecelDecelRT
PSRTRTFAFARTRT
NSRTRTFAFARTRT
NMRTDecelDecelAccelAccelRT
NHDecelDecelDecelAccelAccelAccel
Table 2. Performance summary of the ablation study for the proposed arm force estimation model.
Table 2. Performance summary of the ablation study for the proposed arm force estimation model.
MethodDirectionMedian ErrorMean ErrorMaximum Error
AGA-TGNNFroce_x1.8322.69311.228
Froce_y1.8892.57910.482
Froce_z1.5861.8069.896
Fixed-Adjacency Matrix TGNNFroce_x3.2483.78417.469
Froce_y2.7763.43717.545
Froce_z2.3612.94816.166
LSTM-onlyFroce_x4.6645.13617.583
Froce_y3.7024.32716.898
Froce_z2.9083.47115.392
Table 3. Performance statistics of different arm force estimation models.
Table 3. Performance statistics of different arm force estimation models.
MethodDirectionMedian ErrorMean ErrorMaximum Error
AGA-TGNNFroce_x1.8252.61611.388
Froce_y1.7252.56110.805
Froce_z1.6262.09010.237
FOSFroce_x2.7303.26411.726
Froce_y3.2013.63111.302
Froce_z2.0032.30813.034
LSTMFroce_x4.0125.44311.346
Froce_y4.3465.02812.143
Froce_z2.6922.99712.068
Deep-LSTMFroce_x2.0822.85412.191
Froce_y2.1812.88411.974
Froce_z1.6992.33211.473
Table 4. Performance statistics of different control models.
Table 4. Performance statistics of different control models.
MethodGPRIL Model ADC ModelThe Proposed ModelThe Ablation Model
Collaborative Comfort0.951.091.121.02
Collaborative Smoothness0.940.951.030.93
Assembly Time (s)19.218.418.119.7
Assembly Error (mm)0.620.510.420.61
Assembly Success Rate90%93.3%100%86.7%
Table 5. Performance statistics of different participants.
Table 5. Performance statistics of different participants.
ParticipantCollaborative ComfortCollaborative SmoothnessAssembly Time (s)Assembly Error (mm)Assembly Success Rate
Participant 11.131.0117.60.47100%
Participant 21.091.0118.30.39100%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, Y.; Tang, N.; Li, Z.; Sun, H. Methodology for Human–Robot Collaborative Assembly Based on Human Skill Imitation and Learning. Machines 2025, 13, 431. https://doi.org/10.3390/machines13050431

AMA Style

Zhou Y, Tang N, Li Z, Sun H. Methodology for Human–Robot Collaborative Assembly Based on Human Skill Imitation and Learning. Machines. 2025; 13(5):431. https://doi.org/10.3390/machines13050431

Chicago/Turabian Style

Zhou, Yixuan, Naisheng Tang, Ziyi Li, and Hanlei Sun. 2025. "Methodology for Human–Robot Collaborative Assembly Based on Human Skill Imitation and Learning" Machines 13, no. 5: 431. https://doi.org/10.3390/machines13050431

APA Style

Zhou, Y., Tang, N., Li, Z., & Sun, H. (2025). Methodology for Human–Robot Collaborative Assembly Based on Human Skill Imitation and Learning. Machines, 13(5), 431. https://doi.org/10.3390/machines13050431

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop