Next Article in Journal
Research on AGC Nonlinear Compensation Control for Electro-Hydraulic Servo Pump Control of a Lithium Battery Pole Strip Mill
Next Article in Special Issue
Novel Triplet Loss-Based Domain Generalization Network for Bearing Fault Diagnosis with Unseen Load Condition
Previous Article in Journal
Experimental Assessment of a Decentralized Control Strategy for a Back-to-Back Modular Multilevel Converter Operating in Low-Frequency AC Transmission
Previous Article in Special Issue
Wear Prediction of Tool Based on Modal Decomposition and MCNN-BiLSTM
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimization of Smart Textiles Robotic Arm Path Planning: A Model-Free Deep Reinforcement Learning Approach with Inverse Kinematics

1
School of Mechanical Engineering, Tiangong University, Tianjin 300387, China
2
Engineering Teaching Practice Training Center, Tiangong University, Tianjin 300387, China
3
School of Electronics & Information Engineering, Tiangong University, Tianjin 300387, China
4
Innovation College, Tiangong University, Tianjin 300387, China
*
Author to whom correspondence should be addressed.
Processes 2024, 12(1), 156; https://doi.org/10.3390/pr12010156
Submission received: 8 November 2023 / Revised: 2 January 2024 / Accepted: 5 January 2024 / Published: 9 January 2024

Abstract

:
In the era of Industry 4.0, optimizing the trajectory of intelligent textile robotic arms within cluttered configuration spaces for enhanced operational safety and efficiency has emerged as a pivotal area of research. Traditional path-planning methodologies predominantly employ inverse kinematics. However, the inherent non-uniqueness of these solutions often leads to varied motion patterns in identical settings, potentially leading to convergence issues and hazardous collisions. A further complication arises from an overemphasis on the tool center point, which can cause algorithms to settle into suboptimal solutions. To address these intricacies, our study introduces an innovative path-planning optimization strategy utilizing a model-free, deep reinforcement learning framework guided by inverse kinematics experience. We developed a deep reinforcement learning algorithm for path planning, amalgamating environmental enhancement strategies with multi-information entropy-based geometric optimization. This approach specifically targets the challenges outlined. Extensive experimental analyses affirm the enhanced optimality and robustness of our method in robotic arm path planning, especially when integrated with inverse kinematics, outperforming existing algorithms in terms of safety. This advancement notably elevates the operational efficiency and safety of intelligent textile robotic arms, offering a groundbreaking and pragmatic solution for path planning in real-world intelligent knitting applications.

1. Introduction

The deepening evolution of the Industry 4.0 era is progressively reshaping the face of industrial production, with a pronounced impact in areas like building control [1] and the textile industry [2]. Textile processes, in particular, stand at the forefront of this change. In the realm of intelligent computerized flat knitting, robotic arms bear about 30% of the operational load, undertaking tasks such as yarn fetching, feeding, and knitting. This shift necessitates innovative approaches to robotic arm path planning [3]. The complexity inherent in these tasks not only raises the difficulty of planning and execution but also increases the likelihood of errors, thereby adversely affecting both efficiency and overall production [3]. Traditional path-planning methodologies, which are heavily dependent on offline programming and online instruction, prove inadequate for complex knitting operations, demanding significant labor and time investments [4,5,6]. Activities like yarn fetching and feeding require multiple adjustments in arm posture, each mandating a reevaluation and replanning of the path. Consequently, the pursuit of efficient, accurate, and safe path planning for knitting robotic arms has emerged as a vital area of research within the sphere of automated and intelligent manufacturing.
In industrial settings, the trajectory planning of robotic arms typically employs forward and inverse kinematic techniques [7]. Forward kinematics involves manipulating the end effector, TCP, to execute specific tasks by modulating joint angles, velocities, and accelerations. In contrast, inverse kinematics centers on the TCP’s path planning, translating this planned trajectory from the path space to the robotic arm joints’ operational space through inverse kinematic processes [8]. This paper focuses on exploring model-free path-planning strategies for knitting robotic arms, underpinned by experiential knowledge in inverse kinematics.
DRL distinguishes itself from traditional reinforcement learning methods [9] by adeptly navigating high-dimensional, continuous state and action spaces, thereby exhibiting enhanced representational and generalization skills. This capability enables DRL to identify latent patterns and structures in intricate tasks, facilitating the learning of highly non-linear and complexly associated decisions in unfamiliar environments, as demonstrated in the soft robotics sector [10]. DRL primarily operates through two paradigms: value-function-based and policy-gradient-based reinforcement learning algorithms [11]. The former concentrates on learning the value functions of states or state–action pairs, and evaluating action or state values under prevailing policies. The inherent continuity of action spaces and the infinite variety of action values pose substantial challenges in handling continuous domain issues, with Q-Learning and Deep Q-Network (DQN) being notable examples. Conversely, policy-gradient-based methods, aimed at directly learning policies, are more apt for continuous domains. They produce a probability distribution for each potential action in a given state, refining the policy through parameter optimization to enhance the probabilities of actions yielding higher rewards. Representative algorithms include TRPO [12], DDPG [13], A2C [14], HER [15], PPO [16], SAC [17], and TD3 [18].
These algorithms have all shown remarkable efficacy in path planning for target points. Nevertheless, the non-uniqueness and multiplicity inherent in inverse kinematics can lead to various solutions in the robotic arm joints’ operational space [19], creating “multiple solution conflicts”. This issue may result in the same processing method being applied to disparate joint states, leading to impractical motion outcomes and compromised stability. Moreover, an excessive emphasis on the TCP point could trap algorithms in local optima, potentially causing unsafe movements, increased mechanical wear, and higher energy consumption. This scenario underscores a lack of robustness and an inability to realize optimal global outcomes [20].
In tackling the challenges posed by the multiplicity of inverse kinematics solutions and the tendency towards local optima due to an excessive focus on the TCP point, this study innovatively introduces Environment Augmentation (EA) and Multi-Information Entropy Geometric Optimization (MIEGO) methodologies within the realm of DRL algorithms. The EA strategy, in particular, enriches an environmental state by incorporating the angles of each robotic arm joint and treats the stability of solutions as an additional reward factor. This technique effectively guides agents towards exploring paths with enhanced stability.
To circumvent the “local optima trap” associated with an overemphasis on the TCP point, we propose the MIEGO approach. This method synergizes the TCP optimization issue with broader global optimization goals, creating a comprehensive multi-objective optimization framework. By shifting the focus from a narrow TCP-centric view to a wider multi-objective perspective, the likelihood of falling into local optima is substantially reduced. In optimizing these multi-objective elements, the study leverages information geometric techniques to regulate multi-information entropy, thereby not only advancing the efficacy of deep reinforcement learning (DRL) algorithms but also amplifying their convergence properties.
Subsequently, the paper delineates the principal work, offering an overview of existing policy-gradient-based reinforcement learning algorithms, followed by an exhaustive delineation of both the EA strategy and MIEGO method. Empirical assessments are conducted to thoroughly validate each algorithm. Moreover, the simultaneous implementation of these enhancements in the DDPG and SAC algorithms reveals that their integration effectively harmonizes aspects like path length, motion safety, and equipment longevity.

2. Main Work

In this study, we developed and implemented innovative EA strategies and MIEGO methods within the framework of DRL algorithms. The primary contributions are outlined as follows:
  • The design and analysis of specific industrial scenarios with existing DRL methods: We targeted a specialized application scenario involving intelligent textile robotic arms in textile production lines (depicted in Figure 1), establishing a virtual twin system for simulation training to gather insightful learning data. Furthermore, an extensive examination of current policy-gradient-based reinforcement learning algorithms [21], including TRPO, DDPG, PPO, SAC, and TD3, was undertaken, offering a detailed analysis of their distinct features.
  • The introduction of novel strategies to enhance DRL and their experimental validation: To address the challenges of “multiple solution conflicts” in inverse kinematics and the tendency towards local optima due to a concentrated focus on the TCP point, this study introduces novel EA strategies and MIEGO methods tailored for DRL algorithms. Empirical studies were conducted using the virtual twin model of the intelligent textile robotic arm scenario within VREP 3.6.2 software. Our study contrasted these advancements with classical DRL algorithms. The findings indicate a substantial enhancement in the performance of DRL algorithms across several crucial metrics, attributable to the integration of the methodologies proposed herein.

3. Reinforcement Learning Algorithms

This section delves into existing policy-gradient-based reinforcement learning algorithms, which were selected as foundational algorithms for the textile robotic arms discussed in this study. These algorithms are explored and experimented upon in conjunction with the strategies introduced in this research.

3.1. Trust Region Methods

Trust Region Policy Optimization (TRPO), introduced by Schulman et al. in 2015 [12], represents a sophisticated method in the domain of reinforcement learning, designed to optimize control policies with a guarantee of monotonic advancements. TRPO’s primary principle involves identifying a trajectory that enhances policy efficacy while ensuring minimal deviations in policy alterations. This approach is exceptionally effective for the optimization of extensive, non-linear policy frameworks.
Within the TRPO framework, a policy, π , is conceptualized as a probability distribution governing the selection of an action, a , in a given state, s , expressed as π a s . The overarching aim of this policy is the maximization of the expected cumulative rewards. TRPO primarily focuses on optimizing the following objective function:
m a x i m i z e θ   E s ~ p θ , a ~ π θ π θ a s π θ a s A θ s , a
Here, θ and θ represent the policy parameters of the current and previous iterations, respectively. A θ s , a is the advantage function for state s and action a under policy π θ .
In addition, Trust Region Policy Optimization (TRPO) employs a Kullback–Leibler (KL) divergence constraint to ensure that the magnitude of policy updates remains moderate [12]. The KL divergence is a measure used to quantify the difference between two probability distributions. In this context, it is utilized to quantify the variation in a policy before and after an update. Specifically, TRPO controls the magnitude of policy updates through the following KL divergence constraint:
s . t . E s ~ p θ D K L π θ · s δ K L
Here, D K L represents the Kullback–Leibler (KL) divergence, and π θ · s denotes the probability distribution obtained under state s using the policy with parameters θ . The value δ K L is a predefined threshold. By employing this approach, TRPO can effectively control the magnitude of changes during policy updates, preventing excessively drastic alterations. This ensures the stability and reliability of the learning process.
Proximal Policy Optimization (PPO), proposed by Schulman et al. in 2017 [16], is an improvement of TRPO. PPO simplifies the optimization process of TRPO and enhances sample utilization efficiency. PPO optimizes a policy through the following objective function:
r t θ = π θ a t s t π o l d a t s t  
L θ = E ^ min r t θ A ^ t ,   c l i p r t θ ,   1 ϵ ,   1 + ϵ A ^ t
In this context, r t θ is the probability ratio, representing the likelihood of selecting an identical action under both the new and the old policies. A ^ t is the estimated value of the advantage function, and ϵ is a small constant utilized to limit the magnitude of policy updates. The ‘clip’ function confines the probability ratio, r t θ , within predetermined bounds of 1 ϵ and 1 + ϵ . This constraint on the probability ratio aids in preventing excessive oscillations during policy updates, thereby ensuring stability and reliability in the learning process. When the value of r t θ falls below 1 ϵ , the ‘clip’ function adjusts it to 1 ϵ ; conversely, when it exceeds 1 + ϵ , it is corrected to 1 + ϵ . Within these boundaries, the value of r t θ remains unchanged.
Equation (3) defines the probability ratio, r t θ , which measures the ratio of the probabilities of choosing the same action under new and old policies. This ratio is the foundation for calculating the objective function of the PPO algorithm. Following this, Equation (4) presents the core objective function, L θ , of the PPO algorithm, utilizing the probability ratio, r t θ , from Equation (3), in conjunction with the estimated value of the advantage function, A ^ t , and the clip function, to optimize the policy.
PPO utilizes this approach to balance exploration and exploitation while diminishing reliance on extensive datasets, enhancing its suitability for practical applications.

3.2. Deterministic Policy Methods

In addition to stochastic policies, the Deep Deterministic Policy Gradient (DDPG) emerges as a deterministic policy gradient approach [13], extending the success of DQN to the realm of continuous control. The DDPG adheres to the following update mechanism:
Q s t ,   a t = r s t ,   a t + γ Q s t + 1 , π θ s t + 1
This equation conveys that the estimated value of the function Q, given the current state, s t , and action, a t , is the sum of the immediate reward, r s t ,   a t , and the discounted value of future expected returns. The discount factor, γ , plays a critical role in weighting the significance of immediate versus future rewards. The term Q s t + 1 , π θ s t + 1 denotes the anticipated return associated with the optimal action as per policy θ in the subsequent state s t + 1 . DDPG operates as an off-policy actor–critic algorithm adept at formulating effective policies across a spectrum of tasks.
The Twin Delayed DDPG (TD3) builds upon the DDPG [18], aiming to mitigate some of its inherent constraints. TD3 utilizes a dual-structured framework with two actor networks and two critic networks. These networks are co-trained to synergize and optimize the agent’s performance. The twin actor networks are designed to yield disparate action outputs for each state, subsequently appraised by the dual critic networks [22]. Such an arrangement allows TD3 to more accurately navigate environmental uncertainties and curtail the propensity for value overestimations typical in DDPG.

3.3. Entropy-Regularized Policy Gradient

Soft Actor–Critic (SAC) [17] is a sophisticated off-policy gradient technique that forms a nexus between DDPG and stochastic policy optimization strategies. SAC integrates the concept of clipped Double Q-learning, with its maximum entropy DRL objective function articulated as:
J π = t = 0 T E s t , a t ~ ρ π [ r s t , a t + α H ( π ( · | s t ) ) ]
In this context, r s t , a t denotes the immediate reward obtained from executing action a t in state s t . H ( π ( · | s t ) symbolizes the policy’s entropy, reflecting the inherent randomness or uncertainty within a specific state. The α parameter serves as a crucial balance factor, calibrating the significance of the reward against entropy, thus effectively managing the policy’s exploration–exploitation dynamics. The core objective of this approach is to optimize the policy π , maximize expected returns, and facilitate the optimal choice of action a t in any given state, s t . SAC’s objective function uniquely features entropy regularization, aiming to optimize a policy through a strategic equilibrium between entropy and anticipated returns. Entropy here quantifies the level of stochasticity in policy decisions, mirroring the critical balance between exploration and exploitation. Elevating entropy levels not only fosters increased exploration but also expedites the learning trajectory. Crucially, this mechanism also safeguards against a policy’s convergence to suboptimal local extremums.

4. Environment Augmentation Strategy

In the field of robotics, trajectory planning for tasks is a pivotal endeavor. It encompasses the movement of objects from one point to another or the execution of intricate maneuvers. For this purpose, robots require an efficient trajectory planning strategy. Utilizing DRL for trajectory planning presents a highly promising approach. It leverages the dynamics of Markov Decision Processes (MDPs) to derive optimal strategies from multifaceted environments. Such a strategy enables robots to make optimal decisions based on real-time status during task execution, thereby crafting trajectories that fulfill specified criteria.
In conventional path-planning algorithms integrating inverse kinematics, intelligent textile robotic arms typically align directly above the center of a target yarn spool, subsequently inserting their grippers into the central hole of the spool and then opening them. However, this method can result in collisions between the robot arm and the upper inside of the yarn spool, as illustrated in Figure 2a. Such collisions can become unacceptably hazardous due to environmental variations. To counter this, we introduced an EA strategy tailored to refining the path-planning algorithm of DRL under the paradigm of inverse kinematics integration. This approach addresses the issue of overconcentration on the tool center point (TCP) and the consequent neglect of environmental data, thereby facilitating safer and more efficient trajectory planning. By incorporating MIEGO, the focus is expanded from a singular TCP issue to a broader spectrum of multi-objective optimization, leveraging environmental data to circumvent scenarios of local optima, as shown in Figure 2b.
It is crucial first to clarify that in DRL, the definition of the environmental state is of utmost importance. In typical reinforcement-learning-based trajectory planning tasks, the environmental state encompasses: (1) the position of the TCP of the end effector and (2) relative positional data between the TCP and the target location. Typically, the state s t for trajectory planning is defined as:
s t = T C P , δ , D o n e
In this framework, the T C P is represented as a vector in the global coordinate system, capturing the position of the end effector. The term δ represents the offset vector between the end effector and its targeted object. The boolean variable D o n e is set to 1 when the end effector successfully reaches its target position; otherwise, it remains 0.
Throughout the task execution phase, the DRL model makes action decisions based on both the prevailing state and the strategy it has learned. To ensure adaptability across various robots and end effectors, the action a t is defined in terms of the end effector’s translational movement:
a t = x , y , z
This approach to defining states and actions steers the continuous modification in the position and orientation of the robot’s end TCP, crafting a seamless motion trajectory. However, despite facilitating the autonomous planning of task trajectories by the agent, the inherent multiplicity in inverse kinematics suggests multiple potential configurations of robotic arm joint angles (states) for any given TCP position and orientation. Figure 3 depicts multiple-solution conflict scenarios, a consequence of inverse kinematics, in two distinct cases.
In the realm of DRL, models typically formulate a distinct policy (action) for each specified state. However, a complication arises when joint angles are not factored in and states are solely defined by the TCP position. This can lead to different states of the robotic arm corresponding to the same TCP position, yet the DRL model may produce identical policies. Such a scenario introduces several challenges:
1.
For robotic arms with identical TCP positions but varying joint states, the ideal action could vary. Generating the same policy for these divergent states by the DRL model might compromise task execution efficiency or even result in task failure.
2.
In tasks necessitating maneuvers in complex settings or confined spaces, the specific joint angle configurations of the robotic arm can critically influence task success. Policies derived solely from TCP positions may fall short of meeting these nuanced requirements.
3.
Given that a robotic arm’s state is influenced by both the TCP position and joint angles, overlooking joint angles means the state cannot comprehensively represent a robot’s actual condition, potentially impeding the learning outcomes.
To surmount these DRL challenges, this study introduces a novel EA strategy, as delineated in Figure 4. This strategy centers around enhancing the state representation of the robotic arm within its environment, synergized with a thoughtfully constructed reward function. Initially, the enhanced state and the predefined reward function are fed into the agent. This step is followed by updating the actor–critic network parameters within the agent. Subsequently, the deep network, utilizing these updated parameters, computes the information for the robotic arm’s subsequent target pose (TCP). Post-inverse kinematics processing, these data inform the action angles for the robotic arm’s joints. These angles are then communicated to the robotic arm, directing it to perform the planned actions. This iterative process facilitates the ongoing refinement of the robotic arm’s control strategy.
In DRL, the State Augmentation strategy involves introducing the states of the robotic arm’s joint angles into the environmental state. This approach provides a more accurate reflection of a robotic arm’s actual state, thereby generating more effective strategies. The state s t is defined as:
s t = T C P , θ , δ , D o n e
Here, the definitions of T C P , δ , and D o n e remain consistent with earlier descriptions, while θ represents the angles of the robotic arm’s joints, providing a more accurate representation of the robot’s current state.
Given that the robotic arm used in the scenarios of this paper is the six-DOF ER10-2000 model (EFORT Intelligent Equipment Co., Ltd., Foshan, China), the state incorporates the rotational angles of the robotic arm’s six axes to offer additional information for learning. In DRL, the effectiveness of model learning often depends on the amount of available information. Introducing joint angles provides more data for the model, potentially aiding in learning better strategies. Furthermore, for robots with redundant degrees of freedom, multiple joint angle configurations can achieve the same TCP position. Incorporating joint angles allows the strategy to better utilize these redundant degrees of freedom, potentially resulting in a greater variety of actions.
This paper also uses the change in inverse solutions as a reference index for measuring the stability of solutions. This index serves as an additional reward element, guiding the agent to prioritize exploring paths with higher stability and achieving smoother trajectories that avoid collisions and enhance efficiency.
Initially, θ t R 6 represents the robotic arm’s joint angles at the current time step, and θ t 1 R 6 represents the joint angles at the previous time step. The corresponding planned TCP positions are q t R 3 and q t 1 R 3 . The following formula is established:
θ t = f q t
With the corresponding differential relationship:
θ ˙ t = f q t t = J q t
Here, J q t R 6 is the Jacobian matrix for a six-DOF robotic arm, representing the rotational velocity of each joint. For differential mobile robotic arms, it signifies the rate of change in angles or the rotational acceleration of each joint.
To measure the stability of solutions without excessively rewarding minor changes, this paper employs a variant of the sigmoid function for the nonlinear mapping of stability rewards, defined as:
S = 1 1 + e k J d
Here, k is a positive parameter controlling the slope of the function, and d is a parameter controlling the center of the function.
When J is close to 0, S will be a small value. As J increases, S will smoothly increase. The advantage of this function is that it does not overly reward very small changes, and as the change magnitude reaches a certain level, the reward increases but at a slower rate.
If the stability measure dominates the reward function, the agent might minimize the change in solutions during learning to maximize rewards, leading to local optima or inaction. To avoid this, the paper introduces a weight parameter and incorporates this stability measure into the overall reward function:
R = R b a s e α · 1 1 + e k J d
Here, R b a s e is the basic reward, such as the reward for reaching the target position or other task-related rewards; α is a weight parameter, controlling the importance of the stability measure in the overall reward.

5. Multi-Information Entropy Geometric Optimization Method

This chapter delves into the core aspects of the MIEGO method. This includes establishing multi-objective optimization problems, incorporating information entropy, and utilizing natural gradients for geometric optimization. Figure 5 illustrates the basic framework of this method.

5.1. Construction of Multi-Objective Optimization Problems

The primary goal of robotic arm path-planning tasks is to identify viable motion paths that satisfy a set of predefined objectives within given constraints. These objectives typically encompass minimizing motion time, reducing energy consumption, enhancing safety, optimizing motion smoothness, improving precision, adhering to specific constraints, maximizing work efficiency, and minimizing mechanical stress [23]. In this chapter, we initially provide a detailed definition of these objectives and then explore how to construct and solve a multi-objective optimization problem encompassing these goals.

5.1.1. Motion Time

Motion time refers to the total time required for a robotic arm to move from its initial state to the target state. In intelligent textile manufacturing environments, tasks assigned to textile robotic arms often need to be completed within a certain timeframe to ensure process continuity and production efficiency. We define a time cost function, f 1 X , to quantify the motion time of the robotic arm. Specifically, f 1 X can be defined as the total time needed to move from the initial state to the target state. However, the per-step motion time, t , obtained through different DRL algorithms may vary due to differences in step sizes [24]. Therefore, we define motion time as the sum of the ratios of the step lengths to their velocities:
f 1 X = i = 1 n 1 x i x i + 1 v i
Here, X = x 1 , x 2 , x 3 , , x n represents a path composed of n points from the initial state to the target state. x i i = 1,2 , , n describes the vector of the robotic arm’s TCP, including its position and rotation state. v i is the motion velocity at each step. When the TCP sequence, X , of the robotic arm minimizes f 1 X , we say that the optimal motion time path has been found.

5.1.2. Energy Consumption

Energy consumption is another objective considered for sustainable development and energy conservation. We define an energy consumption cost function, f 2 X , to quantify the energy usage of the robotic arm. Similar to f 1 X , f 2 X can be defined as the total energy consumed by the robotic arm in moving from the initial state to the target state [25]. We define energy consumption as the total angular mileage of joint rotations:
f 2 X = i = 1 n 1 j = 1 6 θ i + 1 , j θ i , j
where X is the same as mentioned above, and θ i , j represents the absolute angle of joint j in the i-th step of the sequence, X. When the TCP sequence, X, of the robotic arm minimizes f 2 X , we say that the optimal energy consumption path has been found.

5.1.3. Safety

In many scenarios where robotic arms share a workspace with humans or other machines, safety is an indispensable consideration. We define a safety cost function, f 3 X , based on factors such as the robotic arm’s range of motion, velocity, acceleration, and minimum distance to people or objects to quantify the safety of the robotic arm:
f 3 X = w 1 C r + w 2 C v + w 3 C a + w 4 C d
Generally, the safety cost should be a non-negative function that decreases as safety increases. Moreover, it should be as smooth as possible, meaning its derivatives are continuous across the entire domain, as many optimization algorithms, like gradient descent, require the function’s derivatives [26]. Based on this analysis, we define the following sub-functions for the safety cost function: C r is the cost function for the range of motion:
C r = i = 1 6 exp θ m a x , i θ i θ i θ m i n , i σ 2
Here, θ i ( i = 1,2 , 3,4 , 5,6 ) represents the angle of the i-th joint, with a safe motion range set as [ θ m i n , θ m a x ] . This function rapidly increases as joint angles approach their limits, with σ being a parameter that regulates the growth rate of the function.
C v and C a are cost functions for velocity and acceleration, respectively:
C v = i = 1 n θ ˙ i v m a x , i 2
C a = i = 1 n θ ¨ i a m a x , i 2
Here, θ ˙ i and θ ¨ i represent the absolute angular velocity and acceleration of the i-th joint, respectively, with v m a x and a m a x denoting the maximum velocity and acceleration for each joint. The cost increases rapidly when the velocity or acceleration exceeds its maximum allowable value.
C d is the cost function for the range of motion:
C d = 1 1 + exp d m i n d s a f e δ
Here, d m i n is the minimum distance between the robotic arm and the closest person or object in the environment, d s a f e is a predefined safe distance, and δ is a parameter that adjusts the growth rate of the function. The function is monotonic, increasing in value as the distance between the robotic arm and other objects or humans in the environment decreases.
Therefore, when the TCP sequence, X, of the robotic arm minimizes f 3 X , we say that the optimal safety path has been found.

5.1.4. Motion Smoothness

We aim for the robotic arm’s motion to be as smooth as possible to minimize shocks and vibrations, thereby enhancing the quality and stability of operations. We define a smoothness cost function, f 4 X , to quantify the motion smoothness of the robotic arm:
f 4 X = i = 2 n 1 x i + 1 2 x i + x i 1 2
Here, x i is defined as in previous sections. This formula essentially calculates the sum of the squares of the discrete acceleration of the TCP trajectory. In cases of smooth motion, the acceleration should be as minimal as possible; hence, the cost function decreases with smoother motion. The advantage of this method is its direct computability from the sequence of TCP points, facilitating network optimization in conjunction with inverse kinematics [27]. Therefore, when the TCP sequence, X , of the robotic arm minimizes f 4 X , we say that the optimal smooth path has been found.
Based on the definitions above, we can construct a multi-objective optimization problem to simultaneously satisfy all these objectives under given constraints. Specifically, we seek to find a solution, X, that minimizes the following objective function, F X :
F X = w 1 f 1 x + w 2 f 2 x + w 3 f 3 x + w 4 f 4 x
Here, w i represents weight coefficients used to balance the importance of different objectives. In practical applications, the choice of w i typically depends on specific task requirements and environmental conditions. For instance, if safety is prioritized over motion time, one might choose w 3 > w 1 ; if motion smoothness is more critical than energy consumption, then w 4 > w 2 could be selected.

5.2. Information Entropy and Geometric Optimization

In the process of seeking optimal solutions, we introduce a crucial concept: information entropy. Originally proposed by Claude Shannon in his study of information transmission in communication systems, information entropy serves as a significant tool to measure the uncertainty or randomness of information.

5.2.1. Mathematical Definition of Information Entropy

Information entropy is a concept widely used in probability theory and information theory, measuring the uncertainty of a random variable. For a random variable, X , its information entropy, H X , is defined as:
H X = P x i log P x i
Here, P x i represents the probability of the random variable X assuming the value x i . Since the calculation of information entropy is based on a probability distribution, for continuous random variables, their probability density functions are required to compute the information entropy.

5.2.2. Application of Information Entropy in Robotic Arm Path Planning

In our context, robotic arm path planning is viewed as finding the optimal solution within all possible solution spaces. Each potential solution can be considered a random event, with its occurrence probability represented by the corresponding objective function value. Therefore, the uncertainty of solutions can be gauged by calculating the information entropy of the solution space.
Generally, greater information entropy indicates a higher uncertainty of the solution, and vice versa. In practical applications, we often seek solutions with high certainty, i.e., lower information entropy. However, in multi-objective optimization problems, conflicting objectives may lead to multiple solutions that satisfy all objectives, forming a Pareto optimal set. In such cases, we aim to find solutions that not only meet all objectives but also possess high certainty, that is, the solutions with the highest information entropy in the Pareto optimal set.
Here, we introduce a new objective function, f 5 X , to quantify the information entropy of solutions:
f 5 X = P x i log P x i
In this equation, P x i represents the probability of the solution x i , which can be indicated by the objective function value. Thus, our multi-objective optimization problem becomes finding a solution X that minimizes the following objective function, F X :
F X = w 1 f 1 x + w 2 f 2 x + w 3 f 3 x + w 4 f 4 x w 5 f 5 x
In this reformulated optimization problem, we seek solutions that not only satisfy all objectives but also have the maximum information entropy.

5.2.3. Information Entropy in DRL

In DRL, policy networks typically represent the behavioral strategies of robots. The output of a policy network is a probability distribution, indicating the likelihood of executing various actions in a given state [28]. During training, the goal is to find a policy that maximizes cumulative rewards. However, in practical applications, it is often desirable for a robot’s behavior to exhibit a degree of randomness, enabling exploration in unknown environments [29]. One study [17] suggests increasing the randomness of a policy by maximizing its information entropy.
Thus, an information entropy term is incorporated into the objective function of reinforcement learning as follows:
J θ = E R s , a β   H π a s ; θ
Here, R s , a represents the reward function, H π a s ; θ is the policy’s information entropy, and β is a hyperparameter controlling the balance between rewards and information entropy.
This approach not only makes the robot’s behavior more random but also stabilizes the policy. During training, too low an information entropy might lead to overfitting, whereas too high an entropy could degrade policy performance. Therefore, by maximizing information entropy, policy stability is enhanced while ensuring performance.

5.2.4. Natural Gradient Geometric Optimization

Conventional gradient methods, such as gradient ascent or descent, often consider only the gradient direction at the current position, overlooking the local geometric structure of the parameter space. This can lead to excessively large or small steps in parameter updates, resulting in unstable learning or low efficiency. Hence, this paper employs “natural gradients” for optimization, a method that explicitly considers the Riemannian structure of the policy parameter space, focusing more on the local geometric structure of the parameter space.
Specifically, the natural gradient method first defines a Riemannian metric, setting the local distance in the parameter space. Then, it updates parameters according to this metric, maintaining the smoothness of policy changes throughout the update process.
The natural gradient update rule is applied as follows:
θ     θ + α G 1 θ J θ
In this formula, α is the learning rate, θ J θ is the gradient of the objective function, J θ , with respect to parameters θ , and G is the policy’s Fisher information matrix. G 1 θ J θ represents the natural gradient.
By combining information entropy and natural gradient methods, we can make the robot’s behavior more random and better consider the geometric structure of the policy parameter space, leading to more effective optimization.

6. Experimental Analysis

This chapter first establishes a virtual twin experimental environment for the intelligent textile robotic arm, as illustrated in Figure 6, and defines the task objectives. Subsequently, this paper employs four policy-gradient-based DRL algorithms, DDPG, TD3, SAC, and PPO, and integrates our EA strategy and MIEGO method into these algorithms. A series of experiments are then conducted to evaluate the proposed optimization strategies and methods, demonstrating their superior performance.

6.1. Intelligent Textile Robotic Arm and Its Virtual Twin

The intelligent textile robotic arm model involves the geometric structure of industrial robot arms, joint parameters, and kinematic modeling. Table 1 and Table 2 and Figure 7 present the parameters and motion range of the real-world textile robot hand ER10-2000. Based on these parameters, a model of the textile robot hand is developed in VREP on a 1:1 scale ratio, as shown in Figure 8, achieving a high-fidelity simulation of the real-world scenario.
To acquire more learnable data in realistic textile yarn-gripping robotic hand scenarios, this study constructs a virtual twin system. This system accurately replicates the 1:1 kinematic model of the robotic arm and hand, enabling extensive training within the simulation environment. This approach allows for the collection of a substantial amount of high-quality training data, which can be directly applied to the trajectory generation of the actual robot.
The introduction of the virtual twin system not only optimizes the efficiency and effectiveness of DRL training for the robotic arm and hand but also reduces the costs and safety risks associated with physical training. By simulating different production scenarios and environmental conditions, the virtual twin system can provide more comprehensive and accurate training data for the actual robot, testing its robustness and reliability in various settings. Furthermore, safety constraints can be tested and optimized in the virtual environment, ensuring the safety and stability of the robot in actual operations. Thus, the virtual twin technology offers a more efficient and reliable solution for the textile production and processing industry.
This paper utilizes VREP as the simulation platform, specifically V-REP PRO EDU Version 3.6.2. This software provides a powerful simulation environment, supporting a variety of robot models and algorithms and facilitating high-quality training of the robotic arm and hand in the virtual environment. Using VREP, this study can quickly set up and test the virtual twin system, laying a solid foundation for the practical application of the robotic arm and hand.

6.2. Experimental Setup

Figure 9 displays the continuous trajectory planning scenario for the textile robotic hand within the virtual twin platform, with the task described as follows.
At the start of the task, the robotic arm is positioned at the initial location, point A, and its vicinity. To the right of the robotic hand is a table with various obstacles of different sizes and dimensions, which will be detailed in subsequent chapters. Also placed on the table is a yarn spool, the object to be grasped and relocated. Above the yarn spool is point B, where the robotic hand’s end effector needs to be positioned vertically downward during grasping. The end effector inserts its grippers into the central hollow of the yarn spool and opens them, lifting the spool using friction before placing it in the designated area, point C, on the left platform with the correct posture. To generalize the problem, only one yarn spool in the middle is selected on the virtual twin platform, but it can be positioned in different areas before each simulation to accommodate more possibilities.
Assume the end effector of the robotic hand is initially at point A, and the center of the yarn spool is located above point B (and its vicinity). To successfully lift the spool, the end effector of the robotic hand must first move a short distance along the −x axis, then move along the −y axis, continue moving along the −x axis, and approach the yarn spool with the appropriate posture before lifting it along the z-axis. It is foreseeable that if the robotic hand moves directly from point A to point B, the robotic arm’s joints (or links) will inevitably collide with the beams in the environment. Therefore, the grasp–place task is divided into the following steps, with the complete trajectory composed of continuous segments:
1.
The yarn spool appears at any location on the table, and the TCP of the robotic arm moves from the initial position, A, along trajectory segment 1 to the preparation position, B, with the appropriate posture. Point B is located directly above the center of the yarn spool along the z-axis and may appear anywhere in Region 1.
2.
The robotic arm moves along trajectory segment 2 from the preparation position, B, to the placement position, C (which can be randomly specified in Region 2), retracts its gripper, and places the yarn spool on the platform.
3.
The arm returns to the vicinity of the initial position, A, from the placement position, C.
In this paper’s experiments, the six-DOF intelligent textile robotic arm model on the virtual twin platform is used to complete the task along trajectory 1. In this environment, the coordinates of point A (the starting point) are (1.3019, 0.0319, 1.5801), and those of point B (the center of the yarn spool) are (0.7500, −1.3000, 1.2000). Therefore, the distance between A and B is approximately 0.7228 m. A step length, a 0,0.03 , is chosen, so the straight-line distance between A and B requires at least 24 steps (all distance units are in meters).

6.3. Experimental Results

In our experiments, we compared the performance of PPO, SAC, DDPG, and TD3 and their corresponding environment-augmented strategy algorithms, EA-PPO, EA-SAC, EA-DDPG, and EA-TD3. The specific parameters for each algorithm are detailed in Table 3.
The training comparison data presented in Figure 10 show that both PPO and EA-PPO were successful in learning appropriate paths over the course of training, ultimately converging to similar reward levels. Notably, due to the requirement of EA-PPO to explore the environment more extensively, its convergence rate was relatively slower. For SAC and EA-SAC, their performance was generally similar, but the step count results indicate that EA-SAC required more steps in the process of exploring suitable paths. The difference between DDPG and EA-DDPG was more pronounced, with EA-DDPG spending more time exploring quality paths due to the influence of inverse solution stability rewards and the uncertainty of environmental information. Additionally, EA-DDPG exhibited a larger error band compared to DDPG, indicating increased fluctuation in reward values across multiple experiments due to the EA strategy. Since TD3 is an improved version of DDPG, in the experiments, relative to DDPG, both TD3 and EA-TD3 achieved convergence faster, although EA-TD3 still required more steps to find ideal paths.
Figure 11 reveals the impact of the EA strategy on the stability metric of the solutions, especially the stability measure referenced by the S value. As training progressed, a gradual decrease in the S value was observed. This trend indicates that the algorithms produced more stable solutions over time. Furthermore, this stability was reflected in the smoothness of the search paths. In other words, as the S value decreased, the exploration paths became smoother. This is particularly important in practical applications, as smoother paths typically mean more predictable and controllable decision-making processes.
Figure 12 illustrates significant differences between the environment-augmented strategy versions of PPO, SAC, DDPG, and TD3 and their original DRL algorithms across various performance indicators. At first glance, the paths planned by the versions with the EA strategy seem longer with extended execution times, giving an initial impression of lower efficiency. However, this reflects a key feature of the EA strategy: a deeper and more comprehensive exploration of the environment. The strategy aims to thoroughly analyze the environmental structure, uncover potential risks, and thus provide more comprehensive and refined decision making.
The experiments validate the effectiveness of our EA strategy. During training, our strategy continually optimizes stability indicators, producing higher-quality solutions. Ultimately, this enhancement in stability and improvement in path smoothness provide strong support for the practical application of our EA strategy.
Furthermore, the EA strategy emphasizes robust control policies, prioritizing precise and cautious movements over speed. This strategy not only reduces joint speeds (except for EA-PPO, where all environment-augmented versions have lower joint speeds than their original versions) but also enhances inverse kinematics stability. By choosing more robust and precise inverse kinematics solutions, the environment-augmented versions can achieve smoother path planning and more stable joint movements, effectively reducing energy consumption. Specifically, the EA strategy brings benefits in terms of energy saving and also helps to reduce the overall lead time. Although this strategy may result in more steps and longer execution times in the initial stages of a task, it effectively reduces repetitions or adjustments due to environmental changes or imprecise path planning over the entire task cycle. This ability to reduce repetitive actions, in the long run, not only lowers energy consumption but also reduces the overall completion time of a task. In other words, the EA strategy enhances efficiency and safety in the long term by improving the accuracy and stability of path planning, thereby avoiding operational failures and safety incidents due to environmental changes or misinterpretations [30].
In summary, while the EA strategy may show weaker performance in terms of steps and execution time compared to the original algorithms, its deep exploration and understanding of the environment, along with its enhanced focus on operational safety and stability, reveal its significant advantages and resilience in dealing with complex environments and unknown risks.
Additionally, this paper selects the widely applied DDPG and SAC algorithms in continuous space path planning as benchmarks for comparison and conducts an in-depth analysis of the path planning results. We designed and implemented an improved algorithm combining the EA strategy and MIEGO method (referred to as the EM algorithm), specifically for training on the first path. Comparative observations of the final paths planned by these algorithms reveal the effectiveness and innovation of the EM algorithm in path optimization, further confirming the practical value of the EA strategy in intelligent textile robotic arm path planning.
Figure 13 shows the paths planned by four different algorithms, from left to right: EM-SAC, SAC, DDPG, and EM-DDPG. It is observed that the paths planned by EM-SAC and EM-DDPG are more circuitous and maintain a greater distance from potential obstacles. This strategy reduces safety risks that could arise due to minor environmental changes, electrical control factors, etc., during actual movement. More importantly, the paths planned by these two algorithms are more streamlined, better suited for smooth motor operation, and can effectively avoid frequent motor starts and stops, thus extending the motor’s lifespan [31].
Figure 14 presents the operational conditions of joint angles in the optimal paths planned by the original algorithms and the EM algorithm. In the figure, (a) and (b) correspond to the EM-SAC and SAC algorithms, while (c) and (d) correspond to DDPG and EM-DDPG, respectively. In comparison, it is noticeable that the variation curves in (a) and (d) are significantly smoother and exhibit less jitter, almost negligible. This phenomenon indicates that the paths planned through the EM algorithm have a higher degree of smoothness. This feature of high smoothness means that in practical applications, the robotic arm’s movement will be considerably less jittery, thus achieving more precise positioning. Additionally, smoother motion paths can reduce the robotic arm’s instantaneous acceleration, thereby mitigating wear and tear and prolonging its lifespan. Furthermore, smoother movement paths also reduce the robotic arm’s demand for electricity, consequently lowering energy consumption [4]. Therefore, the excellent performance of the EM algorithm in this regard further confirms its superior performance and broad application prospects in real-world applications.
In conclusion, the EA strategy’s ability to deeply explore and understand the environment, ensuring operational safety and stability, makes it highly adaptable and superior for tasks involving complex environments and unknown risks. The integration of the EA strategy with the MIEGO method (EM algorithm) in path planning effectively balances factors such as path length, motion safety, and equipment durability, providing an effective strategy and methodology for achieving efficient and safe robotic arm movements in complex environments.

7. Conclusions

In this paper, we successfully addressed a series of challenges associated with inverse kinematics integrated path-planning algorithms in the context of knitting robotic arms through the innovative application of the EA strategy and MIEGO method. These challenges include the non-uniqueness and multiple solution characteristics of inverse kinematics, as well as issues related to local optima caused by an excessive focus on the TCP. Our experimental results revealed that the EA strategy effectively guided agents in conducting deep environmental exploration, enhancing the comprehensiveness and safety of path planning. Additionally, our MIEGO method successfully shifted the focus from the singular issue of TCP to multi-objective optimization, effectively avoiding the risk of falling into local optima.
By integrating the EA strategy and MIEGO method, we proposed an improved algorithm (referred to as the EM algorithm) applicable to the widely used DDPG and SAC algorithms in continuous space path planning. The experimental results showed that the EM algorithm not only planned more circuitous paths, steering clear of potential obstacles and reducing safety risks, but also produced more streamlined paths conducive to smooth motor operation, thereby effectively extending the motor’s lifespan. This research provides new perspectives and methodologies for path planning for intelligent textile robotic arms, offering robust support for future advancements in intelligent manufacturing and automation technologies. Future research directions include the further optimization of the EA strategy and MIEGO method to suit a broader range of industrial applications, the exploration of integration with other advanced machine learning techniques to enhance the intelligence level of path planning, and the extension of our research methods to other types of robotic arms and automation equipment. Additionally, adaptability to noise and uncertainty in real industrial environments remains an important area for future research.

Author Contributions

Conceptualization, D.Z.; Methodology, D.Z.; Software, Z.D.; Validation, Z.D.; Formal analysis, Z.D.; Investigation, W.L.; Resources, D.Z.; Data curation, W.L.; Writing—original draft, S.Z.; Writing—review & editing, D.Z. and S.Z.; Visualization, W.L.; Supervision, Y.D.; Project administration, D.Z.; Funding acquisition, D.Z. and Y.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Tianjin Science and Technology Bureau, grant number 20YDTPJC00740. This research was funded by the Ministry of Education of the People’s Republic of China, grant number 220603032241503.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Fu, Q.; Li, Z.; Ding, Z.; Chen, J.; Luo, J.; Wang, Y.; Lu, Y. ED-DQN: An event-driven deep reinforcement learning control method for multi-zone residential buildings. Build. Environ. 2023, 242, 110546. [Google Scholar] [CrossRef]
  2. Chen, Z.; Xing, M. Upgrading of textile manufacturing based on Industry 4.0. In Proceedings of the 5th International Conference on Advanced Design and Manufacturing Engineering, Shenzhen, China, 19–20 September 2015; Atlantis Press: Amstelkade, AV, USA, 2015; pp. 2143–2146. [Google Scholar]
  3. Wang, J.; Xu, C.; Zhang, J.; Zhong, R. Big data analytics for intelligent manufacturing systems: A review. J. Manuf. Syst. 2022, 62, 738–752. [Google Scholar] [CrossRef]
  4. Sanchez, V.; Walsh, C.J.; Wood, R. Textile technology for soft robotic and autonomous garments. Adv. Funct. Mater. 2021, 31, 2008278. [Google Scholar] [CrossRef]
  5. Wang, J.; Liang, F.; Zhou, H.; Yang, M.; Wang, Q. Analysis of Position, pose and force decoupling characteristics of a 4-UPS/1-RPS parallel grinding robot. Symmetry 2022, 14, 825. [Google Scholar] [CrossRef]
  6. She, Q.; Hu, R.; Xu, J.; Liu, M.; Xu, K.; Huang, H. Learning high-DOF reaching-and-grasping via dynamic representation of gripper-object interaction. ACM Trans. Graf. 2022, 41, 1–14. [Google Scholar] [CrossRef]
  7. Niku, S.B. Introduction to Robotics: Analysis, Systems, Applications; Prentice Hall: Englewood Cliffs, NJ, USA, 2001. [Google Scholar]
  8. Hroncová, D.; Miková, Ľ.; Prada, E.; Rákay, R.; Sinčák, P.J.; Merva, T. Forward and inverse robot model kinematics and trajectory planning. In Proceedings of the 2022 20th International Conference on Mechatronics-Mechatronika (ME), Pilsen, Czech Republic, 7–9 December 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–9. [Google Scholar]
  9. Zhang, C.; Zhou, L.; Li, Y. Pareto optimal reconfiguration planning and distributed parallel motion control of mobile modular robots. In IEEE Transactions on Industrial Electronics; IEEE: Piscataway, NJ, USA, 2023. [Google Scholar]
  10. Qu, J.; Yuan, Q.; Li, Z.; Wang, Z.; Xu, F.; Fan, Q.; Zhang, M.; Qian, X.; Wang, X.; Xu, M.; et al. All-in-one strain-triboelectric sensors based on environment-friendly ionic hydrogel for wearable sensing and underwater soft robotic grasping. Nano Energy 2023, 111, 108387. [Google Scholar] [CrossRef]
  11. Linh, K.; Cox, J.; Buiyan, T.; Lambrecht, J. All-in-one: A drl-based control switch combining state-of-the-art navigation planners. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 2861–2867. [Google Scholar]
  12. Schulman, J.; Levine, S.; Abbeel, P.; Jordan, M.; Moritz, P. Trust region policy optimization. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015; PMLR: London, UK, 2015; pp. 1889–1897. [Google Scholar]
  13. Lillicrap, T.P.; Hunt, J.J.; Pritzel, A.; Heess, N.; Erez, T.; Tassa, Y.; Silver, D.; Wierstra, D. Continuous control with deep reinforcement learning. arXiv 2015, arXiv:1509.02971. [Google Scholar]
  14. Mnih, V.; Badia, A.P.; Mirza, M.; Graves, A.; Lillicrap, T.; Harley, T.; Silver, D.; Kavukcuoglu, K. Asynchronous methods for deep reinforcement learning. In Proceedings of the International Conference on Machine Learning, New York, NY, USA, 20–22 June 2016; PMLR: London, UK, 2016; pp. 1928–1937. [Google Scholar]
  15. Andrychowicz, M.; Wolski, F.; Ray, A.; Schneider, J.; Fong, R.; Welinder, P.; McGrew, B.; Tobin, J.; Pieter Abbeel, O.; Zaremba, W. Hindsight experience replay. arXiv 2017, arXiv:1707.01495v1. [Google Scholar]
  16. Schulman, J.; Wolski, F.; Dhariwal, P.; Radford, A.; Klimov, O. Proximal policy optimization algorithms. arXiv 2017, arXiv:1707.06347. [Google Scholar]
  17. Haarnoja, T.; Zhou, A.; Abbeel, P.; Levine, S. Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. In Proceedings of the International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; PMLR: London, UK, 2018; pp. 1861–1870. [Google Scholar]
  18. Fujimoto, S.; Hoof, H.; Meger, D. Addressing function approximation error in actor-critic methods. In Proceedings of the International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; PMLR: London, UK, 2018; pp. 1587–1596. [Google Scholar]
  19. Reiter, A.; Müller, A.; Gattringer, H. On higher order inverse kinematics methods in time-optimal trajectory planning for kinematically redundant manipulators. IEEE Trans. Ind. Inform. 2018, 14, 1681–1690. [Google Scholar] [CrossRef]
  20. Morel, G.; Valckenaers, P.; Faure, J.-M.; Pereira, C.E.; Diedrich, C. Manufacturing plant control challenges and issues. Control. Eng. Pract. 2007, 15, 1321–1331. [Google Scholar] [CrossRef]
  21. Sivanathan, K.; Vinayagam, B.; Samak, T.; Samak, C. Decentralized motion planning for multi-robot navigation using deep reinforcement learning. In Proceedings of the 2020 3rd International Conference on Intelligent Sustainable Systems (ICISS), Thoothukudi, India, 3–5 December 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 709–716. [Google Scholar]
  22. Dugas, D.; Nieto, J.; Siegwart, R.; Chung, J.J. Navrep: Unsupervised representations for reinforcement learning of robot navigation in dynamic human environments. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 7829–7835. [Google Scholar]
  23. Karaman, S.; Frazzoli, E. Sampling-based algorithms for optimal motion planning. Int. J. Rob. Res. 2011, 30, 846–894. [Google Scholar] [CrossRef]
  24. Lee, C. Robot arm kinematics, dynamics, and control. Computer 1982, 15, 62–80. [Google Scholar] [CrossRef]
  25. Hao, W.G.; Leck, Y.Y.; Hun, L.C. 6-DOF PC-Based Robotic Arm (PC-ROBOARM) with efficient trajectory planning and speed control. In Proceedings of the 2011 4th International Conference on Mechatronics (ICOM), Kuala Lumpur, Malaysia, 17–19 May 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 1–7. [Google Scholar]
  26. Park, S.-O.; Lee, M.C.; Kim, J. Automation; Systems. Trajectory planning with collision avoidance for redundant robots using jacobian and artificial potential field-based real-time inverse kinematics. Int. J. Control. Autom. Syst. 2020, 18, 2095–2107. [Google Scholar] [CrossRef]
  27. Huang, J.; Hu, P.; Wu, K.; Zeng, M. Optimal time-jerk trajectory planning for industrial robots. Mech. Mach. Theory 2018, 121, 530–544. [Google Scholar] [CrossRef]
  28. Ahmed, Z.; Le Roux, N.; Norouzi, M.; Schuurmans, D. Understanding the impact of entropy on policy optimization. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019; PMLR: London, UK, 2019; pp. 151–160. [Google Scholar]
  29. Xu, G.; Meng, Z.; Li, S.; Sun, Y. Collision-free trajectory planning for multi-robot simultaneous motion in preforms weaving. Robotica 2022, 40, 4218–4237. [Google Scholar] [CrossRef]
  30. Zacharias, F.; Schlette, C.; Schmidt, F.; Borst, C.; Rossmann, J.; Hirzinger, G. Making planned paths look more human-like in humanoid robot manipulation planning. In Proceedings of the 2011 IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 1192–1198. [Google Scholar]
  31. Zacharias, F.; Sepp, W.; Borst, C.; Hirzinger, G. Using a model of the reachable workspace to position mobile manipulators for 3-d trajectories. In Proceedings of the 2009 9th IEEE-RAS International Conference on Humanoid Robots, Paris, France, 7–9 December 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 55–61. [Google Scholar]
Figure 1. The present state of textile production lines.
Figure 1. The present state of textile production lines.
Processes 12 00156 g001
Figure 2. Unintended collisions stemming from local optima. (a) Scenario 1; (b) scenario 2.
Figure 2. Unintended collisions stemming from local optima. (a) Scenario 1; (b) scenario 2.
Processes 12 00156 g002
Figure 3. Phenomenon of multiple-solution conflicts. (a) Scenario 1: normal solution; (b) scenario 1: abnormal solution; (c) scenario 2: normal solution; and (d) scenario 2: abnormal solution.
Figure 3. Phenomenon of multiple-solution conflicts. (a) Scenario 1: normal solution; (b) scenario 1: abnormal solution; (c) scenario 2: normal solution; and (d) scenario 2: abnormal solution.
Processes 12 00156 g003aProcesses 12 00156 g003b
Figure 4. Environmental enhancement strategies.
Figure 4. Environmental enhancement strategies.
Processes 12 00156 g004
Figure 5. Schematic of MIEGO method.
Figure 5. Schematic of MIEGO method.
Processes 12 00156 g005
Figure 6. Virtual twin system.
Figure 6. Virtual twin system.
Processes 12 00156 g006
Figure 7. Motion range of robotic arm ER10-2000.
Figure 7. Motion range of robotic arm ER10-2000.
Processes 12 00156 g007
Figure 8. Constructed model of robotic arm and hand.
Figure 8. Constructed model of robotic arm and hand.
Processes 12 00156 g008
Figure 9. Continuous trajectory planning task for textile robotic hand.
Figure 9. Continuous trajectory planning task for textile robotic hand.
Processes 12 00156 g009
Figure 10. Training comparison between standard algorithms and EA strategy algorithms. (a) Comparison of PPO and EA-PPO training curves; (b) Comparison of SAC and EA-SAC training curves; (c) Comparison of DDPG and EA- DDPG training curves; and (d) Comparison of TD3 and EA- TD3 training curves.
Figure 10. Training comparison between standard algorithms and EA strategy algorithms. (a) Comparison of PPO and EA-PPO training curves; (b) Comparison of SAC and EA-SAC training curves; (c) Comparison of DDPG and EA- DDPG training curves; and (d) Comparison of TD3 and EA- TD3 training curves.
Processes 12 00156 g010aProcesses 12 00156 g010b
Figure 11. Stability reward variation curve of the solutions.
Figure 11. Stability reward variation curve of the solutions.
Processes 12 00156 g011
Figure 12. Radar chart comparing the performance of standard algorithms and EA strategy algorithms.
Figure 12. Radar chart comparing the performance of standard algorithms and EA strategy algorithms.
Processes 12 00156 g012
Figure 13. Final paths planned by original algorithms and EM algorithm.
Figure 13. Final paths planned by original algorithms and EM algorithm.
Processes 12 00156 g013
Figure 14. Operational conditions of joint angles in optimal paths planned by original algorithms and EM algorithm. (a) EM-SAC; (b) SAC; (c) DDPG; (d) EM-DDPG.
Figure 14. Operational conditions of joint angles in optimal paths planned by original algorithms and EM algorithm. (a) EM-SAC; (b) SAC; (c) DDPG; (d) EM-DDPG.
Processes 12 00156 g014aProcesses 12 00156 g014b
Table 1. Overall parameters of robotic arm ER10-2000.
Table 1. Overall parameters of robotic arm ER10-2000.
ModelER10-2000
Number of Axes6 axis
Wrist Load Capacity10 kg
Repeatability±0.06 mm
Weight221 kg
Maximum Reach2019 mm
Drive TypeAC Servo Motor
Table 2. Detailed data of robotic arm and hand ER10-2000.
Table 2. Detailed data of robotic arm and hand ER10-2000.
Wrist Allowable TorqueWrist Allowable Inertia TorqueMaximum Single-Axis SpeedMotion Range of Each Axis
Joint 1--170°/s±175°
Joint 2--150°/s+64°/−142°
Joint 3--146°/s+165°/−81°
Joint 422 N·m1 kg·m2360°/s±178°
Joint 522 N·m1 kg·m2360°/s±128°
Joint 616 N·m0.3 kg·m2550°/s±720°
Table 3. Key parameters of the employed DRL algorithms.
Table 3. Key parameters of the employed DRL algorithms.
PPOSACDDPGTD3
learning_rate0.00030.00030.0010.001
n_steps2048---
batch_size64256100100
gamma0.990.990.990.99
tau-0.0050.0050.005
action_noise--GaussianGaussian
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, D.; Ding, Z.; Li, W.; Zhao, S.; Du, Y. Optimization of Smart Textiles Robotic Arm Path Planning: A Model-Free Deep Reinforcement Learning Approach with Inverse Kinematics. Processes 2024, 12, 156. https://doi.org/10.3390/pr12010156

AMA Style

Zhao D, Ding Z, Li W, Zhao S, Du Y. Optimization of Smart Textiles Robotic Arm Path Planning: A Model-Free Deep Reinforcement Learning Approach with Inverse Kinematics. Processes. 2024; 12(1):156. https://doi.org/10.3390/pr12010156

Chicago/Turabian Style

Zhao, Di, Zhenyu Ding, Wenjie Li, Sen Zhao, and Yuhong Du. 2024. "Optimization of Smart Textiles Robotic Arm Path Planning: A Model-Free Deep Reinforcement Learning Approach with Inverse Kinematics" Processes 12, no. 1: 156. https://doi.org/10.3390/pr12010156

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop