Next Article in Journal
The Automatic Detection of Cognition Using EEG and Facial Expressions
Previous Article in Journal
Optimal Use of Titanium Dioxide Colourant to Enable Water Surfaces to Be Measured by Kinect Sensors
Article

A Multitasking-Oriented Robot Arm Motion Planning Scheme Based on Deep Reinforcement Learning and Twin Synchro-Control

by 1,2, 1,2,*, 1,2, 1,2 and 1,2
1
Intelligent Robotics Institute, School of Mechatronical Engineering, Beijing Institute of Technology, 5 Nandajie, Zhongguancun, Haidian, Beijing 100081, China
2
Beijing Advanced Innovation Center for Intelligent Robots and Systems, Beijing 100081, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(12), 3515; https://doi.org/10.3390/s20123515
Received: 7 April 2020 / Revised: 15 June 2020 / Accepted: 17 June 2020 / Published: 21 June 2020
(This article belongs to the Section Electronic Sensors)
Humanoid robots are equipped with humanoid arms to make them more acceptable to the general public. Humanoid robots are a great challenge in robotics. The concept of digital twin technology complies with the guiding ideology of not only Industry 4.0, but also Made in China 2025. This paper proposes a scheme that combines deep reinforcement learning (DRL) with digital twin technology for controlling humanoid robot arms. For rapid and stable motion planning for humanoid robots, multitasking-oriented training using the twin synchro-control (TSC) scheme with DRL is proposed. For switching between tasks, the robot arm training must be quick and diverse. In this work, an approach for obtaining a priori knowledge as input to DRL is developed and verified using simulations. Two simple examples are developed in a simulation environment. We developed a data acquisition system to generate angle data efficiently and automatically. These data are used to improve the reward function of the deep deterministic policy gradient (DDPG) and quickly train the robot for a task. The approach is applied to a model of the humanoid robot BHR-6, a humanoid robot with multiple-motion mode and a sophisticated mechanical structure. Using the policies trained in the simulations, the humanoid robot can perform tasks that are not possible to train with existing methods. The training is fast and allows the robot to perform multiple tasks. Our approach utilizes human joint angle data collected by the data acquisition system to solve the problem of a sparse reward in DRL for two simple tasks. A comparison with simulation results for controllers trained using the vanilla DDPG show that the designed controller developed using the DDPG with the TSC scheme have great advantages in terms of learning stability and convergence speed. View Full-Text
Keywords: deep reinforcement learning; twin synchro-control; humanoid robot deep reinforcement learning; twin synchro-control; humanoid robot
Show Figures

Figure 1

MDPI and ACS Style

Liu, C.; Gao, J.; Bi, Y.; Shi, X.; Tian, D. A Multitasking-Oriented Robot Arm Motion Planning Scheme Based on Deep Reinforcement Learning and Twin Synchro-Control. Sensors 2020, 20, 3515. https://doi.org/10.3390/s20123515

AMA Style

Liu C, Gao J, Bi Y, Shi X, Tian D. A Multitasking-Oriented Robot Arm Motion Planning Scheme Based on Deep Reinforcement Learning and Twin Synchro-Control. Sensors. 2020; 20(12):3515. https://doi.org/10.3390/s20123515

Chicago/Turabian Style

Liu, Chuzhao, Junyao Gao, Yuanzhen Bi, Xuanyang Shi, and Dingkui Tian. 2020. "A Multitasking-Oriented Robot Arm Motion Planning Scheme Based on Deep Reinforcement Learning and Twin Synchro-Control" Sensors 20, no. 12: 3515. https://doi.org/10.3390/s20123515

Find Other Styles
Note that from the first issue of 2016, MDPI journals use article numbers instead of page numbers. See further details here.

Article Access Map by Country/Region

1
Back to TopTop