Next Article in Journal / Special Issue
An Improved Reinforcement Learning System Using Affective Factors
Previous Article in Journal
Psychophysiological Methods to Evaluate User’s Response in Human Robot Interaction: A Review and Feasibility Study
Previous Article in Special Issue
Computationally Efficient Adaptive Type-2 Fuzzy Control of Flexible-Joint Manipulators
Article

Reinforcement Learning in Robotics: Applications and Real-World Challenges

Department of Advanced Robotics, Istituto Italiano di Tecnologia, via Morego 30, 16163 Genova, Italy
*
Author to whom correspondence should be addressed.
Based on “Kormushev, P.; Calinon, S.; Caldwell, D.G.; Ugurlu, B. Challenges for the Policy Representation When Applying Reinforcement Learning in Robotics. In Proceedings of WCCI 2012 IEEE World Congress on Computational Intelligence, Brisbane, Australia, 10–15 June 2012”.
Robotics 2013, 2(3), 122-148; https://doi.org/10.3390/robotics2030122
Received: 4 June 2013 / Revised: 24 June 2013 / Accepted: 28 June 2013 / Published: 5 July 2013
(This article belongs to the Special Issue Intelligent Robots)
In robotics, the ultimate goal of reinforcement learning is to endow robots with the ability to learn, improve, adapt and reproduce tasks with dynamically changing constraints based on exploration and autonomous learning. We give a summary of the state-of-the-art of reinforcement learning in the context of robotics, in terms of both algorithms and policy representations. Numerous challenges faced by the policy representation in robotics are identified. Three recent examples for the application of reinforcement learning to real-world robots are described: a pancake flipping task, a bipedal walking energy minimization task and an archery-based aiming task. In all examples, a state-of-the-art expectation-maximization-based reinforcement learning is used, and different policy representations are proposed and evaluated for each task. The proposed policy representations offer viable solutions to six rarely-addressed challenges in policy representations: correlations, adaptability, multi-resolution, globality, multi-dimensionality and convergence. Both the successes and the practical difficulties encountered in these examples are discussed. Based on insights from these particular cases, conclusions are drawn about the state-of-the-art and the future perspective directions for reinforcement learning in robotics. View Full-Text
Keywords: reinforcement learning; robotics; learning and adaptive systems reinforcement learning; robotics; learning and adaptive systems
Show Figures

Graphical abstract

MDPI and ACS Style

Kormushev, P.; Calinon, S.; Caldwell, D.G. Reinforcement Learning in Robotics: Applications and Real-World Challenges. Robotics 2013, 2, 122-148. https://doi.org/10.3390/robotics2030122

AMA Style

Kormushev P, Calinon S, Caldwell DG. Reinforcement Learning in Robotics: Applications and Real-World Challenges. Robotics. 2013; 2(3):122-148. https://doi.org/10.3390/robotics2030122

Chicago/Turabian Style

Kormushev, Petar, Sylvain Calinon, and Darwin G. Caldwell 2013. "Reinforcement Learning in Robotics: Applications and Real-World Challenges" Robotics 2, no. 3: 122-148. https://doi.org/10.3390/robotics2030122

Find Other Styles

Article Access Map by Country/Region

1
Only visits after 24 November 2015 are recorded.
Back to TopTop