Reinforcement Learning in Robotics: Applications and Real-World Challenges†
AbstractIn robotics, the ultimate goal of reinforcement learning is to endow robots with the ability to learn, improve, adapt and reproduce tasks with dynamically changing constraints based on exploration and autonomous learning. We give a summary of the state-of-the-art of reinforcement learning in the context of robotics, in terms of both algorithms and policy representations. Numerous challenges faced by the policy representation in robotics are identified. Three recent examples for the application of reinforcement learning to real-world robots are described: a pancake flipping task, a bipedal walking energy minimization task and an archery-based aiming task. In all examples, a state-of-the-art expectation-maximization-based reinforcement learning is used, and different policy representations are proposed and evaluated for each task. The proposed policy representations offer viable solutions to six rarely-addressed challenges in policy representations: correlations, adaptability, multi-resolution, globality, multi-dimensionality and convergence. Both the successes and the practical difficulties encountered in these examples are discussed. Based on insights from these particular cases, conclusions are drawn about the state-of-the-art and the future perspective directions for reinforcement learning in robotics. View Full-Text
Share & Cite This Article
Kormushev, P.; Calinon, S.; Caldwell, D.G. Reinforcement Learning in Robotics: Applications and Real-World Challenges. Robotics 2013, 2, 122-148.
Kormushev P, Calinon S, Caldwell DG. Reinforcement Learning in Robotics: Applications and Real-World Challenges. Robotics. 2013; 2(3):122-148.Chicago/Turabian Style
Kormushev, Petar; Calinon, Sylvain; Caldwell, Darwin G. 2013. "Reinforcement Learning in Robotics: Applications and Real-World Challenges." Robotics 2, no. 3: 122-148.