Next Article in Journal
DFC-LSTM: A Novel LSTM Architecture Integrating Dynamic Fractal Gating and Chaotic Activation for Value-at-Risk Forecasting
Previous Article in Journal
Nonlinear Stochastic Adaptive Control of PEMFC Temperature Management System
Previous Article in Special Issue
Frequency-Domain Hybrid Model for EEG-Based Emotion Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

Progressive Policy Learning: A Hierarchical Framework for Dexterous Bimanual Manipulation

Department of Mechanical, Robotics and Energy Engineering, Dongguk University, 30, Pildong-ro 1gil, Jung-gu, Seoul 04620, Republic of Korea
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(22), 3585; https://doi.org/10.3390/math13223585 (registering DOI)
Submission received: 19 September 2025 / Revised: 24 October 2025 / Accepted: 7 November 2025 / Published: 8 November 2025

Abstract

Dexterous bimanual manipulation remains a challenging task in reinforcement learning (RL) due to the vast state–action space and the complex interdependence between the hands. Conventional end-to-end learning struggles to handle this complexity, and multi-agent RL often faces limitations in stably acquiring cooperative movements. To address these issues, this study proposes a hierarchical progressive policy learning framework for dexterous bimanual manipulation. In the proposed method, one hand’s policy is first trained to stably grasp the object, and, while maintaining this grasp, the other hand’s manipulation policy is progressively learned. This hierarchical decomposition reduces the search space for each policy and enhances both the connectivity and the stability of learning by training the subsequent policy on the stable states generated by the preceding policy. Simulation results show that the proposed framework outperforms conventional end-to-end and multi-agent RL approaches. The proposed method was demonstrated via sim-to-real transfer on a physical dual-arm platform and empirically validated on a bimanual cube manipulation task.
Keywords: reinforcement learning; robot manipulation; artificial intelligence; machine learning; dexterous robotic hand reinforcement learning; robot manipulation; artificial intelligence; machine learning; dexterous robotic hand
Graphical Abstract

Share and Cite

MDPI and ACS Style

Lee, K.-W.; Lee, J.-W.; Kim, S.; Lim, S.-C. Progressive Policy Learning: A Hierarchical Framework for Dexterous Bimanual Manipulation. Mathematics 2025, 13, 3585. https://doi.org/10.3390/math13223585

AMA Style

Lee K-W, Lee J-W, Kim S, Lim S-C. Progressive Policy Learning: A Hierarchical Framework for Dexterous Bimanual Manipulation. Mathematics. 2025; 13(22):3585. https://doi.org/10.3390/math13223585

Chicago/Turabian Style

Lee, Kang-Won, Jung-Woo Lee, Seongyong Kim, and Soo-Chul Lim. 2025. "Progressive Policy Learning: A Hierarchical Framework for Dexterous Bimanual Manipulation" Mathematics 13, no. 22: 3585. https://doi.org/10.3390/math13223585

APA Style

Lee, K.-W., Lee, J.-W., Kim, S., & Lim, S.-C. (2025). Progressive Policy Learning: A Hierarchical Framework for Dexterous Bimanual Manipulation. Mathematics, 13(22), 3585. https://doi.org/10.3390/math13223585

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop