Next Article in Journal
Bio-Membrane-Based Nanofiber Scaffolds: Targeted and Controlled Carriers for Drug Delivery—An Experimental In Vivo Study
Previous Article in Journal
DE-YOLOv13-S: Research on a Biomimetic Vision-Based Model for Yield Detection of Yunnan Large-Leaf Tea Trees
Previous Article in Special Issue
Energy-Efficient Fall-Detection System Using LoRa and Hybrid Algorithms
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DT-Loong: A Digital Twin Simulation Framework for Scalable Data Collection and Training of Humanoid Robots

1
Unmanned Vehicle Research Center, China North Vehicle Research Institute, Beijing 100072, China
2
The National and Local Co-Build Humanoid Robotics Innovation Centre, Shanghai 201203, China
3
Humanoid Robot (Shanghai) Co., Ltd., Shanghai 201203, China
4
Shanghai Jiao Tong University, Shanghai 200240, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Biomimetics 2025, 10(11), 725; https://doi.org/10.3390/biomimetics10110725
Submission received: 18 August 2025 / Revised: 12 October 2025 / Accepted: 14 October 2025 / Published: 1 November 2025
(This article belongs to the Special Issue Bio-Inspired Flexible Sensors)

Abstract

Recent advances in bionic intelligence are reshaping humanoid-robot design, demonstrating unprecedented agility, dexterity and task versatility. These breakthroughs drive an increasing need for large scale and high-quality data. Current data generation methods, however, are often expensive and time-consuming. To address this, we introduce Digital Twin Loong (DT-Loong), a digital twin system that combines a high-fidelity simulation environment with a full-scale virtual replica of the humanoid robot Loong, a bionic robot encompassing biomimetic joint design and movement mechanism. By integrating optical motion capture and human-to-humanoid motion re-targeting technologies, DT-Loong generates data for training and refining embodied AI models. We showcase the data collected from the system is of high quality. DT-Loong also proposes a Priority-Guided Quadratic Optimization algorithm for action retargeting, which achieves lower time delay and enhanced mapping accuracy. This approach enables real-time environmental feedback and anomaly detection, making it well-suited for monitoring and patrol applications. Our comprehensive framework establishes a foundation for humanoid robot training and further digital twin applications in humanoid robots to enhance their human-like behaviors through the emulation of biological systems and learning processes.

1. Introduction

Over the past few years, bionic intelligence and humanoid robotics have progressed at an exceptional pace. By mimicking biological morphology, materials and behavioral intelligence, humanoid robotics can yield more naturalistic locomotion, dexterous manipulation and adaptive interaction [1].
However, training such robots hinges on massive and diverse data to unlock advanced intelligent behaviors performed by humanoid robots. Teleoperation, passive observation from human videos, and hand-held grippers [2,3] are three common approaches to obtain reliable training data. While current efforts have made many robot training datasets open-source [4,5], these datasets are often limited to fixed scenarios, tasks or embodiments and are collected with high cost.
Simulation is an alternative to acquiring real-world robot data. It can alleviate issues associated with fixed scenario setups and reliance on physical robots, and thus saving both time and money. However, challenges remain in bridging the sim-to-real gaps [6] for practical applications.
Another limitation of current studies is that most robot platforms are fixed [2], primarily designed for specific table-top manipulation tasks. Consequently, there lacks sufficient data on loco-manipulation [7,8], whole-body control of humanoid robots [9,10,11], and interactions between virtual and real environments [12].
In this paper, we present DT-Loong, a multi-functional digital twin framework for humanoid robots Figure 1. By replicating real-world environments and the bionic robot Loong equipped with an IMU (Inertia Measurement Unit) and visual sensors, we build a digital environment using UE5 (Unreal Engine 5). A MoCap (Motion Capture) system with a delicately designed retargeting algorithm is used to convert human data into humanoid robot data for further training. Beyond data collection, DT-Loong also serves as a platform to test the applicability of embodied AI models, promoting the development of more bio-inspired intelligence. In addition, when Loong is deployed in monitoring and patrolling scenarios, DT-Loong can predict the Loong ’s actions and alert human operators of a surveillance system.
In summary, the main contributions of this paper are:
  • We propose DT-Loong, a digital twin system capable of low-cost, large-scale, and full-body humanoid robot data collection from human visual demonstrations, supporting both model optimization and practical monitoring applications.
  • Through the design of a retargeting algorithm and visualization techniques, we demonstrate that the data generated by DT-Loong is of high quality and strong usability, and can facilitate the training of more naturalistic robot intelligence.
  • We develop an alarming and anomaly detection framework that can trigger alerts in real-time, assisting human operators in surveillance scenarios.

2. Related Work

2.1. Data Collection for Humanoid Robots

Teleoperation is the most widely adopted method for imitation learning of robots. By deploying VR/AR controllers [13,14,15], MoCap suits [16], exoskeletons [17], haptic force feedback [18], etc., human knowledge and experience can be transferred to remotely controlled robots. Major contributions in this field include ALOHA series [2,19,20], OpenTelevision [21], HumanPlus [22], BiDex [23].
While task-space and upper-body teleoperation have attracted more academic attention, whole-body teleoperation remains significantly more challenging. These difficulties stem from differences in kinematics, the need to fully exploit the shared morphology between humans and robots and the tradeoff between similarity and feasibility [24]. Amartya Purushottam [25] develops a whole-body teleoperation frame work and a kinematic retargeting strategy for a wheeled humanoid. H2O [26] introduces a learning-based real-time whole-body teleoperation approach for a full-sized humanoid robot using an RGB camera, and OmniH2O [27] extends this approach to achieve high-precision dexterous loco-manipulation.
MoCap systems serve as a powerful tool to transfer the natural styles and synergies of human behaviors to humanoid robots. DexCap [28] provides an effective way to collect data for dexterous manipulation using a portable MoCap system. Building upon previous research, our approach integrates a MoCap device with a retargeting algorithm to enable whole-body teleoperation of a fully-sized bionic humanoid, Loong, designed with a biomimetic configuration. The resulting comprehensive framework is well-suited for large-scale humanoid data collection and model training.

2.2. Digital Twin in Robotics

The concept of digital twin was first brought up by Michael Grieves [29]. Digital twin systems have been widely used in industrial applications [30]; however, research on their use in humanoid robotics remains limited. RoboTwin [31] creates a diverse data generator for dual-arm robots using 3D generative foundation models. DTPAAL [32] demonstrates the capabilities of a digital-twinned Pepper robot in assisting elderly individuals at home. Jon Skerlj [33] proposes a safe-by-design digital twin framework for humanoid service robots.
One important application of digital twin in robotics is system monitoring, where robot path-planning [34] and collision detection play crucial roles. A clothoid approximation approach [35] using Bézier curves is developed to enhance path accuracy and planning efficiency. An efficient approach [36] based on Separating Axis Theorem (SAT) is applied to the real humanoid robot, DRC-HUBO+. Our work presents a vivid replica of a real-world environment and the robot Loong. The system DT-Loong provides a platform to assist human operators in real-world scenarios, including guarding, monitoring, and anomaly detection.

3. System Design

3.1. Overall Architecture

As shown in Figure 2, the DT-Loong system is composed of five layers. The upper two layers deal with higher-level functions and applications of the system, while the lower layers focus on core services and infrastructure.
The application layer represents the system’s usage, including data collection from visual demonstrations, real-time monitoring for patrol tasks, and algorithm validation to enhance humanoid behaviors. The functional layer manages essential features such as scenario setup, orchestration, task arrangement, etc. The service layer provides core system services, establishing foundational capabilities related to entities, processes, simulations, and associated functionalities. The data layer is responsible for managing data resources, and the infrastructure layer provides the necessary hardware and software environment.
The overall workflow is depicted in Figure 1. Based on practical applications, DT-Loong can be divided into three subsystems, as shown in Figure 3. This modular design enhances the independence of each subsystem and improves the system’s maintainability and scalability.

3.2. MoCap Data Acquisition Subsystem

Targeted at preparing data for humanoid imitation learning, the MoCap data acquisition subsystem employs optical motion capture technology to obtain Cartesian-space coordinates that represent human body motions, which are then mapped into the joint-space coordinates of Loong. After integrating these data into DT-Loong, the system outputs IMU data, camera data, and humanoid joint data in hdf5 files for subsequent robot model training.

3.3. Real-Time Monitoring Subsystem

In commercial buildings, deploying physical robots such as Loong for patrolling and monitoring provides a practical alternative to human guards. The real-time monitoring subsystem collects sensor data, joint data, and video feeds, which are transmitted to DT-Loong. Within DT-Loong, the virtual replica mirrors Loong’s movements, allowing the system to display its status and send warnings upon detection of falls, collisions, or immobilization.

3.4. Model Validation Subsystem

For widely used humanoid embodied AI models such as VLMs [37] and VLAs [38], the model validation subsystem works as a cost-effective platform to evaluate their applicability within the high-fidelity simulation environment of DT-Loong. The system can generate simulated sensor data using the USceneCaptureComponent in UE5. Furthermore, the model under test can control the virtual Loong perform movements and produce pose and displacement data, enabling performance validation through intuitive observation.

4. Implementation

4.1. Digital Twin Environment Establishment

The virtual environment and Loong is shown in Figure 4. To establish basic scenes for the practical application of DT-Loong, we apply UE5 together with Maya and Blender to construct fundamental indoor and outdoor environments incorporating weather condition simulation, high-fidelity entities, and simulation task setups. For the primary application of patrolling and monitoring, in addition to Loong, we build entities that represent a commercial hall, security devices and potential intruders.
Loong https://www.openloong.org.cn/cn/documents/robot/product_introduction (accessed on 13 October 2025) is an open-source bionic humanoid robot with 43 degrees of freedom (DoF), a height of 1.85 m, and a weight of 40 kg. Featuring a human-like joint design focused on the hip, waist, knee, and ankle regions, Loong is capable of fast walking, agile obstacle avoidance, stable uphill/downhill traversal, and robust impact resistance. To digitally represent Loong, we simplify the full-scale 3D model by reducing the number of faces and vertices and replicate its sensor setups. Specifically, two stereo cameras (ZED 2) are placed at Loong’s eyes, two blind-spot cameras (RealSense D435i, Intel Corporation, Santa Clara, CA, USA) are mounted at the middle of the body, and an IMU (Xsens MTi-630, Xsens, Enschede, The Netherlands) is positioned in the chest area.
In addition to physical simulation, we develop task-specific services and functions. PostgreSQL is employed for data acquisition, processing, storage, and database management. We also implement entity and task management, special event simulation, as well as visualization and warning mechanisms.
Our methods for addressing key challenges, such as human-to-humanoid retargeting and task scheduling, are detailed in Section 4.2 and Section 4.3. The three subsystems outlined in Section 3 rely on the integration and coordination of these fundamental components.

4.2. Robot Retargeting

Data acquisition is the primary objective of DT-Loong. We carefully design a MoCap system combining optical markers and an IMU to capture full-body human motion. A motion retargeting algorithm then maps these movements to the humanoid robot, considering joint limits and kinematics, ensuring that the motions are robot-executable and the resulting data are suitable for training embodied AI models.
As depicted in Figure 5, our MoCap system is set up in a 100 m2 square room, with sixteen MC1300 infrared cameras mounted on the truss around the ceiling. We place 53 markers on the human body to capture motion data. The data collected from these markers are then passed to the MoCap module for human-to-humanoid retargeting.
To convert human motion data into humanoid joint state parameters, we adopt a hierarchical QP (PGQO (Priority-Guided Quadratic Optimization)) formulation. Each task is defined by equality and/or inequality constraints with respect to the decision variable x. Tasks sharing the same priority i are vertically stacked into a composite task T i :
T i : A i x b i , D i x f i ,
Here, A i x b i denotes equality-type tracking objectives that will be enforced in a least-squares sense when forming the QP at priority i, and D i x f i denotes inequality constraints for the same priority. The optimization variable x represents joint angular velocities. Variables A, b, D, f are defined in Table 1. Specifically, J 1 is the Jacobian of end-effector pose error, J 2 is the Jacobian of body-segment direction, and J 3 is the Jacobian of the manipulability index [39]. The vector b is the PD term associated with pose error, and I is the identity matrix.

4.2.1. Single-Priority QP

At priority i, we solve the following weighted least-squares QP with nonnegative slack v i :
min x , v i 1 2 λ i A i x b i 2 + 1 2 μ i v i 2 s . t . D i x f i v i , v i 0 ,
with scalar weights λ i , μ i > 0 . The slack v i vanishes whenever the inequalities are feasible; if not, v i > 0 provides the minimal relaxation required to retain feasibility.

4.2.2. Hierarchical Optimization

To guarantee task prioritization, we solve a sequence of QPs. At priority i, we optimize over the set of optimal solutions of all higher priorities:
h i ( x k * , v k * ) = arg min x , v A i x b i W i 2 + v V i 2 s . t . D k x f k v k * , k < i , A k ( x x k * ) = 0 , k < i , v 0 ,
where ( x k * , v k * ) is the solution at priority k < i .

4.2.3. Canonical QP Form

Let ξ i = ( z i , v i ) . Then
min ξ i 1 2 ξ i H i ξ i + g i ξ i s . t . D ˜ i ξ i f ˜ i ,
with
H i = Z i 1 A ˜ i A ˜ i Z i 1 0 0 V i V i , g i = Z i 1 A ˜ i ( A ˜ i x i 1 * b ˜ i ) ,
D ˜ i = D 1 Z i 1 0 D i 1 Z i 1 0 D i Z i 1 I 0 I , f ˜ i = f 1 D 1 x i 1 * + v 1 * f i 1 D i 1 x i 1 * + v i 1 * .

4.2.4. Null-Space Recursion

To preserve higher-priority equalities, we compute the null-space basis recursively:
Z 0 = I , A i , pre = A i Z i 1 , Z i = Z i 1 N ( A i , pre ) ,
where N ( · ) returns an orthonormal basis of the null space.

4.3. Task Scheduling

We design a task scheduling system that manages tasks based on priority. It features task parameter editing, task instance execution, task lifecycle management, cross-task scheduling, and script management.
There are three sources of tasks: coordination tasks, which directly interrupt current tasks; to-do scenario tasks, which can be selected from the queue or suspended pool for execution; and pre-configured autonomous actions. We manage these tasks to enable unified scheduling and state transitions.
Additionally, we perform serialization and deserialization to streamline and accelerate the process. The event serialization function efficiently organizes and encodes complex event data structures generated by the event editing module, transforming them into a serialized format for easy storage. Specifically, the events are encoded into JSON string format, with RapidJSON used for encoding and parsing the JSON. During the actual training phase, the serialized event information is decoded and restored. Through deserialization, the original event model and attributes are reconstructed, ensuring that the events are executed precisely within their preset temporal logic framework.

5. Experimental Results

5.1. Scenario Setup

To evaluate the applicability of DT-Loong, we design multiple scenarios to rigorously test its performance. These include an indoor environment where Loong is deployed for patrolling and monitoring, as illustrated in Figure 4, and an outdoor setting where Loong is applied to capture human motions and perform Tai Chi exercises, as depicted in Figure 8. These diverse environments enable a practical assessment of the system’s adaptability and robustness across a variety of real-world contexts.

5.2. Status Visualization

DT-Loong provides a user-friendly interface that allows users to easily observe sensor data, humanoid movements and the virtual Loong’s interactions with the environment. Figure 6 presents the observation window, displaying feeds from the two stereo cameras, two blind-spot cameras, and data gathered from the IMU. Figure 7 shows three sample joint angle curves of Loong, which are captured from a video demo of the virtual Loong practicing Tai Chi, as shown in Figure 8. These visualizations facilitate a clearer understanding of the complex dynamics involved in the Tai Chi movement sequences.

5.3. Real-World Applications

Indoor spaces such as banks, offices and shopping malls typically require a large number of guards to maintain order and ensure safety. Deploying Loong in these environments with the support of DT-Loong provides a practical alternative to reduce human labor. Once the virtual Loong falls or collides with an object (such as an intruder), or remains static for an extended period, our DT-Loong system will send out warnings to alert human operators to handle the situation, as shown in Figure 9. Figure 10 illustrates the virtual representation of Loong mimicking the walking pose of its physical counterpart. It demonstrates the effectiveness of our monitoring subsystem and highlights its applicability in practical surveillance tasks.

6. Discussion and Conclusions

In this study, we propose a novel digital twin framework for humanoid data collection, status monitoring, and algorithm evaluation. By incorporating an optical MoCap system and a hierarchical quadratic programming method, our approach addresses the challenges of large-scale full-body humanoid data collection in a cost-effective manner. While the monitoring system currently emphasizes Loong’s locomotion, it does not yet include mapping of human hand motions. Another limitation is that, as the current work focuses on system functionality and practical applications, quantitative evaluations of data quality and mapping performance are not fully addressed. Future efforts will be put into enhancing system responsiveness through improved data compression and networking techniques. Additionally, by developing a more fine-grained physical simulation mechanism–improving simulation speed, diversity and accuracy, DT-Loong could serve as a comprehensive digital twin platform for training, evaluating, and deploying bionic humanoid robots across diverse operational scenarios.

Author Contributions

Conceptualization, Y.L. (Yongyao Li) and Y.R.; Writing—original draft preparation, Y.L. (Yang Li); Writing—review and editing, Y.L. (Yufei Liu), Y.L. (Yongyao Li), J.D. and Y.L. (Yang Li); Methodology, Y.R., Y.L. (Yang Li), J.D. and Y.L. (Yongyao Li); Formal analysis, Y.L. (Yang Li), J.D. and Y.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number 52205035.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding authors.

Acknowledgments

The authors sincerely acknowledge the collaborative efforts and dedication of all team members involved in this work. Special thanks are extended to Lei Jiang for his significant contributions to the overall project design and guidance throughout the research. We also express our gratitude to the corresponding author, Yufei Liu, for his invaluable leadership, coordination, and support in bringing this study to completion. The success of this research is a testament to the combined commitment and expertise of everyone involved.

Conflicts of Interest

Authors Yufei Liu, Yang Li, Jinda Du, Yanjie Rui and Yongyao Li were employed by the company Humanoid Robot (Shanghai) Co., Ltd. The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Tong, Y.; Liu, H.; Zhang, Z. Advancements in Humanoid Robots: A Comprehensive Review and Future Prospects. IEEE/CAA J. Autom. Sin. 2024, 11, 301–328. [Google Scholar] [CrossRef]
  2. Zhao, T.; Kumar, V.; Levine, S.; Finn, C. Learning Fine-Grained Bimanual Manipulation with Low-Cost Hardware. In Proceedings of the Robotics: Science and Systems XIX, Daegu, Republic of Korea, 10–14 July 2023. [Google Scholar] [CrossRef]
  3. Chi, C.; Xu, Z.; Pan, C.; Cousineau, E.; Burchfiel, B.; Feng, S.; Tedrake, R.; Song, S. Universal Manipulation Interface: In-The-Wild Robot Teaching Without In-The-Wild Robots. In Proceedings of the Robotics: Science and Systems XX, Delft, The Netherlands, 15–19 July 2024. [Google Scholar] [CrossRef]
  4. O’Neill, A.; Rehman, A.; Maddukuri, A.; Gupta, A.; Padalkar, A.; Lee, A.; Pooley, A.; Gupta, A.; Mandlekar, A.; Jain, A.; et al. Open X-Embodiment: Robotic Learning Datasets and RT-X Models: Open X-Embodiment Collaboration0. In Proceedings of the 2024 IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–17 May 2024; pp. 6892–6903. [Google Scholar] [CrossRef]
  5. Khazatsky, A.; Pertsch, K.; Nair, S.; Balakrishna, A.; Dasari, S.; Karamcheti, S.; Nasiriany, S.; Srirama, M.; Chen, L.; Ellis, K.; et al. DROID: A Large-Scale In-The-Wild Robot Manipulation Dataset. In Proceedings of the Robotics: Science and Systems XX, Delft, The Netherlands, 15–19 July 2024. [Google Scholar] [CrossRef]
  6. Zhao, W.; Queralta, J.P.; Westerlund, T. Sim-to-Real Transfer in Deep Reinforcement Learning for Robotics: A Survey. In Proceedings of the 2020 IEEE Symposium Series on Computational Intelligence (SSCI), Canberra, ACT, Australia, 1–4 December 2020; pp. 737–744. [Google Scholar] [CrossRef]
  7. Murooka, M.; Kumagai, I.; Morisawa, M.; Kanehiro, F.; Kheddar, A. Humanoid Loco-Manipulation Planning Based on Graph Search and Reachability Maps. IEEE Robot. Autom. Lett. 2021, 6, 1840–1847. [Google Scholar] [CrossRef]
  8. Rigo, A.; Chen, Y.; Gupta, S.K.; Nguyen, Q. Contact Optimization for Non-Prehensile Loco-Manipulation via Hierarchical Model Predictive Control. In Proceedings of the 2023 IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; pp. 9945–9951. [Google Scholar] [CrossRef]
  9. Sentis, L.; Khatib, O. A whole-body control framework for humanoids operating in human environments. In Proceedings of the 2006 IEEE International Conference on Robotics and Automation (ICRA 2006), Orlando, FL, USA, 15–19 May 2006; pp. 2641–2648. [Google Scholar] [CrossRef]
  10. Gienger, M.; Janssen, H.; Goerick, C. Task-oriented whole body motion for humanoid robots. In Proceedings of the 5th IEEE-RAS International Conference on Humanoid Robots, Tsukuba, Japan, 5 December 2005; pp. 238–244. [Google Scholar] [CrossRef]
  11. Cheng, X.; Ji, Y.; Chen, J.; Yang, R.; Yang, G.; Wang, X. Expressive Whole-Body Control for Humanoid Robots. In Proceedings of the Robotics: Science and Systems XX, Delft, The Netherlands, 15–19 July 2024. [Google Scholar] [CrossRef]
  12. Burdea, G. Invited review: The synergy between virtual reality and robotics. IEEE Trans. Robot. Autom. 1999, 15, 400–410. [Google Scholar] [CrossRef]
  13. Zhang, T.; McCarthy, Z.; Jow, O.; Lee, D.; Chen, X.; Goldberg, K.; Abbeel, P. Deep Imitation Learning for Complex Manipulation Tasks from Virtual Reality Teleoperation. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, QLD, Australia, 21–25 May 2018; IEEE: New York, NY, USA, 2018; pp. 5628–5635. [Google Scholar] [CrossRef]
  14. Bates, T.; Ramirez-Amaro, K.; Inamura, T.; Cheng, G. On-line simultaneous learning and recognition of everyday activities from virtual reality performances. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 3510–3515. [Google Scholar] [CrossRef]
  15. Kaigom, E.G.; Roßmann, J. Developing virtual testbeds for intelligent robot manipulators—An eRobotics approach. In Proceedings of the 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems, Chicago, IL, USA, 14–18 September 2014; pp. 1589–1594. [Google Scholar] [CrossRef]
  16. Liu, H.; Xie, X.; Millar, M.; Edmonds, M.; Gao, F.; Zhu, Y.; Santos, V.J.; Rothrock, B.; Zhu, S.C. A glove-based system for studying hand-object manipulation via joint pose and force sensing. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 6617–6624. [Google Scholar] [CrossRef]
  17. Fang, H.; Fang, H.S.; Wang, Y.; Ren, J.; Chen, J.; Zhang, R.; Wang, W.; Lu, C. AirExo: Low-Cost Exoskeletons for Learning Whole-Arm Manipulation in the Wild. In Proceedings of the 2024 IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–17 May 2024; pp. 15031–15038. [Google Scholar] [CrossRef]
  18. Zhang, H.; Hu, S.; Yuan, Z.; Xu, H. DOGlove: Dexterous Manipulation with a Low-Cost Open-Source Haptic Force Feedback Glove. In Proceedings of the Robotics: Science and Systems 2025, Los Angeles, CA, USA, 21–25 June 2025. [Google Scholar] [CrossRef]
  19. Aldaco, J.; Armstrong, T.; Baruch, R.; Bingham, J.; Chan, S.; Dwibedi, D.; Finn, C.; Florence, P.; Goodrich, S.; Gramlich, W.; et al. ALOHA 2: An Enhanced Low-Cost Hardware for Bimanual Teleoperation. arXiv 2024, arXiv:2405.02292. [Google Scholar]
  20. Fu, Z.; Zhao, T.Z.; Finn, C. Mobile ALOHA: Learning Bimanual Mobile Manipulation using Low-Cost Whole-Body Teleoperation. In Proceedings of the 8th Conference on Robot Learning, Munich, Germany, 6–9 November 2024; Agrawal, P., Kroemer, O., Burgard, W., Eds.; Proceedings of Machine Learning Research. PMLR: New York, NY, USA, 2024; Volume 270, pp. 4066–4083. [Google Scholar]
  21. Cheng, X.; Li, J.; Yang, S.; Yang, G.; Wang, X. Open-TeleVision: Teleoperation with Immersive Active Visual Feedback. In Proceedings of the 8th Conference on Robot Learning, Munich, Germany, 6–9 November 2024; Agrawal, P., Kroemer, O., Burgard, W., Eds.; Proceedings of Machine Learning Research. PMLR: New York, NY, USA, 2024; Volume 270, pp. 2729–2749. [Google Scholar]
  22. Fu, Z.; Zhao, Q.; Wu, Q.; Wetzstein, G.; Finn, C. HumanPlus: Humanoid Shadowing and Imitation from Humans. In Proceedings of the 8th Annual Conference on Robot Learning, Munich, Germany, 6–9 November 2024. [Google Scholar]
  23. Shaw, K.; Li, Y.; Yang, J.; Srirama, M.K.; Liu, R.; Xiong, H.; Mendonca, R.; Pathak, D. Bimanual Dexterity for Complex Tasks. In Proceedings of the 8th Annual Conference on Robot Learning, Munich, Germany, 6–9 November 2024. [Google Scholar]
  24. Penco, L.; Clement, B.; Modugno, V.; Mingo Hoffman, E.; Nava, G.; Pucci, D.; Tsagarakis, N.G.; Mouret, J.B.; Ivaldi, S. Robust Real-Time Whole-Body Motion Retargeting from Human to Humanoid. In Proceedings of the 2018 IEEE-RAS 18th International Conference on Humanoid Robots (Humanoids), Beijing, China, 6–9 November 2018; pp. 425–432. [Google Scholar] [CrossRef]
  25. Purushottam, A.; Xu, C.; Jung, Y.; Ramos, J. Dynamic Mobile Manipulation via Whole-Body Bilateral Teleoperation of a Wheeled Humanoid. IEEE Robot. Autom. Lett. 2024, 9, 1214–1221. [Google Scholar] [CrossRef]
  26. He, T.; Luo, Z.; Xiao, W.; Zhang, C.; Kitani, K.; Liu, C.; Shi, G. Learning Human-to-Humanoid Real-Time Whole-Body Teleoperation. In Proceedings of the 2024 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Abu Dhabi, United Arab Emirates, 14–18 October 2024; pp. 8944–8951. [Google Scholar] [CrossRef]
  27. He, T.; Luo, Z.; He, X.; Xiao, W.; Zhang, C.; Zhang, W.; Kitani, K.M.; Liu, C.; Shi, G. OmniH2O: Universal and Dexterous Human-to-Humanoid Whole-Body Teleoperation and Learning. In Proceedings of the 8th Annual Conference on Robot Learning, Munich, Germany, 6–9 November 2024. [Google Scholar]
  28. Wang, C.; Shi, H.; Wang, W.; Zhang, R.; Fei-Fei, L.; Liu, K. DexCap: Scalable and Portable Mocap Data Collection System for Dexterous Manipulation. In Proceedings of the Robotics: Science and Systems XX, Delft, The Netherlands, 15–19 July 2024. [Google Scholar] [CrossRef]
  29. Grieves, M. Virtually Perfect: Driving Innovative and Lean Products Through Product Lifecycle Management; Space Coast Press: Rockledge, FL, USA, 2011. [Google Scholar]
  30. Tao, F.; Zhang, H.; Liu, A.; Nee, A.Y.C. Digital Twin in Industry: State-of-the-Art. IEEE Trans. Ind. Inform. 2019, 15, 2405–2415. [Google Scholar] [CrossRef]
  31. Mu, Y.; Chen, T.; Peng, S.; Chen, Z.; Gao, Z.; Zou, Y.; Lin, L.; Xie, Z.; Luo, P. RoboTwin: Dual-Arm Robot Benchmark with Generative Digital Twins (Early Version). In Proceedings of the Computer Vision—ECCV 2024 Workshops, Milan, Italy, 29 September–4 October 2024; Del Bue, A., Canton, C., Pont-Tuset, J., Tommasi, T., Eds.; Springer: Cham, Switerland, 2025; pp. 264–273. [Google Scholar] [CrossRef]
  32. Cascone, L.; Nappi, M.; Narducci, F.; Passero, I. DTPAAL: Digital Twinning Pepper and Ambient Assisted Living. IEEE Trans. Ind. Inform. 2022, 18, 1397–1404. [Google Scholar] [CrossRef]
  33. Škerlj, J.; Hamad, M.; Elsner, J.; Naceri, A.; Haddadin, S. Safe-By-Design Digital Twins for Human-Robot Interaction: A Use Case for Humanoid Service Robots. In Proceedings of the 2024 IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–17 May 2024; pp. 13369–13375. [Google Scholar] [CrossRef]
  34. Allen, J.; Dorosh, R.; Ninatanta, C.; Allen, A.; Shui, L.; Yoshida, K.; Luo, J.; Luo, M. Modeling and Experimental Verification of a Continuous Curvature-Based Soft Growing Manipulator. IEEE Robot. Autom. Lett. 2024, 9, 3594–3600. [Google Scholar] [CrossRef]
  35. Chen, Y.; Cai, Y.; Zheng, J.; Thalmann, D. Accurate and Efficient Approximation of Clothoids Using Bézier Curves for Path Planning. IEEE Trans. Robot. 2017, 33, 1242–1247. [Google Scholar] [CrossRef]
  36. Lee, I.; Lee, K.K.; Sim, O.; Woo, K.S.; Buyoun, C.; Oh, J.H. Collision detection system for the practical use of the humanoid robot. In Proceedings of the 2015 IEEE-RAS 15th International Conference on Humanoid Robots (Humanoids), Seoul, Republic of Korea, 3–5 November 2015; pp. 972–976. [Google Scholar] [CrossRef]
  37. Akiyama, S.; Dossa, R.F.J.; Arulkumaran, K.; Sujit, S.; Johns, E. Open-loop VLM Robot Planning: An Investigation of Fine-tuning and Prompt Engineering Strategies. In Proceedings of the First Workshop on Vision-Language Models for Navigation and Manipulation at ICRA 2024, Seoul, Republic of Korea, 3–5 November 2024. [Google Scholar]
  38. Kim, M.J.; Pertsch, K.; Karamcheti, S.; Xiao, T.; Balakrishna, A.; Nair, S.; Rafailov, R.; Foster, E.P.; Sanketi, P.R.; Vuong, Q.; et al. OpenVLA: An Open-Source Vision-Language-Action Model. In Proceedings of the 8th Conference on Robot Learning, Munich, Germany, 6–9 November 2024; Agrawal, P., Kroemer, O., Burgard, W., Eds.; Proceedings of Machine Learning Research. PMLR: New York, NY, USA, 2024; Volume 270, pp. 2679–2713. [Google Scholar]
  39. Dufour, K.; Suleiman, W. On integrating manipulability index into inverse kinematics solver. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; IEEE: New York, NY, USA, 2017; pp. 6967–6972. [Google Scholar]
Figure 1. System overview. DT-Loong is a novel framework that leverages a motion capture module to record human motion and map it into Loong’s joint space. The digital twin replica then updates and outputs joint space states and sensor feedback, supporting model training and performance evaluation. DT-Loong also enables real-world implementations such as monitoring tasks and serves as an adaptable testbed for optimizing embodied AI models, making it a versatile tool for both research and practical deployment.
Figure 1. System overview. DT-Loong is a novel framework that leverages a motion capture module to record human motion and map it into Loong’s joint space. The digital twin replica then updates and outputs joint space states and sensor feedback, supporting model training and performance evaluation. DT-Loong also enables real-world implementations such as monitoring tasks and serves as an adaptable testbed for optimizing embodied AI models, making it a versatile tool for both research and practical deployment.
Biomimetics 10 00725 g001
Figure 2. System architecture illustrating the hierarchical layers from infrastructure to application, detailing the core components involved in visual training, real-time monitoring, and model validation.
Figure 2. System architecture illustrating the hierarchical layers from infrastructure to application, detailing the core components involved in visual training, real-time monitoring, and model validation.
Biomimetics 10 00725 g002
Figure 3. Subsystems of DT-Loong. (a) The MoCap data acquisition subsystem collects human motion data and generates corresponding humanoid joint data for robot training. (b) The real-time monitoring subsystem monitors robot movements by processing sensor, joint, and video data. (c) The model validation subsystem uses the humanoid robot’s data to test and improve AI models.
Figure 3. Subsystems of DT-Loong. (a) The MoCap data acquisition subsystem collects human motion data and generates corresponding humanoid joint data for robot training. (b) The real-time monitoring subsystem monitors robot movements by processing sensor, joint, and video data. (c) The model validation subsystem uses the humanoid robot’s data to test and improve AI models.
Biomimetics 10 00725 g003
Figure 4. Virtual environment and humanoid.
Figure 4. Virtual environment and humanoid.
Biomimetics 10 00725 g004
Figure 5. MoCap system. (a) MoCap setup. (b) MoCap demonstration.
Figure 5. MoCap system. (a) MoCap setup. (b) MoCap demonstration.
Biomimetics 10 00725 g005
Figure 6. Sensor board displaying camera and IMU information (demo). It aggregates live views from the ZED2 stereo RGB camera (left/right) and the Intel RealSense D435i (left/right and depth), together with IMU readouts–angular velocity, linear acceleration, and orientation—from an Xsens MTi-630.
Figure 6. Sensor board displaying camera and IMU information (demo). It aggregates live views from the ZED2 stereo RGB camera (left/right) and the Intel RealSense D435i (left/right and depth), together with IMU readouts–angular velocity, linear acceleration, and orientation—from an Xsens MTi-630.
Biomimetics 10 00725 g006
Figure 7. Sample joint angle curves from Tai Chi practices locating at Loong ’s arm, waist, and head.
Figure 7. Sample joint angle curves from Tai Chi practices locating at Loong ’s arm, waist, and head.
Biomimetics 10 00725 g007
Figure 8. Movements of the virtual Loong performing Tai Chi. It illustrates natural, human-like actions produced by our retargeting algorithm.
Figure 8. Movements of the virtual Loong performing Tai Chi. It illustrates natural, human-like actions produced by our retargeting algorithm.
Biomimetics 10 00725 g008
Figure 9. System alert.
Figure 9. System alert.
Biomimetics 10 00725 g009
Figure 10. Application of the monitoring subsystem.
Figure 10. Application of the monitoring subsystem.
Biomimetics 10 00725 g010
Table 1. Optimization tasks.
Table 1. Optimization tasks.
TasksPriorityAbDf
Upper body end-effectors 6D pose tracking0 J 1 K p E r r + K i 0 t E r r ( r ) d r I I q ˙ m a x q ˙ m i n
Whole body segment orientation tracking1 J 2 K p i E r r i + K i i 0 t E r r i ( r ) d r nonenone
Manipulability index1 J 3 K p E r r i + K i 0 t E r r i ( r ) d r nonenone
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, Y.; Li, Y.; Du, J.; Rui, Y.; Li, Y. DT-Loong: A Digital Twin Simulation Framework for Scalable Data Collection and Training of Humanoid Robots. Biomimetics 2025, 10, 725. https://doi.org/10.3390/biomimetics10110725

AMA Style

Liu Y, Li Y, Du J, Rui Y, Li Y. DT-Loong: A Digital Twin Simulation Framework for Scalable Data Collection and Training of Humanoid Robots. Biomimetics. 2025; 10(11):725. https://doi.org/10.3390/biomimetics10110725

Chicago/Turabian Style

Liu, Yufei, Yang Li, Jinda Du, Yanjie Rui, and Yongyao Li. 2025. "DT-Loong: A Digital Twin Simulation Framework for Scalable Data Collection and Training of Humanoid Robots" Biomimetics 10, no. 11: 725. https://doi.org/10.3390/biomimetics10110725

APA Style

Liu, Y., Li, Y., Du, J., Rui, Y., & Li, Y. (2025). DT-Loong: A Digital Twin Simulation Framework for Scalable Data Collection and Training of Humanoid Robots. Biomimetics, 10(11), 725. https://doi.org/10.3390/biomimetics10110725

Article Metrics

Back to TopTop