Autonomous Grasping of Deformable Objects with Deep Reinforcement Learning: A Study on Spaghetti Manipulation
Abstract
1. Introduction
2. Literature Reviews
2.1. Automation and Robotics in Food Handling
2.2. Deep Reinforcement Learning for Robotic Manipulation
2.3. Robotic Handling of Deformable and Delicate Objects
2.4. Multi-Sensor Integration in Robotic Systems
2.5. Applications and Limitations of Robotics in Food Automation
3. Methodology
3.1. Noodle-like Grasping Problem
3.2. Problem Formulation
- is the RGB image providing visual features including lighting conditions;
- is the depth image, representing spatial geometry;
- is the segmentation mask of the target object;
- is the grasped weight of spaghetti from selected action ;
- is the total weight of spaghetti in the tray excluding the tray weight.
3.3. System Overview
3.3.1. Hardware System
3.3.2. Software System
4. Noodle-like Grasping Framework
5. Experimental Setup and Results
5.1. Experiment Setup
Algorithm 1 Self-learning Algorithm for Spaghetti Grasping |
|
5.2. Results
6. Conclusions and Discussion
7. Future Works
- Training an agent for the spaghetti grasping task is challenging and time-consuming. When introducing a new set of configuration parameters, the agent must typically learn from scratch, which can be inefficient. To address this limitation, offline deep reinforcement learning (offline DRL) can be employed to provide the agent with a foundational understanding of noodle-like grasping environments. By leveraging a pretrained policy obtained through offline DRL, the learning process in new environments can be significantly accelerated, resulting in faster convergence and improved learning curves.
- Another promising direction for future work is the integration of auxiliary learning to enhance grasp performance. Specifically, an auxiliary task could be introduced to predict grasp success probability alongside the main reinforcement learning objective. This parallel learning objective would allow the network to learn richer representations, improve sample efficiency, and provide an additional confidence signal during both training and inference. Such auxiliary prediction could also be used to guide action selection or filter suboptimal grasps, thereby improving the overall robustness and reliability of the system.
- Sim-to-real transfer can also reduce the reliance on extensive real-world data collection. Training robotic agents entirely in simulation offers faster iteration and safer exploration, but discrepancies between simulated and real environments—commonly referred to as the reality gap—pose significant challenges. To address this, techniques such as domain randomization, domain adaptation, and fine-tuning with a small amount of real data can be explored. Incorporating more realistic physics and visual textures in simulation, as well as leveraging pretrained perception modules, could further narrow the gap. Ultimately, a robust sim-to-real strategy would enable more scalable and cost-effective deployment of robotic grasping systems in real-world applications.
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
DRL | Deep reinforcement learning |
RL | Reinforcement learning |
DCNN | Deep convolution neural network |
TD | Temporal difference |
DNN | Deep neural network |
YOLO | You Only Look Once |
MDP | Markov decision process |
KF | Kalman filter |
DoF | Degree of freedom |
ROS | Robot operating system |
SAC | Soft Actor-Critic |
FCNN | Fully connected neural network |
References
- Kaur, N.; Sharma, A. Robotics and Automation in Manufacturing Processes. In Intelligent Manufacturing; CRC Press: Boca Raton, FL, USA, 2025; pp. 97–109. [Google Scholar]
- Dzedzickis, A.; Subačiūtė-Žemaitienė, J.; Šutinys, E.; Samukaitė-Bubnienė, U.; Bučinskas, V. Advanced applications of industrial robotics: New trends and possibilities. Appl. Sci. 2021, 12, 135. [Google Scholar] [CrossRef]
- Ono, K.; Hayashi, T.; Fujii, M.; Shibasaki, N.; Sonehara, M. Development for industrial robotics applications. IHI Eng. Rev. 2009, 42, 103–107. [Google Scholar]
- de Guzman, P. How a Train Bento Box is Made in Japan. 2021. Available online: https://www.youtube.com/watch?v=eBPsaa0_RtQ&t=348s (accessed on 20 May 2025).
- Ummadisingu, A.; Takahashi, K.; Fukaya, N. Cluttered food grasping with adaptive fingers and synthetic-data trained object detection. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 8290–8297. [Google Scholar]
- Endo, G.; Otomo, N. Development of a Food Handling Gripper Considering an Appetizing. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Stockholm, Sweden, 16–21 May 2016; IEEE: Piscataway, NJ, USA, 2016. [Google Scholar]
- Chen, Y.L.; Cai, Y.R.; Cheng, M.Y. Vision-based robotic object grasping—a deep reinforcement learning approach. Machines 2023, 11, 275. [Google Scholar] [CrossRef]
- Du, G.; Wang, K.; Lian, S.; Zhao, K. Vision-based robotic grasping from object localization, object pose estimation to grasp estimation for parallel grippers: A review. Artif. Intell. Rev. 2021, 54, 1677–1734. [Google Scholar] [CrossRef]
- Dewi, T.; Risma, P.; Oktarina, Y. Fruit sorting robot based on color and size for an agricultural product packaging system. Bull. Electr. Eng. Inform. 2020, 9, 1438–1445. [Google Scholar] [CrossRef]
- Wang, Z.; Makiyama, Y.; Hirai, S. A soft needle gripper capable of grasping and piercing for handling food materials. J. Robot. Mechatronics 2021, 33, 935–943. [Google Scholar] [CrossRef]
- Kazarian, K. Robots Reach for Food Processing. Available online: https://www.foodengineeringmag.com/articles/100208-robots-reach-for-food-processing (accessed on 18 April 2022).
- Sutton, R.S. Reinforcement Learning: An Introduction; A Bradford Book; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
- Watkins, C.J.; Dayan, P. Q-learning. Mach. Learn. 1992, 8, 279–292. [Google Scholar] [CrossRef]
- Mohammed, M.Q.; Chung, K.L.; Chyi, C.S. Pick and place objects in a cluttered scene using deep reinforcement learning. Int. J. Mech. Mechatron. Eng. 2020, 20, 50–57. [Google Scholar]
- Guillaume, A. rl-taxi: Reinforcement Learning for Taxi Cab v3. 2023. Available online: https://github.com/gandroz/rl-taxi (accessed on 20 May 2025).
- de Lange, E. Escape from a Maze Using Reinforcement Learning. 2022. Available online: https://github.com/erikdelange/Reinforcement-Learning-Maze (accessed on 20 May 2025).
- Prasenjit, K. GridWorld: Gridworld Environment Creator for Testing RL Algorithms. 2024. Available online: https://github.com/prasenjit52282/GridWorld (accessed on 20 May 2025).
- Low, J.H.; Khin, P.M.; Han, Q.Q.; Yao, H.; Teoh, Y.S.; Zeng, Y.; Li, S.; Liu, J.; Liu, Z.; y Alvarado, P.V.; et al. Sensorized reconfigurable soft robotic gripper system for automated food handling. IEEE/ASME Trans. Mechatron. 2021, 27, 3232–3243. [Google Scholar] [CrossRef]
- Franco, L.; Turco, E.; Bo, V.; Pozzi, M.; Malvezzi, M.; Prattichizzo, D.; Salvietti, G. The double-scoop gripper: A tendon-driven soft-rigid end-effector for food handling exploiting constraints in narrow spaces. In Proceedings of the 2024 IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–17 May 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 4170–4176. [Google Scholar]
- Wang, Z.; Furuta, H.; Hirai, S.; Kawamura, S. A scooping-binding robotic gripper for handling various food products. Front. Robot. AI 2021, 8, 640805. [Google Scholar] [CrossRef] [PubMed]
- Tulapornpipat, W. Picking-and-Placing Food with Vacuum Pad and Soft Gripper for Industrial Robot Arm. Master’s Thesis, Chulalongkorn University, Bangkok, Thailand, 2020. [Google Scholar]
- Welch, G.; Bishop, G. An introduction to the kalman filter. In Proceedings of the SIGGRAPH, Course, Los Angeles, CA, USA, 12–17 August 2001; Volume 8, p. 41. [Google Scholar]
- Ahmed, M.; Qari, K.; Kumar, R.; Lall, B.; Kherani, A. Vision-based human wrist localization and with Kalman-filter backed stabilization for Bilateral Teleoperation of Robotic Arm. Procedia Comput. Sci. 2024, 235, 264–273. [Google Scholar] [CrossRef]
- Sun, R.; Wu, C.; Zhao, X.; Zhao, B.; Jiang, Y. Object Recognition and Grasping for Collaborative Robots Based on Vision. Sensors 2023, 24, 195. [Google Scholar] [CrossRef] [PubMed]
- Prasad, S. Application of robotics in dairy and food industries: A review. Int. J. Sci. Environ. Technol. 2017, 6, 1856–1864. [Google Scholar]
- Candemir, A.; Can, F.C. Pick &Place Task Implementation of a Scara Manipulator via Robot Operating System and Machine Vision. Available online: https://www.kalaharijournals.com/resources/JUNE-49.pdf (accessed on 20 May 2025).
- Wang, Z.; Hirai, S.; Kawamura, S. Challenges and opportunities in robotic food handling: A review. Front. Robot. AI 2022, 8, 789107. [Google Scholar] [CrossRef] [PubMed]
- Urrea, C.; Kern, J. Recent Advances and Challenges in Industrial Robotics: A Systematic Review of Technological Trends and Emerging Applications. Processes 2025, 13, 832. [Google Scholar] [CrossRef]
- Haarnoja, T.; Zhou, A.; Abbeel, P.; Levine, S. Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. In Proceedings of the International Conference on Machine Learning, PMLR, Stockholm, Sweden, 10–15 July 2018; pp. 1861–1870. [Google Scholar]
- Chen, K.; Pang, J.; Wang, J.; Xiong, Y.; Li, X.; Sun, S.; Feng, W.; Liu, Z.; Shi, J.; Ouyang, W.; et al. Hybrid task cascade for instance segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4974–4983. [Google Scholar]
Gripper Type | Design | Actuation Method | Suitable Food Types | Features |
---|---|---|---|---|
Dual-Scoop Gripper | Scoop-shaped concave jaws | Pneumatic | Fried chicken, Onigiri | High adaptability to shape |
Parallel Gripper | Finray four-finger design | Stepping Motor | Sausage-like | Strong grip for firm items |
Suction Gripper | Vacuum suction | Pneumatic | Flat or sealed surfaces | Minimal contact, hygienic |
Item | Value |
---|---|
Operating System | Linux Ubuntu 20.04 LTS |
Processor | 12th Gen Intel® Core™ i9-12900K×24 |
Memory | 64 GB |
Graphic Card | Nvidia GeForce RTX3090 24 GB |
Disk Capacity | 4 TB |
Model | Trial | Average Grasped Weight (g) | Standard Deviation (g) | Percent Error (%) |
---|---|---|---|---|
SAC + Augmentation | First | 50.2408 | 5.6022 | 0.4816 |
Second | 48.1213 | 5.2384 | 3.7573 | |
SAC | First | 45.0747 | 14.0797 | 9.8504 |
Second | 42.1971 | 15.7234 | 15.6056 |
Noodles | Average Grasped Weight (g) | Standard Deviation (g) | Percent Error (%) |
---|---|---|---|
Udon | 50.2472 | 5.8464 | 0.4944 |
Champon | 48.7638 | 5.9325 | 2.4723 |
Soba | 36.211 | 60.932 | 27.576 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Gamolped, P.; Koomklang, N.; Mowshowitz, A.; Hayashi, E. Autonomous Grasping of Deformable Objects with Deep Reinforcement Learning: A Study on Spaghetti Manipulation. Robotics 2025, 14, 113. https://doi.org/10.3390/robotics14080113
Gamolped P, Koomklang N, Mowshowitz A, Hayashi E. Autonomous Grasping of Deformable Objects with Deep Reinforcement Learning: A Study on Spaghetti Manipulation. Robotics. 2025; 14(8):113. https://doi.org/10.3390/robotics14080113
Chicago/Turabian StyleGamolped, Prem, Nattapat Koomklang, Abbe Mowshowitz, and Eiji Hayashi. 2025. "Autonomous Grasping of Deformable Objects with Deep Reinforcement Learning: A Study on Spaghetti Manipulation" Robotics 14, no. 8: 113. https://doi.org/10.3390/robotics14080113
APA StyleGamolped, P., Koomklang, N., Mowshowitz, A., & Hayashi, E. (2025). Autonomous Grasping of Deformable Objects with Deep Reinforcement Learning: A Study on Spaghetti Manipulation. Robotics, 14(8), 113. https://doi.org/10.3390/robotics14080113