Optimized Design and Deep Vision-Based Operation Control of a Multi-Functional Robotic Gripper for an Automatic Loading System
Abstract
1. Introduction
2. Automatic Loading System
2.1. System Architecture and Composition
- 1.
- Standardized Load-Carrying Platform Module: This can include mining flatbed trucks, cargo trucks, trains, and ships.
- 2.
- Modular Transport Container Module: Designed in various forms to carry different materials. As shown on the left side of Figure 1a, each container is equipped with eight corner fittings—four at the bottom for securing the container, and four at the top for lifting and transport.
- 3.
- Locking Module: Includes manually adjustable twist locks and fully automatic twist locks. The manual twist locks are mounted on the standardized platform and can raise the locking head to secure the transport container. When locking is not required, the lock head can be lowered to convert the platform into a standard flatbed. The fully automatic twist locks are used to interconnect stacked containers.
- 4.
- Overhead Crane and Spreader Module: Responsible for loading and unloading the transport containers onto and from the standardized platform, as well as for stacking containers within the warehouse.
- 5.
- Multi-Sensor Recognition and Localization Module: Comprising four cameras, an IMU (Inertial Measurement Unit), and UWB (Ultra Wide Band) positioning modules, LD150 from Haoru Tech (Dalian, China). The cameras detect the four bottom corners of the containers and the handles of the manual twist locks within the work zone, providing feedback for automated operation. The IMU and UWB modules estimate the container’s position and orientation outside the working zone.
- 6.
- Multi-Robot Arm Collaborative Loading and Unloading Module: Installed in the working zone, consisting of four robotic arms and multifunctional grippers. This module facilitates the automated transfer of containers between the platform and the storage area.
2.2. Loading and Unloading Process
2.2.1. Loading Process
2.2.2. Unloading Process
3. Optimized Design
3.1. Tasks of the Multi-Robot Arm Collaborative Loading and Unloading Module
- Handling fully automatic twist locks;
- Locking and unlocking manual adjustable twist locks;
- Correcting the posture of the transport container;
- Positioning and securing the transport container.
3.2. Operation Process
3.2.1. Locking Procedure
- 1.
- Initial Grasping: The robotic arm moves from its initial position to a designated position where the gripper closes to securely grasp the twist lock.
- 2.
- Alignment: The robotic arm adjusts its posture and executes a translational motion to align the twist lock with the mounting port of the corner piece.
- 3.
- Insertion: The robotic arm performs a vertical motion along the z-axis, raising the twist lock to a certain height until it enters the corner piece.
- 4.
- Rotation: The robotic arm rotates the twist lock counterclockwise by about its own central axis, which is parallel to the z-axis (viewed from above).
- 5.
- Release: The gripper opens to release the twist lock, allowing it to be securely installed within the corner piece.
- 6.
- Return: The robotic arm returns to its initial position.
3.2.2. Unlocking Procedure
- 1.
- Approach: The robotic arm moves from its initial position to a designated location with a specific posture to approach and grasp the twist lock.
- 2.
- Grasping: The gripper closes to securely hold the twist lock.
- 3.
- Rotation: The robotic arm rotates the twist lock clockwise by about its own central axis, which is parallel to the z-axis (viewed from above).
- 4.
- Extraction: The robotic arm performs a downward motion along the z-axis, lowering the twist lock until it exits the corner piece.
- 5.
- Placement: The robotic arm moves the twist lock to a designated placement location, where the gripper opens to release the twist lock.
- 6.
- Return: The robotic arm returns to its initial position.
3.3. Optimized Design of the Multi-Functional Robotic Gripper
3.3.1. Type I Gripper (90° Between Gripper and Container Side)
3.3.2. Type II Gripper (0° Between Gripper and Container Side)
3.3.3. Type III Gripper (30° Between Gripper and Container Side)
4. Sensing Based on Deep Vision
4.1. Improved YOLOv5 Object Detection Algorithm
- Lightweight Feature Extraction Network: Reduces model parameters, ensuring efficient processing.
- Dynamic Resolution Input Strategy: Adapts to different computational constraints across various scenarios.
- Channel Attention Mechanism: Improves the discriminative capability of feature representations.
4.2. Multi-Feature Fusion Recognition
- Corner piece (Corner);
- Flatbed truck edge (Edge);
- Flatbed truck twist lock horizontal component (Lift_lock_1);
- Flatbed truck twist lock vertical component (Lift_lock_2);
- Twist lock Orientation (F/B/L/R);
- Automatic twist lock (Auto_lock).
- Original coordinates of Auto_lock, Lift_lock_1, Lift_lock_2, and Edge are spatially registered and transformed to match the Corner coordinate system.
- Kalman filtering is applied for temporal prediction and dynamic weighted fusion of observations, effectively suppressing sensor noise and motion blur.
- For Corner features, depth information is derived from the centroid depth of the bounding box’s four corner pixel points.
- For other features, the minimum depth value within the bounding box is utilized, balancing geometric precision and robustness against occlusion.
- A depth thresholding mechanism is implemented to filter erroneous data.
5. Operation Control
5.1. Kinematics Model
5.1.1. Forward Kinematics
- is the joint angle.
- is the link offset.
- is the link length.
- is the twist angle.
5.1.2. Inverse Kinematics
- Position: Solving for , , and based on the position vector .
- Orientation: Solving for , , and based on the rotation matrix .
5.2. Jacobian Matrix
5.3. Trajectory Planning
5.4. Control Scheme
6. Simulation and Experiments
6.1. Simulations
6.2. Experiments
7. Discussion
7.1. Advantages
- End-to-end autonomy. The tight coupling of perception, planning, and control allows the system to complete the entire twist lock workflow without human supervision.
- Industrial-scale robustness. Extensive simulation and full-scale experiments confirm a success rate and mm mean positional error under varying illumination, dust, and occlusions—requirements typical of real warehouses.
- Modular, transferable design. Thanks to the containerized architecture and parameterized gripper CAD model, the solution can be re-scaled to different container sizes or robot brands with minimal re-engineering effort.
- Computational efficiency. The vision stack runs at FPS and the whole control loop at ms, meeting real-time constraints on standard off-the-shelf hardware.
7.2. Limitations
- Environment-specific calibration. Fixed-pose cameras must be re-calibrated if the working volume or robot base is moved; future work could employ active depth sensors on the end-effector to reduce this dependency.
- Lighting sensitivity. Although multi-feature fusion improves robustness, extreme back-lighting still degrades corner detection. Adaptive exposure control or event cameras are promising remedies.
- Limited generalisation beyond twist locks. The gripper fingertips are optimized for the ISO standard locking head; handling irregular bulk cargo would require replaceable jaws or soft robotic add-ons.
- Compute-heavy training. Building the 3200-image dataset and training the improved YOLOv5 once demand a high-end GPU cluster; incremental learning strategies could mitigate this overhead.
7.3. Implications for Industrial Deployment
8. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Kim, D.; TabkhPaz, M.; Park, S.S.; Lee, J. Development of a vision-based automated hole assembly system with quality inspection. Manuf. Lett. 2023, 35, 64–73. [Google Scholar] [CrossRef]
- Kang, L.; Seo, J.T.; Kim, S.H.; Kim, W.J.; Yi, B.J. Design and implementation of a multi-function gripper for grasping general objects. Appl. Sci. 2019, 9, 5266. [Google Scholar] [CrossRef]
- Gümbel, P.; Dröder, K. Design of Highly Repeatable and Multi-Functional Grippers for Precision Handling with Articulated Robots. In Proceedings of the 2024 IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–17 May 2024; pp. 1971–1977. [Google Scholar]
- Cramer, J.; Decraemer, B.; Afzal, M.R.; Kellens, K. A user-friendly toolkit to select and design multi-purpose grippers for modular robotic systems. Procedia CIRP 2024, 127, 2–7. [Google Scholar] [CrossRef]
- Ai, S.; Chen, Y. Multifunctional gripper design at the end of the robot. Instrumentation 2022, 9, 4. [Google Scholar]
- Kladovasilakis, N.; Sideridis, P.; Tzetzis, D.; Piliounis, K.; Kostavelis, I.; Tzovaras, D. Design and development of a multi-functional bioinspired soft robotic actuator via additive manufacturing. Biomimetics 2022, 7, 105. [Google Scholar] [CrossRef]
- Yildiz, B.S.; Pholdee, N.; Bureerat, S.; Yildiz, A.R.; Sait, S.M. Robust design of a robot gripper mechanism using new hybrid grasshopper optimization algorithm. Expert Syst. 2021, 38, e12666. [Google Scholar] [CrossRef]
- Nguyen, T.V.; Huynh, N.T.; Vu, N.C.; Kieu, V.N.; Huang, S.C. Optimizing compliant gripper mechanism design by employing an effective bi-algorithm: Fuzzy logic and ANFIS. Microsyst. Technol. 2021, 27, 3389–3412. [Google Scholar] [CrossRef]
- Pinskier, J.; Wang, X.; Liow, L.; Xie, Y.; Kumar, P.; Langelaar, M.; Howard, D. Diversity-Based Topology Optimization of Soft Robotic Grippers. Adv. Intell. Syst. 2024, 6, 2300505. [Google Scholar] [CrossRef]
- Lee, L.Y.; Syadiqeen, O.A.; Tan, C.P.; Nurzaman, S.G. Closed-structure compliant gripper with morphologically optimized multi-material fingertips for aerial grasping. IEEE Robot. Autom. Lett. 2021, 6, 887–894. [Google Scholar] [CrossRef]
- Sun, Y.; Liu, Y.; Pancheri, F.; Lueth, T.C. Larg: A lightweight robotic gripper with 3-d topology optimized adaptive fingers. IEEE/ASME Trans. Mechatronics 2022, 27, 2026–2034. [Google Scholar] [CrossRef]
- Zablocki, É.; Ben-Younes, H.; Pérez, P.; Cord, M. Explainability of deep vision-based autonomous driving systems: Review and challenges. Int. J. Comput. Vis. 2022, 130, 2425–2452. [Google Scholar] [CrossRef]
- Ghasemieh, A.; Kashef, R. Towards explainable artificial intelligence in deep vision-based odometry. Comput. Electr. Eng. 2024, 115, 109127. [Google Scholar] [CrossRef]
- Wang, R.; Wang, S.; Du, S.; Xiao, E.; Yuan, W.; Feng, C. Real-time soft body 3D proprioception via deep vision-based sensing. IEEE Robot. Autom. Lett. 2020, 5, 3382–3389. [Google Scholar] [CrossRef]
- Choi, H.; Lee, J.; Yi, Y.; Na, H.; Kang, K.; Kim, T. Deep vision-based occupancy counting: Experimental performance evaluation and implementation of ventilation control. Build. Environ. 2022, 223, 109496. [Google Scholar] [CrossRef]
- Wang, Z.; Jin, L.; Wang, S.; Xu, H. Apple stem/calyx real-time recognition using YOLO-v5 algorithm for fruit automatic loading system. Postharvest Biol. Technol. 2022, 185, 111808. [Google Scholar] [CrossRef]
- Zou, W.; Zhang, D.; Lee, D.J. A new multi-feature fusion based convolutional neural network for facial expression recognition. Appl. Intell. 2022, 52, 2918–2929. [Google Scholar] [CrossRef]
- Jiang, M.; Yin, S. Facial expression recognition based on convolutional block attention module and multi-feature fusion. Int. J. Comput. Vis. Robot. 2023, 13, 21–37. [Google Scholar] [CrossRef]
- Li, H.; Vong, C.M.; Wong, P.K.; Ip, W.F.; Yan, T.; Choi, I.C.; Yu, H.H. A multi-feature fusion method for image recognition of gastrointestinal metaplasia (GIM). Biomed. Signal Process. Control 2021, 69, 102909. [Google Scholar] [CrossRef]
- Hu, F.; Qian, M.; He, K.; Zhang, W.a.; Yang, X. A novel multi-feature fusion network with spatial partitioning strategy and cross-attention for armband-based gesture recognition. IEEE Trans. Neural Syst. Rehabil. Eng. 2024, 32, 3878–3890. [Google Scholar] [CrossRef]
- Liu, C.; Xu, X. AMFF: A new attention-based multi-feature fusion method for intention recognition. Knowl. Based Syst. 2021, 233, 107525. [Google Scholar] [CrossRef]
- Zhang, Y.; Guo, Z.; Wu, J.; Tian, Y.; Tang, H.; Guo, X. Real-time vehicle detection based on improved yolo v5. Sustainability 2022, 14, 12274. [Google Scholar] [CrossRef]
- Shahria, M.T.; Sunny, M.S.H.; Zarif, M.I.I.; Ghommam, J.; Ahamed, S.I.; Rahman, M.H. A comprehensive review of vision-based robotic applications: Current state, components, approaches, barriers, and potential solutions. Robotics 2022, 11, 139. [Google Scholar] [CrossRef]
Parameters | Value |
---|---|
Controller-CPU | Core i7-12700 |
Controller-GPU | RTX4060 |
Number of DoF | 6 (arm) + 1 (gripper) |
K | |
D | |
Method | Success Rate | Accuracy |
---|---|---|
Single-feature | 88.26% | 3.7 mm |
Multi-feature | 95.72% | 1.2 mm |
Method | Efficiency | Accuracy | Control Cycle |
---|---|---|---|
With deep vision | 63.4% | 15.3 mm | 35 ms |
Without deep vision | 97.1% | 1.2 mm | 8 ms |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Wang, Y.; Guo, S.; Zhang, J.; Ding, H.; Zhang, B.; Cao, A.; Sun, X.; Zhang, G.; Tian, S.; Chen, Y.; et al. Optimized Design and Deep Vision-Based Operation Control of a Multi-Functional Robotic Gripper for an Automatic Loading System. Actuators 2025, 14, 259. https://doi.org/10.3390/act14060259
Wang Y, Guo S, Zhang J, Ding H, Zhang B, Cao A, Sun X, Zhang G, Tian S, Chen Y, et al. Optimized Design and Deep Vision-Based Operation Control of a Multi-Functional Robotic Gripper for an Automatic Loading System. Actuators. 2025; 14(6):259. https://doi.org/10.3390/act14060259
Chicago/Turabian StyleWang, Yaohui, Sheng Guo, Jinliang Zhang, Hongbo Ding, Bo Zhang, Ao Cao, Xiaohu Sun, Guangxin Zhang, Shihe Tian, Yongxu Chen, and et al. 2025. "Optimized Design and Deep Vision-Based Operation Control of a Multi-Functional Robotic Gripper for an Automatic Loading System" Actuators 14, no. 6: 259. https://doi.org/10.3390/act14060259
APA StyleWang, Y., Guo, S., Zhang, J., Ding, H., Zhang, B., Cao, A., Sun, X., Zhang, G., Tian, S., Chen, Y., Ma, J., & Chen, G. (2025). Optimized Design and Deep Vision-Based Operation Control of a Multi-Functional Robotic Gripper for an Automatic Loading System. Actuators, 14(6), 259. https://doi.org/10.3390/act14060259