Next Article in Journal
Improving Tandem Fluency Through Utilization of Deep Learning to Predict Human Motion in Exoskeleton
Previous Article in Journal
A Review of Electroactive Polymers in Sensing and Actuator Applications
Previous Article in Special Issue
A Low-Cost Prototype of a Soft–Rigid Hybrid Pneumatic Anthropomorphic Gripper for Testing Tactile Sensor Arrays
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimized Design and Deep Vision-Based Operation Control of a Multi-Functional Robotic Gripper for an Automatic Loading System

1
Robotics Research Center, Beijing Jiaotong University, Beijing 100044, China
2
Huaneng Coal Technology Research Co., Ltd., Beijing 100070, China
3
Zhalainuoer Coal Industry Co., Ltd., Hulunbuir 021410, China
4
Tangshan Research Institute, Beijing Jiaotong University, Tangshan 063000, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Actuators 2025, 14(6), 259; https://doi.org/10.3390/act14060259
Submission received: 9 April 2025 / Revised: 15 May 2025 / Accepted: 20 May 2025 / Published: 23 May 2025

Abstract

This study presents an optimized design and vision-guided control strategy for a multi-functional robotic gripper integrated into an automatic loading system for warehouse environments. The system adopts a modular architecture, including standardized platforms, transport containers, four collaborative 6-DOF robotic arms, and a multi-sensor vision module. Methodologically, we first developed three gripper prototypes, selecting the optimal design (30° angle between the gripper and container side) through workspace and interference analysis. A deep vision-based recognition system, enhanced by an improved YOLOv5 algorithm and multi-feature fusion, was employed for real-time object detection and pose estimation. Kinematic modeling and seventh-order polynomial trajectory planning ensured smooth and precise robotic arm movements. Key results from simulations and experiments demonstrated a 95.72% success rate in twist lock operations, with a positioning accuracy of 1.2 mm. The system achieved a control cycle of 35 ms, ensuring efficiency compared with non-vision-based methods. Practical implications include enabling fully autonomous container handling in logistics, reducing labor costs, and enhancing operational safety. Limitations include dependency on fixed camera setups and sensitivity to extreme lighting conditions.

1. Introduction

With the rapid advancement of intelligent manufacturing and automated logistics, the demand for flexible, efficient, and unmanned material handling systems has become increasingly critical in industrial applications. Traditional loading and unloading processes in warehouses are often labor-intensive and error-prone, resulting in reduced operational efficiency and increased safety risks. In response to these challenges, robotic-based automatic loading systems have emerged as a promising solution to achieve high-precision, high-reliability, and intelligent logistics operations. This type of system contains multiple operational tasks, usually requiring the use of a robotic arm and a multi-functional robotic gripper to complete, and the gripper needs to be optimized and designed. The system generally adopts vision-based control, and due to the complexity of the environment, multiple feature fusion techniques need to be carried out to ensure stable and robust operation. For example, Kim et al. proposed a novel mechatronics system capable of automating a peg-in-hole assembly process and inspecting the quality of the assembly with vision [1].
In the multi-functional mechanical gripper field, Kang et al. designed and implemented a multi-function gripper for grasping general objects, aiming to enhance versatility and flexibility in industrial automation applications [2]. Gümbel and Dröder proposed a design for highly repeatable and multi-functional grippers intended for precision handling with articulated robots [3]. Cramer et al. developed a user-friendly toolkit to select and design multi-purpose grippers for modular robotic systems [4]. Ai and Chen proposed a multifunctional gripper design at the end of the robot for versatile object manipulation [5]. Kladovasilakis et al. developed a multi-functional bioinspired soft robotic actuator using additive manufacturing techniques [6].
In the optimized design field, Yildiz et al. introduced a robust robot gripper mechanism design using a new hybrid grasshopper optimization algorithm [7]. Nguyen et al. optimized compliant gripper mechanism design by employing a bi-algorithm combining fuzzy logic and ANFIS [8]. Pinskier et al. applied diversity-based topology optimization techniques to the design of soft robotic grippers [9]. Lee et al. presented a closed-structure compliant gripper with morphologically optimized multi-material fingertips for aerial grasping [10]. Sun et al. developed LARG, a lightweight robotic gripper with 3D topology-optimized adaptive fingers [11].
In the vision-based control field, Zablocki et al. reviewed the explainability of deep vision-based autonomous driving systems and identified future challenges [12]. Ghasemieh and Kashef explored explainable artificial intelligence for deep vision-based odometry systems [13]. Wang et al. proposed a real-time deep vision-based method for soft body 3D proprioception [14]. Choi et al. conducted an experimental evaluation of deep vision-based occupancy counting for ventilation control [15]. Wang et al. applied the YOLO-v5 algorithm for real-time recognition of apple stem/calyx in automatic fruit loading systems [16].
In the multi-feature fusion field, Zou et al. developed a new multi-feature fusion convolutional neural network for facial expression recognition [17]. Jiang and Yin proposed a facial expression recognition approach using convolutional block attention and multi-feature fusion [18]. Li et al. presented a multi-feature fusion method for recognizing gastrointestinal metaplasia in medical images [19]. Hu et al. designed a novel multi-feature fusion network with spatial partitioning and cross-attention for gesture recognition using armband signals [20]. Liu and Xu introduced AMFF, an attention-based multi-feature fusion method for human intention recognition [21].
Inspired by port automation and container handling techniques, this study proposes a modular, containerized, and vision-guided robotic system for automated loading and unloading in warehouse scenarios. The system integrates standardized container platforms, modular transport units, multi-degree-of-freedom robotic manipulators, and a deep vision sensing module to perform autonomous material transfer tasks. Central to the system is the design of a multi-functional mechanical gripper capable of handling twist locks and performing precise positioning adjustments under complex spatial constraints.
To enhance system intelligence and operational autonomy, the proposed solution incorporates a multi-sensor vision system utilizing RealSense depth cameras. This enables real-time object recognition, spatial localization, and grasp planning, providing robust perception support for manipulation tasks. A coordinated control strategy for multiple robotic arms further improves task efficiency and ensures collision-free operations.
The objectives of this work are threefold: (1) to develop a mechanical gripper design optimized for both grasping and fine manipulation of twist locks; (2) to establish a vision-guided control framework for real-time object detection and pose estimation; and (3) to validate the system’s performance in simulation and experimental settings. This paper presents the system architecture, mechanical design methodology, motion planning strategies, and experimental results, demonstrating the feasibility and effectiveness of the proposed approach in enabling intelligent and autonomous loading operations in industrial warehouse environments.
Compared with existing state-of-the-art methods, our approach introduces several practical innovations that contribute to its superior performance. First, the geometry-aware gripper design improves workspace coverage and reduces mechanical interference, resulting in more reliable grasping during complex operations. Second, by fusing multiple 88 complementary visual features, the perception module achieves enhanced robustness and precision under varying environmental conditions. Third, the lightweight deep learning model enables real-time inference on standard hardware, ensuring smooth system integration without specialized computation resources. Lastly, the use of high-order motion 92 planning techniques ensures smoother and more stable manipulation trajectories, which benefits both alignment accuracy and mechanical durability. These combined improvements allow our system to deliver consistently higher efficiency and reliability in real-world automated loading tasks.

2. Automatic Loading System

2.1. System Architecture and Composition

Inspired by the automated container loading and unloading processes used in ports, this project proposes an unmanned automated system suitable for warehouse environments, as illustrated in Figure 1. The simulation prototype is established by using ROS Noetic Ninjemys Gazebo 11.X in Ubuntu 20.04 Focal Fossa. The system adopts a containerized standard storage format similar to that of port logistics, enabling digital management of warehouse goods by recording the contents and locations of each container.
The system is composed of six main modules:
1.
Standardized Load-Carrying Platform Module: This can include mining flatbed trucks, cargo trucks, trains, and ships.
2.
Modular Transport Container Module: Designed in various forms to carry different materials. As shown on the left side of Figure 1a, each container is equipped with eight corner fittings—four at the bottom for securing the container, and four at the top for lifting and transport.
3.
Locking Module: Includes manually adjustable twist locks and fully automatic twist locks. The manual twist locks are mounted on the standardized platform and can raise the locking head to secure the transport container. When locking is not required, the lock head can be lowered to convert the platform into a standard flatbed. The fully automatic twist locks are used to interconnect stacked containers.
4.
Overhead Crane and Spreader Module: Responsible for loading and unloading the transport containers onto and from the standardized platform, as well as for stacking containers within the warehouse.
5.
Multi-Sensor Recognition and Localization Module: Comprising four cameras, an IMU (Inertial Measurement Unit), and UWB (Ultra Wide Band) positioning modules, LD150 from Haoru Tech (Dalian, China). The cameras detect the four bottom corners of the containers and the handles of the manual twist locks within the work zone, providing feedback for automated operation. The IMU and UWB modules estimate the container’s position and orientation outside the working zone.
6.
Multi-Robot Arm Collaborative Loading and Unloading Module: Installed in the working zone, consisting of four robotic arms and multifunctional grippers. This module facilitates the automated transfer of containers between the platform and the storage area.
The system comprises four six-degree-of-freedom (6-DOF) robotic arms, specifically BRTIRUS1510A from Borunte (Shanghai, China), symmetrically mounted on both sides of the workspace. Each robotic arm is equipped with a two-finger gripper, RM-GB-17-80-2-1 from Robustmotion (Guangdong, China), to facilitate grasping operations. Additionally, four RealSense D415 depth cameras from Intel (Santa Clara, CA, USA) are individually positioned on the base of arms to acquire environmental data, providing real-time feedback for task execution by the robotic arms. The single robotic arm with a gripper and a depth camera is shown in Figure 2.

2.2. Loading and Unloading Process

2.2.1. Loading Process

The procedure for loading a transport container onto the load-carrying platform is as follows:
When the overhead crane hoists the transport container above the platform from a designated position in the warehouse, the multi-camera system captures the poses of the four bottom corner fittings of the container relative to the robotic arms’ base coordinate frame. Using this information, the system calculates the precise poses of the fully automatic twist locks mounted at the bottom of the container.
Then, the multi-robot arm collaborative module removes the fully automatic twist locks based on the estimated poses. Afterward, the robotic arms correct the container’s posture and align it with the platform. The overhead crane then lowers the container onto the platform.
Next, the multi-camera system detects the pose of the manual twist lock handles located on the platform. The robotic arms subsequently lift the locking heads according to the handle poses to securely fix the container in place.

2.2.2. Unloading Process

The procedure for unloading a transport container from the load-carrying platform is as follows:
The multi-camera system first detects the pose of the manual twist lock handles on the platform. The multi-robot arm module then lowers the locking heads to release the container.
The overhead crane lifts the container to a certain height. The multi-camera system again captures the poses of the four bottom corner fittings of the container relative to the robotic arms’ base frame, enabling calculation of the poses of the fully automatic twist locks.
Subsequently, the multi-robot arm module attaches the fully automatic twist locks back onto the container according to the estimated poses. Finally, the overhead crane transfers the container from above the platform to a designated storage position in the warehouse for stacking.

3. Optimized Design

3.1. Tasks of the Multi-Robot Arm Collaborative Loading and Unloading Module

Based on the operational workflow described above, the multi-robot arm collaborative loading and unloading module is designed to perform four main tasks:
  • Handling fully automatic twist locks;
  • Locking and unlocking manual adjustable twist locks;
  • Correcting the posture of the transport container;
  • Positioning and securing the transport container.
Among these tasks, handling the fully automatic twist locks is the most technically challenging and demands the highest precision from the multi-camera recognition and localization system. Therefore, the subsequent analysis focuses primarily on the task of handling fully automatic twist locks.

3.2. Operation Process

The operation process of the multi-functional mechanical gripper contains locking and unlocking the fully twist lock procedure.

3.2.1. Locking Procedure

The locking procedure consists of the following steps:
1.
Initial Grasping: The robotic arm moves from its initial position to a designated position where the gripper closes to securely grasp the twist lock.
2.
Alignment: The robotic arm adjusts its posture and executes a translational motion to align the twist lock with the mounting port of the corner piece.
3.
Insertion: The robotic arm performs a vertical motion along the z-axis, raising the twist lock to a certain height until it enters the corner piece.
4.
Rotation: The robotic arm rotates the twist lock counterclockwise by 60 ° about its own central axis, which is parallel to the z-axis (viewed from above).
5.
Release: The gripper opens to release the twist lock, allowing it to be securely installed within the corner piece.
6.
Return: The robotic arm returns to its initial position.

3.2.2. Unlocking Procedure

The unlocking procedure involves the following steps:
1.
Approach: The robotic arm moves from its initial position to a designated location with a specific posture to approach and grasp the twist lock.
2.
Grasping: The gripper closes to securely hold the twist lock.
3.
Rotation: The robotic arm rotates the twist lock clockwise by 60 ° about its own central axis, which is parallel to the z-axis (viewed from above).
4.
Extraction: The robotic arm performs a downward motion along the z-axis, lowering the twist lock until it exits the corner piece.
5.
Placement: The robotic arm moves the twist lock to a designated placement location, where the gripper opens to release the twist lock.
6.
Return: The robotic arm returns to its initial position.
Loading and unloading a twist lock is the opposite process. The three flag states are shown in Figure 3.

3.3. Optimized Design of the Multi-Functional Robotic Gripper

To meet the requirements of the four core functions of the multi-robot arm module, an optimized design of a multi-functional mechanical gripper has been developed.
An embedded two-finger gripper is used to grip the fully automatic twist locks. A cylindrical structure fixed above the two fingers is employed to rotate and lift the handle of the manual retractable twist lock. A right-angle structure, formed by the two fingers and a rectangular block on the side, is used to position and secure the four vertical corners of the transport container. Universal wheels at the fingertip ends help align the container and prevent unintended movement or frictional dragging between the container and the fingers.
To accommodate different gripping angles required for the fully automatic twist locks, three types of embedded two-finger grippers have been designed.

3.3.1. Type I Gripper (90° Between Gripper and Container Side)

The first gripper design, in which the angle between the gripper and the short side of the container is 90°, is illustrated in Figure 4a. The gripping posture for locking the twist lock is shown in Figure 4b.
When gripping from the outside, one of the extreme reachable positions is shown in Figure 4c. Although no interference occurs with the container in either extreme posture, the manipulator reaches the limit of its workspace. If the distance between the manipulator and the twist lock is too large, the gripper may fail to reach the required position, resulting in failure to complete the gripping action.
When gripping from the inside, the extreme posture is shown in Figure 4d. In this case, interference occurs between the robot’s lower arm and the container.

3.3.2. Type II Gripper (0° Between Gripper and Container Side)

The second gripper design, where the gripper is parallel to the short side of the container (0° angle), is shown in Figure 5a, and its twist lock gripping posture is shown in Figure 5b.
Two extreme positions for gripping the twist lock using this design are shown in Figure 5c,d. While both positions avoid interference with the container, in one posture, the robotic arm partially extends beneath the container, which increases the risk of collision during motion. Moreover, if the container is relatively low, rotation of joint θ 5 may cause the lower arm of the robot to interfere with the container.

3.3.3. Type III Gripper (30° Between Gripper and Container Side)

The third gripper design, ultimately selected for deployment, features a 30° angle between the gripper and the short side of the container. The model is shown in Figure 6a, and its gripping posture is shown in Figure 6b.
The two extreme gripping positions for this design are illustrated in Figure 6c,d. In both cases, there is no interference with the container, and the robotic arms do not extend beneath the container. Furthermore, both extreme positions are comfortably within the robot’s workspace limits. Based on these advantages, the third gripper design was adopted as the final solution.
During the optimization process, both the workspace of the robotic arms and potential interferences between the arms and the container were thoroughly considered. Based on these analyses, the third design of the embedded two-finger gripper was selected, as the first two designs encountered limitations due to workspace constraints and physical interference [2]. As shown in Figure 7, there exist interference or workspace limitations in the Type I and II grippers. While the Type III gripper has the same maximum angle space before and after the 60° rotation operation, note that there are four arms that carry out the operation at the four corners of the container.
As a result, the final optimized multi-functional mechanical gripper design fulfills all four functional requirements, as shown in Figure 8.

4. Sensing Based on Deep Vision

The perception system aims to identify the fully automatic twist lock and determine its pose. It utilizes a YOLOv5-based algorithm to detect the planar coordinates ( x , y ) , depth point cloud information to obtain z-axis data, and geometric relationships among multiple target features for fusion recognition. This robust recognition and localization approach ensures reliable operation even under varying lighting conditions or when the target is temporarily lost.

4.1. Improved YOLOv5 Object Detection Algorithm

The proposed approach employs an improved object detection algorithm based on the YOLOv5 architecture [22], characterized by optimized network structure and training strategies that significantly reduce computational resource consumption while maintaining high detection accuracy.
Specifically, the algorithm achieves enhanced efficiency through the following techniques:
  • Lightweight Feature Extraction Network: Reduces model parameters, ensuring efficient processing.
  • Dynamic Resolution Input Strategy: Adapts to different computational constraints across various scenarios.
  • Channel Attention Mechanism: Improves the discriminative capability of feature representations.
Due to the fixed camera and workspace, the scene is also fixed and singular. The recognized features only have a small correlation between Corner and container types. Other features are not on the container. Thus, the dataset is mainly affected by time and lighting. To enhance robustness, a specialized dataset comprising 3200 high-resolution images was constructed under different time periods and lighting conditions. The number of images of the train, validation, and test set are 2200, 500, and 500, respectively. This dataset encompasses various illumination conditions, target scales, and complex backgrounds. Data augmentation and rigorous annotation procedures are applied to significantly improve the algorithm’s detection accuracy and generalization performance.

4.2. Multi-Feature Fusion Recognition

A high-precision recognition method for containers and flatbed trucks is proposed, utilizing six heterogeneous features to achieve robust localization and motion decision-making under complex conditions. The feature system includes the following:
  • Corner piece (Corner);
  • Flatbed truck edge (Edge);
  • Flatbed truck twist lock horizontal component (Lift_lock_1);
  • Flatbed truck twist lock vertical component (Lift_lock_2);
  • Twist lock Orientation (F/B/L/R);
  • Automatic twist lock (Auto_lock).
The Corner feature serves as the primary constraint, while other auxiliary features are hierarchically integrated to enable core functionalities such as automatic twist lock locking/unlocking and flatbed truck twist lock manipulation.
During loading and unloading of a fully automatic twist lock onto or from a Modular Transport Container Module operations, motion commands for the robotic arm are generated based on the combined three-dimensional coordinate calculations of Corner and the automatic twist lock orientation. The localization precision is further optimized through multi-source data fusion, involving the following steps:
  • Original coordinates of Auto_lock, Lift_lock_1, Lift_lock_2, and Edge are spatially registered and transformed to match the Corner coordinate system.
  • Kalman filtering is applied for temporal prediction and dynamic weighted fusion of observations, effectively suppressing sensor noise and motion blur.
For depth calculation, a modal-based strategy is introduced:
  • For Corner features, depth information is derived from the centroid depth of the bounding box’s four corner pixel points.
  • For other features, the minimum depth value within the bounding box is utilized, balancing geometric precision and robustness against occlusion.
  • A depth thresholding mechanism is implemented to filter erroneous data.
The recognition results are shown in Figure 9. The system successfully identifies multiple features, with the primary target being the Corner. Since the container structure is pre-designed, other features maintain consistent geometric relationships with the Corner. When the Corner feature is compromised, its pose can be indirectly calculated through these related features.

5. Operation Control

5.1. Kinematics Model

5.1.1. Forward Kinematics

The forward kinematics of a six-degree-of-freedom (6-DOF) robotic arm determines the position and orientation of the end-effector given the joint angles. The general forward kinematics equation can be represented as:
T = i = 1 6 T i = T 1 T 2 T 3 T 4 T 5 T 6
Each transformation matrix T i is defined using the Denavit–Hartenberg (DH) convention:
T i = cos θ i sin θ i cos α i sin θ i sin α i a i cos θ i sin θ i cos θ i cos α i cos θ i sin α i a i sin θ i 0 sin α i cos α i d i 0 0 0 1
where:
  • θ i is the joint angle.
  • d i is the link offset.
  • a i is the link length.
  • α i is the twist angle.
The final transformation matrix T provides the position and orientation of the end-effector in the base frame. The transformation matrix T relates the position and orientation of the end-effector as follows:
T = R P 0 1 = n x o x a x p x n y o y a y p y n z o z a z p z 0 0 0 1
n x = s 234 c 1 s 6 c 6 s 1 s 5 c 234 c 1 c 5 n y = c 6 c 1 s 5 + c 234 c 5 s 1 + s 1 s 6 n z = c 234 s 6 s 234 c 5 c 6 o x = s 6 s 1 s 5 c 234 c 5 c 1 + s 234 c 1 c 6 o y = s 234 s 1 c 6 s 6 c 1 s 5 + c 234 s 1 c 5 o z = c 234 c 6 + s 234 c 5 s 6 a x = c 5 s 1 + c 234 c 1 s 5 a y = c 234 s 1 s 5 c 1 c 5 a z = s 234 s 5 p x = d 6 c 5 s 1 + c 234 c 1 s 5 + a 2 c 1 c 2 + a 4 c 1 c 234 + a 3 c 1 c 23 p y = d 6 c 234 s 1 s 5 c 1 + a 4 s 1 c 234 + a 3 s 1 c 23 + a 2 c 2 s 1 p z = d 1 a 3 s 23 a 2 s 2 a 4 s 4 23 d 6 s 5 s 234
where s i , c i , s i j , c i j , s i j k and c i j k donate sin θ i , cos θ i , sin ( θ i + θ j ) , cos ( θ i + θ j ) , sin ( θ i + θ j + θ k ) and cos ( θ i + θ j + θ k ) , respectively. P is the position vector p x p y p z T , and R is the rotation matrix representing orientation (corresponding to α , β , γ using Euler angles or RPY angles or other representations). Using RPY angles, it yields
β = a t a n 2 n z , n x 2 + n y 2 α = a t a n 2 n y , n x γ = a t a n 2 o z , a z
This can be summarized as follows:
X = F K ( Q )
where X = p x p y p z α β γ T is the end-effector pose, Q = θ 1 θ 2 θ 3 θ 4 θ 5 θ 6 T is the vector of joint angles, and F K is the forward kinematics function.

5.1.2. Inverse Kinematics

Inverse kinematics involves determining the joint angles θ i given the desired end-effector pose T . This process is generally more complicated than forward kinematics and often requires numerical methods.
The desired transformation matrix can be divided into a rotational part R and a translational part P as Equation (3). Solving the inverse kinematics involves breaking down the problem into position and orientation components:
  • Position: Solving for θ 1 , θ 2 , and θ 3 based on the position vector P .
  • Orientation: Solving for θ 4 , θ 5 , and θ 6 based on the rotation matrix R .
θ 1 = arctan ± p y d 6 a y p x d 6 a x θ 5 = arctan ± n x s 1 n y c 1 2 + o x s 1 o y c 1 2 a x s 1 a y c 1 θ 6 = arctan o x s 1 o y c 1 n x s 1 + n y c 1 θ 2 = arctan ± A 2 + B 2 4 a 2 2 p 2 A 2 + B 2 2 arctan A B θ 234 = θ 2 + θ 3 + θ 4 = a z a x c 1 + a y s 1 θ 23 = arctan d 1 p z d 6 s 5 s 234 a 4 s 234 a 2 s 2 p x c 1 + p y s 1 a 4 c 234 d 6 s 5 c 234 a 2 c 2 θ 3 = θ 23 θ 2 θ 4 = θ 234 θ 23
where A = p x c 1 + p y s 1 a 4 c 234 d 6 s 5 c 234 , B = d 1 p z d 6 s 5 s 234 a 4 s 234 . To avoid the robotic arm being in a singular position, it is necessary to ensure s i n θ 5 0 .
This process can be summarized as:
Q = I K ( X )
where I K is the inverse kinematics function.
Analytical solutions can be obtained for simple robotic arms, while numerical methods such as Jacobian Inverse or Jacobian Transpose methods are employed for complex configurations.

5.2. Jacobian Matrix

The Jacobian matrix J relates the joint velocities Q ˙ to the end-effector velocity X ˙ :
X ˙ = J Q ˙
The Jacobian matrix can be derived by differentiating the forward kinematics equations with respect to the joint angles.
J 11 = d 6 c 1 s 4 s 5 s 1 a 2 + a 4 c 23 + d 4 s 23 + a 3 c 2 + d 6 c 23 s 45 / 2 + d 6 s 23 c 5 d 6 s 45 c 23 / 2 J 12 = c 1 c 2 d 6 c 3 c 5 c 4 s 3 s 5 + d 4 c 3 a 4 s 3 s 2 a 3 + d 6 c 5 s 3 + c 3 c 4 s 5 + a 4 c 3 + d 4 s 3 J 13 = c 1 s 23 a 4 + d 6 c 4 s 5 c 23 d 4 + d 6 c 5 J 14 = d 6 s 5 c 4 s 1 c 1 c 2 c 3 s 4 + c 1 s 2 s 3 s 4 J 15 = d 6 c 5 s 1 s 4 d 6 c 1 c 2 s 3 s 5 d 6 c 1 c 3 s 2 s 5 d 6 c 1 c 4 c 5 s 2 s 3 + d 6 c 1 c 2 c 3 c 4 c 5 J 16 = 0 J 21 = c 1 a 2 + a 4 c 23 + d 4 s 23 + a 3 c 2 + d 6 c 23 s 45 / 2 + d 6 s 23 c 5 d 6 s 4 c 5 s 5 c 4 c 23 / 2 + d 6 s 1 s 4 s 5 J 22 = s 1 c 2 d 6 c 3 c 5 c 4 s 3 s 5 + d 4 c 3 a 4 s 3 s 2 a 3 + d 6 c 5 s 3 + c 3 c 4 s 5 + a 4 c 3 + d 4 s 3 J 23 = s 1 s 23 a 4 + d 6 c 4 s 5 c 23 d 4 + d 6 c 5 J 24 = d 6 s 5 c 1 c 4 + c 2 c 3 s 1 s 4 s 1 s 2 s 3 s 4 J 25 = d 6 c 2 c 3 c 4 c 5 s 1 d 6 c 2 s 1 s 3 s 5 d 6 c 3 s 1 s 2 s 5 d 6 c 1 c 5 s 4 d 6 c 4 c 5 s 1 s 2 s 3 J 26 = 0 J 31 = 0 J 32 = a 3 c 2 + a 4 c 2 c 3 + d 4 c 2 s 3 + d 4 c 3 s 2 a 4 s 2 s 3 + d 6 c 2 c 5 s 3 + d 6 c 3 c 5 s 2 + d 6 c 2 c 3 c 4 s 5 d 6 c 4 s 2 s 3 s 5 J 33 = a 4 c 23 + d 4 s 23 + d 6 s 23 c 5 + d 6 c 23 c 4 s 5 J 34 = d 6 s 23 s 4 s 5 J 35 = d 6 c 2 c 3 s 5 s 2 s 3 s 5 + c 2 c 4 c 5 s 3 + c 3 c 4 c 5 s 2 J 36 = 0 J 41 = 0 J 42 = s 1 J 43 = s 1 J 44 = s 23 c 1 J 45 = c 4 s 1 + s 4 c 1 s 2 s 3 c 1 c 2 c 3 J 46 = s 5 s 1 s 4 c 4 c 1 s 2 s 3 c 1 c 2 c 3 + c 5 c 1 c 2 s 3 + c 1 c 3 s 2 J 51 = 0 J 52 = c 1 J 53 = c 1 J 54 = s 23 s 1 J 55 = s 4 s 1 s 2 s 3 c 2 c 3 s 1 c 1 c 4 J 56 = c 5 c 2 s 1 s 3 + c 3 s 1 s 2 s 5 c 1 s 4 + c 4 s 1 s 2 s 3 c 2 c 3 s 1 J 61 = 1 J 62 = 0 J 63 = 0 J 64 = c 23 J 65 = s 23 s 4 J 66 = s 23 c 4 s 5 c 23 c 5

5.3. Trajectory Planning

To ensure smooth and vibration-free motion of the robot, abrupt changes in joint torques must be avoided. This requires the velocity and acceleration of each joint and the end-effector pose to change continuously over time. To achieve this, a seventh-order polynomial trajectory is adopted in Cartesian space. This approach guarantees the continuity of position, velocity, acceleration, and jerk.
X ( t ) = k 0 + k 1 t + k 2 t 2 + k 3 t 3 + k 4 t 4 + k 5 t 5 + k 6 t 6 + k 7 t 7 X ˙ ( t ) = k 1 + 2 k 2 t + 3 k 3 t 2 + 4 k 4 t 3 + 5 k 5 t 4 + 6 k 6 t 5 + 7 k 7 t 6 X ¨ ( t ) = 2 k 2 + 6 k 3 t + 12 k 4 t 2 + 20 k 5 t 3 + 30 k 6 t 4 + 42 k 7 t 5 X ( t ) = 6 k 3 + 24 k 4 t + 60 k 5 t 2 + 120 k 6 t 3 + 210 k 7 t 4
where the coefficients are given by:
k 0 = X 0 k 1 = X ˙ 0 k 2 = 1 2 X ¨ 0 k 3 = 1 6 X 0 k 4 = 210 ( X f X 0 ) t f [ ( 30 X ¨ 0 15 X ¨ f ) + ( 4 X 0 + X f ) t f 2 + 120 X ˙ 0 + 90 X ˙ f ] 6 t f 4 k 5 = 168 ( X f X 0 ) t f [ ( 20 X ¨ 0 14 X ¨ f ) t f + ( 2 X 0 + X f ) t f 2 + 90 X ˙ 0 + 78 X ˙ f ] 2 t f 5 k 6 = 420 ( X f X 0 ) t f [ ( 45 X ¨ 0 39 X ¨ f ) t f + ( 4 X 0 + 3 X f ) t f 2 + 216 X ˙ 0 + 204 X ˙ f ] 6 t f 6 k 7 = 120 ( X f X 0 ) t f [ ( 12 X ¨ 0 12 X ¨ f ) t f + ( X 0 + X f ) t f 2 + 60 X ˙ 0 + 60 X ˙ f ] 6 t f 7
with boundary conditions:
X ( t 0 ) = X 0 X ˙ ( t 0 ) = X ˙ 0 X ¨ ( t 0 ) = X ¨ 0 X ( t 0 ) = X 0 X ( t f ) = X f X ˙ ( t f ) = X ˙ f X ¨ ( t f ) = X ¨ f X ( t f ) = X f
where X 0 = p x 0 p y 0 p z 0 α 0 β 0 γ 0 T and X f = p x f p y f p z f α f β f γ f T represent the initial and final positions of the trajectory, and t f is the total duration of the trajectory planning.

5.4. Control Scheme

The operation control is based on vision positioning and feedback [23]. The control block diagram of the system is shown in Figure 10. The system employs a depth camera and UWB sensors to detect the position x , y , z of the corner component, while an IMU sensor measures its attitude angles α , β , γ . Based on the acquired position and orientation X c , the desired end-effector pose of the robotic arm X d is calculated by the pose transformation relationship between operational pose and recognition pose T c d . The difference between the desired pose and the current pose X 0 of the robotic arm Δ X = X d X 0 is scaled by a PD controller to determine the pose increment K Δ X + D Δ X ˙ for the next time step. Consequently, the desired pose for the next moment is obtained as X 0 + Δ X . Using this pose, the joint driving angles θ i are calculated through inverse kinematics, allowing the robotic arm to perform the required operation. Here, the current pose X 0 is calculated via the forward kinematics based on the current joint angles θ i . And, the relationship of the pose increment of the end-effector and the joint increment is Δ X = J Δ Q .
The advantage of this control method lies in its adaptability: when the end-effector is far from the desired pose, it enables a rapid approach to save time; when the end-effector is close to the target pose, it allows a slow approach to avoid collision with the manipulated object.

6. Simulation and Experiments

To verify the feasibility and effectiveness of the proposed system, a design and control method, a simulation system, and a prototype system were built for experimental validation. The simulation and experiment parameters are shown in Table 1. Some parameters such as K and D are fine-tuned by trials. The parameters of the desired trajectories are produced through Equation (12).

6.1. Simulations

Based on the previously designed automatic loading system hardware, the optimized design of a multi-functional mechanical gripper, and the deep vision-based control algorithms, a dynamic simulation model was constructed in Gazebo, as shown in Figure 1a. Subsequently, simulation experiments were conducted on the loading and unloading of a fully automatic twist lock onto or from a Modular Transport Container Module, as illustrated in Figure 11a,b, respectively. The pose curves of the multi-functional mechanical gripper during loading and unloading are shown in Figure 12a–d, respectively.

6.2. Experiments

The full-scale prototype of automatic loading system is implemented as shown in Figure 1b. Subsequently, prototype experiments were conducted on the loading and unloading of a fully automatic twist lock onto or from a Modular Transport Container Module, as illustrated in Figure 13a,b, respectively.
The perception results of the depth camera during the loading and unloading process are shown in Figure 14a,b, respectively. It can be found that many features are recognized, including the target Corner feature, and multi-feature fusion is employed to improve the stability and robustness of positioning. The comparative experiments about the difference between single-feature and multi-feature perception methods are shown in Table 2. Meanwhile, the object detection algorithm based on the improved YOLOv5 is compared with traditional YOLOv5. The reasoning speed of improved YOLOv5 is 450FPS, which is faster than that of traditional YOLOv5, that is, 120FPS.
The position and attitude angle curves of the multi-functional mechanical gripper during the loading process are shown in Figure 15a,b, respectively. As shown in the figure, the fully automated twist lock loading process using Arm 1 can be divided into six stages: initial grasping, rotating and moving beneath the corner piece, lifting the twist lock into the corner piece, rotating and releasing the twist lock, and returning the gripper to the initial position. In the initial grasping stage, the manipulator remains at its starting position, with the gripper closed to securely hold the twist lock. During the manipulation stage, the end-effector gripper rotates to an angle of 70 ° along the z-axis, carrying the twist lock to a position directly beneath the corner piece. The manipulator then moves a certain distance along the negative y-axis to align the twist lock with the bottom hole of the corner piece. Subsequently, it moves upward along the positive z-axis. The end-effector then rotates from 70 ° to 130 ° along the z-axis. The gripper opens to release the twist lock into the corner piece. Finally, the gripper moves downward and returns to the initial position, completing the installation process.
The position and attitude angle curves of the multi-functional mechanical gripper during the unloading process are shown in Figure 15c,d, respectively. As shown in the figure, the fully automated twist lock unloading process using Arm 1 can be divided into six stages: approaching and reaching beneath the twist lock from the initial position, closing the gripper to grasp the twist lock, rotating to unload the twist lock, lowering to extract the twist lock, and placing it at the designated position before returning to the initial position. In the initial positioning stage, the manipulator starts from its home position. During the manipulation stage, the manipulator first moves to the rear-right side of the twist lock. The gripper then moves beneath the rear-right of the twist lock and rotates to an angle of 130 ° along the z-axis. It continues moving to a position directly beneath the twist lock and then moves upward along the positive z-axis. The gripper closes to securely grasp the twist lock and lifts it up by 3 mm. The end-effector then rotates from 130 ° to 70 ° along the z-axis. The gripper releases slightly and moves downward to extract the twist lock. Finally, the manipulator moves the twist lock to a designated location, opens the gripper to release it, and returns to the initial position, completing the unloading process.
To highlight the significance of deep vision-based operation control, comparative experiments are implemented between with (closed-loop) and without (open-loop) deep vision. The comparative results of efficiency, accuracy, and control cycle are shown in Table 3. The efficiency is reflected by the success rate of operation. The accuracy means the error of position between the gripper and the automatic twist lock. The control cycle donates the system latency or real-time constraints. The target perception positioning algorithm takes up most of the time and the GPU is employed to accelerate the process to ensure real-time performance.

7. Discussion

To help readers better appreciate the practical significance of our work, this section summarizes the advantages, limitations, and implications of the proposed system.

7.1. Advantages

  • End-to-end autonomy. The tight coupling of perception, planning, and control allows the system to complete the entire twist lock workflow without human supervision.
  • Industrial-scale robustness. Extensive simulation and full-scale experiments confirm a 95.72 % success rate and 1.2 mm mean positional error under varying illumination, dust, and occlusions—requirements typical of real warehouses.
  • Modular, transferable design. Thanks to the containerized architecture and parameterized gripper CAD model, the solution can be re-scaled to different container sizes or robot brands with minimal re-engineering effort.
  • Computational efficiency. The vision stack runs at > 400 FPS and the whole control loop at 35 ms, meeting real-time constraints on standard off-the-shelf hardware.

7.2. Limitations

  • Environment-specific calibration. Fixed-pose cameras must be re-calibrated if the working volume or robot base is moved; future work could employ active depth sensors on the end-effector to reduce this dependency.
  • Lighting sensitivity. Although multi-feature fusion improves robustness, extreme back-lighting still degrades corner detection. Adaptive exposure control or event cameras are promising remedies.
  • Limited generalisation beyond twist locks. The gripper fingertips are optimized for the ISO standard locking head; handling irregular bulk cargo would require replaceable jaws or soft robotic add-ons.
  • Compute-heavy training. Building the 3200-image dataset and training the improved YOLOv5 once demand a high-end GPU cluster; incremental learning strategies could mitigate this overhead.

7.3. Implications for Industrial Deployment

Deploying the system in production warehouses could cut manual labor by more than 50 % and reduce error-induced downtime. Moreover, the design principles, including gripper–workspace co-optimization, feature-level perceptual fusion, and high-order trajectory planning, are applicable to other logistics operations such as palletizing or autonomous container stacking.

8. Conclusions

In this paper, an intelligent automatic loading system with a multi-functional mechanical gripper and deep vision-based control framework was proposed and implemented. The system effectively combines containerized transport logic with robotic automation, enabling unmanned loading and unloading in warehouse environments. Through optimized gripper design and multi-robot coordination, the system successfully completes complex tasks such as twist lock manipulation and container posture adjustment. The integration of a multi-sensor vision system ensures accurate pose estimation and improves operational reliability. Experimental validations confirm that the proposed solution achieves high efficiency and precision, marking a significant step toward fully autonomous industrial logistics systems. Future work will focus on enhancing the generalization capability of the vision module and expanding system scalability for larger scale deployments.

Author Contributions

Conceptualization, G.C. and Y.W.; methodology, G.C., Y.W., J.Z., B.Z., X.S., J.M. and S.T.; software, H.D., A.C., G.Z. and Y.C.; validation, H.D., A.C., G.Z. and Y.C.; formal analysis, H.D., A.C., G.Z. and Y.C.; investigation, G.C., Y.W., J.Z., B.Z., X.S., J.M. and S.T.; resources, G.C. and Y.W.; data curation, H.D., A.C., G.Z. and Y.C.; writing—original draft preparation, G.C.; writing—review and editing, G.C. and S.G.; visualization, H.D., A.C., G.Z. and Y.C.; supervision, G.C. and S.G.; project administration, G.C., Y.W. and S.G.; funding acquisition, G.C., Y.W. and S.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Huaneng Group Headquarters Science and Technology Project “Research and Development of Standardized Material Carrying Platform and Intelligent Loading System” (HNKJ22-HF130), Central Government Guides Local Science and Technology Development Fund Projects (246Z1813G), National Key Research and Development Program of China (2022YFB4701600), Beijing Natural Science Foundation (L243013), A New Generation Information Technology Innovation Project of China University Research Innovation Fund (2023IT098) and Fundamental Research Funds for the Central Universities (2024XKRC069).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

Authors Yaohui Wang and Xiaohu Sun were employed by the company Huaneng Coal Technology Research Co., Ltd. Authors Jinliang Zhang, Bo Zhang, Shihe Tian and Jixuan Ma were employed by the company Zhalainuoer Coal Industry Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Kim, D.; TabkhPaz, M.; Park, S.S.; Lee, J. Development of a vision-based automated hole assembly system with quality inspection. Manuf. Lett. 2023, 35, 64–73. [Google Scholar] [CrossRef]
  2. Kang, L.; Seo, J.T.; Kim, S.H.; Kim, W.J.; Yi, B.J. Design and implementation of a multi-function gripper for grasping general objects. Appl. Sci. 2019, 9, 5266. [Google Scholar] [CrossRef]
  3. Gümbel, P.; Dröder, K. Design of Highly Repeatable and Multi-Functional Grippers for Precision Handling with Articulated Robots. In Proceedings of the 2024 IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–17 May 2024; pp. 1971–1977. [Google Scholar]
  4. Cramer, J.; Decraemer, B.; Afzal, M.R.; Kellens, K. A user-friendly toolkit to select and design multi-purpose grippers for modular robotic systems. Procedia CIRP 2024, 127, 2–7. [Google Scholar] [CrossRef]
  5. Ai, S.; Chen, Y. Multifunctional gripper design at the end of the robot. Instrumentation 2022, 9, 4. [Google Scholar]
  6. Kladovasilakis, N.; Sideridis, P.; Tzetzis, D.; Piliounis, K.; Kostavelis, I.; Tzovaras, D. Design and development of a multi-functional bioinspired soft robotic actuator via additive manufacturing. Biomimetics 2022, 7, 105. [Google Scholar] [CrossRef]
  7. Yildiz, B.S.; Pholdee, N.; Bureerat, S.; Yildiz, A.R.; Sait, S.M. Robust design of a robot gripper mechanism using new hybrid grasshopper optimization algorithm. Expert Syst. 2021, 38, e12666. [Google Scholar] [CrossRef]
  8. Nguyen, T.V.; Huynh, N.T.; Vu, N.C.; Kieu, V.N.; Huang, S.C. Optimizing compliant gripper mechanism design by employing an effective bi-algorithm: Fuzzy logic and ANFIS. Microsyst. Technol. 2021, 27, 3389–3412. [Google Scholar] [CrossRef]
  9. Pinskier, J.; Wang, X.; Liow, L.; Xie, Y.; Kumar, P.; Langelaar, M.; Howard, D. Diversity-Based Topology Optimization of Soft Robotic Grippers. Adv. Intell. Syst. 2024, 6, 2300505. [Google Scholar] [CrossRef]
  10. Lee, L.Y.; Syadiqeen, O.A.; Tan, C.P.; Nurzaman, S.G. Closed-structure compliant gripper with morphologically optimized multi-material fingertips for aerial grasping. IEEE Robot. Autom. Lett. 2021, 6, 887–894. [Google Scholar] [CrossRef]
  11. Sun, Y.; Liu, Y.; Pancheri, F.; Lueth, T.C. Larg: A lightweight robotic gripper with 3-d topology optimized adaptive fingers. IEEE/ASME Trans. Mechatronics 2022, 27, 2026–2034. [Google Scholar] [CrossRef]
  12. Zablocki, É.; Ben-Younes, H.; Pérez, P.; Cord, M. Explainability of deep vision-based autonomous driving systems: Review and challenges. Int. J. Comput. Vis. 2022, 130, 2425–2452. [Google Scholar] [CrossRef]
  13. Ghasemieh, A.; Kashef, R. Towards explainable artificial intelligence in deep vision-based odometry. Comput. Electr. Eng. 2024, 115, 109127. [Google Scholar] [CrossRef]
  14. Wang, R.; Wang, S.; Du, S.; Xiao, E.; Yuan, W.; Feng, C. Real-time soft body 3D proprioception via deep vision-based sensing. IEEE Robot. Autom. Lett. 2020, 5, 3382–3389. [Google Scholar] [CrossRef]
  15. Choi, H.; Lee, J.; Yi, Y.; Na, H.; Kang, K.; Kim, T. Deep vision-based occupancy counting: Experimental performance evaluation and implementation of ventilation control. Build. Environ. 2022, 223, 109496. [Google Scholar] [CrossRef]
  16. Wang, Z.; Jin, L.; Wang, S.; Xu, H. Apple stem/calyx real-time recognition using YOLO-v5 algorithm for fruit automatic loading system. Postharvest Biol. Technol. 2022, 185, 111808. [Google Scholar] [CrossRef]
  17. Zou, W.; Zhang, D.; Lee, D.J. A new multi-feature fusion based convolutional neural network for facial expression recognition. Appl. Intell. 2022, 52, 2918–2929. [Google Scholar] [CrossRef]
  18. Jiang, M.; Yin, S. Facial expression recognition based on convolutional block attention module and multi-feature fusion. Int. J. Comput. Vis. Robot. 2023, 13, 21–37. [Google Scholar] [CrossRef]
  19. Li, H.; Vong, C.M.; Wong, P.K.; Ip, W.F.; Yan, T.; Choi, I.C.; Yu, H.H. A multi-feature fusion method for image recognition of gastrointestinal metaplasia (GIM). Biomed. Signal Process. Control 2021, 69, 102909. [Google Scholar] [CrossRef]
  20. Hu, F.; Qian, M.; He, K.; Zhang, W.a.; Yang, X. A novel multi-feature fusion network with spatial partitioning strategy and cross-attention for armband-based gesture recognition. IEEE Trans. Neural Syst. Rehabil. Eng. 2024, 32, 3878–3890. [Google Scholar] [CrossRef]
  21. Liu, C.; Xu, X. AMFF: A new attention-based multi-feature fusion method for intention recognition. Knowl. Based Syst. 2021, 233, 107525. [Google Scholar] [CrossRef]
  22. Zhang, Y.; Guo, Z.; Wu, J.; Tian, Y.; Tang, H.; Guo, X. Real-time vehicle detection based on improved yolo v5. Sustainability 2022, 14, 12274. [Google Scholar] [CrossRef]
  23. Shahria, M.T.; Sunny, M.S.H.; Zarif, M.I.I.; Ghommam, J.; Ahamed, S.I.; Rahman, M.H. A comprehensive review of vision-based robotic applications: Current state, components, approaches, barriers, and potential solutions. Robotics 2022, 11, 139. [Google Scholar] [CrossRef]
Figure 1. Simulation and experimental prototypes of the unmanned warehouse automated loading and unloading system.
Figure 1. Simulation and experimental prototypes of the unmanned warehouse automated loading and unloading system.
Actuators 14 00259 g001aActuators 14 00259 g001b
Figure 2. The single robotic arm with a gripper and a depth camera.
Figure 2. The single robotic arm with a gripper and a depth camera.
Actuators 14 00259 g002
Figure 3. Three flag states in the operation process. (a) The state before insertion in the locking procedure or after extraction in the unlocking procedure; (b) the state before rotation in the locking procedure or after rotation in the unlocking procedure; (c) the state after rotation in the locking procedure or before rotation in the unlocking procedure.
Figure 3. Three flag states in the operation process. (a) The state before insertion in the locking procedure or after extraction in the unlocking procedure; (b) the state before rotation in the locking procedure or after rotation in the unlocking procedure; (c) the state after rotation in the locking procedure or before rotation in the unlocking procedure.
Actuators 14 00259 g003
Figure 4. Type I Embedded Two-Finger Gripper and Its Limitations.
Figure 4. Type I Embedded Two-Finger Gripper and Its Limitations.
Actuators 14 00259 g004
Figure 5. Type II Embedded Two-Finger Gripper and Its Limitations.
Figure 5. Type II Embedded Two-Finger Gripper and Its Limitations.
Actuators 14 00259 g005
Figure 6. Type III Embedded Two-Finger Gripper and Its Limitations.
Figure 6. Type III Embedded Two-Finger Gripper and Its Limitations.
Actuators 14 00259 g006
Figure 7. Comparison among three types of gripper.
Figure 7. Comparison among three types of gripper.
Actuators 14 00259 g007
Figure 8. The final optimized multi-functional mechanical gripper design. (a) Locking and unlocking manual adjustable twist locks; (b) Correcting the posture of the transport container 1; (c) correcting the posture of the transport container 2; (d) positioning and securing the transport container 1; (e) positioning and securing the transport container 2.
Figure 8. The final optimized multi-functional mechanical gripper design. (a) Locking and unlocking manual adjustable twist locks; (b) Correcting the posture of the transport container 1; (c) correcting the posture of the transport container 2; (d) positioning and securing the transport container 1; (e) positioning and securing the transport container 2.
Actuators 14 00259 g008
Figure 9. Recognition results showing multiple detected features. The primary target is the Corner, while other features provide supplementary information for pose estimation.
Figure 9. Recognition results showing multiple detected features. The primary target is the Corner, while other features provide supplementary information for pose estimation.
Actuators 14 00259 g009
Figure 10. The control scheme of deep vision-based operation control.
Figure 10. The control scheme of deep vision-based operation control.
Actuators 14 00259 g010
Figure 11. The simulation experiment of operating the fully automatic twist lock by a multi-functional mechanical gripper.
Figure 11. The simulation experiment of operating the fully automatic twist lock by a multi-functional mechanical gripper.
Actuators 14 00259 g011
Figure 12. The pose curves of the multi-functional mechanical gripper in simulations. (a) The position curves during the loading process; (b) the attitude angle curves during the loading process; (c) the position curves during the unloading process; (d) the attitude angle curves during the unloading process.
Figure 12. The pose curves of the multi-functional mechanical gripper in simulations. (a) The position curves during the loading process; (b) the attitude angle curves during the loading process; (c) the position curves during the unloading process; (d) the attitude angle curves during the unloading process.
Actuators 14 00259 g012aActuators 14 00259 g012b
Figure 13. The prototype experiment of operating the fully automatic twist lock by a multi-functional mechanical gripper.
Figure 13. The prototype experiment of operating the fully automatic twist lock by a multi-functional mechanical gripper.
Actuators 14 00259 g013
Figure 14. The perception results of the depth camera.
Figure 14. The perception results of the depth camera.
Actuators 14 00259 g014
Figure 15. The pose curves of the multi-functional mechanical gripper in experiments. (a) The position curves during the loading process; (b) the attitude angle curves during the loading process; (c) the position curves during the unloading process; (d) the attitude angle curves during the unloading process.
Figure 15. The pose curves of the multi-functional mechanical gripper in experiments. (a) The position curves during the loading process; (b) the attitude angle curves during the loading process; (c) the position curves during the unloading process; (d) the attitude angle curves during the unloading process.
Actuators 14 00259 g015aActuators 14 00259 g015b
Table 1. Comparative results between single-feature and multi-feature perception method.
Table 1. Comparative results between single-feature and multi-feature perception method.
ParametersValue
Controller-CPUCore i7-12700
Controller-GPURTX4060
Number of DoF6 (arm) + 1 (gripper)
K 0.02 0.01 0.01 0.005 0.005 0.009
D 0.0015 0.0011 0.0013 0.00035 0.00033 0.00087
T c d 1 0 0 50 0 1 0 0 0 0 1 120 0 0 0 1 m m
Table 2. Comparative results between single-feature and multi-feature perception method.
Table 2. Comparative results between single-feature and multi-feature perception method.
MethodSuccess RateAccuracy
Single-feature88.26%3.7 mm
Multi-feature95.72%1.2 mm
Table 3. Comparative results between with and without deep vision.
Table 3. Comparative results between with and without deep vision.
MethodEfficiencyAccuracyControl Cycle
With deep vision63.4%15.3 mm35 ms
Without deep vision97.1%1.2 mm8 ms
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Y.; Guo, S.; Zhang, J.; Ding, H.; Zhang, B.; Cao, A.; Sun, X.; Zhang, G.; Tian, S.; Chen, Y.; et al. Optimized Design and Deep Vision-Based Operation Control of a Multi-Functional Robotic Gripper for an Automatic Loading System. Actuators 2025, 14, 259. https://doi.org/10.3390/act14060259

AMA Style

Wang Y, Guo S, Zhang J, Ding H, Zhang B, Cao A, Sun X, Zhang G, Tian S, Chen Y, et al. Optimized Design and Deep Vision-Based Operation Control of a Multi-Functional Robotic Gripper for an Automatic Loading System. Actuators. 2025; 14(6):259. https://doi.org/10.3390/act14060259

Chicago/Turabian Style

Wang, Yaohui, Sheng Guo, Jinliang Zhang, Hongbo Ding, Bo Zhang, Ao Cao, Xiaohu Sun, Guangxin Zhang, Shihe Tian, Yongxu Chen, and et al. 2025. "Optimized Design and Deep Vision-Based Operation Control of a Multi-Functional Robotic Gripper for an Automatic Loading System" Actuators 14, no. 6: 259. https://doi.org/10.3390/act14060259

APA Style

Wang, Y., Guo, S., Zhang, J., Ding, H., Zhang, B., Cao, A., Sun, X., Zhang, G., Tian, S., Chen, Y., Ma, J., & Chen, G. (2025). Optimized Design and Deep Vision-Based Operation Control of a Multi-Functional Robotic Gripper for an Automatic Loading System. Actuators, 14(6), 259. https://doi.org/10.3390/act14060259

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop