Next Article in Journal
Effective Thermal Diffusivity Measurement Using Through-Transmission Pulsed Thermography: Extending the Current Practice by Incorporating Multi-Parameter Optimisation
Previous Article in Journal
Research on the Assembly of Square Parts on Circular Production Lines Using Admittance Control for Six-DOF Robots
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Vision-Assisted Design of End Effector Pose in Robotic Mixed Depalletizing of Heterogeneous Cargo

1
Department of Industrial Engineering, Universidad de los Andes, Bogotá 111711, Colombia
2
Integra S.A., Pereira 660003, Colombia
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(4), 1137; https://doi.org/10.3390/s25041137
Submission received: 4 December 2024 / Revised: 29 January 2025 / Accepted: 8 February 2025 / Published: 13 February 2025
(This article belongs to the Section Sensors and Robotics)

Abstract

:
Automated depalletizing systems aim to offer continuous and efficient operation in warehouse logistics, reducing cycle times and contributing to worker safety. However, most commercially available depalletizing solutions are designed primarily for highly homogeneous cargo arranged in orthogonal configurations. This paper presents a real-time approach for depalletizing heterogeneous pallets with boxes of varying sizes and arbitrary orientations, including configurations where the topmost surfaces of boxes are not necessarily parallel to each other. To accomplish this, we propose an algorithm that leverages deep learning-based machine vision to determine the size, position, and orientation of boxes relative to the horizontal plane of a robot arm from sparse depth data. Using this information, we implement a path planning method that generates collision-free trajectories to enable precise box grasping and placement onto a production line. Validation through both simulated and real-world experiments demonstrates the feasibility and accuracy of this approach in complex industrial settings, highlighting potential improvements in the efficiency and adaptability of automated depalletizing systems.

1. Introduction

Automated depalletizing solutions consist of systematically removing goods or materials from a pallet to place them on a production line [1]. They can operate continuously, significantly reducing cycle times compared to manual methods and enhancing production rates. By performing repetitive tasks with high accuracy and consistency, these systems minimize human error and can handle hazardous or heavy materials, reducing the risk of workplace injuries. Furthermore, these systems can be reprogrammed to adapt to changes in production lines or to handle new products without the need for extensive structural reconfiguration [2].
Most existing depalletizing systems, such as the one demonstrated by Etde [3], perform well with homogeneous and consistently orthogonal pallet configurations, but exhibit increased error rates when these conditions are not fully met. This paper explores the design and implementation of an automated depalletizing system tailored for heterogeneous pallets, starting by presenting the general depalletizing scheme, which outlines the system configuration and operational workflow. Subsequently, we present a deep learning-based computer vision algorithm designed to locate boxes in three-dimensional space, accounting for their position and orientation, using sparse depth information obtained from passive stereo cameras. This approach enables scene understanding of pallets composed of tilted boxes. Following this, the process for calculating the end effector orientation is detailed, ensuring alignment with the tilted surfaces of the boxes for precise handling. Finally, the paper examines the motion planning of the robotic manipulator, emphasizing the development of safe and efficient trajectories to perform the depalletizing process.
The paper is structured in five sections, starting with a literature review in Section 2, where an overview of the depalletizing process is given, discussing the main system requirements, relevant research in the field as well as the identified knowledge gaps in pallet and package recognition. Section 3 details the proposed methodology and materials used in the system, including the definition of the work area of the robotic manipulator, the method used to identify box parameters (locations, orientations and sizes), an algorithm for estimating the end effector pose, and the proposed approach for path planning. Section 4 presents the results, including a description of the experimental testing of the prototype in a laboratory environment and a detailed analysis of these results. Finally, Section 5 draws the relevant conclusions, summarizes the impact of the proposed depalletizing solution, and proposes knowledge gaps that may be addressed in future works.

2. Literature Review

The literature review is organized into four subsections. The first subsection describes the automatic depalletizing process and examines its industrial applications. The second subsection focuses on computer vision, specifically areas focused on object detection and pose estimation in conditions typical of depalletizing scenarios. The third subsection discusses mechanical designs that support depalletizing solutions, such as gripper designs and elevated conveyor belts. Finally, we identify knowledge gaps relevant to this review, highlighting areas that justify our contributions.

2.1. Automated Depalletizing

Generally speaking, the automated depalletizing process is composed of a system that includes an autonomous mobile robot, a robotic arm, a lifting mechanism, and a visual sensor, as indicated by the authors in [4,5]. The literature suggests depalletizing systems are designed mainly to extract items from homogeneous pallets. Such is the case of the algorithm developed by [6], which was designed to stack packages in a staggered arrangement to ensure structural stability. Chiaravalli in [7] proposed a solution that employs a multisensor vision system and a force-controlled collaborative robot to detect the boxes on the pallet. More than one robot can be employed in the depalletizing process, as proposed by Mingyang [8], who showed an ABB industrial robot and a KUKA mobile manipulator, expanding the depalletizing area. Additionally, Aleotti [9] ensured the accurate execution of tasks by implementing force measurement and robot admittance control.
Some applications that highlight the importance of the depalletizing process can be found in the healthcare sector, as demonstrated by Opaspilai [10] using a robotic system to handle products before dispensing medication through a SCARA robot, or in the food sector, where Arpenti [11] presents an approach for the detection, recognition, and position estimation of boxes of various types in a robotic cell. Caccavale in [12] proposes a robotic depalletizing system specifically designed to integrate seamlessly into supermarket logistics processes and Kalapyshina presents an innovative robotic system specifically designed to depalletize glass containers in an automated food packaging environment, known as DP3 [13].

2.2. Depalletizing-Focused Machine Vision

Conventional computer and machine vision techniques face difficulties in quickly discerning packing patterns in heterogeneous stacking configurations, especially in scenarios with additional complications related to light, whether light is scarce, or there is overexposure due to an overly bright source of light, or glare artifacts due to direct lighting on objects with high reflectivity. Some of these complications have been addressed by Holz in [14], which presents a system designed for depalletizing that includes a comprehensive process for accurate object detection and localization through a multi-resolution surface model approach; by Kim [15], who proposes a method based on deep learning and image processing for efficient detection of cluttered and diverse packet surfaces, employing RGBD technology; and by Costa [16], who investigates the integration of artificial intelligence in industrial robots for automated depalletizing. To address scarce or unstable ambient lighting, Li [17] proposes a two-step method to detect relevant objects. First, the region-growing method is used to extract the whole target region from the original image. Then, a region proposal is achieved by the Probabilistic Hough Transform (PPHT). However, morphological operations required to implement the method proposed by Li are highly sensitive to parameter tuning, and offer poor adaptability.

2.3. Mechanical Grasping in Pallets

Precise end effector positioning on industrial robots is essential to ensure stability and accuracy during the handling of boxes on tilted pallets. In this context, large positioning errors can arise due to disturbances that are difficult to compensate for using conventional control methods [18]. In non-orthogonal environments, any misalignment in the position of the end effectors can lead to incorrect stacking, affecting cargo stability and increasing the risk of damage to the packages. Furthermore, precise positioning contributes to operational safety by reducing the likelihood of operator injury due to falling boxes. This technical aspect is therefore essential to improve efficiency and safety in automated handling operations in the industry.
Regarding the design of the grip for robotic handling of boxes, several strategies ensure robust and safe gripping of work objects, adapting to the geometry and properties of the object, as shown in [19,20]. Among the most common techniques, the lateral or pincer grip applies pressure on opposite sides of the box and is suitable for regular-sized objects; the suction grip, using suction cups, allows handling boxes with smooth surfaces [21]; and the top or clamp grip holds the box from the top, providing stability for medium or large objects. Also notable are the corner grip, which offers additional contact points for small boxes, and the base grip, ideal for lifting heavy boxes. In addition, multi-finger grippers provide adaptability by adjusting to different orientations and shapes [22], while the combined suction and pincer grip provide additional stability on irregular-sized objects. The choice of grip type depends on factors such as box size, weight, material, and weight distribution, as well as spatial restrictions [23] and even sensor accuracy [24].
To address low grasping accuracy and high failure rates, Guo et al. [25] designed a 3D vision-based robotic palletizing system using the Shape-NCC algorithm. Similarly, Fontanelli in [26] presented a flexible and adaptable gripper designed specifically for robotic depalletizing, thus facilitating the handling of complex-shaped boxes. For its part, Tanaka in [27] proposed a vacuum suction end effector designed specifically for depalletizing robots in distribution centers, offering increased load capacity. Similarly, the difficulties of optimizing the speed of depalletizing compared to the annual process have been observed, as presented by Eto in [28].
To the best of our knowledge, no other work addresses the depalletizing of heterogeneous pallets with tilted boxes in non-orthogonal arrangements, as detailed in [24]. Furthermore, the literature is limited concerning the incorporation of workspace filters to optimize inverse kinematics calculations, also highlighted in [16]. Another identified area of improvement in depalletizing processes is the consideration of only the width and length of the packages, obtained by simple observation of the topmost surfaces, discarding height information. This limitation, pointed out in [28], is addressed in this study.

3. Materials and Methods

3.1. Proposed Depalletizing Scheme

In the proposed scheme, a Universal Robots UR10 cobot is positioned on a platform near a conveyor belt and a movable pallet base. The UR10 is a serial manipulator, a type of robot that can reach various positions and orientations within a working area approximately defined by a spherical volume, as illustrated in Figure 1b.
A depth camera composed of two lateral infrared sensors and a central RGB sensor, located on the end effector of the manipulator, is positioned to capture the entire pallet to obtain its physical dimensions and location within the working space of the robotic manipulator. To estimate these values, it is necessary to move the robotic manipulator to the pallet identification pose. This pose will only be used once per depalletizing cycle. This same pose is used to determine the maximum height of the boxes per stratum of the pallet. Once the height of the pallet is known, the end effector of the robotic manipulator is positioned on the pallet, allowing a top view of the boxes. This pose will be called a box acquisition pose. Each time the end effector is positioned in the acquisition pose, the sensor scans for the pose of a selected box, and this information is processed to start the path-planning procedure, which is divided into three stages: Picking, Moving, and Placing.
The Picking stage starts with the origin pose obtained by the box acquisition model, where a position and an orientation are defined for the end effector to pick the box, along with a joint configuration that will reach said position and orientation. The Moving stage executes the sequence of movements that make up the path of the robotic manipulator. It starts with calculating the center of the box to move the gripper to the position of the box, hold it, lift it, and move it toward the conveyor belt. It is configured to connect the Picking and Placing positions in an ergonomic way for the manipulator. Finally, the Placing stage consists of lowering and dropping the box on the belt in a fixed position and returning to the identification pose of the box, completing a depalletizing cycle per box. This cycle is repeated until the camera detects the empty pallet, indicating that the process has finished, using a message.

3.2. Box Acquisition

Imaging sensors provide little inherent understanding of the scenes they capture. In the box acquisition stage, the boxes must be identified in both the projected UV space defined by pixels and the three-dimensional coordinates. To achieve this, the following distinct problems must be addressed: box detection, pose estimation, pose estimation, and hand–eye calibration.

3.2.1. Box Detection

Locating an object of interest within a digital image by estimating the coordinates and size of its bounding box is a computer vision task known as object detection. Given that boxes on pallets are typically near each other, bounding boxes can only separate different objects in the scene if they are rectangular and have their horizontal and vertical axes aligned with those of the image. This also assumes that no occlusion occurs, as boxes that overlap each other make feature matching more challenging. To address the boxes where one or several of the conditions cannot be guaranteed, the box detection algorithm must be able to identify either non-rectangular polygons or specific pixels corresponding to each object detection. This is formally known as an instance segmentation problem. Examples of successful instance segmentation results are displayed in Figure 2.
We employ the detection and segmentation architectures proposed by Ultralytics on their YOLO series of models [29] trained on synthetic data. A portion of the mentioned data intentionally replicate challenging lighting conditions, with cluttering and glare in particular being factors that strongly influence the performance of the algorithm. A total of 1500 image sets replicating the data captured by the camera array (left grayscale, right grayscale and center color) were generated and then split into train/validation/test following a 60/20/20 ratio. Samples of generated data can be seen in Figure 3. Each detection is used to extract a cropped patch from the image that serves as an input for the next stage of visual processing.

3.2.2. Pose Estimation from Corner Position

We propose a heatmap-based corner detection algorithm inspired by the pallet center detection method presented by Nvidia, where synthetic data are used to train a UNet segmentation model with a Resnet-18 backbone. The model architecture used in our work is presented in Figure 4. The segmentation model was trained for 100 epochs on the expanded crops produced by the detection step.
A low-resolution pass through the segmentation network produces coarse predictions for the corner locations that are then used to formulate region proposals, where the pixel prediction is refined by combining the heatmap output with traditional edge detection procedures. This produces a two-dimensional array where each value represents the likelihood of the indices for said value to correspond to the UV coordinates of a pixel containing one of the four corners of the topmost face. As each corner is detected, the array is masked on its UV coordinates so that nearby pixels of similar intensity are not confused for a new corner. This procedure is repeated for the RGB image captured by the central image sensor as well as the greyscale infrared images from the stereoscopic array. A simple ordering algorithm based on the angle of vectors for each corner relative to the centroid of all corners allows for matching the features on all images. Examples of corner heatmap predictions and refined corner position estimations are shown in Figure 5.
With the variation of the relative x pixel coordinates for each corner between the left and right infrared images, it is possible to calculate a sparse disparity value, which can then be related to a numerical depth value according to the relationship presented in Equation (1).
d e p t h [ m m ] = f x [ p x ] r v [ m m ] d i s p a r i t y [ p x ]
In this equation, f x is the focal distance on the horizontal plane expressed in pixels, obtained from a camera calibration procedure, and r v is a reference value given by the distance between the two cameras, which for the OAK-D Lite is equal to 75 mm. The greater the disparity, the closer the distance to the camera. With known depths, it is possible to obtain the three-dimensional coordinates for each corner relative to the camera, which can then be used to obtain the absolute size of the topmost face of the box and the coordinates of its center.

3.2.3. Hand–Eye Calibration

To obtain the motion plan in which the robot picks up the box from the pallet, the coordinates and orientation must be expressed relative to the coordinate frame of the base of the robot, not relative to the camera. The problem of finding a homogeneous transformation matrix H c a m e e that transforms the coordinates with respect to (w.r.t) the camera to coordinates w.r.t the robotic manipulator is known as the hand–eye calibration problem. This variation, where the camera is mounted on the end effector, is called eye-in-hand calibration. Once obtained, the calibration matrix will be valid as long as the configuration of the camera relative to the end effector remains stationary. Marker-based procedures consist of capturing physical patterns from several different poses selected by a human worker; the discernible key points of the pattern remain visible but the viewpoints are also sufficiently dissimilar to reduce ambiguity when solving for the H c a m e e and H o b j e c t c a m in Equation (2).
H o b j e c t r o b o t = H c a m r o b o t H o b j e c t c a m = H e e r o b o t H c a m e e H o b j e c t c a m
We propose a flexible method that simplifies the calibration procedure and does not rely on externally calibrated patterns. As three-dimensional data are already available from the corner position estimation step, the number of captures needed to solve for all variables in Equation (2) is reduced. This also allows for the automation of the acquisition of new viewpoints, requiring only a static box that fits entirely within the field of view of the robot. After an initial capture of 3D coordinates, it is assumed that the pose of the box w.r.t the end effector is exactly equal to the pose of the box w.r.t the camera. A series of known transformations are applied equally to both the camera and the end effector by moving the robot to different configurations that alter both the position and the orientation of the end effector. Considering the known transformation matrices, the system results in a matrix equation of the form:
A X = X B
A and B are known transformations, from the base of the robot to the end effector and from the detected boxes to the origin of the camera, respectively, while X is the unknown transformation from the camera to the end effector: the hand–eye calibration matrix. This is a particular form of Sylvester’s equation, and for transformation, matrices can be solved using methods such as Tsai–Lenz. After obtaining the hand–eye calibration matrix, the transformation of the box with the highest z coordinate is set as the origin pose for the path planning stages.

3.3. Estimating the End Effector Pose in Depalletizing Processes

In depalletizing scenarios where the boxes that compose the pallet are not aligned with the coordinate axes of the robotic manipulator base, it is crucial to estimate the rotation angle necessary to align the end effector plane of the manipulator with the plane of the box. This alignment ensures the best possible pose to effectively grasp the box and carry out the depalletizing process using pick-and-place techniques. Below, we present the pseudo-code of the proposed Algorithm 1 to estimate the position of the end effector in a depalletizing process from a heterogeneous pallet.
Algorithm 1 Box orientation estimation
     Input:  C b (4 × 3 box corner coordinate matrix)
     Output:  R b (3 × 3 rotation matrix defining box orientation)
1:
P r [ 0 , 0 , 1 , 0 ]
▹ Define reference plane at the origin (z = 0)
2:
M b s u b d i v i d e ( C b )
▹ Generate plane mesh points
3:
P b l e a s t _ s q u a r e s ( M b )
▹ Compute plane equation for the box surface
4:
V r ( P r × P b )
▹ Compute rotation vector from cross-product
5:
R b Rodrigues ( V r )
▹ Convert rotation vector to rotation matrix
6:
return  R b
The first step is to establish the horizontal reference plane P r (line 1). Next, the plane corresponding to the top face of the box is defined using the mesh M b , generated from a 3 × 3 grid and the vertices of the box (line 3). Then, the normals of planes P r and M b are determined, and the axis of rotation is calculated using the cross-product between these normals (line 4). Finally, this rotation vector is normalized to calculate the rotation matrix that will allow the planes to be aligned using Rodrigues’s formula (line 5).
To estimate the angle of rotation between the plane of the box and the plane of the base of the robotic manipulator, the first step is to define the planes from the x, y, and z coordinates of the top vertices of the box corners. With these vertices, additional points have been calculated that form a 3 × 3 grid, called “the initial points”. Simultaneously, a reference plane is established parallel to the robot base plane, taking as reference the first upper vertex of the box. From the dimensions of the box, a 3 × 3 grid orthogonal to the axes of the robotic manipulator, called “the endpoints”, is calculated. Each of these two sets of points represents a plane, and the objective is to find the rotation angles on the different axes between these two planes.
The next step in finding these angles is to estimate the constants of each of the planes using a system of linear equations based on the least squares method. This method minimizes the sum of squares of the differences (residuals) between the values predicted by the plane and the actual values, thus fitting the parameters of the plane to the observed points. The general equation of the plane is:
A x + B y + C z + D = 0
Next, the normal vectors of the planes are calculated through the cross-product of vectors. This normal vector is essential to determine the required rotation; in case the planes are parallel (i.e., when the normal vectors are collinear), the angle of rotation will be zero.
Once the rotation vector needed to align the planes has been determined, Rodrigues’s formula is used to obtain the rotation matrix around the axis defined by the vector and the calculated angle. See Equation (5). It is important to take into account the order of multiplications of the rotation matrices, since rotations in three-dimensional space do not commute (the order affects the final result) [30,31].
Rodrigues’s method is used to find the rotation matrix in three dimensions from a rotation vector and a rotation angle. This technique is particularly useful when one wishes to rotate a vector or an object around an arbitrary axis in 3D. In Figure 6, the rotation axis is obtained by the cross-product between the normal vectors of the initial and final planes. If the angle between the two normal vectors is zero, it means that the vectors are collinear [32].
v = v c o s ( θ ) + ( k × v ) s e n ( θ ) + k ( k × v ) ( 1 c o s ( θ ) )
where v′ is the resultant vector after rotation, k is the axis of rotation, a unit vector, θ , is the angle of rotation in radians, and k × v is the cross-product. Rodrigues’s formula can also be expressed in a matrix formula:
R ( θ ) = I + s i n ( θ ) K + ( 1 c o s ( θ ) ) K 2
where R( θ ) is the resultant rotation matrix, I is the 3 × 3 identity matrix, and K is the matrix of the vector version of the cross-product of the rotation axis k, also known as the antisymmetric matrix or Lie matrix, given by:
K = 0 k z k y k z 0 k x k y k x 0
Here, k x , k y , and k z are the components of the vector k.
To validate the obtained rotation matrix, the rotation angles are decomposed using Euler’s convention in X Y Z order. Subsequently, a reverse process is applied, where the calculated rotation angles are used to reconstruct the rotation matrix and, consequently, the homogeneous transformation matrix. These calculated angles are sent to the robotic manipulator to orient the end effector so that it is parallel to the box surfaces, thus ensuring efficient and accurate depalletizing.

3.4. Motion Planning

The motion planning process for the automated depalletizing system relies on the definition of a series of specific end effector poses that serve as waypoints in a blended trajectory. Successive waypoints are constructed from the location of boxes and known static geometry information of the workspace to avoid collisions. Once the vision module correctly identifies a box and estimates its center, it signals the path planning module to start the planning procedure. This procedure combines the PRM (Probabilistic Roadmap) algorithm from [33] and the A* algorithm as implemented in [34], optimized for the static environment of robotic depalletizing. Initially, the PRM generates a configuration graph by sampling random points in the workspace of the manipulator, discarding those that collide with obstacles such as nearby boxes or the operational area boundaries. Valid points are connected using the CKDTree nearest neighbors search algorithm, forming an undirected graph that represents the robot’s possible trajectories. From this graph, the A* algorithm is employed to find the shortest path between the picking and placing positions, avoiding collisions, and considering cumulative costs.
Once the path is computed, it is transformed from the Cartesian space to the robotic configuration space of the manipulator using inverse kinematics, which results in a sequence of waypoints. These waypoints represent critical intermediate positions through which the manipulator must traverse to complete the planned trajectory. The PRM-generated graph can be reused across multiple cycles without recomputing. Additionally, combining PRM and A* allows adaptation to different pallet configurations and static obstacles. A grid search process is performed to fine-tune the PRM planner parameters (such as the number of samples, connection radius, and the number of neighbors). In each iteration, a set of parameters is generated from a pre-determined distribution and then used to construct the PRM graph. Following this, A* is used to attempt to find a viable path. If a path is found, its cost is calculated based on the sum of the number of samples and the path length. If this cost is lower than previous ones, the parameters and path are saved. The proposed approach enhances sampling, as explained in [35], by directing nodes toward the goal. This involves guided sampling to generate more points near the initial, final, and intermediate points, reducing the execution time of the PRM.

4. Results

The results obtained are presented below and are structured in three main stages. First, the robotic cell configuration system used to carry out the experimental test is described. In the second stage, the virtual and experimental tests performed to validate the accuracy of the artificial vision methods used in locating and estimating the position of the boxes are detailed. Finally, in the third stage, the depalletizing path is presented, divided into the stages defined in the “Proposed Depalletizing Scheme” section, together with the analysis of the variables of time, distance, and average linear velocity. In addition, the behavior of the segmented path for each coordinate axis in the depalletizing process of a box is examined.

4.1. System Configuration

The system incorporates a UR10 robot, together with a VGC 10 gripper and an OAK-D Lite RGB-D camera attached to the robot end effector link. The configuration of the system is limited to holding the boxes by the topmost face; this represents a reduction in the possible solutions of the grasping problem.
The parallel slit of the complete pallet is 40 × 34 × 130 cm. The pallet is composed of nine levels of boxes, where each different level can exhibit one of the palletization sequences shown in Figure 7. The dimensions and the weight of the boxes are presented in Table 1.
The position units have been defined in meters, and the Cartesian coordinates corresponding to the positions of the boxes and the end effector are expressed relative to the frame of reference of the base of the robotic manipulator. This choice aims to standardize the process between the various modules that make up the system. A view of the system may be found in Figure 1a. The origin of the frame of reference ( x , y , z , r x , r y , r z ) is located at the base of the robot.

4.2. Box Acquisition

To validate the accuracy of the computer vision methods used to locate and estimate the pose of boxes from data acquired by the RGB-D sensors, both virtual and experimental tests were performed. For the object detection task, box recall and accuracy were measured for a validation dataset constructed in the same randomized manner as the training dataset, where the detection model achieves perfect mean average precision scores. No false positives or false negatives were encountered in real-world detection experiments.

4.2.1. Corner Detection

In the box of the corner position estimation step, the proposed heatmap regression network was pitted against FAST-CPDA [36], a corner detector based on morphological operations. The latter was chosen according to the performance evaluation presented by Zhang et al. [37], as it offered a reasonable balance between speed and accuracy, even when compared to more recent corner detection methods.
Our heatmap regressor and the FAST-CDPA algorithms were used to detect corner positions on 1000 images that had already been cropped to the region of interest containing boxes with an additional border margin of 20 pixels. All cropped images were scaled to the regressor input size of 512 × 512 pixels. Unlike the regressor, which by design will always output four corner coordinates (resulting in precision and recall being equal), FAST-CDPA occasionally produces a greater or lower number of corner detections. Whenever the distance between any ground truth corner and the nearest detected corner was greater than 5 pixels (the size of the Gaussian kernel used to produce ground truth heatmaps), a false negative was recorded. In a similar manner, every detected corner that is more than 5 pixels away from all ground truth corners was registered as a false positive. Average position error is computed only for true positives, and is reported along with precision and recall for each method in Table 2 for both methods.
Under the mentioned conditions, the average position error for the UNet-based heatmap regressor was 1.26 pixels, while FAST-CDPA exhibited an average position error of 0.91 pixels. However, precision and recall performance are significantly lower for the FAST-CDPA algorithm. Average computation time was 52 ms for the heatmap regressor while running on an NVidia Quadro T600 GPU; FAST-CDPA on average took 36 ms on an Intel Core i7-12700 CPU. These results show that the proposed method achieves competitive corner detection performance with FAST-CDPA. From the results presented in Figure 8, it can be inferred that the proposed method is more robust to lighting disturbances, irregular box edges and clutter.

4.2.2. Box Pose Estimation

The accuracy and precision of three-dimensional corner position estimation were evaluated by placing a box in a fixed position and manually configuring the robot so that the gripper was aligned with the center of the box. As it was the box most prone to orientation errors due to having the shortest edges, box c1 was chosen for this experiment to evaluate performance in a worst-case scenario. Joint angles were recorded on this configuration to obtain a ground truth for the transformation of the box relative to the rest of the robot.
Following the acquisition of the ground truth orientation and position, these were estimated by placing the robot in 12 different configurations that ensured significantly different orientations and translations of the end effector and sensor relative to the box. These configurations resulted in multiple recorded viewpoints, shown for the left infrared sensor in Figure 9, which approximates the views of boxes that could feasibly be encountered in a heterogeneous pallet.
From these viewpoints, the size and pose of the box were estimated using the segmentation network, the disparity-to-depth algorithm, and Rodrigues’s formula. Table 3 summarizes the results of this experiment.
The average position error is 2.45 cm, while the average orientation error is 2.76 degrees. These results communicate the accuracy of both the box pose estimation algorithm and the hand–eye calibration algorithm. The former is in line with passive stereo techniques used in depalletizing systems and is within the acceptable tolerance of vacuum grippers like the VGC10 [24], while the latter is comparable to other automated hand–eye calibration procedures using closely spaced stereo cameras [38]. The processing of the images through the detection and segmentation model takes 62 ms on average on a computer running a Linux-based operating system with a Core i7-12700 CPU, 64 GB of DDR4 RAM an NVidia Quadro T600 with 4 GB of GDDR6 VRAM. This equates to approximately 15 frames per second (FPS), considerably faster than recent approaches based on dense depth [15] that operate between 1 and 2 FPS.

4.3. Motion Planning

The experimental test in the laboratory was carried out taking into account the motion stages established in the Proposed Depalletizing Scheme, which are picking, moving, and placing, as illustrated in Figure 10.

4.3.1. Velocity and Acceleration Limits

To define the kinematic limits for acceleration and velocity of the robotic manipulator, a sensitivity analysis was conducted. In this exercise, the distance variable was fixed, while the angular velocity limit for every joint was set to values between 0.5 and 3 rad/s and the angular acceleration to values between 0.1 and 2.1 rad/s². As shown in Figure 11, when varying acceleration within the established range, the time curve exhibited an analogous behavior regardless of the changes in velocity and acceleration. From an acceleration of 1.5 rad/s² onward, the trajectory execution time stabilizes. Regarding velocity, no significant differences in time behavior were observed for values of 2, 2.5 and 3 rad/s. However, it was found that velocities above 2.5 rad/s resulted in abrupt acceleration and deceleration of the joints. Consequently, an angular velocity of 2 rad/s was established, achieving the required operational stability.

4.3.2. Waypoint-Based Path Planning

A simulated experiment in a virtual environment was performed to establish a comparison between the basic PRM path planning algorithm combined with A* and an improved version using directed sampling. The initial parameters were explored using the following limits; the number of samples ranged from 100 to 200, the connection radius from 0.05 to 2, and the number of neighbors from 3 to 50. Fifty replicates were performed for each algorithm. The results are presented in Table 4.
For the basic PRM A*, three waypoints were generated to define the trajectory, with an average time of 11.59 s. Meanwhile, for the improved PRM A*, five waypoints were created in an average time of 8.49 s. The PRM used parameters generated by the improved algorithm, with the number of samples set to 100, a connection radius of 0.96, and 26 neighbors. Path planning used these values in subsequent experiments.

4.3.3. Traveled Distance and Path Execution Time

The trajectory is established in Cartesian space concerning the coordinate system located at the base of the robotic manipulator. During the depalletizing exercise, it was observed that the displacements of the end effector are distributed on the x-axis from −0.11 to 0.07 m, on the y-axis from −0.88 to 0.62 m, and on the z-axis from −0.7 to 0.6 m. The mean traversed distance was 2.82 meters, with trajectories completed in 6.73 s on average, excluding the time spent activating and deactivating the gripper. More time and distance data are available in Appendix A Table A1.
Figure 12 illustrates the behavior of the segmented path on each coordinate axis. In the graphs of the positions on the x and z axes, it is possible to identify the picking and placement stages on the plateaus, representing the moments in which the end effector remained stationary. It can also be observed that the position on the y axis remains constant, suggesting a reduction in the distance traveled.

5. Conclusions

Automated depalletizing of heterogeneous cargo requires a sufficient understanding of the workspace and path planning adapted to box surfaces exhibiting arbitrary orientations concerning the coordinate frame of the robot. We introduce a machine vision algorithm that can accurately retrieve three-dimensional pose and semantic information from sparse stereo-depth information obtained from RGB-D images. This algorithm detects and identifies box dimensions, position, and orientation at a faster rate than methods that rely on processing dense depth data, with greater adaptability to challenging lighting conditions and occlusions than high-performing morphological methods. We also propose a path-planning trajectory method based on probabilistic roadmaps that can perform automated depalletizing when boxes are stacked at tilted angles, implementing a grasp design procedure that accounts for their orientation and obstacles resulting from pallet geometry. The algorithms presented in this paper were validated experimentally on a robotic work cell composed of a UR10 robotic manipulator, a vacuum end effector (VGC10 gripper), an OAK-D Lite camera, and a conveyor belt. The results showed that the vision algorithm produces an average position error of 2.45 cm at distances ranging from 40 cm to 150 cm, displaying enough accuracy to successfully deconstruct pallets using the proposed path planning strategy. This strategy uses Rodrigues’s formula to adjust the gripper orientation according to the orientation of the topmost face of a box while using known robot constraints to reduce the computational complexity of path planning calculations.

Author Contributions

Conceptualization, C.M. and D.Á.; Data curation, S.V.; Formal analysis, S.V. and A.M.M.; Funding acquisition, C.M.; Methodology, S.V. and J.C.M.; Project administration, R.B.; Resources, R.B.; Software, S.V.; Supervision, D.Á.; Validation, S.V., J.C.M. and C.M.; Visualization, J.C.M.; Writing—original draft, S.V. and J.C.M.; Writing—review and editing, A.M.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was made possible thanks to funding from PATRIMONIO AUTÓNOMO FONDO NACIONAL DE FINANCIAMIENTO PARA LA CIENCIA, LA TECNOLOGÍA Y LA INNOVACIÓN FRANCISCO JOSÉ DE CALDAS and the support of the Universidad de Los Andes, Integra S.A. and OR4. The authors would also like to thank the Vice Presidency of Research and Creations Publication Fund at Universidad de los Andes for its financial support.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original data presented in the study are openly available at https://github.com/jcmartinez10/SEMA (accessed on 6 December 2024).

Conflicts of Interest

Author César Augusto Marín-Moreno and Rubén Bolaños were employed by the company Integra S.A. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
A IArtificial Intelligence
IoTInternet of Things
YOLOYou Only Look Once
PPHTProbabilistic Hough Transform
DOFDegree of freedom

Appendix A

Appendix A Table A1 presents the times in seconds associated with each action of the depalletizing path for each of the 31 boxes that make up the pallet. This table documents the specific duration of each stage of the process, providing a detailed perspective of the time invested in each action and facilitating the identification of possible areas of improvement in the robot programming. In a similar manner to the previous table, Appendix A Table A2 collects the distances traveled by the end effector of the robot during each trajectory step, as well as the total accumulated distance for the depalletizing of each box. The analysis of these values is essential to evaluate path efficiency and energy consumption in the robotic system, providing key information to improve both path planning and the layout of the pallet in future works.
Table A1. Mechanical timing results (all in seconds).
Table A1. Mechanical timing results (all in seconds).
IdentifyDown to PalletOn GripperMovingOff GripperReturnSum
11.94390.53212.79160.33272.01867.6189
21.20790.53262.22460.33251.80366.1012
31.93610.53332.2230.33331.92246.9481
42.2790.53232.2240.33152.01987.3866
52.22390.53242.1430.3331.80347.0357
62.25490.53332.01660.33151.82766.9639
72.48770.53322.15220.33281.92287.4287
82.71970.53252.1520.33272.01957.7564
92.65620.53292.02390.33131.99637.5406
102.47940.53292.07240.33221.80337.2202
112.51980.53281.92780.33191.82717.1394
122.7110.53332.07160.33181.92427.5719
133.01570.53212.07170.3322.01937.9708
142.95150.53321.92730.33311.99487.7399
152.65580.53272.00740.33211.80387.3318
163.06420.53222.01660.33141.9727.9164
172.68780.5321.84020.33151.82757.219
183.11170.53291.83910.33311.99427.811
193.23180.53171.82460.33261.99487.9155
202.80790.5321.92810.33231.82727.4275
212.80660.53361.72770.33251.82757.2279
223.27180.53261.82370.33231.99497.9553
233.14300.53321.91260.33171.92377.8442
243.13550.53221.72730.33261.92387.6514
253.37560.53281.82350.33281.82697.8916
263.59020.53231.82350.3321.99538.2733
273.40670.53351.72730.33271.80297.8031
283.6160.53261.72780.33261.97048.1794
293.54310.53341.92040.33181.92268.2513
303.53630.53271.72690.33331.82747.9566
313.77560.53261.72810.33271.99528.3642
Mean 2.84340.53271.97250.33231.9147.5949
Sd 0.58070.00050.22270.00060.08390.4789
Table A2. Mechanical distance results (all in meters).
Table A2. Mechanical distance results (all in meters).
Down to PalletMovingReturnSum
0.32021.51651.13112.9678
0.21331.31421.08732.6148
0.27711.31541.10182.6943
0.37661.31651.13112.8242
0.40531.20811.08732.7007
0.42631.0461.08892.5612
0.44171.20921.10182.7527
0.50961.21031.13112.851
0.50311.04781.12182.6727
0.52371.08811.08732.6991
0.5410.9261.08892.5559
0.55211.08921.10182.7431
0.60751.09031.13112.8289
0.6030.92781.12182.6526
0.64260.96811.08732.698
0.68660.97471.11462.7759
0.65750.8061.08892.5524
0.70920.80781.12182.6388
0.80990.8451.12182.7767
0.76420.9351.08892.7881
0.77520.7651.08892.6291
0.86630.8451.12182.8331
0.89690.921.10092.9178
0.9080.761.10092.7689
1.12370.8451.08893.0576
1.15390.8451.12183.1207
1.190.761.08733.0373
1.21280.7651.11463.0924
1.19210.931.10093.223
1.30980.7651.08893.1637
1.33560.7651.12183.2224
Mean0.72690.98731.10562.8198
Sd0.32070.20450.01670.1992

References

  1. Jocas, M.; Kurrek, P.; Zoghlami, F.; Gianni, M.; Salehi, V. Ai-based learning approach with consideration of safety criteria on example of a depalletization robot. In Proceedings of the Design Society: International Conference on Engineering Design; Cambridge University Press: Cambridge, UK, 2019; Volume 1, pp. 2041–2050. [Google Scholar]
  2. Setchi, R.M.; Lagos, N. Reconfigurability and reconfigurable manufacturing systems: State-of-the-art review. In Proceedings of the 2nd IEEE International Conference on Industrial Informatics, 2004. INDIN’04. 2004, Berlin, Germany, 24–26 June 2004; pp. 529–535. [Google Scholar]
  3. Sonoda, A.; Onoue, K.; Shimomura, N.; Kegasa, Y.; Hashimoto, Y.; Inada, T.; Enomoto, M.; Miyamoto, Y. Development of Palletizing and Depalletizing Robot Systems. [Palletize depalletize robot system no kaihatsu]. Kawasaki Tech. Rev. J. 1994, 13–19. [Google Scholar]
  4. Baldassarri, A.; Innero, G.; Di Leva, R.; Palli, G.; Carricato, M. Development of a mobile robotized system for palletizing applications. In Proceedings of the 2020 25th IEEE International Conference on Emerging Technologies and Factory Automation (ETFA), Vienna, Austria, 8–11 September 2020; Volume 1, pp. 395–401. [Google Scholar]
  5. Zaccaria, F.; Baldassarri, A.; Palli, G.; Carricato, M. A mobile robotized system for depalletizing applications: Design and experimentation. IEEE Access 2021, 9, 96682–96691. [Google Scholar] [CrossRef]
  6. Lee, J.D.; Chang, C.H.; Cheng, E.S.; Kuo, C.C.; Hsieh, C.Y. Intelligent robotic palletizer system. Appl. Sci. 2021, 11, 12159. [Google Scholar] [CrossRef]
  7. Chiaravalli, D.; Palli, G.; Monica, R.; Aleotti, J.; Rizzini, D.L. Integration of a multi-camera vision system and admittance control for robotic industrial depalletizing. In Proceedings of the 2020 25th IEEE International Conference on Emerging Technologies and Factory Automation (ETFA), Vienna, Austria, 8–11 September 2020; Volume 1, pp. 667–674. [Google Scholar]
  8. Mingyang, G.; Zhen, L.L.; Hettiarachchi, R. Multi-robot cooperation for mixed depalletizing. In Proceedings of the 2022 IEEE 17th Conference on Industrial Electronics and Applications (ICIEA), Chengdu, China, 16–19 December 2022; pp. 1528–1533. [Google Scholar]
  9. Aleotti, J.; Baldassarri, A.; Bonfè, M.; Carricato, M.; Chiaravalli, D.; Di Leva, R.; Fantuzzi, C.; Farsoni, S.; Innero, G.; Lodi Rizzini, D.; et al. Toward future automatic warehouses: An autonomous depalletizing system based on mobile manipulation and 3d perception. Appl. Sci. 2021, 11, 5959. [Google Scholar] [CrossRef]
  10. Opaspilai, P.; Vongbunyong, S.; Dheeravongkit, A. Robotic System for Depalletization of Pharmaceutical Products. In Proceedings of the 2021 7th International Conference on Engineering, Applied Sciences and Technology (ICEAST), Pattaya, Thailand, 1–3 April 2021; pp. 133–138. [Google Scholar] [CrossRef]
  11. Arpenti, P.; Caccavale, R.; Paduano, G.; Fontanelli, G.A.; Lippiello, V.; Villani, L.; Siciliano, B. RGB-D recognition and localization of cases for robotic depalletizing in supermarkets. IEEE Robot. Autom. Lett. 2020, 5, 6233–6238. [Google Scholar] [CrossRef]
  12. Caccavale, R.; Arpenti, P.; Paduano, G.; Fontanellli, A.; Lippiello, V.; Villani, L.; Siciliano, B. A flexible robotic depalletizing system for supermarket logistics. IEEE Robot. Autom. Lett. 2020, 5, 4471–4476. [Google Scholar] [CrossRef]
  13. Kalapyshina, I.I.; Nuzhdin, K.A. Design and Implementation the Robotics System for Glass Containers Depalletizing. In Proceedings of the I4SDG Workshop 2021: IFToMM for Sustainable Development Goals 1; Springer: Berlin/Heidelberg, Germany, 2022; pp. 299–306. [Google Scholar]
  14. Holz, D.; Topalidou-Kyniazopoulou, A.; Stückler, J.; Behnke, S. Real-time object detection, localization and verification for fast robotic depalletizing. In Proceedings of the 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–2 October 2015; pp. 1459–1466. [Google Scholar]
  15. Kim, S.; Truong, V.D.; Lee, K.H.; Yoon, J. Revolutionizing Robotic Depalletizing: AI-Enhanced Parcel Detecting with Adaptive 3D Machine Vision and RGB-D Imaging for Automated Unloading. Sensors 2024, 24, 1473. [Google Scholar] [CrossRef]
  16. Costa, J.P.G. Robotic System for Depalletization 2023. Available online: https://repositorio-aberto.up.pt/bitstream/10216/152053/2/636770.pdf (accessed on 1 December 2024).
  17. Li, J.; Kang, J.; Chen, Z.; Cui, F.; Fan, Z. A workpiece localization method for robotic de-palletizing based on region growing and PPHT. IEEE Access 2020, 8, 166365–166376. [Google Scholar] [CrossRef]
  18. Zou, X.; Ye, M.; Luo, C.; Xiong, J.; Luo, L.; Wang, H.; Chen, Y. Fault-tolerant design of a limited universal fruit-picking end-effector based on vision-positioning error. Appl. Eng. Agric. 2016, 32, 5–18. [Google Scholar]
  19. Cutkosky, M.R. Robotic Grasping and Fine Manipulation; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2012; Volume 6. [Google Scholar]
  20. Fantoni, G.; Santochi, M.; Dini, G.; Tracht, K.; Scholz-Reiter, B.; Fleischer, J.; Lien, T.K.; Seliger, G.; Reinhart, G.; Franke, J.; et al. Grasping devices and methods in automated production processes. CIRP Ann. 2014, 63, 679–701. [Google Scholar] [CrossRef]
  21. Jaiswal, A.; Kumar, B. Vacuum cup grippers for material handling in industry. Int. J. Innov. Sci. Eng. Technol. 2017, 4, 187–194. [Google Scholar]
  22. Ciocarlie, M.; Hicks, F.M.; Holmberg, R.; Hawke, J.; Schlicht, M.; Gee, J.; Stanford, S.; Bahadur, R. The Velo gripper: A versatile single-actuator design for enveloping, parallel and fingertip grasps. Int. J. Robot. Res. 2014, 33, 753–767. [Google Scholar] [CrossRef]
  23. Marwan, Q.M.; Chua, S.C.; Kwek, L.C. Comprehensive review on reaching and grasping of objects in robotics. Robotica 2021, 39, 1849–1882. [Google Scholar] [CrossRef]
  24. Kim, S.; Lee, K.H.; Kim, C.; Yoon, J. Vision-centric 3D point cloud technique and custom gripper process for parcel depalletisation. J. Intell. Manuf. 2024, 1–17. [Google Scholar] [CrossRef]
  25. Guo, A.; Jing, H.; Ge, C. Design and implementation of robot depalletizing system based on 3D vision. In Proceedings of the Fourth International Conference on Sensors and Information Technology (ICSI 2024), Xiamen, China, 5–7 January 2024; Volume 13107, pp. 392–397. [Google Scholar]
  26. Fontanelli, G.A.; Paduano, G.; Caccavale, R.; Arpenti, P.; Lippiello, V.; Villani, L.; Siciliano, B. A reconfigurable gripper for robotic autonomous depalletizing in supermarket logistics. IEEE Robot. Autom. Lett. 2020, 5, 4612–4617. [Google Scholar] [CrossRef]
  27. Tanaka, J.; Ogawa, A.; Nakamoto, H.; Sonoura, T.; Eto, H. Suction pad unit using a bellows pneumatic actuator as a support mechanism for an end effector of depalletizing robots. ROBOMECH J. 2020, 7, 1–30. [Google Scholar] [CrossRef]
  28. Eto, H.; Nakamoto, H.; Sonoura, T.; Tanaka, J.; Ogawa, A. Development of automated high-speed depalletizing system for complex stacking on roll box pallets. J. Adv. Mech. Des. Syst. Manuf. 2019, 13, JAMDSM0047. [Google Scholar] [CrossRef]
  29. Jocher, G.; Qiu, J.; Chaurasia, A. Ultralytics YOLO, 2023. Available online: https://github.com/ultralytics/ultralytics (accessed on 15 November 2024).
  30. Koks, D. Explorations in Mathematical Physics: The Concepts Behind an Elegant Language; Springer: Berlin/Heidelberg, Germany, 2006; Volume 1. [Google Scholar]
  31. Dai, J.S. Euler–Rodrigues formula variations, quaternion conjugation and intrinsic connections. Mech. Mach. Theory 2015, 92, 144–152. [Google Scholar] [CrossRef]
  32. Barrientos, A.; Álvarez, M.; Hernández, J.; Del Cerro, J.; Rossi, C. Modelado de Caden as Cinemáticas mediante Matrices de Desplazamiento. Una alternativa al método de Denavit-Hartenberg. Rev. Iberoam. Autom. Inform. Ind. 2012, 9, 371–382. [Google Scholar] [CrossRef]
  33. Kavraki, L.E.; Svestka, P.; Latombe, J.C.; Overmars, M.H. Probabilistic roadmaps for path planning in high-dimensional configuration spaces. IEEE Trans. Robot. Autom. 1996, 12, 566–580. [Google Scholar] [CrossRef]
  34. Duchoň, F.; Babinec, A.; Kajan, M.; Beňo, P.; Florek, M.; Fico, T.; Jurišica, L. Path planning with modified a star algorithm for a mobile robot. Procedia Eng. 2014, 96, 59–69. [Google Scholar] [CrossRef]
  35. LaValle, S.M.; Kuffner, J.J., Jr. Randomized kinodynamic planning. Int. J. Robot. Res. 2001, 20, 378–400. [Google Scholar] [CrossRef]
  36. Awrangjeb, M.; Lu, G.; Fraser, C.S.; Ravanbakhsh, M. A Fast Corner Detector Based on the Chord-to-Point Distance Accumulation Technique. In Proceedings of the 2009 Digital Image Computing: Techniques and Applications, Melbourne, VIC, Australia, 1–3 December 2009; pp. 519–525. [Google Scholar] [CrossRef]
  37. Zhang, Y.; Zhong, B.; Sun, X. A Benchmark for the Evaluation of Corner Detectors. Appl. Sci. 2022, 12, 11984. [Google Scholar] [CrossRef]
  38. Antonello, M.; Gobbi, A.; Michieletto, S.; Ghidoni, S.; Menegatti, E. A fully automatic hand-eye calibration system. In Proceedings of the 2017 European Conference on Mobile Robots (ECMR), Paris, France, 6–8 September 2017; pp. 1–6. [Google Scholar] [CrossRef]
Figure 1. Workspace definition.
Figure 1. Workspace definition.
Sensors 25 01137 g001
Figure 2. Segmentation results for a non-orthogonal view of a constructed pallet.
Figure 2. Segmentation results for a non-orthogonal view of a constructed pallet.
Sensors 25 01137 g002
Figure 3. Synthetic data generated for the purpose of training the object detection and segmentation algorithm. Detection bounding boxes are drawn in blue.
Figure 3. Synthetic data generated for the purpose of training the object detection and segmentation algorithm. Detection bounding boxes are drawn in blue.
Sensors 25 01137 g003
Figure 4. UNet architecture with a ResNet-18 backbone used in this paper to perform heatmap-based corner predictions.
Figure 4. UNet architecture with a ResNet-18 backbone used in this paper to perform heatmap-based corner predictions.
Sensors 25 01137 g004
Figure 5. Heat map predictions for corner locations (top) from input images captured by the camera array (bottom). Brighter pixel values indicate a higher probability of a corner existing at the related coordinates.
Figure 5. Heat map predictions for corner locations (top) from input images captured by the camera array (bottom). Brighter pixel values indicate a higher probability of a corner existing at the related coordinates.
Sensors 25 01137 g005
Figure 6. Visual representation of Rodrigues’s formula. The cross-product of the normal vector for the red plane by that of the horizontal plane results in the axis of rotation for the plane, displayed as the black vector on the figure.
Figure 6. Visual representation of Rodrigues’s formula. The cross-product of the normal vector for the red plane by that of the horizontal plane results in the axis of rotation for the plane, displayed as the black vector on the figure.
Sensors 25 01137 g006
Figure 7. Different types of levels expected during depalletizing. Starting clockwise from (af) every level increases in heterogeneity.
Figure 7. Different types of levels expected during depalletizing. Starting clockwise from (af) every level increases in heterogeneity.
Sensors 25 01137 g007
Figure 8. Qualitative comparison of our corner detection method with the high-performing FAST-CDPA. Blue circles represent the ground truth and green circles represent the predictions.
Figure 8. Qualitative comparison of our corner detection method with the high-performing FAST-CDPA. Blue circles represent the ground truth and green circles represent the predictions.
Sensors 25 01137 g008
Figure 9. Different viewpoints used in the pose estimation experiment.
Figure 9. Different viewpoints used in the pose estimation experiment.
Sensors 25 01137 g009
Figure 10. Stages of the trajectory.
Figure 10. Stages of the trajectory.
Sensors 25 01137 g010
Figure 11. Speed and acceleration sensitivity test.
Figure 11. Speed and acceleration sensitivity test.
Sensors 25 01137 g011
Figure 12. Evolution of end effector position by axis while reaching different waypoints.
Figure 12. Evolution of end effector position by axis while reaching different waypoints.
Sensors 25 01137 g012
Table 1. Box specifications (all in centimeters).
Table 1. Box specifications (all in centimeters).
BoxWidthLengthHeightQuantityWeight [g]
c112151210326.2
c21619129474.6
c315392431262.8
c416251210588.0
c5193264600.6
Table 2. Corner detection metrics.
Table 2. Corner detection metrics.
FAST-CPDAOurs
Average position error [px]0.911.26
Precision0.7910.939
Recall0.8370.939
Average processing time [ms]3652
Table 3. Position and orientation error.
Table 3. Position and orientation error.
PosePosition Error [m]Orientation Error [°]
10.01833.026
20.00912.6753
30.02234.8284
40.04463.3983
50.10912.1888
60.02072.2291
70.0121.4445
80.00812.7595
90.00940.8262
100.00455.352
110.01282.1061
120.02382.3623
Average0.02452.7663
Table 4. Comparison of PRM A* Basic and Improved for the estimation of PRM parameters.
Table 4. Comparison of PRM A* Basic and Improved for the estimation of PRM parameters.
DistributionNum SamplesConnection RadiusNum NeighborsTime (s)Total Samples GeneratedWaypoints
Random centered Basic1001.142611,591203
Improved Random1000.96268.491355
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Valero, S.; Martinez, J.C.; Montes, A.M.; Marín, C.; Bolaños, R.; Álvarez, D. Machine Vision-Assisted Design of End Effector Pose in Robotic Mixed Depalletizing of Heterogeneous Cargo. Sensors 2025, 25, 1137. https://doi.org/10.3390/s25041137

AMA Style

Valero S, Martinez JC, Montes AM, Marín C, Bolaños R, Álvarez D. Machine Vision-Assisted Design of End Effector Pose in Robotic Mixed Depalletizing of Heterogeneous Cargo. Sensors. 2025; 25(4):1137. https://doi.org/10.3390/s25041137

Chicago/Turabian Style

Valero, Sebastián, Juan Camilo Martinez, Ana María Montes, Cesar Marín, Rubén Bolaños, and David Álvarez. 2025. "Machine Vision-Assisted Design of End Effector Pose in Robotic Mixed Depalletizing of Heterogeneous Cargo" Sensors 25, no. 4: 1137. https://doi.org/10.3390/s25041137

APA Style

Valero, S., Martinez, J. C., Montes, A. M., Marín, C., Bolaños, R., & Álvarez, D. (2025). Machine Vision-Assisted Design of End Effector Pose in Robotic Mixed Depalletizing of Heterogeneous Cargo. Sensors, 25(4), 1137. https://doi.org/10.3390/s25041137

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop