Next Article in Journal
DBO-AWOA: An Adaptive Whale Optimization Algorithm for Global Optimization and UAV 3D Path Planning
Previous Article in Journal
Electrochemical Sensor Based on DNA Aptamers Immobilized on V2O5/rGO Nanocomposite for the Sensitive Detection of Hg(II)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A New Pallet-Positioning Method Based on a Lightweight Component Segmentation Network for AGV Toward Intelligent Warehousing

College of Mechanical and Electronic Engineering, Nanjing Forestry University, Nanjing 210037, China
*
Authors to whom correspondence should be addressed.
Sensors 2025, 25(7), 2333; https://doi.org/10.3390/s25072333
Submission received: 18 February 2025 / Revised: 4 April 2025 / Accepted: 5 April 2025 / Published: 7 April 2025
(This article belongs to the Section Sensing and Imaging)

Abstract

:
In human–robot hybrid intelligent warehouses, pallets often come in various shapes and sizes, posing challenges for AGVs to automate pallet picking. This, in turn, reduces the overall operational efficiency of the warehouse. To address this issue, this paper proposes a lightweight component segmentation network using a dual-attention mechanism to achieve precise segmentation of the pallet’s stringer board and accurate localization of the pallet slots. To overcome the challenge of redundant computations in existing semantic segmentation models, which are unable to balance spatial details and high-level semantic information, this network utilizes a dual-branch attention mechanism within an encoder–decoder architecture to effectively capture spatial details. On this basis, a residual structure is introduced to reduce redundant network parameters, addressing issues like vanishing and exploding gradients during training. Due to the lack of a public pallet image segmentation dataset, the network was tested using a custom-made dataset. The results show that by extracting intermediate-, low-, and high-level features from dual-branch input images and integrating them to construct multi-scale images, precise segmentation of various types of pallets can be achieved with limited annotated images. Furthermore, to comprehensively evaluate the model’s robustness, additional pallet localization experiments were conducted under varying illumination conditions and background noise levels. The results demonstrate that the proposed method can effectively identify and locate multi-category pallet targets while maintaining high segmentation accuracy under different lighting conditions and background interferences, verifying the model’s robustness in complex warehousing environments. Compared to the traditional model, the proposed model in this paper achieves a 10.41% improvement in accuracy and a 32.8% increase in image processing speed. The segmentation network we proposed is used for pallet-positioning experiments and has achieved good positioning results in pallet images taken from different distances and angles.

1. Introduction

The intelligent warehouse is a critical component of the supply chain in intelligent manufacturing, which cannot only reduce labor costs, minimize material loss, and enhance production safety, but also serve as a crucial guarantee for achieving intelligent production of products [1]. As the demand for global industrial upgrading continues to increase, many warehouses are facing the need for intelligent transformation and upgrades [2]. Using automated guided vehicle (AGV) forklifts in conjunction with pallets for automated material handling and storage is the main form of operation in current intelligent warehouses [3]. However, due to cost and geographical constraints, most intelligent warehouses have been upgraded from traditional manual warehouses. Consequently, many warehouses still operate in a human–robot hybrid mode, leading to random placement and orientation of pallets within the warehouse, which poses challenges for AGVs in automatically picking up and placing goods [4].
To accurately locate the position and posture of the pallet is a prerequisite for AGV to pick and place goods. Currently, the methods commonly used in the industry are QR code-assisted positioning and LiDAR positioning [5]. The QR code-based positioning method can effectively determine the type of goods and the position of the pallet [6]. However, in a human–robot hybrid warehouse, pallet placement angles and positions are not only random, but the types of pallets are also diverse. On the one hand, attaching QR codes consumes a large amount of labor and material costs. On the other hand, QR codes cannot provide information on the angle deviation of the pallet. When the surface of the pallet with the QR code is not facing the AGV directly, the AGV cannot even recognize and locate the pallet position [7]. For the LiDAR positioning method, apart from its high cost, there is also the challenge that when pallet types are inconsistent and the sizes of the insertion holes vary, the forklift finds it difficult to accurately locate the pallet slots for picking and placing goods [8]. Therefore, how to quickly and cost-effectively locate randomly placed pallets and achieve precise insertion and retrieval of goods by AGV forklifts are urgent problems that need to be solved in the field of smart warehousing [9].
To achieve the lowest cost, the image-based pallet location and recognition method has become a research hotspot in the field of intelligent warehouses. A pallet pose recognition algorithm based on an RGB-D camera was developed [10]. Image-based pallet-positioning methods primarily analyze image characteristics to detect pallets by identifying distinctive features within an image. Methods that utilize geometric [11] and color features [12] are proposed for pallet detection; these approaches employ mathematical morphology to define the edge points and corner points of the pallets. However, relying on color or geometric thresholds to distinguish pallets from backgrounds is unreliable, as actual warehouse environments often contain diverse types and colors of both pallets and backgrounds. The ROBOLIFT system detects pallets using region-growing algorithms [13]. However, this system fails to maintain positioning accuracy and reliability when dealing with different types of pallets or random placement angles. Subsequent research studies have proposed detecting the edges where pallets contact the ground directly and estimating their orientation; however, this method performs poorly in cases where edge features are not clearly defined.
An improved UPerNet network was developed to achieve pallet recognition in complex warehouse environments [14]. Although the researchers optimized the network to segment pallets in complex scene images, the algorithm is still limited to overall pallet recognition. Precise insertion of the forklift into the pallet after localization requires additional methods. Precise insertion of the pallet by an AGV forklift after localization still requires additional methods. For example, images and point clouds captured by an RGB-D camera were utilized to localize the pallet slots of the pallet [15]. First, the front face of the pallet was segmented using the YOLOv5 network. Subsequently, depth information and a point cloud segmentation algorithm were employed to establish specific segmentation criteria for determining the center coordinates of the pallet slots. This information was subsequently used to guide the forklift in transporting the pallet. This method is mainly suitable for situations where the forklift is facing the pallet directly and requires prior knowledge of the pallet’s features and dimensions. When the angle between the pallet and the AGV is random or the type of pallet changes, this method may produce significant errors. On the other hand, in human–robot hybrid warehouses, AGVs need to quickly assess the posture of target goods and free up resources to respond to the sudden appearance of people or other vehicles to prevent safety accidents. Currently, there are few reports on the optimization of response time and computational resources for pallet localization detection networks.
In human–robot hybrid intelligent warehouses, how to achieve pallet precise positioning through the obtained visual information is an urgent problem that needs to be solved in the field of intelligent warehousing [16]. Among these challenges, the precise segmentation of the target pallet stringer board is crucial for successfully completing pallet positioning for AGVs. To achieve accurate and efficient localization of pallet posture and slots, while ensuring the network has a degree of generalization ability, this work presents a lightweight component segmentation network based on a dual-attention mechanism (LCS-Net). The network integrates dual-attention mechanisms to enhance segmentation accuracy and optimizes the backbone to improve computational efficiency. Comprehensive experimental validation demonstrates the model’s ability to generalize across various pallet types. These innovations enable LCS-Net to improve segmentation accuracy and computational speed, addressing critical issues such as pallet positioning and slot identification in human–robot hybrid intelligent warehouse environments. Experimental results show that LCS-Net excels at accurately locating pallets of different types, meeting the real-time demands of complex warehouse environments. By combining multi-scale semantic information and focusing on the weights of different channels and spatial positions, the network improves component segmentation performance. Summing the outputs of the two attention modules further enhances feature representation. Additionally, by adaptively learning feature information and capturing rich contextual dependencies, the network can effectively tackle scene segmentation tasks, further improving segmentation performance.

2. Related Works

In human–robot hybrid intelligent warehouses, where pallet position deviations are significant, it is necessary to obtain precise location information of the target pallet through supplementary visual information. How to utilize this visual information to detect and accurately locate pallets is an urgent issue that needs to be addressed in the field of smart warehousing. In many applications, such as image recognition, target tracking, and robot navigation, image segmentation is a necessary prerequisite. Currently, there are still some urgent issues to be resolved in the application of image segmentation.

2.1. Blurry Edge Segmentation and Slow Segmentation Speed

In recent years, the development of deep learning techniques has brought new breakthroughs to image segmentation [17], and many studies have evaluated the performance of these algorithms. Among them, the FCN is used to extract buildings from high-resolution images [18], but FCN only uses high-level feature maps for pixel classification and low-level feature maps are discarded, resulting in a limited ability to process small targets. To address this issue, some semantic segmentation models that focus on handling low-level feature points have emerged, such as U-Net [19] and SegNet [20]. These semantic segmentation models adopt an encoder–decoder architecture, where the encoding region is used for feature extraction and the decoding region restores the extracted feature map to the original image scale. During the process of restoring feature maps, the decoding region fuses with the low-level feature maps of the encoding region to obtain feature information at different levels and improve segmentation accuracy. However, the high-level semantic features lack sufficient detailed information, resulting in poor segmentation results. In order to obtain advanced semantic information to solve the problem of classification errors, ResNet [21] gradually extracts advanced semantic information by continuously deepening the network. Later, the DeepLabv3+ network became popular [22]. It combines semantic information from different scales of receptive fields to improve the classification ability of the model. However, such networks rely on downsampling to obtain advanced semantic information, ignoring spatial details, and having problems such as blurred segmentation edge detection. Obtaining spatial detail information and high-level semantic information is always contradictory, but for solving the problem of inaccurate edge contour segmentation, the spatial detail information of images is equally important. In order to address related defects, dual-branch networks have been proposed [23], which separately extract processing spatial detail information through new branches and design feature fusion modules to fuse advanced semantic information and spatial detail information. This is very effective for segmenting edge contours, but it also brings about greater computational complexity. Solving the problem of segmentation accuracy types often requires more computational resources and longer network training times. Ref. [24] developed a preliminary method for image segmentation by introducing the convolutional block attention module (CBAM) into DeepLabv3+, which improved the original network’s shortcomings in rough edge target segmentation. The authors applied the improved network to autonomous driving semantic segmentation scenarios, but it still failed to solve the problem of slow network training speeds. Ref. [25] introduced depthwise separable convolution instead of ordinary dilated convolution in hetero receptive field splicing ASPP to accelerate model training speed. However, this improved network primarily addresses the weak detail representation ability of DeepLabv3+, which suffers from mis-segmentation issues. It still has shortcomings in real-time engineering applications.

2.2. Small-Sample Semantic Segmentation

With the rapid development of deep learning technology, breakthrough progress has been made in semantic segmentation methods using deep learning. However, the segmentation performance of such methods overly relies on a large amount of annotated training data. In addition, the high cost of per-pixel annotation greatly limits the development of image semantic segmentation, which is also unrealistic in some practical applications [26]. The small-sample semantic segmentation method mainly adopts a dual-branch network structure; it uses techniques [27] to extract the middle-, low-, and high-level features from the dual-branch input image, and then fuses and constructs multi-scale features. Based on this, an iterative optimization few-shot semantic segmentation model is proposed, and the efficiency of the proposed model is verified on the PASCAL and COCO datasets.
The quality of deep learning models depends on the size and quality of the dataset. The data in the warehousing field have problems such as a limited number of annotations, uneven sample distribution, and common data missing phenomena. Therefore, it is necessary to apply small-sample semantic segmentation research to the field of warehousing.

2.3. Generalization of the Segmentation Model

In the fields of autonomous driving and smart healthcare, automatic image segmentation has been extensively studied. The most advanced image segmentation models are typically based on structures such as fully convolutional networks (FCNs) and U-Net. These models have achieved high-precision segmentation results across various automatic segmentation tasks. In the implementation of automatic prediction tasks using deep learning neural networks, excellent performance often relies on the assistance of accurately labeled data for training. Therefore, deep models are data-driven methods that heavily depend on labeled information. In the medical field, researchers have found that although deep learning models demonstrate advanced performance on benchmark data, the lack of multi-center generalization leads to difficulties in applying these methods to real clinical problems [28]. In the field of intelligent warehousing, the types and shapes of pallets that need to be segmented vary widely. Automatic segmentation methods that heavily rely on data labeling often struggle to transfer learned knowledge from one center to another. As a result, models can only rely on their inherent self-awareness and generalization capabilities to handle previously unseen target domain data. This challenge is the primary focus of research on model generalization. Therefore, addressing the model generalization issue in the segmentation of multi-center intelligent warehousing scenarios is a practical problem closely related to computer-aided pallet localization.
In human–robot collaborative scenarios where pallet position deviations are significant, it is necessary to obtain precise location information of the target pallet through supplementary visual information. How to use this visual information to detect and accurately locate pallets is an urgent problem that needs to be solved in the field of smart warehousing. The warehousing field is characterized by complex application scenarios and a wide variety of target objects. The effectiveness of deep learning models depends on the size and quality of the dataset. However, in the warehousing field, the data often suffer from limited labeling, imbalanced sample distribution, and widespread data missing issues. Therefore, further improvements to existing object segmentation models are needed for practical application. To summarize, the following contributions are presented in this paper:
LCS-Net is proposed to address the issues of limited annotation quantity, uneven sample distribution, and common data missing in redundant computing and warehousing due to the inability of existing real-time semantic segmentation network models to balance spatial detail information and advanced semantic information. Key points include the following:
  • Obtaining spatial detail information through a dual-branch attention network and optimizing the weights of deep features to solve the problem of inaccurate pallet edge segmentation.
  • Extracting the middle-, low-, and high-level features of the dual-branch input image and fusing them to construct multi-scale images, which guide precise segmentation of unknown new classes using limited annotated images.
  • Introducing residual structures and inverted residual structures in the backbone network to reduce redundancy in network parameters. During the image generation phase, a noise component is introduced, increasing data diversity and further enhancing the model’s generalization ability.

3. Methodology

3.1. Using the Segmentation Network to Achieve Pallet Positioning

The process for pallet positioning is shown in Figure 1. Using the LCS-Net to segment the pallet stringer board, further image processing is performed on the pallet stringer board. To achieve the positioning of the center point of the pallet stringer board, it is necessary to first calibrate the extrinsic parameters of the camera in the warehouse working environment. This will determine the transformation relationship between the pixel coordinate system and the pallet stringer board coordinate system, generating a conversion formula between the two coordinate systems. By processing the image, the minimum enclosing rectangle of the segmented pallet stringer board can be obtained, from which the pixel coordinates of the center of the stringer board are derived. Using the coordinate transformation formula, the three-dimensional coordinates of the center point of the pallet stringer board can be obtained.

3.2. Basic Architecture of LCS-Net

The pallet stringer board is usually the same color as the other surfaces of the pallet, making the boundaries of these surfaces in the image blurry and difficult to distinguish. Therefore, segmenting individual stringer boards is challenging. This phenomenon is particularly noticeable in oblique and isometric views. To enhance edge recognition capabilities and maintain the scalability of network functions, the network proposed in this paper adopts an encoder–decoder structure. LCS-Net is shown in Figure 2. In the encoder backbone network, an improved Xception is used as the feature extractor, and a dual-attention module is used for multi-scale field perception. The position attention module and channel attention module operate in parallel within the dual-attention module. Firstly, the improved backbone network Xception extracts the input image and divides it into two parts: shallow features and deep features. The deep features are downsampled 16 times and then convolved with an expansion rate of 2 by 3 × 3. The encoding region of this article adopts a dual-branch network structure, which processes spatial detail information and high-level semantic information respectively, achieving high-precision and efficient real-time segmentation of pallet components, they are then separately fed into the position attention module and the channel attention module for processing. The position attention module aggregates the features of each position by weighting the sum of all position features, promoting classification accuracy between different local features. The channel attention module allocates appropriate attention weights to each channel based on the interdependence between channel maps, improving segmentation accuracy. Shallow features are downsampled by a factor of four before entering the decoding region of the network.

3.3. Optimization of Xception Network

Improving the segmentation accuracy of network models often increases the training time and computational resources of the network. In human–robot hybrid intelligent warehouses, AGVs need to make decisions in the shortest time while minimizing the demand for computational resources. Therefore, while considering segmentation accuracy, the scale and segmentation speed of the model also need to be further improved. This article improves Xception; the improved Xception uses depthwise separable convolutions instead of max pooling layers, allowing the network to extract feature information at any resolution and increase the receptive field. As the neural network deepens, the number of features that the network can extract also increases. However, this also leads to problems such as gradient explosion and vanishing, making it difficult to train the network. Therefore, the improved Xception network adds batch normalization and H-Swish residual structures after each 3 × 3 depth separable convolution. Batch normalization first processes the same batch of data then applies scaling coefficients and offsets to correct it. This stabilizes the data from the preceding layers and reduces the influence of the parameters from the earlier layers on those of the subsequent layers. The input data of the backbone network are sequentially passed through the entry flow layer, middle flow layer, and exit layer to obtain the final output; the network structure is shown in Figure 3.
The entry flow consists of two layers of regular convolutions and three block structures. Each block has three depthwise separable convolution layers and residual connection channels are added. The residual structure is shown in Figure 4. Each depthwise separable convolution is composed of depthwise convolution and pointwise convolution. The channel separation calculation is first performed using 3 × 3 convolution, followed by feature merging using 1 × 1 convolution. After each convolution, batch normalization is performed, and the network output is adjusted using the mean and standard deviation on small batches to accelerate the convergence speed of the network. The middle flow has a total of 16 units, and the structure of each unit is the same as the block structure in the entry flow, with the difference being the number of convolution kernels. There are two blocks in the exit flow, and the structure of block 2-1 is consistent with the previous block 1. There is no residual connection in block 2-2.

3.4. Dual-Branch Network Structure

The classic encoding decoding architecture network combines different scales of receptive field semantic information to improve the classification ability of the model, but such networks rely on downsampling to obtain advanced semantic information, ignoring spatial detail information. In order to segment the pallet stringer board from similar textures, the encoding region of this article adopts a dual-branch network structure, which processes spatial detail information and high-level semantic information respectively, achieving high-precision and efficient real-time segmentation of pallet components. The position attention module is shown in Figure 5. The position attention module can simulate rich contextual relationships between global features, thereby enhancing similar features at different positions and improving the semantic segmentation ability [29].
In the position attention module, local feature A from Figure 5 is first input into the convolutional layer to obtain feature matrices B, C, and D, B , C , D R C × H × W . Then, we perform a series of operations such as reshaping, transposing, and multiplying matrices B and C to obtain T, and finally use the softmax layer to calculate the position attention S R N × N :
S j i = e x p B i · C j i = 1 N e x p B i · C j
where N = H × W represents the number of pixels; B i represents the i-th position element in matrix B; C j   represents the j-th position element of matrix C; S i j represents the influence factor of the i-th position on the j-th position. Equation (1) indicates that the more similar the feature representations of two different positions are, the greater the correlation between them. At the same time, the generated feature matrix D R C × H × W is reshaped, and then multiplied with the position attention S matrix. The resulting result is reshaped again, multiplied with the scale parameter α , and summed with the feature map A to obtain the final output matrix F R C × H × W :
F j = α i = 1 N S j i D i + A j
where D i is the i-th element in matrix D; α is a learning parameter that gradually learns to allocate more weights from 0. Equation (2) shows that the result feature F for each position is the weighted sum of all position features and the original features. Therefore, the position attention module weights the features in each channel, allowing the model to focus on global contextual information and integrate broader contextual information into local features, promoting similar semantic features and improving the model’s semantic segmentation ability.
The channel attention module is shown in Figure 6. By utilizing the interrelationships between feature maps of different channels, it is possible to highlight the interrelated feature maps and promote specific semantic features. Therefore, it is necessary to explore the features of different channels [30].
In the channel attention module, the channel attention map X is directly calculated from the original feature matrix A. Firstly, the feature matrix is reshaped to obtain A R C × N , N = H × W . Then, A prime is multiplied by its transpose matrix, and X is obtained through the softmax layer [31]:
X i j = e x p A × A T j i = 1 C e x p A i × A T j
where A i represents the i-th position element value of the A matrix, A T j represents the j-th element value of the A i transpose matrix, and X i j represents the influence factor of the i-th channel on the j-th channel. Then, we transpose X and multiply it by the A matrix, reshape the result, multiply it with the training parameter β , and add it to the matrix A to obtain the final result F R C × H × W :
F j = β i = 1 C x i j A i + A j
where β is the training parameter, gradually learning weights from 0. From Equation (4), it can be concluded that the final feature of each channel is the weighted sum of all channel features and the original channel features. Therefore, the channel attention module can use the interdependence between channel graphs to enhance the feature representation of specific semantics.

3.5. Focal Loss Function

The characteristics of the pallet stringer board are similar to the other surfaces of the pallet, and there are a large number of other surfaces occupying pixels in the dataset. The detection accuracy is generally lower than that of medium and large target samples, and there are many problems such as missed judgments and false detections when detecting samples. To address the issue of class imbalance and improve model focus on small or difficult-to-detect regions during training, this paper adopts focal loss as the segmentation network’s loss function [32].
Focal loss dynamically adjusts the loss contribution of each sample by down-weighting the loss for well-classified examples, allowing the model to focus more on hard-to-detect regions, such as pallet slots that may be partially occluded or located at extreme angles. This characteristic makes it particularly suitable for our task, as precise slot localization directly impacts the AGV’s accuracy in pallet handling. To solve the problem of insufficient sample learning during model training, this paper uses focal loss as the loss function of the segmentation network.
p t = p ^ i f   y = 1 1 p ^ o t h e r w i s e
F L p t = α t 1 p t γ log p t
The p t in Equation (5) reflects the degree of closeness to label y, and the larger the p t value, the closer it is to category y; 1 p t γ is the modulation factor. The more accurate the classification, the closer p t approaches 1, and the closer the modulation factor approaches 0. The p t also reflects the difficulty of classification. The larger the p t , the higher the confidence level of classification, indicating that the sample is easier to classify. The smaller the p t , the lower the confidence level of classification, indicating that the sample is more difficult to classify. In Formula (6), α t is the weighting factor, which can suppress the imbalance in the number of positive and negative samples. By using γ , the imbalance in the number of simple and difficult-to-distinguish samples can be controlled.

4. Ablation Experiment and Analysis

4.1. Datasets and Evaluation Criteria

4.1.1. Datasets Preparation and Processing

The dataset used in this paper consists of images of pallets collected from warehouse scenarios. This dataset includes commonly used pallets in handling scenes, as shown in Figure 7, such as universal flat pallets for intermodal transport and steel flat pallets for railway freight. The dataset is categorized into single-type and multi-type pallet datasets.
We label the pallet dataset by dividing it into three categories: background, stringer board, and pallet slots. In this study, the pallet dataset was constructed using the PASCAL VOC dataset format and annotated based on JSON annotation files. In this format, background pixels are represented as 0, the pallet stringer board is denoted by 1, and the pallet slots area is denoted by 2. The original pallet dataset consists of 500 images, covering commonly used standard pallet types, such as universal flat pallets for intermodal transport and steel flat pallets for railway freight. Due to the small dataset size, which may lead to overfitting and could affect the model’s generalization ability, data augmentation techniques such as rotation, cropping, noise addition, and stitching were applied only to the training set, expanding it to 3000 images. The augmentation methods for the original images are shown in Figure 8. After preprocessing, we manually filtered out images without targets that were generated during the data augmentation process. As the original images vary in resolution, all images were resized to 512 × 512 pixels. These augmented images enable the model to better adapt to various real-world scenarios, thereby improving training effectiveness and ultimately enhancing performance.

4.1.2. Performance Evaluation

In order to verify the effectiveness and stability of the improved model, this study used the average intersection-to-union (mIoU) ratio and average pixel accuracy (mPA) as indicators to evaluate the network segmentation performance. The confusion matrix is used to evaluate the classification performance of the model. The four elements in the confusion matrix are as follows: true positive (TP), false positive (FP), true negative (TN), and false negative (FN).
Mean intersection union ratio (mIoU): The ratio between the intersection, the union of predicted results, and true values for each category is summed and then averaged to measure the segmentation accuracy of the model.
m I o U = 1 M i = 1 M T P T P + F N + F P
Mean pixel accuracy (mPA): The ratio of the number of correctly recognized pixels in each category to the total number of pixels in that category is summed and then averaged to measure the model’s pixel classification accuracy for each category.
m P A = T P + T N T P + T N + F P + F N
Among them, TP is the number of correctly predicted target image pixels by the network, FP is the number of incorrectly predicted target image pixels by the network, TN is the number of correctly predicted background pixels by the network, and FN is the number of incorrectly predicted background pixels by the network.

4.1.3. Implementation Details

The LCS-Net proposed in this paper is trained end-to-end. It accurately identifies the stringer board and pallet slots of pallets at the pixel level, laying a solid technical foundation for achieving better localization results subsequently. All designed experiments were conducted on a Windows 10 operating system. The server environment used in the experiment included an Intel (R) Core (TM) i9-13900HX CPU, an NVIDIA GeForce RTX 4060 GPU, with 8 GB of graphics memory, and was configured with Python 3.9 + Python 2.1.2 + Cuda12.1. The input images to the model were cropped to a size of 512 × 512 pixels. The training objective was to minimize the loss function.
During training, the batch size was set to 8, and the gradient descent algorithm solved the minimum value by descending along the gradient. Due to the large number of model parameters, using the gradient descent algorithm can lead to excessive computational load. The random gradient descent algorithm can reduce the computational load, but the convergence speed is slow and easily affected by noisy samples, leading to the model falling into local optima. Therefore, the Adam algorithm was used during the training process to adaptively adjust the learning rate according to the gradient size and converge to the optimal solution faster [33].
For model training, the dataset was split into 80% for training, 10% for validation, and 10% for testing. The input images were resized to 512 × 512 pixels. The training objective was to minimize the loss function. The model was trained for 100 epochs. In addition, for the Adam optimizer when applied, hyperparameters β1 and β2 were set to 0.9 and 0.999. To optimize the model’s performance, the Adam optimizer adaptively adjusted the learning rate and ensured faster convergence. Early stopping was also employed during training to prevent overfitting, with training halted if the validation loss did not improve after a certain number of consecutive epochs. For the loss function, focal loss was employed to address the issue of class imbalance. α t is a weighting factor set to 1 to balance the class imbalance. γ is the focusing parameter, set to 1 to focus more on harder-to-classify samples. This setup helps reduce the effect of easily classified examples, thereby emphasizing the harder ones during training, especially for class imbalance problems.

4.2. Ablation Studies

4.2.1. Backbone Network Ablation Experiment

To verify the impact of the data augmentation methods employed in this study on the segmentation model’s generalization ability, the segmentation network proposed in this paper was trained on both the original data and the data augmented with the methods described. The training results are shown in Table 1.
The experimental results show that by enhancing and expanding the original dataset, all indicators on the test dataset improved. The intersection ratio of the stringer board increased by 12%, the intersection ratio of the pallet slots increased by 16%, and the average intersection ratio increased by 9.12%.
To verify the effectiveness of the improved backbone network, MobileNetv2 [34], Xception [35], and the proposed backbone network were used as the backbone networks of the LCS-Net proposed in this paper. Ablation experiments were conducted to verify the performance of the model. The evaluation indicators include the average intersection-to-union (mIoU) ratio and average frame rate (FPS). To ensure the reliability of the experimental results, each network was trained with 100 epochs under the same dataset and experimental environment. The training results are shown in Table 2.
From Table 2, it can be seen that the improved Xception backbone network proposed in this paper has an average intersection-to-union ratio that is 3.09% higher than the original Xception in the DeepLabv3+ network compared to the mIoU. When using the improved Xception proposed in this article as the backbone network, the processing speed is also faster compared to MobileNetv2. Therefore, considering both the processing speed and accuracy of the model, the backbone network proposed in this paper is the optimal choice.

4.2.2. Evaluating the Impact of Attention Mechanisms

To further validate the choice of the position attention and channel attention modules in LCS-Net, we added a comparative experiment by replacing these modules. The spatial attention module and channel attention module effectively capture the feature dependencies in the spatial and channel dimensions, which is crucial for accurately identifying the pallet slot areas. The spatial attention module enhances the network’s focus on spatial relationships, aiding in more precise pallet slot localization, while the channel attention module highlights key feature maps, improving the network’s ability to distinguish between pallets and background noise. To further validate the role of the spatial attention module and channel attention module in the pallet segmentation task, an ablation study is designed as follows:
In Table 3, Backbone specifies the backbone network used for each model. All models in the experiments utilize the Xception backbone to ensure consistency in the architecture. Baseline indicates the model without any attention mechanism, Channel-Only indicates the model using only the channel attention module, Position-Only indicates the model using only the position attention module, and CBAM means that the variant applies the CBAM, which combines both channel and spatial attention mechanisms. ‘Position and Channel’ refers to the complete model that uses both the channel and position attention mechanisms.
As shown in Table 3, the ablation study evaluates the impact of different attention mechanisms on segmentation performance. The results indicate that attention mechanisms effectively improve segmentation accuracy, although different approaches lead to varying trade-offs in computational efficiency. Furthermore, the experimental results demonstrate that the proposed dual-attention mechanism, which combines position and channel attention, achieves the best performance, providing optimal segmentation accuracy while maintaining high inference speed.
The ablation experiment results show that the proposed attention mechanism combination in this paper has significant advantages in both segmentation accuracy and processing speed, meeting the real-time segmentation requirements of AGVs.

4.2.3. Single-Type Pallet Comparative Experiment

In order to verify the advantages of the LCS-Net proposed in this article, compared to existing classical segmentation models, the classical segmentation network U-Net, PSPNet, DeepLabv3+, Fast-SCNN, and SegFormer. U-Net employs an encoder–decoder architecture, which effectively captures local features and performs excellently in medical image segmentation tasks, especially for small-sample datasets. PSPNet introduces a pyramid pooling module, which captures contextual information at different scales, making it highly suitable for segmenting images with multi-scale objects and enhancing global semantic information. DeepLabv3+ is based on deep dilated convolutions, effectively expanding the receptive field, and incorporates an encoder–decoder structure, making it ideal for high-precision segmentation tasks in complex scenes. Fast-SCNN is a lightweight network, specifically designed for real-time applications. It delivers good segmentation results while ensuring high inference speed, making it suitable for deployment in resource-constrained environments. SegFormer is based on a Transformer architecture, capable of handling multi-scale features and contextual information, and exhibits strong segmentation capabilities, particularly in complex backgrounds.
According to Table 4, the proposed model achieves the highest mIoU of 88.16%, surpassing all other models. The results indicate that the proposed model not only improves segmentation accuracy but also ensures high computational efficiency, making it well-suited for real-time applications. The segmentation effect of each network is shown in Figure 9.
From Figure 9, it can be seen that all six segmentation networks can effectively extract the pallet stringer board and pallet slots, but there are differences in the morphology of the extracted areas. The LCS-Net proposed in this article can accurately segment the pallet stringer board and the edge of the socket, effectively reducing edge burrs and false detections. At the same time, it significantly shortens the single image inference time. Compared with classical segmentation networks, it can better balance segmentation accuracy and speed.
To validate and assess the performance of the segmentation network, a validation set is used during training to calculate the segmentation network loss. This ensures that the detector maintains overall performance throughout training, preventing overfitting and preserving generalization capability. The training results are shown in Figure 10, indicating that the segmentation model proposed in this paper can detect most pallet objects with high precision.

4.2.4. Multi-Type Pallet Comparative Experiment

In human–robot hybrid intelligent warehouses, various types of pallets frequently appear, with differing pallet slots sizes and center distances. Therefore, the network must have a certain degree of generalization ability to accurately locate the centers of the pallet slots. This study aims to verify the superiority of the segmentation model proposed herein over existing classical segmentation models in detecting multiple types of pallets. To achieve this, we conducted comparative experiments using the classic segmentation network U-Net, PSPNet, DeepLabv3+, Fast-SCNN, SegFormer, and the improved model proposed in this paper. After 100 epochs of training, we obtained evaluation metrics for the models on various datasets containing different types of pallets, as shown in Table 5.
From Table 5, the results demonstrate that our model performs competitively in terms of segmentation accuracy, with an mIoU of 78.02%. It achieves excellent performance in pallet slot segmentation, comparable to the highest-performing models. While SegFormer leads in segmentation accuracy, our model offers a better balance between segmentation performance and inference speed. Fast-SCNN is the fastest model, but its segmentation accuracy is lower than that of others. Compared to other classical networks, our model achieves a balance between segmentation accuracy and speed. The segmentation results of our network on different types of pallets are illustrated in Figure 11.
From Figure 11, it can be observed that the LCS-Net accurately segments the pillar surfaces and plug-hole edges of various types of pallets. This model effectively reduces edge artifacts and false detection occurrences. Moreover, it significantly reduces the inference time per image compared to classic segmentation networks. Thus, it achieves a good balance between segmentation accuracy and speed.
To validate and assess the performance of the segmentation network, we use a validation set to calculate the segmentation network loss. This ensures that the detector maintains overall performance throughout training, preventing overfitting and preserving generalization capability. The training results are shown in Figure 12, indicating that the segmentation model proposed in this paper can detect most pallet objects with high precision.

4.3. Location of Pallet Based on LCS-Net

4.3.1. Positioning the Center Point of the Pallet Stringer Board

To achieve the positioning of the center point of the pallet stringer board, it is necessary to first calibrate the extrinsic parameters of the camera in the warehouse working environment. This will determine the transformation relationship between the pixel coordinate system and the pallet stringer board coordinate system, generating a conversion formula between the two coordinate systems. By processing the image, the minimum enclosing rectangle of the segmented pallet stringer board can be obtained, from which the pixel coordinates of the center of the stringer board are derived. Using the coordinate transformation formula, the three-dimensional coordinates of the center point of the pallet stringer board can be obtained.
To obtain the transformation parameters between the pixel coordinate system and the pallet coordinate system, this paper uses Zhang’s calibration method to calibrate the camera. By using the Calibration Toolbox, the camera’s intrinsic parameters, including the effective focal length and lens distortion coefficients, are obtained. The distortion coefficients are used to correct image distortion caused by lens distortion. The camera’s extrinsic matrix is calibrated using a chessboard calibration plate, resulting in the camera’s extrinsic matrix [RPalletToCamera|TPalletToCamera]. Based on the pinhole imaging model and the obtained camera extrinsic and intrinsic parameters, the transformation relationship between the pixel coordinate system and the pallet coordinate system is established. The position of the center of the pallet stringer board in the pallet coordinate system can be expressed as X , Y , Z . The transformation relationship between the position of a point on the pallet stringer board in the pallet coordinate system and its projected point coordinates is given by the following:
X Y Z 1 = Z C R P a l l e t T o C a m e r a   T P a l l e t T o C a m e r a 0 1 1 f d x 0 u 0 0 0 f d y v 0 0 0 0 1 0 u v 1
where f is the effective focal length, d x and d y represent the physical length of a pixel on the camera sensor in the x and y directions, respectively, and u 0 and v 0 represent the coordinates of the center of the camera sensor in the pixel coordinate system.
In actual warehouse scenarios, it is necessary to accurately locate the centroid of the pallet stringer board to ensure accuracy and stability during handling. The center of the pallet stringer board can be determined by finding the center of the minimum enclosing rectangle. By using the segmented mask of the pallet stringer board, the rotating caliper method is applied to iteratively compute the minimum enclosing rectangle of the contour point set for each target instance mask. First, the convex hull algorithm is used to obtain the convex hull E of the stringer board contour. Based on the convex hull, the minimum area enclosing the rectangle is found. By comparing the areas of the enclosing rectangles, the one with the smallest area is selected as the minimum enclosing rectangle of the stringer board contour. During the enumeration process, the top-left vertex (x1, y1), length l, and width w of the enclosing rectangle are recorded. The pixel center coordinates O are defined as x 1 + l 2 , y 1 + w 2 . After obtaining the pixel center of the pallet stringer board, the coordinate transformation formula is used to convert the pixel coordinates of the stringer board into the three-dimensional coordinate information (X, Y, Z) of the center point of the pallet stringer board.

4.3.2. Evaluation Experiment of Pallet-Positioning Accuracy Under Different Distance

The equipment used in this experiment is shown in Figure 13. The experiment varied the distance and angle between the camera and the pallet. The distance was adjusted between 1 m and 3 m, with specific distances set at 1 m, 1.25 m, 1.5 m, 1.75 m, 2 m, 2.25 m, 2.5 m, 2.75 m, and 3 m, to test the pallet-positioning accuracy at different distances.

4.3.3. Evaluation Experiment of Pallet-Positioning Accuracy Under Different Offset Angles

Considering that the pallet’s position may not be fixed in human–robot hybrid scenarios, different pallet offset angles were selected as experimental variables. The pallet was placed 1.5 m away from the camera, and the pallet’s offset angle was adjusted in 10° increments within a range from −30° to 30°. The experimental setup is illustrated in Figure 14.

4.3.4. Evaluation Experiment Result Analysis

The distance between the camera and the pallet is the variable in this experiment and is denoted by z . The variable x represents the horizontal relative displacement based on the pallet, and y represents the vertical relative distance between the camera and the pallet. The absolute errors at different distances, the mean absolute error (MAE), and the standard deviations (STDs) are shown in Table 6.
The results indicate that at various experimental distances, the maximum absolute errors for horizontal displacement, vertical displacement, distance, and angle are 9.19 mm 7.59 mm, 12.15 mm, and 1.68°, respectively. The minimum absolute errors are 1.45 mm, 1.42 mm, 0.95 mm, and 0.43°.
The angle between the camera and the pallet is the variable in this experiment and is denoted by β. The variable x represents the horizontal relative displacement based on the pallet, and y represents the vertical relative distance between the camera and the pallet. The results are shown in Table 7.
The results indicate that at various experimental distances, the maximum absolute errors for horizontal displacement, vertical displacement, distance, and angle are 3.78 mm, 5.78 mm, 6.54 mm, and 2.08°, respectively. The minimum absolute errors are 0.93 mm, 0.52 mm, 1.76 mm, and 0.79°. The experimental results indicate that the proposed pallet-positioning algorithm can recognize and locate pallets within a 30-degree deviation angle, with an average absolute error of within 1 cm for distance and 3° for angle.

4.4. Validation of Generalization Ability Across Different Environments

4.4.1. Illumination and Noise Joint Interference Experiment

Building upon the original distance and angle variation experiments, this section further investigates the impact of illumination changes and background noise on the pallet localization accuracy of the LCS-Net model to comprehensively evaluate its robustness. The experimental variable includes different illumination intensities and background complexities. Normal illumination simulates a standard indoor warehouse environment, low illumination simulates nighttime or insufficient lighting conditions, and high illumination simulates direct sunlight or light reflections. Three noise levels (simple, moderately complex, and complex) are set by adjusting noise parameters to simulate three warehouse scenarios: simple background, moderately complex background, and complex background. To obtain more realistic robustness data, the interaction effects of illumination and background noise are comprehensively considered, generating nine combination groups. Fix the pallet at a distance of 1 m and an angle of 0°, gradually change the lighting intensity and background complexity, capture images, and use LCS-Net to segment the pallet position, as shown in Figure 15.

4.4.2. Different Environments Result Analysis

The angle between the camera and the pallet is the variable in this experiment, denoted by β. The variable x represents the horizontal relative displacement based on the pallet, and y represents the vertical relative distance between the camera and the pallet. The results are shown in Table 8.
The table shows the absolute errors in the x, y, and z directions and the angle β under different lighting conditions and background complexities. Under normal illumination and simple background and low illumination and complex background, the maximum error is far below 10mm. Under normal illumination, the maximum error is 1.20 mm, and under low lighting and a complex background, the maximum error is 1.70 mm. The maximum angle error is 1.10°.
Even under challenging conditions such as low illumination and complex backgrounds, the maximum errors remain within the acceptable range for AGV operations, ensuring reliable positioning performance. LCS-Net’s robustness and stability guarantee consistent performance in practical applications, making it suitable for AGV pallet positioning in various environments.

5. Conclusions

We present a component segmentation network, LCS-Net, to address the precise positioning of multiple types of pallets, which uses a dual-branch attention mechanism to obtain spatial detail information and optimize the weights of deep features to solve the problem of inaccurate edge segmentation of pallets. The position attention module selectively aggregates the features of each position by weighting and summing up the features of all positions; The channel attention module obtains the interdependence between different feature channels, strengthens the learning of important channel features, and improves the semantic segmentation effect. Adding the outputs of two attention modules can further improve feature representation and enhance segmentation performance.
Test results indicate that our model outperforms existing segmentation methods in terms of average accuracy, processing speed, and generalization capability. The LCS-Net we proposed achieves an mIoU of 88.16% on the single-class pallet dataset, representing a 4.89% improvement over the classic segmentation networks DeepLabv3+. On the multi-class pallet dataset, the mIoU is improved by 10.11% compared to the classic segmentation network DeepLabv3+, and the average pixel accuracy is increased by 10.41%. We conducted a visual localization experiment for pallet picking, where AGV was guided by a camera to perform picking tasks for pallets placed in unknown locations. These localization experiments were conducted at various distances and angles. The results show that the LCS-Net-based visual localization method can effectively recognize and locate multi-category pallet targets in the warehousing field. Moreover, in experiments considering interference factors such as illumination changes and background noise, LCS-Net still maintains high accuracy, further proving the model’s stability and robustness in complex environments. Compared to using specialized data sources like depth data for pallet localization, our proposed pallet localization algorithm meets the accuracy and speed requirements necessary for practical production tasks.
Although this study focuses on pallet positioning in intelligent warehousing scenarios, the system can be easily transplanted to other scenarios by changing the dataset and setting corresponding loss functions. It can achieve precise detection and positioning of randomly placed objects in human–robot hybrid intelligent warehouse scenarios using component segmentation, such as robots, empowering new industries and providing technical support for highly intelligent related fields.

Author Contributions

B.W.: writing—review and editing, funding acquisition, project administration, resources. S.W.: writing—original draft, investigation, methodology. Y.L.: writing—review and editing, methodology, resources. Y.Y.: writing—review and editing, conceptualization, methodology. D.J.: resources, supervision. M.Q.: resources, supervision. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Jiangsu Province Modern Agricultural Machinery Equipment and Technology Demonstration and Promotion program (NJ2023-16).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data will be made available upon request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kim, B.-i.; Graves, R.J.; Heragu, S.S.; Onge, A.S. Intelligent agent modeling of an industrial warehousing problem. IIE Trans. 2002, 34, 601–612. [Google Scholar] [CrossRef]
  2. Zhao, Z.; Cheng, J.; Liang, J.; Liu, S.; Zhou, M.; Al-Turki, Y. Order picking optimization in smart warehouses with human-robot collaboration. IEEE Internet Things J. 2024, 11, 16314–16324. [Google Scholar] [CrossRef]
  3. Kubasakova, I.; Kubanova, J.; Benco, D.; Kadlecová, D. Implementation of Automated Guided Vehicles for the Automation of Selected Processes and Elimination of Collisions between Handling Equipment and Humans in the Warehouse. Sensors 2024, 24, 1029. [Google Scholar] [CrossRef] [PubMed]
  4. Solman, K.N. Analysis of interaction quality in human–machine systems: Applications for forklifts. Appl. Ergon. 2002, 33, 155–166. [Google Scholar] [CrossRef] [PubMed]
  5. Baglivo, L.; Biasi, N.; Biral, F.; Bellomo, N.; Bertolazzi, E.; Da Lio, M.; De Cecco, M. Autonomous pallet localization and picking for industrial forklifts: A robust range and look method. Meas. Sci. Technol. 2011, 22, 085502. [Google Scholar] [CrossRef]
  6. Seelinger, M.; Yoder, J.-D. Automatic visual guidance of a forklift engaging a pallet. Robot. Auton. Syst. 2006, 54, 1026–1038. [Google Scholar] [CrossRef]
  7. Lin, R.; Huang, H.; Li, M. An automated guided logistics robot for pallet transportation. Assem. Autom. 2021, 41, 45–54. [Google Scholar] [CrossRef]
  8. Mohamed, I.S.; Capitanelli, A.; Mastrogiovanni, F.; Rovetta, S.; Zaccaria, R. Detection, localisation and tracking of pallets using machine learning techniques and 2D range data. Neural Comput. Appl. 2020, 32, 8811–8828. [Google Scholar]
  9. Leng, J.; Wang, D.; Shen, W.; Li, X.; Liu, Q.; Chen, X. Digital twins-based smart manufacturing system design in Industry 4.0: A review. J. Manuf. Syst. 2021, 60, 119–137. [Google Scholar]
  10. Vu, V.-D.; Hoang, D.-D.; Tan, P.X.; Nguyen, V.-T.; Nguyen, T.-U.; Hoang, N.-A.; Phan, K.-T.; Tran, D.-T.; Vu, D.-Q.; Ngo, P.-Q. Occlusion-robust pallet pose estimation for warehouse automation. IEEE Access 2024, 12, 1927–1942. [Google Scholar] [CrossRef]
  11. Cai, W.; Song, Y.; Duan, H.; Xia, Z.; Wei, Z. Multi-feature fusion-guided multiscale bidirectional attention networks for logistics pallet segmentation. Comput. Model. Eng. Sci. 2022, 131, 1539–1555. [Google Scholar] [CrossRef]
  12. Chen, G.; Peng, R.; Wang, Z.; Zhao, W. Pallet Recognition and Localization Method for Vision Guided Forklift. In Proceedings of the 2012 8th International Conference on Wireless Communications, Networking and Mobile Computing, Shanghai, China, 21–23 September 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 1–4. [Google Scholar]
  13. Garibott, G.; Masciangelo, S.; Ilic, M.; Bassino, P. Robolift: A vision guided autonomous fork-lift for pallet handling. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS’96, Osaka, Japan, 8 November 1996; IEEE: Piscataway, NJ, USA, 1996; pp. 656–663. [Google Scholar]
  14. Carata, S.V.; Ghenescu, M.; Mihaescu, R. Real-Time Detection of Unrecognized Objects in Logistics Warehouses Using Semantic Segmentation. Mathematics 2023, 11, 2445. [Google Scholar] [CrossRef]
  15. Fang, D.; Han, J.; Shao, J.; Zhao, C. Pallet Detection and Localization Based on RGB Image and Point Cloud Data for Automated Forklift. In Proceedings of the 2024 7th International Conference on Advanced Algorithms and Control Engineering (ICAACE), Shanghai, China, 1–3 March 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 642–647. [Google Scholar]
  16. Bao, J.; Guo, D.; Li, J.; Zhang, J. The modelling and operations for the digital twin in the context of manufacturing. Enterp. Inf. Syst. 2019, 13, 534–556. [Google Scholar]
  17. Minaee, S.; Boykov, Y.; Porikli, F.; Plaza, A.; Kehtarnavaz, N.; Terzopoulos, D. Image segmentation using deep learning: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3523–3542. [Google Scholar]
  18. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2015), Boston, MA, USA, 7–12 June 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 3431–3440. [Google Scholar]
  19. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, Part III 18, 2015. Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  20. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [PubMed]
  21. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2016), Las Vegas, NV, USA, 27–30 June 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 770–778. [Google Scholar]
  22. Chen, L.-C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 801–818. [Google Scholar]
  23. Lu, C.; Xia, M.; Qian, M.; Chen, B. Dual-branch network for cloud and cloud shadow segmentation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5410012. [Google Scholar] [CrossRef]
  24. Yang, Z. Semantic segmentation method based on improved DeeplabV3+. In Proceedings of the International Conference on Cloud Computing, Performance Computing, and Deep Learning (CCPCDL 2023), Huzhou, China, 17–19 February 2023; SPIE: Bellingham, WA, USA, 2023; pp. 32–37. [Google Scholar]
  25. Meng, J.; Zhang, L.; Cao, Y.; Zhang, L.; Song, Q. Research on optimization of image semantic segmentation algorithms based on Deeplab v3+. Laser Optoelectron. Prog. 2022, 59, 161–170. [Google Scholar]
  26. Hu, T.; Yang, P.; Zhang, C.; Yu, G.; Mu, Y.; Snoek, C.G. Attention-based multi-context guiding for few-shot semantic segmentation. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), Honolulu, HI, USA, 27 January–1 February 2019; AAAI: Palo Alto, CA, USA, 2019; pp. 8441–8448. [Google Scholar]
  27. Ding, H.; Zhang, H.; Jiang, X. Self-regularized prototypical network for few-shot semantic segmentation. Pattern Recognit. 2023, 133, 109018. [Google Scholar] [CrossRef]
  28. Antonelli, M.; Reinke, A.; Bakas, S.; Farahani, K.; Kopp-Schneider, A.; Landman, B.A.; Litjens, G.; Menze, B.; Ronneberger, O.; Summers, R.M. The medical segmentation decathlon. Nat. Commun. 2022, 13, 4128. [Google Scholar]
  29. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual attention network for scene segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2019), Long Beach, CA, USA, 16–20 June 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 3146–3154. [Google Scholar]
  30. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient channel attention for deep convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2020), Seattle, WA, USA, 14–20 June 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 11534–11542. [Google Scholar]
  31. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Seattle, WA, USA, 14–19 June 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 3–19. [Google Scholar]
  32. Li, X.; Wang, W.; Hu, X.; Li, J.; Tang, J.; Yang, J. Generalized focal loss v2: Learning reliable localization quality estimation for dense object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2021), Nashville, TN, USA, 19–25 June 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 11632–11641. [Google Scholar]
  33. Luo, L.; Xiong, Y.; Liu, Y.; Sun, X. Adaptive gradient methods with dynamic bound of learning rate. arXiv 2019, arXiv:1902.09843. [Google Scholar]
  34. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2018), Salt Lake City, UT, USA, 8–22 June 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 4510–4520. [Google Scholar]
  35. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2017), Honolulu, HI, USA, 21–26 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1251–1258. [Google Scholar]
Figure 1. AGV intelligent pallet forking system framework.
Figure 1. AGV intelligent pallet forking system framework.
Sensors 25 02333 g001
Figure 2. Network structure diagram.
Figure 2. Network structure diagram.
Sensors 25 02333 g002
Figure 3. Schematic diagram of improved Xception network structure.
Figure 3. Schematic diagram of improved Xception network structure.
Sensors 25 02333 g003
Figure 4. Residual structure with depthwise separable convolution.
Figure 4. Residual structure with depthwise separable convolution.
Sensors 25 02333 g004
Figure 5. Position attention module.
Figure 5. Position attention module.
Sensors 25 02333 g005
Figure 6. Channel attention module.
Figure 6. Channel attention module.
Sensors 25 02333 g006
Figure 7. Partial of the pallet dataset.
Figure 7. Partial of the pallet dataset.
Sensors 25 02333 g007
Figure 8. Data enhancement methods: (a) stitching, (b) rotation, (c) noise addition, and (d) cropping.
Figure 8. Data enhancement methods: (a) stitching, (b) rotation, (c) noise addition, and (d) cropping.
Sensors 25 02333 g008
Figure 9. The segmentation effect of different models on pallets.
Figure 9. The segmentation effect of different models on pallets.
Sensors 25 02333 g009
Figure 10. Loss curve and accuracy over iterations measured by the mean intersection over union under single-type pallet.
Figure 10. Loss curve and accuracy over iterations measured by the mean intersection over union under single-type pallet.
Sensors 25 02333 g010
Figure 11. Segmentation results of different models on different pallet types.
Figure 11. Segmentation results of different models on different pallet types.
Sensors 25 02333 g011
Figure 12. Loss curve and accuracy over iterations measured by mean intersection over union under multi-type pallet.
Figure 12. Loss curve and accuracy over iterations measured by mean intersection over union under multi-type pallet.
Sensors 25 02333 g012
Figure 13. Experiments on pallet positioning at different distances.
Figure 13. Experiments on pallet positioning at different distances.
Sensors 25 02333 g013
Figure 14. Experiments on pallet positioning at different angles.
Figure 14. Experiments on pallet positioning at different angles.
Sensors 25 02333 g014
Figure 15. Hybrid experiment at a distance of 1 m.
Figure 15. Hybrid experiment at a distance of 1 m.
Sensors 25 02333 g015
Table 1. Comparison between raw data and expanded data.
Table 1. Comparison between raw data and expanded data.
Data TypeStringer Board IoU/%Pallet Slots IoU/%mIoU/%
Raw data645974.15
Enhanced data767583.27
Table 2. Experimental results of different backbone networks.
Table 2. Experimental results of different backbone networks.
BackboneModelmIoU/%FPS
MobileNetv2DeepLabv3+84.3636.88
XceptionDeepLabv3+83.2732.73
OursDeepLabv3+86.3643.23
Table 3. Performance comparison of LCS-Net with different attention mechanisms.
Table 3. Performance comparison of LCS-Net with different attention mechanisms.
Backbone Attention MechanismmIoU/%FPS
Xception Baseline69.0945.71
Xception Channel-Only75.9543.89
Xception Position-Only73.4844.25
Xception CBAM83.9937.83
Xception Position and Channel87.3743.57
Table 4. Comparison of network evaluation indicators under single-type pallet.
Table 4. Comparison of network evaluation indicators under single-type pallet.
MethodmIoU/%Stringer Board IoU/%Pallet Slots IoU/%mPA/%FPS
U-Net77.87666996.4934.89
PSPNet72.09646787.5945.88
DeepLabv3+83.27767590.1232.73
Fast-SCNN66.45666779.1747.55
SegFormer82.88828390.4828.06
Ours88.16828393.8249.39
Table 5. Comparison of network evaluation indicators under multi-type pallet.
Table 5. Comparison of network evaluation indicators under multi-type pallet.
MethodmIoU/%Stringer Board IoU/%Pallet Slots IoU/%mPA/%FPS
U-Net65.28455973.4834.27
PSPNet70.01555876.7941.04
Deeplabv3+67.91466074.2329.96
Fast-SCNN66.45666779.1743.43
SegFormer78.44727383.7926.84
Ours78.02637384.6439.79
Table 6. The absolute error, mean absolute error, and standard deviation of the camera and pallet at different distances.
Table 6. The absolute error, mean absolute error, and standard deviation of the camera and pallet at different distances.
z/mmAbsolute Visual Positioning Error
x/mmy/mmz/mmβ/°
10002.761.420.950.57
12501.452.652.750.43
15001.874.114.810.97
17502.663.676.290.96
20003.984.743.791.15
22505.865.485.241.68
25004.867.599.431.18
27506.546.418.311.09
30009.197.5112.150.51
MAE4.355.025.750.86
STD2.381.642.900.37
Table 7. The absolute error, mean absolute error, and standard deviation at different pallet offset angles.
Table 7. The absolute error, mean absolute error, and standard deviation at different pallet offset angles.
β/°Absolute Visual Positioning Error
x/mmy/mmz/mmβ/°
−300.930.521.761.67
−201.221.233.431.54
−101.461.213.961.09
01.084.893.220.79
102.783.764.981.98
201.885.274.172.08
303.785.786.541.81
MAE1.733.883.861.57
STD0.981.971.170.41
Table 8. The absolute error, mean absolute error, and standard deviation in different environments.
Table 8. The absolute error, mean absolute error, and standard deviation in different environments.
IlluminationBackgroundAbsolute Visual Positioning Error
x/mmy/mmz/mmβ/°
NormalSimple2.761.420.950.57
NormalModerately complex2.982.531.040.65
NormalComplex3.452.671.200.72
LowSimple4.203.171.350.85
LowModerately complex4.513.521.670.96
LowComplex5.104.271.701.10
HighSimple3.101.511.100.65
HighModerately complex3.782.731.270.71
HighComplex3.973.101.360.75
MAE3.762.771.290.77
STD0.730.850.250.16
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, B.; Wang, S.; Lu, Y.; Yi, Y.; Jiang, D.; Qiao, M. A New Pallet-Positioning Method Based on a Lightweight Component Segmentation Network for AGV Toward Intelligent Warehousing. Sensors 2025, 25, 2333. https://doi.org/10.3390/s25072333

AMA Style

Wu B, Wang S, Lu Y, Yi Y, Jiang D, Qiao M. A New Pallet-Positioning Method Based on a Lightweight Component Segmentation Network for AGV Toward Intelligent Warehousing. Sensors. 2025; 25(7):2333. https://doi.org/10.3390/s25072333

Chicago/Turabian Style

Wu, Bin, Shijie Wang, Yi Lu, Yang Yi, Di Jiang, and Mengmeng Qiao. 2025. "A New Pallet-Positioning Method Based on a Lightweight Component Segmentation Network for AGV Toward Intelligent Warehousing" Sensors 25, no. 7: 2333. https://doi.org/10.3390/s25072333

APA Style

Wu, B., Wang, S., Lu, Y., Yi, Y., Jiang, D., & Qiao, M. (2025). A New Pallet-Positioning Method Based on a Lightweight Component Segmentation Network for AGV Toward Intelligent Warehousing. Sensors, 25(7), 2333. https://doi.org/10.3390/s25072333

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop