Next Article in Journal
Variable Splitting and Fusing for Image Phase Retrieval
Next Article in Special Issue
Clinician and Visitor Activity Patterns in an Intensive Care Unit Room: A Study to Examine How Ambient Monitoring Can Inform the Measurement of Delirium Severity and Escalation of Care
Previous Article in Journal
Addressing Once More the (Im)possibility of Color Reconstruction in Underwater Images
Previous Article in Special Issue
Deep Learning for Generating Time-of-Flight Camera Artifacts
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhanced Self-Checkout System for Retail Based on Improved YOLOv10

1
Department of Information System, Arizona State University, Tempe, AZ 85281, USA
2
Department of Computer Science, University of North Carolina, Chapel Hill, NC 27599, USA
3
Department of Industrial and Systems Engineering, University of Minnesota, Minneapolis, MN 55455, USA
4
Ira A. Fulton Schools of Engineering, Arizona State University, Tempe, AZ 85281, USA
5
Henry Samueli School of Engineering, University of California, Irvine, CA 92617, USA
6
Department of Psychological Sciences, Case Western Reserve University, Cleveland, OH 44106, USA
*
Author to whom correspondence should be addressed.
J. Imaging 2024, 10(10), 248; https://doi.org/10.3390/jimaging10100248
Submission received: 16 August 2024 / Revised: 7 October 2024 / Accepted: 8 October 2024 / Published: 10 October 2024

Abstract

:
With the rapid advancement of deep learning technologies, computer vision has shown immense potential in retail automation. This paper presents a novel self-checkout system for retail based on an improved YOLOv10 network, aimed at enhancing checkout efficiency and reducing labor costs. We propose targeted optimizations for the YOLOv10 model, incorporating the detection head structure from YOLOv8, which significantly improves product recognition accuracy. Additionally, we develop a post-processing algorithm tailored for self-checkout scenarios, to further enhance the application of the system. Experimental results demonstrate that our system outperforms existing methods in both product recognition accuracy and checkout speed. This research not only provides a new technical solution for retail automation but offers valuable insights into optimizing deep learning models for real-world applications.

1. Introduction

1.1. Research Background

Efficient retail operations are critical to ensuring customer satisfaction and maintaining competitive advantage in the fast-paced market today. Self-checkout systems, in particular, have become an essential component of modern retail environments, offering customers a quicker, more convenient way to complete their purchases [1]. These systems help reduce the burden on staff, minimize wait times, and enhance the overall shopping experience. As consumer expectations for speed and convenience rise, the demand for reliable and accurate self-checkout solutions continues to grow. However, the effectiveness of these systems depends on their ability to accurately identify and process a diverse range of products in real time, a challenge that has driven significant research and development in the field of computer vision and deep learning.

1.2. Challenges in Current Self-Checkout Systems

Despite the growing adoption of self-checkout systems in modern retail environments, several challenges persist that can hinder their effectiveness. One of the primary issues is the accurate recognition and differentiation of a wide variety of products, especially in environments with complex backgrounds, varying lighting conditions, or occluded items [2]. Traditional systems often struggle with identifying items that are similar in appearance or those that are irregularly shaped, leading to increased error rates and customer frustration. Additionally, maintaining real-time processing capabilities without compromising accuracy remains a significant hurdle, particularly as the volume of transactions and the diversity of products continue to expand [3]. Handling overlapping objects presents a significant challenge, as current self-checkout systems often struggle when products are partially occluded or stacked, which is a common occurrence when customers place multiple items on the checkout surface. Additionally, these systems face difficulties with sub-optimal projections, particularly when products are positioned irregularly or exposed to varying lighting conditions. Such scenarios can severely compromise recognition accuracy and hinder the overall efficiency of the checkout process [4]. These challenges underscore the need for ongoing research and development in computer vision and deep learning to create more robust and reliable self-checkout systems that can meet the demands of today’s retail industry.

1.3. Overview of Advancements of Deep Learning and Computer Vision in Retail

With the rapid advancement of technology, the wave of artificial intelligence (AI) is profoundly transforming a wide range of industries, including retail [5,6]. Both 2D and 3D vision models are widely used in different scenarios [7,8], and real-time tracking detection can be used on both manned and unmanned objects [9,10]. In the realm of AI, machine learning (ML) has become a powerful tool for tackling complex tasks and enhancing efficiency. Deep learning (DL) and computer vision (CV), two significant subsets of ML, have gained attention in recent years for their precise handling and analysis of visual data even in difficult and remote scenarios [11,12].
The development of powerful convolutional neural networks (CNNs) and real-time object detection algorithms, such as the You Only Look Once (YOLO) series, has enabled machines to process visual data with unprecedented speed and accuracy [13,14]. These technologies have greatly improved the ability to recognize and classify retail products in various environments, even under challenging conditions such as occlusions, varying lighting, and complex backgrounds.

1.4. Purpose of the New System and Advantages

The purpose of the current study is to develop an enhanced self-checkout system that effectively addresses the limitations of existing technologies by leveraging the latest advancements in deep learning and computer vision. Our proposed system, MidState-YOLO-ED, integrates key innovations from YOLOv10 with enhancements inspired by YOLOv8, aiming to significantly improve product recognition accuracy and processing efficiency in real-time retail environments [15,16]. The advantages of this system include its ability to accurately identify a diverse range of retail products, maintain robust performance under challenging conditions, and operate efficiently on resource-constrained devices. These improvements are designed to make the self-checkout process faster, more reliable, and better suited to the demands of modern retail settings.

2. Related Work

2.1. Traditional Object Detection Algorithms

Object detection has long been a central challenge in computer vision, with traditional approaches primarily relying on CNNs to identify and locate objects within images. Among the most notable early models are Faster R-CNN, Single Shot MultiBox Detector (SSD), and the original YOLO series. Before the advent of DL, the field of object detection primarily relied on traditional manual feature engineering techniques, while DL advanced the field of object detection and brought its benefit to retail operators and consumers [17,18]. Solutions from this era typically followed a two-stage process: feature extraction and classification. Besides the advancement of algorithms for exploring data samples [19,20], techniques from image processing and computer vision were used to manually construct and extract prominent features from images, such as Haar feature sets and scale-invariant feature transform (SIFT) [21]. These designed features were then fed into classifiers to identify objects within images. However, the effectiveness of this approach heavily depended on the quality of the feature engineering and often fell short in complex and variable image backgrounds and scenes.
With the widespread adoption of CNNs, a significant leap occurred in object detection. CNNs, with their robust automatic feature-learning capabilities, eliminated the need for cumbersome manual feature design [22,23]. They could directly learn hierarchical and rich feature representations from raw image data, greatly enhancing the accuracy and efficiency of object detection. Faster R-CNN marked a significant milestone by introducing the region proposal network (RPN), which efficiently generates region proposals that are likely to contain objects, reducing the need for exhaustive search methods and improving detection accuracy [24]. However, its two-stage process, which separates region proposal and classification, makes it computationally intensive, limiting its real-time application.
SSD aimed to balance accuracy and speed by eliminating the region proposal stage entirely. Instead, it predicts object classes and bounding boxes directly from feature maps at different scales, allowing for the detection of objects of various sizes in a single pass [25]. While SSD offers improved speed over Faster R-CNN, it still struggles with detecting smaller objects and achieving the highest levels of accuracy [26].
RT-DETR (real-time detection transformer) aims to optimize key components of DETR, achieving real-time and high-precision object detection [27]. It represents an innovative real-time end-to-end object detection model. The RT-DETR model consists of three components: the backbone network, a hybrid encoder, and a transformer decoder equipped with an auxiliary prediction head. The backbone network uses convolutional networks to extract salient features at three different scales, and the hybrid encoder is a key innovation in the RT-DETR model, optimized for the traditional transformer’s inefficiency in handling multi-scale features. The hybrid encoder combines the strengths of convolutional neural networks (CNNs) and transformers, enhancing feature representation quality by decoupling intra-scale feature interaction and cross-scale feature fusion. The transformer decoder is another core component of the RT-DETR model, responsible for generating predictions such as object bounding boxes, categories, and embedding vectors based on the feature maps output by the hybrid encoder. Unlike traditional anchor-based detectors, RT-DETR adopts an end-to-end prediction approach, directly predicting object attributes through a set of learnable queries.
YOLO introduced a groundbreaking approach by framing object detection as a single regression problem, predicting bounding boxes and class probabilities directly from the entire image in one forward pass through the network. This made YOLO exceptionally fast compared to its predecessors, making it suitable for real-time applications. However, early versions of YOLO had limitations in detecting small and overlapping objects and struggled with localization accuracy [28].
These traditional algorithms laid the groundwork for the rapid advancements in object detection seen in later models. They highlighted the trade-offs between speed and accuracy and set the stage for more sophisticated approaches, such as the newer iterations of the YOLO series, which aim to overcome these limitations.

2.2. Development of the YOLO Series

The YOLO series has significantly influenced the field of object detection since its introduction [14]. By reimagining object detection as a single regression problem, YOLO enables real-time processing with relatively high accuracy, making it a groundbreaking approach in computer vision.
YOLOv1 laid the foundation by dividing an input image into a grid and predicting bounding boxes and class probabilities directly from each grid cell [14]. While this approach offered unprecedented speed, it struggled with detecting small and overlapping objects due to the limitations of its grid-based prediction method. In response to these challenges, YOLOv2 introduced anchor boxes, which improved the accuracy of bounding box predictions by allowing the model to predict multiple bounding boxes for each grid cell [29]. YOLOv2 also adopted the Darknet-19 backbone, significantly enhancing its feature extraction capabilities. The introduction of multi-scale training further allowed the model to generalize better across different object sizes and shapes, making YOLOv2 a more robust solution.
Furthermore, YOLOv3 incorporated a multi-scale feature pyramid network (FPN), which enabled better detection of objects at varying scales by merging features from different layers of the network [30]. YOLOv3 also improved class prediction accuracy, enhancing the model’s ability to recognize objects across a broader range of categories. YOLOv4 and YOLOv5 continued to refine the architecture with a focus on optimizing both speed and accuracy [31,32]. These versions integrated better backbone networks, more advanced feature fusion techniques, and improved loss functions. The enhancements made in these iterations further solidified YOLO’s position as a leading framework for real-time object detection, particularly in scenarios requiring a balance between computational efficiency and detection performance.
With the introduction of YOLOv8, the series saw significant architectural innovations, including the use of cross-stage partial network (CSPNet) for more efficient feature extraction and anchor-free detection heads that simplified the model’s design [33]. YOLOv8 also leveraged the SiLU activation function, which facilitated better gradient flow during training, leading to faster convergence and higher accuracy.
The most recent version, YOLOv10, represents a culmination of these advancements, introducing a non-maximum suppression (NMS)-free training strategy that minimizes inference delays [16]. YOLOv10 also employs a dual-label assignment mechanism that enhances both recall and precision, making it the most advanced and capable iteration in the YOLO series.
The continuous evolution of the YOLO series highlights the ongoing efforts to balance speed, accuracy, and efficiency, ensuring that each new version builds on the strengths of its predecessors while addressing their limitations.

2.3. Application of YOLO

The YOLO series has seen widespread adoption across various industries due to its real-time object detection capabilities. In autonomous driving, YOLO is used for detecting pedestrians, vehicles, and other road elements, ensuring timely decisions for safety [34]. In surveillance and security, it aids in monitoring environments, detecting intruders, and analyzing crowd behavior, making it ideal for real-time threat detection [35]. YOLO’s applications can also extend to healthcare, where it assists in detecting abnormalities in medical images [36]. In agriculture, YOLO is employed in precision farming to monitor crop health and detect pests, helping optimize yields [37]. In public service, YOLO is used to enhance automatic pavement distress recognition to assist highway maintenance decision making [38]. In the retail industry, YOLO powers automated checkout systems by accurately identifying products, enhancing customer experience, and streamlining inventory management [2]. The versatility and efficiency of YOLO across these diverse applications highlight its significant impact on real-time object detection across multiple sectors.

3. The Improved MidState-YOLO-ED Network

3.1. Integration of YOLOv8n and YOLOv10

One of the most distinct features of YOLOv10 compared to its predecessors is the elimination of non-maximum suppression (NMS), achieved by introducing a consistent dual-assignment strategy [16]. This strategy involves calling the loss function calculation method on YOLOv8n twice, summing the results, and returning them. The goal of this approach is to address the issue of redundant predictions in post-processing, aligning more closely with the end-to-end direction of the RT-DETR model. However, this modification resulted in less accuracy for many datasets in practical applications. To prevent a loss of precision, the prediction head of YOLOv10 was reverted back to that of YOLOv8n.
YOLOv10 introduces several modifications to enhance efficiency and reduce computational redundancy. These include a lightweight classification head, spatial-channel decoupled downsampling (SCDD), and a rank-based block design (i.e., C2fUIB). SCDD is a two-step process that first adjusts the channel dimensions using point-wise convolution, followed by spatial downsampling using depth-wise convolution. This reduces the number of parameters and helps to minimize information loss during the downsampling process. However, experiments have shown that some information loss still occurs, which can result in reduced latency but not necessarily improved performance.
This study modifies two core components of YOLOv10, reverting them back to their YOLOv8n counterparts and integrating both versions. As a result, the current model has been named MidState-YOLO. The model also incorporates efficient multi-scale attention (EMA) and the C2f-Dual convolution design, ultimately leading to the final model being named the MidState-YOLO-ED network.

3.2. Integration of EMA Attention

To further enhance the expressive capability of the MidState-YOLO network and establish long and short dependency relationships, EMA has been integrated into the neck network. EMA is a parallel attention mechanism used in computer vision tasks to improve model performance and processing speed. Unlike traditional CNNs, EMA adopts a parallel structure to handle input data. This parallel convolution allows faster training of models when dealing with large-scale data and enhances accuracy by enabling parallel processing of features at different scales. In Figure 1, the divided groups are represented by “g”, while “X Avg Pool” denotes one-dimensional horizontal global pooling, and “Y Avg Pool” represents one-dimensional vertical global pooling, respectively. The formula for the average pooling operation is as follows, where Xc(i, j) represents the element at position (i, j):
Z c = 1 H × W j H i W X c i , j
The input to EMA is first grouped and reshaped, redistributing part of the channel dimensions to the batch dimension. This is followed by further subdivision of the channel dimension into multiple sub-features to preserve key information in each channel and optimize the distribution of spatial semantic features [39].
This structure contains two main parallel branches: one branch performs one-dimensional global pooling operations to encode global information, while the other branch performs feature extraction through a 3 × 3 convolution. The output features from these two branches are modulated through a sigmoid function and normalization processes, then merged through a cross-dimensional interaction module to capture pixel-level pairwise relationships. Finally, the sigmoid-modulated output feature maps are used to enhance or weaken the original input features, thus achieving a more refined and effective feature representation. Therefore, EMA not only encodes inter-channel information to adjust the importance of various channels but preserves precise spatial structural details within these channels.

3.3. Lightweight Dual Convolution C2f-Dual Design

In YOLOv8 and YOLOv10, the C2f module integrates both low-level and high-level feature maps, facilitating the capture of gradient information flows. However, with the increasing number of layers in CNNs, semantic information in feature maps tends to be progressively extracted and aggregated, leading to redundancy in deep feature maps. Additionally, due to the weight-sharing mechanism of convolutional layers, convolutional kernel parameters are shared at different positions in the deep feature maps, further exacerbating redundancy. Bottleneck modules, composed of many complex convolutions, significantly increase parameter size and computational complexity.
To address this issue, the C2f-Dual convolution design, improved using dual convolutional kernels (DualConv), significantly reduces computational costs and the number of parameters while also enhancing precision. This improvement involves replacing the C2f modules before the spatial pyramid pooling fast (SPPF) with C2f-Dual modules, as shown in Figure 2. This adaptation not only streamlines the network but also optimizes performance by ensuring that critical spatial and semantic features are efficiently processed and integrated.
DualConv is designed to build lightweight deep neural networks by combining 3 × 3 and 1 × 1 convolution kernels to process the same input feature map channels, optimizing information processing and feature extraction. In DualConv, the 3 × 3 convolution kernels are used to extract spatial features from the feature maps, capturing more spatial information, while the 1 × 1 convolution kernels integrate these features and reduce the model’s parameters. Each group of convolution kernels processes a portion of the input channels independently before the outputs are merged, facilitating efficient flow and integration of information across different feature map channels.
Additionally, DualConv employs group convolution technology to efficiently arrange convolution filters. In group convolution, both input and output feature maps are divided into multiple groups, with each group’s convolution filters processing only a part of the corresponding input feature map. This arrangement allows different kernels within a group to process the same set of input channels in parallel, optimizing information flow and feature extraction efficiency while maintaining the network’s representational capabilities.
Thus, replacing the bottleneck structures in C2f with DualBottleneck enriches gradient flow representation, enhances feature extraction capabilities, and reduces the diversity of false positives and false negatives in network learning. This makes it more suitable for retail commodity object detection scenarios.

4. Experimental Results and Analysis

4.1. Experimental Setup and Parameters

The hardware environment and software configurations used for the experiments are listed in Table 1. During the model training process, the learning rate was set to 0.01, with optimization carried out using stochastic gradient descent (SGD). The momentum parameter was set at 0.937, and the weight decay factor was 0.0005. The batch size used was 32, and the image size was 640 × 640 pixels. The comparative experiments were conducted over 30 epochs, while the ablation studies were carried out over 25 epochs.

4.2. Dataset

This study employs a portion of the retail product (RPC) dataset for training and validation. The RPC dataset, developed by Megvii Technology’s Nanjing Research Institute, is currently the largest product recognition dataset available [40]. It includes up to 200 different product categories and a total of 83,000 images, realistically simulating retail environments and surpassing existing datasets in fidelity. Moreover, it effectively captures the fine-grained characteristics inherent in the automatic check-out (ACO) problem.
The conceptual approach of this study may differ from that of the researchers who collected the RPC dataset. When customers enter a store and place the items they wish to purchase on the checkout counter, an ideal ACO system would automatically identify each product and accurately generate a shopping list in one go, as shown in Figure 3. Thus, ACO is fundamentally a system designed to identify and count the occurrence of each item in any combination of products.
This research posits that there are multiple important metrics for assessing performance on ACO tasks. To ensure accuracy and performance, the images used to train the ACO recognition system should mirror the actual retail checkout environments, which can indeed be simplified and stabilized. Additionally, initial models do not need to exhaust all product combinations to perform ACO tasks; instead, creating random groups of product combinations suffices.
Therefore, for this study, only the checkout configurations from the RPC dataset images are used. The single-product images are not included in the training dataset because they do not perfectly mimic real-life retail scenarios. As a result, only 30,000 out of the 83,000 images from the RPC dataset are utilized for the research. We randomly divided 30,000 images of checkout configurations into training, validation, and test sets in an 8:1:1 ratio. This approach aims to provide a realistic yet controlled set of data that reflects real-world ACO system operations while maintaining manageable complexity and variety in training scenarios. The dataset includes checkout images with varying levels of clutter, capturing the real-world challenge of recognizing products when multiple items are grouped together in a single scene. However, further research will focus on addressing different lighting conditions, either by utilizing additional datasets or through image augmentation techniques. By doing so, it could help reduce potential biases inherent in the dataset, leading to a more robust and generalized model.

4.3. Evaluation Metrics

This paper employs precision (P), recall (R), mean average precision (mAP), number of parameters (Params), floating point operations (FLOPs), and frames per second (FPS) as evaluation metrics, with a set IoU threshold of 0.5. Note that [email protected] denotes the mean average precision when the IoU is set at 0.5, and [email protected]:0.95 indicates the mean average precision when the IoU ranges from 0.5 to 0.95, with a step size of 0.05. The floating point operations indicate the complexity of the algorithm. The specific meanings of other performance metrics are as follows:
P = ( T P T P + F P ) × 100
R = ( T P T P + F N ) × 100
where precision is the probability that a positive sample predicted by the model is indeed a positive sample, and recall is the probability that a positive sample in reality is predicted as positive by the model. There are two important metrics used to evaluate model performance, A P and m A P , such that:
A P = 0 1 p ( r ) d r
m A P = 1 n j j = 1 n j A P j
In Formulas (2) and (3), TP (true positives) denotes positive examples correctly identified as positive by the model; FP (false positives) denotes negative examples incorrectly identified as positive; and FN (false negatives) denotes positive examples incorrectly identified as negative.
In Formulas (4) and (6), mAP represents the mean of the average precision across all object detection categories. AP is the average of precision values at different recall levels. The curve plotted with precision as the y-axis and recall as the x-axis is known as the PR curve. mAP can be calculated as the average area under the PR curves for all categories, n represents the number of instances in a given category, and APj represents the detection precision for category j.
FPS is an important metric for measuring the speed of a model’s image processing capability. It indicates the number of images the model can detect in one second, determining the model’s response speed and real-time performance in practical applications. The higher the FPS value, the more images the model can process within a unit of time, thus indicating a faster detection speed. Specifically:
F P S = F r a m e N u m E l a p s e d T i m e
where FrameNum is the total number of images detected, and ElapsedTime is the total time the model took to perform the detection.

4.4. Ablation Study

To investigate the extent of improvements from the three modification schemes, ablation studies were conducted using YOLOv8-n and YOLOv10-n as baseline networks. These experiments were carried out on the experimental dataset without changing the software and hardware environment, with the only parameter change being the reduction of epochs to 25. As indicated in Table 2, the MidState-YOLO network, which integrates modules from YOLOv8 and YOLOv10, achieved a 23.2% increase in mAP compared to the original baseline network of YOLOv10-n. This suggests that combining two models can harness the advantages of both while avoiding some of their respective shortcomings, thereby enhancing model performance.
Using the lightweight dual convolution C2f_Dual module in the MidState-YOLO network significantly reduced the number of parameters, and all performance metrics showed improvements. The reason for this is the reduction in redundant information in the deep feature maps and a decrease in false positives and false negatives during network learning. The inclusion of the EMA module increased the [email protected] to 84.4%, resulting in a 4.2% gain. This demonstrates that the EMA module effectively captures global information to learn richer semantic features, focusing the network model more on the overall context of retail product targets and enhancing model performance.
The final improved model, MidState-YOLO-ED, showed improvements in all evaluation metrics relative to YOLOv8-n, with precision and recall increasing by 2.3 and 1.6 percentage points, respectively, and mAP reaching 89%. Additionally, the number of parameters and floating point operations were significantly lower.

4.5. Experimental Results and Discussion

The experimental results primarily organize the performance parameters of the trained algorithms, explaining the strengths and weaknesses of each algorithm based on these results, and analyzing the experimental data and actual detection effects. All experiments were conducted under the same configuration settings, and after training, the weight files generated by each algorithm were tested. The algorithms used in the comparative experiments include SSD, Faster-RCNN, YOLOv8-n, YOLOv10-n, and RT-DETR-L. The experimental results are shown in Table 3.
The experimental results indicate that, compared to the SSD and Faster-RCNN algorithms, the YOLO series algorithms and the improvements introduced in this study exhibit superior detection performance. Additionally, since Faster-RCNN is a two-stage algorithm, its applicability is limited by its complexity and the extended duration required for detection, which makes its overall performance inferior to that of the YOLO series.
The comparative results of different algorithms demonstrate that the MidState-YOLO-ED algorithm excels in terms of the number of parameters and floating point operations, with only 3,288,096 parameters and 9.6 GFLOPs. This fully proves the excellent performance of the improved algorithm in terms of lightweight design. The algorithm can process image data quickly and accurately, making it suitable for scenarios requiring fast response. It is also more appropriate for operation in resource-constrained environments such as mobile devices and embedded systems, and it meets real-time requirements [41]. As a current mainstream real-time object detection model, the RT-DETR algorithm performs excellently in terms of accuracy. However, it has similar issues to the Faster-RCNN algorithm, such as a large number of parameters and suboptimal time performance in retail product detection.
Speed tests were conducted on different detection algorithms to evaluate their detection efficiency. The speed tests were conducted in the same GPU environment, and the results showed that the improved MidState-YOLO-ED algorithm surpassed most algorithms in detection speed, achieving an impressive improvement of 71.88 fps compared to the Faster-RCNN algorithm.
Furthermore, key indicators such as recall and mAP for the MidState-YOLO-ED algorithm are higher than those of baseline algorithms, offering a new option for efficient and rapid object detection.
After training the final improved model, MidState-YOLO-ED, for 30 epochs, the prediction results are displayed in Figure 4. The results demonstrate that the algorithm proposed in this paper achieves excellent detection performance while maintaining a lightweight framework.

4.6. Comparative Visualization Analysis

To visually illustrate the effectiveness and superiority of MidState-YOLO-ED, a set of images was selected for testing using YOLOv8-n, YOLOv10-n, and the proposed model. The detection results are shown in Figure 5, where Figure 5a–c in the figure represent the test results of the YOLOv10, YOLOv8, and MidState-YOLO-ED models, respectively. It is clear from the figure that both the YOLOv10 and YOLOv8 models perform poorly in the retail product detection environment, resulting in missed detections. In Figure 5a, although the distribution of products is not dense, the YOLOv10 model still detects a single product as multiple instances, which is a critical issue in the retail product detection field. Figure 5c clearly shows that MidState-YOLO-ED can effectively avoid a series of problems present in the original models.

4.7. Loss Function Analysis

Based on the logs saved during the training process, loss comparison curves for seven models were plotted. Figure 6 (left) represents the training loss, while Figure 6 (right) represents the validation loss. The horizontal and vertical axes represent the number of epochs and the loss value, respectively. For most models, both training loss and validation loss steadily decrease as the training progresses. The loss of the Faster-RCNN model begins to rise after reaching a certain low point, indicating that the Faster-RCNN model is too complex for the retail environment and has a risk of overfitting. Observing the proposed MidState-YOLO-ED model, it demonstrates good stability, with its validation loss remaining relatively stable throughout the training cycle without significant fluctuations. This indicates that the MidState-YOLO-ED model is able to effectively learn data features during training while maintaining good generalization capability, which is crucial for object detection tasks in practical applications.

5. Discussion

The improvements proposed in our system, particularly the integration of efficient multi-scale attention and the C2f-Dual convolution module, directly contribute to reducing these errors by enhancing both spatial and contextual feature extraction. This results in better differentiation between visually similar products, such as different flavors of chocolate or sizes of packaged goods, which are typically challenging for conventional object detection models.
However, our study has limitations. Currently, our system does not explicitly process a full 360-degree rotation of objects. However, the model’s capability to handle diverse viewpoints is enhanced by the use of multi-scale attention mechanisms and augmented data during training. These components help improve the model’s generalization to various angles, including side views and tilted projections. To address the need for 360-degree object recognition more thoroughly, future work could incorporate multi-view learning or synthetic data generation to simulate a wider range of object orientations. This would allow the model to learn how to recognize objects from all possible angles, which is particularly important for products that may be placed on the checkout surface in unpredictable orientations.

6. Conclusions

In this paper, we presented an enhanced self-checkout system using an improved YOLOv10 network. The system significantly advances retail automation by optimizing checkout efficiency and minimizing labor costs. Our adaptations to the YOLOv10 model, integrating features from YOLOv8 and new post-processing algorithms, markedly improves product recognition accuracy, with our experiments demonstrating superior performance over existing systems.
Broader applications in inventory control and customer service will be benefited by this study. Our study shows that AI-driven technologies will play a pivotal role in enhancing consumer experiences and operational efficiency.

Author Contributions

Conceptualization, L.T. and X.L.; methodology, L.T. and X.L.; software, L.T.; validation, X.L.; formal analysis, L.T. and X.L.; data curation, X.L.; writing—original draft preparation, L.T.; writing—review and editing, L.T., S.L., J.G., X.L., L.C. and H.J.; visualization, L.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data included in this manuscript cannot be shared publicly, due to the need to protect the privacy of the included subjects. Data may be shared upon reasonable request to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Orel, F.D.; Kara, A. Supermarket self-checkout service quality, customer satisfaction, and loyalty: Empirical evidence from an emerging market. J. Retail. Consum. Serv. 2014, 21, 118–129. [Google Scholar]
  2. Vats, A.; Anastasiu, D.C. Enhancing retail checkout through video inpainting, yolov8 detection, and deepsort tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 5530–5537. [Google Scholar]
  3. Shoman, M.; Aboah, A.; Morehead, A.; Duan, Y.; Daud, A.; Adu-Gyamfi, Y. A region-based deep learning approach to automated retail checkout. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 3210–3215. [Google Scholar]
  4. Drid, K.; Allaoui, M.; Kherfi, M.L. Object detector combination for increasing accuracy and detecting more overlapping objects. In Proceedings of the International Conference on Image and Signal Processing, Marrakesh, Morocco, 4–6 June 2020; Springer: Cham, Switzerland, 2020; pp. 290–296. [Google Scholar]
  5. Oosthuizen, K.; Botha, E.; Robertson, J.; Montecchi, M. Artificial intelligence in retail: The AI-enabled value chain. Australas. Mark. J. 2021, 29, 264–273. [Google Scholar]
  6. Zheng, Q.; Yu, C.; Cao, J.; Xu, Y.; Xing, Q.; Jin, Y. Advanced Payment Security System: XGBoost, CatBoost and SMOTE Integrated. arXiv 2024, arXiv:2406.04658. [Google Scholar]
  7. Liu, J.; Huang, I.; Anand, A.; Chang, P.H.; Huang, Y. Digital Twin in Retail: An AI-Driven Multi-Modal Approach for Real-Time Product Recognition and 3D Store Reconstruction. In Proceedings of the 2024 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW), Orlando, FL, USA, 16–21 March 2024; pp. 368–373. [Google Scholar]
  8. Lin, Z.; Wang, C.; Li, Z.; Wang, Z.; Liu, X.; Zhu, Y. Neural radiance fields convert 2d to 3d texture. Appl. Sci. Biotechnol. J. Adv. Res. 2024, 3, 40–44. [Google Scholar]
  9. Liu, F.; Wang, X.; Chen, Q.; Liu, J.; Liu, C. SiamMAN: Siamese multi-phase aware network for real-time unmanned aerial vehicle tracking. Drones 2023, 7, 707. [Google Scholar] [CrossRef]
  10. Mokayed, H.; Quan, T.Z.; Alkhaled, L.; Sivakumar, V. Real-time human detection and counting system using deep learning computer vision techniques. In Proceedings of the Artificial Intelligence and Applications, Wuhan, China, 18–20 November 2023; Volume 1, pp. 221–229. [Google Scholar]
  11. He, C.; Li, K.; Zhang, Y.; Tang, L.; Zhang, Y.; Guo, Z.; Li, X. Camouflaged object detection with feature decomposition and edge reconstruction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 22046–22055. [Google Scholar]
  12. Li, H.; Zhang, R.; Pan, Y.; Ren, J.; Shen, F. Lr-fpn: Enhancing remote sensing object detection with location refined feature pyramid network. arXiv 2024, arXiv:2404.01614. [Google Scholar]
  13. Gu, J.; Wang, Z.; Kuen, J.; Ma, L.; Shahroudy, A.; Shuai, B.; Liu, T.; Wang, X.; Wang, G.; Cai, J.; et al. Recent advances in convolutional neural networks. Pattern Recognit. 2018, 77, 354–377. [Google Scholar]
  14. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  15. Wang, G.; Chen, Y.; An, P.; Hong, H.; Hu, J.; Huang, T. UAV-YOLOv8: A small-object-detection model based on improved YOLOv8 for UAV aerial photography scenarios. Sensors 2023, 23, 7190. [Google Scholar] [CrossRef]
  16. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J.; Ding, G. Yolov10: Real-time end-to-end object detection. arXiv 2024, arXiv:2405.14458. [Google Scholar]
  17. Wei, Y.; Tran, S.; Xu, S.; Kang, B.; Springer, M. Deep learning for retail product recognition: Challenges and techniques. Comput. Intell. Neurosci. 2020, 2020, 8875910. [Google Scholar]
  18. Dang, B.; Zhao, W.; Li, Y.; Ma, D.; Yu, Q.; Zhu, E.Y. Real-Time pill identification for the visually impaired using deep learning. arXiv 2024, arXiv:2405.05983. [Google Scholar]
  19. Tarvainen, A.; Valpola, H. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  20. Duan, Y.; Zhao, Z.; Qi, L.; Wang, L.; Zhou, L.; Shi, Y.; Gao, Y. Mutexmatch: Semi-supervised learning with mutex-based consistency regularization. IEEE Trans. Neural Netw. Learn. Syst. 2022, 35, 8441–8455. [Google Scholar] [CrossRef] [PubMed]
  21. Cruz-Mota, J.; Bogdanova, I.; Paquier, B.; Bierlaire, M.; Thiran, J.P. Scale invariant feature transform on the sphere: Theory and applications. Int. J. Comput. Vis. 2012, 98, 217–241. [Google Scholar] [CrossRef]
  22. Tokunaga, H.; Teramoto, Y.; Yoshizawa, A.; Bise, R. Adaptive weighting multi-field-of-view CNN for semantic segmentation in pathology. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 12597–12606. [Google Scholar]
  23. Jin, Y. GraphCNNpred: A stock market indices prediction using a Graph based deep learning system. arXiv 2024, arXiv:2407.03760. [Google Scholar]
  24. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  25. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the 14th European Conference of the Computer Vision (ECCV 2016), Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part I 14. Springer: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  26. Li, Z.; Yang, L.; Zhou, F. FSSD: Feature fusion single shot multibox detector. arXiv 2017, arXiv:1712.00960. [Google Scholar]
  27. Zhao, Y.; Lv, W.; Xu, S.; Wei, J.; Wang, G.; Dang, Q.; Liu, Y.; Chen, J. Detrs beat yolos on real-time object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024; pp. 16965–16974. [Google Scholar]
  28. Jiang, P.; Ergu, D.; Liu, F.; Cai, Y.; Ma, B. A Review of Yolo algorithm developments. Procedia Comput. Sci. 2022, 199, 1066–1073. [Google Scholar] [CrossRef]
  29. Sang, J.; Wu, Z.; Guo, P.; Hu, H.; Xiang, H.; Zhang, Q.; Cai, B. An improved YOLOv2 for vehicle detection. Sensors 2018, 18, 4272. [Google Scholar] [CrossRef]
  30. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  31. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  32. Zhu, X.; Lyu, S.; Wang, X.; Zhao, Q. TPH-YOLOv5: Improved YOLOv5 based on transformer prediction head for object detection on drone-captured scenarios. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 2778–2788. [Google Scholar]
  33. Terven, J.; Córdova-Esparza, D.M.; Romero-González, J.A. A comprehensive review of yolo architectures in computer vision: From yolov1 to yolov8 and yolo-nas. Mach. Learn. Knowl. Extr. 2023, 5, 1680–1716. [Google Scholar] [CrossRef]
  34. Sarda, A.; Dixit, S.; Bhan, A. Object detection for autonomous driving using yolo [you only look once] algorithm. In Proceedings of the IEEE 2021 3rd International Conference on Intelligent Communication Technologies and Virtual Mobile Networks (ICICV), Online, 4–6 February 2021; pp. 1370–1374. [Google Scholar]
  35. Narejo, S.; Pandey, B.; Esenarro Vargas, D.; Rodriguez, C.; Anjum, M.R. Weapon detection using YOLO V3 for smart surveillance system. Math. Probl. Eng. 2021, 2021, 9975700. [Google Scholar] [CrossRef]
  36. Qureshi, R.; Ragab, M.G.; Abdulkader, S.J.; Alqushaib, A.; Sumiea, E.H.; Alhussian, H. A comprehensive systematic review of yolo for medical object detection (2018 to 2023). Authorea Prepr. 2023, 11, 2–31. [Google Scholar]
  37. Li, M.; Zhang, Z.; Lei, L.; Wang, X.; Guo, X. Agricultural greenhouses detection in high-resolution satellite images based on convolutional neural networks: Comparison of faster R-CNN, YOLO v3 and SSD. Sensors 2020, 20, 4938. [Google Scholar] [CrossRef] [PubMed]
  38. Dan, H.C.; Yan, P.; Tan, J.; Zhou, Y.; Lu, B. Multiple distresses detection for Asphalt Pavement using improved you Only Look Once Algorithm based on convolutional neural network. Int. J. Pavement Eng. 2024, 25, 2308169. [Google Scholar] [CrossRef]
  39. Zhong, X.; Liu, X.; Gong, T.; Sun, Y.; Hu, H.; Liu, Q. FAGD-Net: Feature-Augmented Grasp Detection Network Based on Efficient Multi-Scale Attention and Fusion Mechanisms. Appl. Sci. 2024, 14, 5097. [Google Scholar] [CrossRef]
  40. Wei, X.S.; Cui, Q.; Yang, L.; Wang, P.; Liu, L. RPC: A large-scale retail product checkout dataset. arXiv 2019, arXiv:1901.07249. [Google Scholar]
  41. Ouyang, D.; He, S.; Zhang, G.; Luo, M.; Guo, H.; Zhan, J.; Huang, Z. Efficient multi-scale attention module with cross-spatial learning. In Proceedings of the 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2023), Rhodes Island, Greece, 4–10 June 2023; pp. 1–5. [Google Scholar]
Figure 1. Integration of EMA.
Figure 1. Integration of EMA.
Jimaging 10 00248 g001
Figure 2. MidState-YOLO-ED.
Figure 2. MidState-YOLO-ED.
Jimaging 10 00248 g002
Figure 3. Dual convolution C2f-Dual design.
Figure 3. Dual convolution C2f-Dual design.
Jimaging 10 00248 g003
Figure 4. Prediction results.
Figure 4. Prediction results.
Jimaging 10 00248 g004
Figure 5. Comparative visualization analysis.
Figure 5. Comparative visualization analysis.
Jimaging 10 00248 g005
Figure 6. Loss function analysis.
Figure 6. Loss function analysis.
Jimaging 10 00248 g006
Table 1. Hardware and software configuration for experimental environment.
Table 1. Hardware and software configuration for experimental environment.
ConfigurationParameter
GPURTX4080-16G
CPUAMD Ryzen7
Operating SystemWindows11
Deep Learning FrameworksPytorch2.1.1
+cuda12.1
Build SystemPyCharm
Table 2. Ablation study experiments with improved strategies.
Table 2. Ablation study experiments with improved strategies.
ModelPrecisionRecall[email protected][email protected]:0.95ParamsGFLOPs
YOLOv8-n0.8240.8090.8770.6913,371,0249.8
YOLO v10-n0.5510.5950.610.4812,885,8889.2
MidState-YOLO0.7940.7750.8420.6543,405,4569.8
MidState-YOLO+DualConv0.840.8160.8830.6863,251,8569.4
MidState-YOLO+EMA0.8430.8130.8840.6943,408,9289.9
MidState-YOLO-ED0.8470.8250.890.6913,288,0969.6
Table 3. Comparison of experimental data.
Table 3. Comparison of experimental data.
ModelRecall[email protected][email protected]:0.95ParamsGFLOPsFPS
Faster R-CNN0.8990.9950.85541,808,406134.938.01
SSD0.7580.9430.69330,160,46813.458.78
YOLOv8-n0.9810.9920.8693,371,0249.8106.38
YOLO v10-n0.9700.9910.8712,885,8889.2112.36
RT-DETR-L0.9910.9940.87932,394,740104.389.29
MidState-YOLO0.9870.9930.8753,405,4569.8105.26
MidState-YOLO-ED0.9850.9940.8753,288,0969.6109.89
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tan, L.; Liu, S.; Gao, J.; Liu, X.; Chu, L.; Jiang, H. Enhanced Self-Checkout System for Retail Based on Improved YOLOv10. J. Imaging 2024, 10, 248. https://doi.org/10.3390/jimaging10100248

AMA Style

Tan L, Liu S, Gao J, Liu X, Chu L, Jiang H. Enhanced Self-Checkout System for Retail Based on Improved YOLOv10. Journal of Imaging. 2024; 10(10):248. https://doi.org/10.3390/jimaging10100248

Chicago/Turabian Style

Tan, Lianghao, Shubing Liu, Jing Gao, Xiaoyi Liu, Linyue Chu, and Huangqi Jiang. 2024. "Enhanced Self-Checkout System for Retail Based on Improved YOLOv10" Journal of Imaging 10, no. 10: 248. https://doi.org/10.3390/jimaging10100248

APA Style

Tan, L., Liu, S., Gao, J., Liu, X., Chu, L., & Jiang, H. (2024). Enhanced Self-Checkout System for Retail Based on Improved YOLOv10. Journal of Imaging, 10(10), 248. https://doi.org/10.3390/jimaging10100248

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop