Next Article in Journal
Optimization Operation Strategy for Comprehensive Energy System Considering Multi-Mode Hydrogen Transportation
Next Article in Special Issue
A Parallel Prognostic Method Integrating Uncertainty Quantification for Probabilistic Remaining Useful Life Prediction of Aero-Engine
Previous Article in Journal
An Interpretation Method of Gas–Water Two-Phase Production Profile in High-Temperature and High-Pressure Vertical Wells Based on Drift-Flux Model
Previous Article in Special Issue
Urban-Scale Acoustic Comfort Map: Fusion of Social Inputs, Noise Levels, and Citizen Comfort in Open GIS
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automated Dual-Side Leather Defect Detection and Classification Using YOLOv11: A Case Study in the Finished Leather Industry

1
DigitaLLeather Doo, 34000 Kragujevac, Serbia
2
Faculty of Electrical Engineering, Mechanical Engineering and Naval Architecture (FESB), University of Split, R. Boskovica 32, 21000 Split, Croatia
*
Author to whom correspondence should be addressed.
Processes 2024, 12(12), 2892; https://doi.org/10.3390/pr12122892
Submission received: 1 November 2024 / Revised: 10 December 2024 / Accepted: 14 December 2024 / Published: 17 December 2024

Abstract

:
This study explores the optimization of leather defect detection through the advanced YOLOv11 model, addressing long-standing challenges in quality control within the leather industry. Traditional inspection methods, reliant on human accuracy ranging between 70% and 85%, have limited leather utilization rates and contributed to substantial material waste. To overcome these limitations, we developed an automated solution leveraging controlled environmental conditions within a custom-designed light chamber. This research specifically targets common defects in leather, such as insect larvae damage and removal cuts, by analyzing both the grain and flesh sides of the material. The results reveal a notable improvement in detection accuracy on the flesh side, achieving 93.5% for grubs and 91.8% for suckout, compared to 85.8% and 87.1% on the grain side. Classification accuracy further demonstrates the advantage of dual-side analysis, with the flesh side reaching 98.2% for grubs and 97.6% for suckout, significantly outperforming the grain side. The dual-side methodology, combined with YOLOv11’s enhanced capabilities, enables the precise identification of subtle defects and offers a transformative approach to leather defect detection. By integrating cutting-edge AI models with standardized digitization environments, this research presents a scalable, highly efficient solution that reduces human error, optimizes leather utilization, and supports industrial sustainability.

1. Introduction

The traditional processing and use of leather date back to around 7000 BCE. However, with the invention of chrome tanning in the early 19th century, leather production began to industrialize, marking the start of modern leather manufacturing. According to research by International Market Analysis Research and Consulting (IMARC), the company specialized in delivering comprehensive market insights, the global leather industry market size was valued at $390.9 billion in 2023, with an estimated compound annual growth rate (CAGR) of 4.8% from 2024 to 2032 [1]. Despite adopting alternatives like synthetic and vegan leather, genuine animal leather remains widely utilized, with its market estimated at approximately $100 billion in 2023, and an expected growth rate of around 4% per year [2]. Animal hides in use are cow (67%), sheep (12%), pig (11%), goat (10%), and other (0.5%) [3]. Finished leather goods find extensive applications across various industries, including fashion, automotive, furniture, and footwear.
For manufacturers of finished leather hides and products, leather utilization poses significant challenges from business, environmental, and social perspectives. Increasing the utilization of finished leather hides before the cutting process can lead to cost savings, reduced leather waste, decreased inventory requirements, and a lower demand for animal hides. According to the Food and Agriculture Organization of the United Nations (FAO), this high consumption of genuine leather contributes to the death of about 1 billion animals annually worldwide [4].
Leather utilization rates for finished leather vary based on the type of animal hide used and the specific industry or product. This paper focuses on optimizing the defect detection, classification, and segmentation of finished cowhide leather in the automotive industry, which is 17% of the global market for genuine leather products. We conducted research on automotive interior suppliers operating in Serbia. In total, 12 international companies that supply industrial leather have been surveyed: YanFeng (3), Adient (2), Magna (2), Grammer (1), AutoStop Interiors (2), SMP (1), and Passubio SPA (1).
Key findings relevant to this study indicate that the accuracy rate of quality inspection operators for finished leather hides ranges between 70% and 85%, directly affecting the leather utilization rate, which varies from 50% to 85%. Scrap rates following the cutting process range from 1.5% to 10%, while customer complaints are between 0.5% and 3%. Customer complaints and scrap rates further reduce utilization by an additional 6–10% due to losses in the optimization of post-processing leather hides or recalculations in the automatic pattern nesting for leather. The total number of finished leather hides cut daily by these companies ranges between 120 and 350. The average weight of a leather hide is around 25 kg, and the average purchase price is approximately 35 EUR per hide. Considering all these factors, the result is concerning: each of the surveyed companies produces approximately 3400 kg of leather waste daily, amounting to a loss of around 4700 EUR per day. The cause of such a poor process performance is the human factor and the inability to conduct a consistent inspection of industrial leather.
A comprehensive review [5] of over 100 scientific papers highlights the evolution of defect detection and classification methods for industrial leather, ranging from traditional image-processing to state-of-the-art deep-learning approaches. While methods such as Mask R-CNN and U-Net have demonstrated reasonable accuracy, challenges persist in detecting subtle or complex defects [6,7,8]. Recent advancements, particularly YOLO-based models, have emerged as the preferred solution for their superior speed and accuracy in real-time scenarios [9,10]. However, a gap remains in addressing defects on the flesh side of leather—a critical area this study focuses on, introducing a dual-side analysis approach that significantly enhances detection performance. The methods applied for defect detection and classification on finished leather include the following [5]:
  • Traditional image-processing methods achieved an accuracy range of 70% to 85%;
  • Machine-learning-based methods reported accuracies ranging from 80% to 90%;
  • Deep-learning approaches demonstrated higher accuracy levels, often exceeding 90%.
Deep-learning techniques, particularly those based on You Only Look Once (YOLO) models, were identified as the most effective for leather defect inspection and classification among the methods analyzed. YOLO models not only provided higher accuracy rates but also improved processing efficiency compared to traditional methods. YOLOv2 and similar architectures have been noted for their real-time processing capabilities and high detection and classification rates, achieving accuracies of around 92% to 95% in some studies [5]. To the best of our knowledge, the versions of YOLO models applied to defect detection and classification range from v1 to v9 [5,6,7]. In this study, the YOLOv11 model is employed, which was released on 30 September 2024, and currently represents the state of the art in YOLO models, for defect detection and classification in industrial leather.
The key contributions of this work include the development of a dual-side analysis approach, which uniquely emphasizes defect visibility on the flesh side, and the implementation of a controlled digitization chamber to standardize environmental variables such as lighting and dust. By leveraging YOLOv11’s enhanced features—such as anchor-free detection and multi-task capabilities—this study achieves significant improvements in detection accuracy, surpassing previous YOLO models. These advancements demonstrate a practical and scalable solution to automate quality control processes, reduce leather waste, and improve industrial sustainability.

2. Theoretical Background

2.1. Literature Review

In a comprehensive review conducted by Chen, Z. et al. [5], 90 scientific papers published over the last three decades, up to 2022, were analyzed. These studies primarily focused on addressing the challenges of defect detection and classification on the grain side of the leather using image-processing and machine-learning techniques. The defects addressed by these authors are Tick Marks, Hot-Iron Marks, Scabies, Open and Closed Surface Cuts, Scars, Pinholes, Fungal Attack, Chrome Patch, Wrinkles, and Healing Wounds. As this research is centered on the application of deep-learning methodologies, traditional image-processing and machine-learning approaches were not considered in this literature review. Various models have been employed for detection and classification tasks on the grain side of the leather, including Mask R-CNN, U-Net, YOLOv2, AlexNet, SVM, CNN, ResNet50, GAN, and Salthisature models [5]. This work also provides insights into data volume, defect types, and the best-achieved accuracy. For defect detection and classification on the grain side of the leather, the authors utilized datasets ranging from 270 to 15,000 samples across one to nine different defect types, with an accuracy rate ranging from 80% to 96%. Based on this literature review, it has been concluded that a dataset of 300 samples per defect will be sufficient to achieve meaningful results [5].
In the experimental study, Chen, Z. et al. [6] provide a detailed analysis of the performance and application of YOLO (v1–v8) models for defect detection and classification on leather surfaces. The authors consider YOLO a practical option for improving quality control processes in the leather industry. They highlight issues with certain types of defects where YOLO did not perform well, suggesting the potential to detect hard-to-see defects by examining the leather’s reverse side. The defects addressed by these authors are as follows: Cavity, Pinhole, Scratch, Rotten Surface, Growth Line, Healing Wound, Crease, and Bacterial Wounds.
Deep learning has emerged as the most effective approach for solving defect detection and identification problems on the grain side of the leather, as confirmed by the findings of Tabernik, D. et al. [8].
Andrzej Wróbel and Piotr Szymczyk [9] reported achieving an over 95% accuracy in defect detection using the YOLOv5 model, with the requirement of 1000 or more samples per defect. Similarly, Thangakumar, J., et al. [10] applied YOLOv8 for precise defect detection and segmentation on the front side of the wet blue leather, which falls outside the scope of this study, as it pertains to pre-finished leather. However, their findings indicate that the YOLOv8 model demonstrated a high accuracy and reliability in defect detection and segmentation, using less time and fewer training samples. Leveraging prior research on YOLO models, Silva, V. et al. [11] applied YOLOv5 for detection and classification tasks on the grain side of leather. Their work utilized a relatively small dataset of defective and non-defective samples, demonstrating the model’s potential even under limited data conditions.
In contrast, Wang, M. et al. [7] utilized YOLOv9, positioning them as one of the few authors to explore the latest versions of YOLO. Their study compared the performance of Faster-RCNN, VFNet, Detr, and YOLO (v3, v5, and v7) with YOLOv9, where YOLOv9 achieved superior speed and accuracy, reaching a nearly 95% defect detection accuracy on the grain side of the leather. The defects addressed by these authors are as follows: Bubble, Dent, and Broken Glue.
The findings of this study, supported by the prior literature, suggest that datasets containing approximately 300 samples per defect are sufficient to achieve meaningful accuracy levels exceeding 90%. However, it is important to note that dataset size alone does not guarantee perfect accuracy. Achieving 100% accuracy would require significantly larger datasets, greater defect variability, and further model optimization. While YOLOv11 demonstrated robust performance with the dataset used in this study, factors such as image resolution, defect complexity, and annotation quality also play a critical role in determining accuracy limits.
Furthermore, the extensive review highlights that YOLO technology is particularly effective for defect detection and classification challenges. Notably, no previous studies have explored the use of the reverse side of leather for defect detection, demonstrating a unique contribution of this research. All studies reviewed by the authors were conducted in laboratory conditions without environmental control. To address this gap, we have developed a dedicated, controlled environment specifically designed to ensure consistent conditions suitable for industrial applications. This innovative setup represents a significant improvement over traditional methods. A literature review also led us to conclude that no author has yet addressed defects resulting from damage created by insect larvae or from the removal of the hide, most likely because they are barely or not at all visible on the leather surface. Even when visible, they are difficult to classify, as the two classes (grubs and suckout) manifest almost identically on the leather surface.

2.2. YOLO Model

This section provides a summary of the advancements in the YOLO model family over the past year (Table 1). The previous models, from YOLOv1 to YOLOv8, are presented and thoroughly analyzed in an experimental study conducted by Chen Z. et al. [6].
The YOLOv9 model, developed in February 2024, incorporates advanced architectures and optimization techniques to reduce data loss and improve model efficiency. A key advancement over YOLOv8 is the implementation of programmable gradient information (PGI) (Table 2) and the generalized efficient layer aggregation network (GELAN). PGI mitigates information loss by preserving gradient information across network layers, while GELAN facilitates feature extraction at all scales, simplifying computation. Together, these techniques yield improvements over the YOLOv8 model—a 49% reduction in training parameters and a 43% decrease in computational load—making this model suitable for high-performance tasks and resource-constrained devices [12]. In the literature, we have encountered applications of YOLOv9 across various industries. Yang et al. [13] applied YOLOv9 in the case of a target-matching method for visible and infrared images at sea based on decision-level topological relations. Huang, X. et al. [14] used an improved YOLOv9-OREPA model in their study on underwater crack detection. Rizzieri, N. et al. [15] employed YOLOv8 and YOLOv9 for detecting diabetic retinopathy (DR), a severe eye disease secondary to diabetes. Wang, X. et al. [16] utilized the YOLOv9 model for coffee detection and segmentation in the agriculture sector. Bustamante, A. et al. [17] used YOLOv9 for the accurate posture detection of elderly people, aiming to improve monitoring and provide timely alerts in homes and elderly care facilities. Mi, Z., and Yan, W. Q. [18] combined YOLOv9 and the Swin Transformer to develop a complex model for detecting strawberry ripeness. Li, J., Feng [19] applied YOLOv9 to handle various adverse weather conditions. Wan, L. et al. [20] used YOLOv9 for mobile inspections conducted by intelligent tunnel robots, which are instrumental in broadening the inspection reach, economizing on inspection expenditures, and augmenting the operational efficiency of inspections. Xu, W. et al. [21] applied YOLOv9 to violence detection. Although the YOLOv9 model was released at the beginning of 2024, numerous scientific papers already demonstrated its diverse applications in object detection, achieving excellent results in detection, classification, and segmentation.
The YOLOv10 model [22], introduced in May 2024, offers efficiency-oriented advancements, such as the removal of the NMS algorithm during inference. This model employs consistent dual assignments for object label prediction: one generating multiple predictions during training and another selecting the optimal prediction per object. For example, Hussain and Khanam [23] demonstrated YOLOv10’s applicability by achieving a higher precision and reduced inference time in photovoltaic defect detection. Additionally, YOLOv10 employs an enhanced backbone for feature extraction (CSPNet), and, unlike previous versions, it does not use predefined anchors, making it simpler and faster. It also incorporates elements of the self-attention mechanism, helping it focus on the important parts of the image. These improvements enhance the model’s accuracy and efficiency, making YOLOv10 an excellent choice for applications requiring a fast response and high precision [22].
Although the YOLOv10 model was only released in mid-2024, studies published on its applications are already available. Hussain, M., and Khanam, R. [23] focused on improving the quality inspection within the photovoltaic (PV) domain using YOLOv10. Tan, L. et al. [24] used YOLOv10 for a novel self-checkout system for retail, enhancing checkout efficiency and reducing labor costs. Qiu, X. et al. [25] employed YOLOv10 for a lightweight drone target detection algorithm. Liu, W. et al. [26] used YOLOv10 for tomato recognition and for a rapid sorting system. Zhang, C. et al. [27] developed a novel YOLOv10-DECA model for the real-time detection of concrete cracks.
The YOLOv11 model [28], introduced in September 2024, represents a significant leap in balancing speed, accuracy, and versatility. Its modular architecture, composed of the backbone, neck, and head, incorporates innovative features such as the novel C3k2 block in the backbone, which enhances feature extraction through smaller kernel convolutions. The neck integrates advanced modules like SPPF and C2PSA to better handle small or occluded objects. The head supports multi-task outputs, including object detection, segmentation, classification, and tracking.
As shown in Figure 1, the modular structure of YOLOv11 is visually depicted, highlighting the architectural improvements and the interrelation of its components. This figure provides an intuitive understanding of how these innovations contribute to a 22% reduction in parameters compared to YOLOv8, making YOLOv11 an optimal choice for real-time tasks and resource-constrained environments.

3. Methodology

3.1. Environmental Set-Up

For the research, we created a dedicated small-scale environment in the form of a light chamber, which would be suitable for industrial conditions. This environment for digitizing industrial leather cuts is fully enclosed to control lighting and industrial dust—factors that can significantly impact the research. Lighting is essential for achieving adequate photographs for computer vision applications, while industrial dust on the leather poses a problem by concealing tiny perforation defects.
To ensure optimal conditions for defect detection, we constructed a digitization chamber designed to minimize environmental variability. The chamber, measuring 1000 mm × 1000 mm × 1000 mm, features matte white Forex walls to eliminate glare and provide even light distribution. A burgundy background enhances contrast, particularly for defects on the flesh side, while the enclosed environment prevents dust interference, which is critical for detecting irregularities in small surface. The lighting setup, positioned at a 45-degree angle with a color temperature of 4000 K, highlights subtle textures and edges of defects, ensuring maximum visibility for the YOLOv11 model. To manage the technical solution, we developed an operational software system. The surface on which we digitize the leather is burgundy, providing contrast that is beneficial for AI models.
The lighting is LED type, 2100 LUX, with a color temperature of 4000 K, positioned at a 45-degree angle. The industrial camera used for digitizing leather cut parts is The Imaging Source DFK 38UX541 1.1″, 20.3 Megapixels, 18 fps, with a Global Shutter and USB 3.1 output with a resolution of 4504 × 4504 pixels (1:1 aspect ratio) (available online: https://www.theimagingsource.com/en-us/product/industrial/38u/dfk38ux541/ accessed on 1 January 2020). The lens used on the camera is a V1226-MPZ, 12 mm 1″ f/2.6-f/16 with a fixed focal length (available online: https://www.rmaelectronics.com/computar-v1226-mpz/ accessed on 1 January 2020).
To manage the solution, we developed operational software for lighting, camera control, and AI model applications. The software is controlled via a touchscreen monitor Zeus ZUS215MAX located on the front of the chamber. All equipment is connected via the NVIDIA Jetson Orin Nano Developer kit (NVIDIA Bac Giang, Vietnam, Hungary). The prototype developed for this research, which consists of custom-made hardware and software, is named Prototype P2-24 (Figure 2).

3.2. Data Collection and Model Training

The data annotation was carried out using the CVAT.ai software (available online: https://www.cvat.ai/ accessed on 1 January 2020) allowing easy data export in the desired format. The data consist of images of leather pieces, each with two images: front and back (Figure 3 and Figure 4). For each piece of leather, polygonal annotations of defects were made on both sides. As mentioned before, we focus on two defects (grubs and suckout) that are more clearly visible on the back side, in an attempt to improve the overall classification accuracy. These are defects to which no tolerance is applied according to the general standards of leather industry customers; they must be removed without exception. The dataset, carefully created by our team using the previously described prototype system and leather supplied by the AutoStop interiors company (AutoStop interiors, www.autostopinteriors.com), includes a total of 1200 images. Of these, 600 images (300 front and 300 back) correspond to the grubs defect class, and 600 images (300 front and 300 back) represent the suckout defect class, amounting to a total of 1327 annotations. The image resolution is 4504 × 4504. The dataset is split into training, validation, and test sets, which creates the following: 840 images for training, 240 for validation, and 120 for test sets.
We use two fine-tuned YOLOv11 models for detection, trained separately with data from the front and back side of the leather, respectively. Models are trained to recognize defects on both sides of each leather piece with corresponding annotations. The training was conducted on an NVIDIA RTX 4090 GPU (NVIDIA Bac Giang, Vietnam, Hungary), with optimized parameters for epochs, initial learning rate, and batch size, using rigid image augmentations to enhance the model’s robustness.

3.3. Metrics for Evaluation

We used a test set of 120 (60 + 60) images of leather pieces, with both front and back views provided for each piece. Each image is annotated with polygonal defect markings, allowing for precise comparison of defect detection on each side. Across all images, there are a total of 64 grubs and 60 suckout defects annotated. After detecting defects on both sides of the leather, we defined a combined weighted scoring system that applies to the output classification. The experiment tests two approaches:
  • Front-only classification: based solely on the front side;
  • Dual-side combined classification: combines results from both front and back sides.
The selection of scoring weights involves a degree of arbitrariness and is refined through trial and error. However, an informed approach can guide this process, and we have determined that a 0.3/0.7 weighting, favoring the backside, yields the most accurate results when considering these two classes.
CombinedScore = 0.7 × Conffront + 0.3 × Confback

4. Results

We applied the YOLO model on a dataset of 1200 defective samples from both sides of the leather, focusing on grubs and suckout defects, and the results are shown in Table 3. In the case of detection on the grain side, we achieved a satisfactory accuracy of 85.8% for grubs and 87.1% for suckout. However, for defect detection on the flesh side, we achieved significantly better results, with 93.5% for grubs and 91.8% for suckout.
In terms of classification, we obtained relatively low results for grubs on the grain side at 66.8%, while the suckout classification accuracy was 97.6%. For defect classification on the flesh side, we achieved outstanding results of 98.2% for grubs and 97.6% for suckout.
The results demonstrate a notable improvement in defect detection accuracy on the flesh side of the leather, with an overall accuracy increase from 0.85 to 0.93. The F1-Score improved from 0.86 to 0.93, reflecting a more balanced and effective classification. Precision increased from 0.87 to 0.92, indicating fewer false positives on the flesh side, while recall rose from 0.84 to 0.94, showing a higher rate of correctly identified defects. This improvement across all three metrics suggests that the flesh side provides clearer defect signals, allowing for more accurate and reliable detection. In Figure 5, confusion matrices show the performance of the two detectors.
Table 4 evaluates the performance of the YOLOv11 model using four key metrics: Accuracy, Precision, Recall, and F1-Score. Each metric provides a specific insight into the defect detection and classification performance. Accuracy measures the overall correctness of the predictions by considering both true positives and true negatives, giving a comprehensive view of the model’s reliability. It is calculated as follows:
Accuracy = (TP + TN)/(TP + TN + FP + FN)
Precision highlights the system’s ability to avoid false positives by calculating the proportion of true positive defect detections out of all positive predictions. It is defined as follows:
Precision = (TP)/(TP + FP)
Recall emphasizes the system’s capability to minimize false negatives by measuring the proportion of actual defects correctly identified. The formula for Recall is as follows:
Recall = (TP)/(TP + FN)
F1-Score combines Precision and Recall into a harmonic mean, providing a balanced evaluation, especially useful when Precision and Recall show significant differences. The F1-Score is calculated as as follows:
F1-Score = 2 × Precision*Recall)/(Precision + Recall)
In these formulas, TP (True Positive) represents correctly identified defective samples, TN (True Negative) represents correctly identified non-defective samples, FP (False Positive) represents non-defective samples incorrectly classified as defective, and FN (False Negative) represents defective samples incorrectly classified as non-defective. These metrics collectively offer a comprehensive evaluation of the model’s performance, addressing both its reliability and robustness in defect detection.
These metrics were selected due to their relevance in assessing the performance for industrial applications, where both the over-detection and under-detection of defects carry critical consequences. Accuracy was chosen to offer a broad evaluation of the model’s effectiveness in correctly classifying all instances. Precision is essential in reducing unnecessary rejections of leather pieces, preventing waste and minimizing operational costs. Recall is equally critical to ensure that no defective pieces are overlooked, which could compromise product quality and lead to customer dissatisfaction. F1-Score was included to offer a holistic view, balancing Precision and Recall in scenarios where both metrics are equally significant. Together, these metrics provide a comprehensive validation of the YOLOv11 model, ensuring its practicality for defect detection in real-world industrial settings.
Furthermore, Figure 6 illustrates the distribution and relationship between the class probability for the flesh side (blue) and the grain side (green). Scores are sorted in ascending order for both the front and flesh side to facilitate a direct comparison. The difference between the two distributions is significant, indicating a clear separation in scores for the front and flesh sides. This separation may contribute to a more nuanced approach to defect detection, where certain defects are more reliably detected on the flesh side, as confirmed by higher confidence scores for the flesh side.
It is also important to note that the grain side shows much greater variability in classification confidence, reinforcing the hypothesis that weighted scoring could significantly enhance accuracy (Table 5). These results highlight differences in the detection scores for the defects grubs and suckout when comparing the flesh side and grain side of leather pieces. For both classes, the back-side scores are significantly higher than the grain-side scores, with an average difference of around 28.7% for grubs and 18.2% for suckout. This indicates that they are much more discernable on the flesh side compared to the front. The standard deviation on the flesh side is low, showing that flesh side scores are consistently high. In contrast, the front-side scores have a higher standard deviation, reflecting more variability and less consistent detection on the front.
The proposed workflow, illustrated in Figure 7, outlines the process for leveraging YOLOv11 to detect and classify defects in finished leather. The methodology begins with the collection and annotation of leather samples scanned on both the grain and flesh sides, ensuring comprehensive defect coverage. The annotated dataset is divided into training, validation, and test sets, with data augmentation applied to enhance model performance. Separate YOLOv11 models are trained for each side, taking advantage of the model’s advanced architecture for defect detection. Results from the two models are combined using a weighted scoring system, prioritizing defects that are more visible on the flesh side. Finally, evaluation metrics, including Accuracy, Precision, Recall, and F1-Score, are used to assess both the individual and combined model performance, demonstrating the scalability and industrial applicability of the approach.

5. Discussion

The results of this study demonstrate the effectiveness of using YOLOv11 for defect detection and classification on finished leather, especially when analyzing both the grain and flesh sides. The experiment confirms that defect detection on the flesh side yields a higher accuracy and reliability for certain defects, such as grubs (larval damage) and suckout (cut damage). This outcome supports the hypothesis that examining the flesh side can reveal defects otherwise difficult to detect on the grain side, providing a promising approach for optimizing leather utilization.
Table 6 provides a comparative analysis of various YOLO-based models applied to leather defect detection and classification, highlighting their methodologies, accuracies, advantages, and limitations.
Wang et al. [7] introduced the GEI-YOLOv9 model, which employed several innovative components, including the GhostNetV2 backbone to reduce redundant computations, EMA Attention for enhanced small-scale feature extraction, and the Inner-IOU loss function for faster and more accurate bounding box regression. These enhancements allowed the model to achieve a mean average precision (mAP) of 94.7% for defects such as bubbles, dents, and broken glue on the dataset provided by three leather manufacturers in the Pearl River Delta. Although the model demonstrated high precision and recall, its focus on the grain side limited its ability to detect subtle defects that are more prominent on the flesh side.
Wróbel and Szymczyk [9] utilized the YOLOv5 model for general defect detection. Their approach, which did not focus on categorizing specific defect types, achieved an average precision (mAP) of 95%. This simplicity made YOLOv5 an adaptable and computationally efficient choice for real-time applications. However, the lack of categorization in defect classes reduces its suitability for specific industrial requirements, where detailed classification is often necessary. The origin of the dataset in this study is not specified.
Chen et al. [6] conducted an extensive evaluation of YOLOv5, YOLOv6, YOLOv7, and YOLOv8 for detecting various leather defects, including pinholes, scratches, cavities, bacterial wounds, and creases. These models achieved up to an 85.1% accuracy for detecting multiple defect types, showcasing their reliability in diverse scenarios. However, the models faced challenges with detecting smaller defects or irregular shapes and lacked optimization for real-time industrial environments, which limited their applicability in high-speed inspection processes. The origin of the dataset used in this study is a leather production enterprise in Guangdong Province, China.
Thangakumar et al. [10] applied YOLOv8 for precise defect detection and segmentation in pre-finished leather materials. This model achieved a detection accuracy of 92%, demonstrating its effectiveness in handling complex defect shapes and edges. Its focus on pre-finished leather, however, restricted its direct applicability to finished leather inspection, which requires more nuanced detection capabilities.
As illustrated in Table 6, the YOLOv11 model offers distinct advantages over earlier iterations, including YOLOv5, YOLOv6, YOLOv7, and YOLOv8. While prior models achieved accuracies of up to 92% for grain-side defect detection [9], YOLOv11 surpasses this with detection accuracies of 96.85% on the flesh side, owing to its novel C3k2 block architecture and optimized detection capabilities [12,13]. Additionally, YOLOv11’s multi-task output supports simultaneous detection, classification, and segmentation, making it a transformative tool for addressing the challenges of industrial defect inspection [10,11]. These models, while achieving respectable accuracies, faced challenges with certain defect types, particularly those involving subtle surface features. YOLOv11 addresses these challenges through several novel contributions: a dual-side analysis approach that enables detection on both the grain and flesh sides of the leather, a controlled digitization environment that ensures consistency and reduces external variability, and an anchor-free detection mechanism that improves both precision and speed.
The results highlight YOLOv11’s superior performance, with detection accuracies of 93.5% and 91.8% for grubs and suckout defects on the flesh side, compared to the highest accuracy of 92% reported by YOLOv8 in earlier studies. Additionally, the introduction of a weighted scoring system allows YOLOv11 to leverage the unique strengths of each leather side, offering a tailored and adaptable solution for industrial applications. Unlike prior models, YOLOv11 also incorporates multi-task capabilities, enabling simultaneous detection, classification, and segmentation, further enhancing its utility. These advancements collectively position YOLOv11 as a state-of-the-art model, setting a new benchmark for automated leather defect detection and addressing the limitations of previous approaches. The specialized environmental setup, consisting of a light chamber equipped with a 20.3-megapixel industrial camera, played a crucial role in enhancing the detection accuracy by minimizing variables like lighting inconsistencies and dust interference, which are common in industrial environments. This controlled setting proved essential for achieving consistent, reliable defect detection, indicating that adopting similar environments could be beneficial in industrial applications.
The two-class weighted scoring system introduced here also proved advantageous, particularly when detection on one side provided better visibility or detail than the other. By assigning greater weight to the flesh side for specific defect types, this scoring method capitalized on the flesh side’s improved accuracy, as shown by an average accuracy difference of 28.7% for grubs and 18.2% for suckout. This approach highlights the potential of side-specific detection, where each side of the leather can provide unique data, refining the classification process and potentially reducing waste through more precise grading.
The increased precision and recall scores on the flesh side underscore the model’s capability to reduce false positives and false negatives, two critical aspects of automated leather inspection. With precision increasing from 0.87 to 0.92 and recall from 0.84 to 0.94 on the flesh side, these improvements suggest that incorporating backside analysis can significantly reduce inspection errors. Such accuracy has substantial implications for real-world applications, potentially allowing manufacturers to meet stricter quality standards and minimize waste due to misclassification or overlooked defects.
The low detection and classification results on the grain side for grubs and suckout defects are primarily due to the smooth surface texture, which makes subtle defects difficult to discern. The grain side often exhibits surface patterns and glare that can obscure critical features of these defects. For instance, grubs, caused by larval damage, are characterized by uniform circular or elliptical erosion, but these features are often masked by the grain side’s reflective properties. Similarly, suckout defects, which typically appear as linear separations, are less distinguishable against the grain side’s complex texture.
On the flesh side, these defects are much more evident due to its rough texture, which provides better contrast and visibility. Grubs defects appear as clusters of circular or elliptical patterns, while suckout defects are visible as distinct tears or splits, making them easier to classify. These observations, illustrated in Figure 2 and Figure 3, highlight the physical advantages of the flesh side in providing clearer visual cues for defect detection. Additionally, the controlled digitization environment used in this study further enhances the visibility of these features by minimizing glare and ensuring uniform lighting, enabling the YOLOv11 model to achieve a significantly higher detection accuracy on the flesh side.
For this reason, the application of YOLOv11 achieved relatively lower results on the grain side for grubs and suckout defects compared to other studies using YOLO models for different defect classes. However, by applying YOLOv11 to the flesh side, we achieved higher results than previously reported in the literature.
It is essential to recognize that not all defect types may benefit from dual-side analysis. Certain surface characteristics unique to the grain side may not manifest clearly on the flesh side, warranting further research to assess the applicability of this approach across all defect types. Thus, while dual-side analysis shows promise, expanding the classification system to accommodate a wider range of leather defect profiles could enhance its robustness.
The results presented in this study highlight the substantial practical benefits of employing YOLOv11 in the finished leather industry. The proposed approach addresses key challenges in defect detection and classification, offering several advancements that directly enhance operational efficiency, quality control, and sustainability in industrial leather production.
One of the most significant contributions is the high detection accuracy achieved by YOLOv11, particularly on the flesh side of leather, where subtle defects such as suckout and grubs are better detected. With detection accuracies of 91.8% for suckout and 93.5% for grubs, YOLOv11 surpasses traditional manual inspection methods, which typically achieve accuracies of 70–85%. This improved accuracy translates directly to optimized leather utilization, as accurate defect identification allows manufacturers to minimize waste and maximize the usable area of each hide, ultimately improving profitability.
The introduction of dual-side analysis represents another major innovation. Unlike prior approaches that are predominantly focused on the grain side, this study demonstrates the importance of leveraging the flesh side, where certain defects are more visible. By integrating this dual-side methodology, YOLOv11 enhances the comprehensiveness of leather quality inspections.
Furthermore, the system ensures consistency and scalability by automating the inspection process. Automation eliminates human error and ensures uniform defect detection across large-scale production batches, which is crucial for meeting industrial demands. The study also introduces a specialized digitization chamber with controlled lighting and dust elimination, enabling YOLOv11 to operate effectively under real-world conditions. This adaptability ensures the model’s suitability for diverse manufacturing environments.
While the practical benefits of YOLOv11 are clear, potential challenges for implementation must also be acknowledged. Initial investment costs, including the acquisition of digitization chambers, industrial cameras, and computing infrastructure, may pose a barrier for smaller manufacturers. Additionally, data preparation requirements for training the models such as extensive and precise defect annotations are both time-intensive and technically demanding. Furthermore, while YOLOv11 performs exceptionally well for certain defects, such as those on the flesh side, its effectiveness for certain grain-side defects may require further refinement or supplementary approaches. Finally, integration into existing workflows necessitates staff training and system adaptation, which could introduce temporary delays and additional costs.
Despite these challenges, the findings of this study underscore the transformative potential of YOLOv11 for the finished leather industry. By automating defect detection with unprecedented accuracy and scalability, this approach offers a pathway to enhance productivity, reduce costs, and support sustainability in leather manufacturing.
The results of this study firmly establish YOLOv11 as a robust solution for industrial defect detection, achieving a balanced integration of accuracy and speed. The dual-side analysis approach, coupled with a controlled digitization chamber, represents a significant advancement in detecting subtle defects such as grubs and suckout [6,8]. By enabling precise defect identification and reducing leather waste, this methodology not only enhances operational efficiency but also supports broader sustainability goals [7]. Future work could explore expanding this approach to a wider range of defect types and other industrial applications, reinforcing its potential for a transformative impact across various sectors [10,13]. Recent advances in deep learning and computer vision, as exemplified by YOLOv11, offer a pathway to streamline leather inspection processes, reduce human error, and improve leather utilization. Future studies could expand the range of detectable defects, validate this approach across larger datasets, and adapt the methodology to other stages of leather processing. Practical considerations for industrial applications should also be considered, as both sides of the leather would need to be digitized, possibly simultaneously, to avoid errors. A potential solution could involve using a transparent surface for dual-side digitization, although this would require entirely new studies focused on the development of appropriate hardware and environments for digitization. The results demonstrate that a dataset of approximately 300 samples per defect provides sufficient variability for YOLOv11 to achieve high accuracy levels. Combined with the dual-side analysis and controlled digitization environment, this threshold ensures robust performance across diverse defect types, addressing the inconsistencies noted in prior studies.

6. Conclusions

This work is the result of an experimental study within a scientific innovation project. For the purposes of the experiment, dedicated hardware was developed, and the latest technology of the YOLO model series (YOLOv11) was used for defect recognition and localization on the grain and flesh side of industrial leather. Experimental validation confirmed that the application of YOLO on the flesh side of industrial leather led to the improved detection and classification of defects whose damage is more evident on the flesh side of industrial leather. For the experiment, we used the grubs defect, which is caused by damage inflicted by larvae on the flesh side of industrial leather, and suckout, which results from cuts during the separation of the hide from the animal’s body. The data used for training and validating the experiment were more than sufficient to achieve an appropriate accuracy in detection and classification.
In this study, we addressed critical challenges in leather defect detection and classification by offering novel methodologies and practical solutions tailored to industrial applications. The key contributions of this work are summarized as follows:
  • Dual-side analysis approach: This study introduces a novel dual-side analysis method, emphasizing defect detection on both the grain and flesh sides of leather. This approach uniquely highlights defects such as grubs and suckout, which are more discernible on the flesh side, improving overall detection rates.
  • Controlled digitization chamber: A dedicated digitization environment was developed to standardize imaging conditions by controlling variables such as lighting and dust. This innovation ensures consistency and reliability in defect detection, making the methodology robust for industrial-scale applications.
  • Application of YOLOv11’s advanced features: The study leverages YOLOv11’s enhanced capabilities, including anchor-free detection and multi-task learning for simultaneous detection, classification, and segmentation. These features enable the model to achieve superior accuracy and efficiency, surpassing previous YOLO models and other traditional methods.
  • Improved detection accuracy and scalability: The proposed approach demonstrates significant improvements in detection accuracy, achieving high performance with a dataset of approximately 300 samples per defect, providing a scalable and practical solution for industrial adoption.
  • Contribution to sustainability: By automating the quality control process and reducing leather waste through accurate defect detection, this study supports sustainability goals within the leather manufacturing industry.
The primary limitation of this study lies in its inapplicability to other types of defects commonly observed in industrial leather, particularly those that are either absent or minimally visible on the flesh side of the leather. Certain defects on the grain side exhibit unique surface characteristics that do not translate effectively to the flesh side, thereby limiting the scope of the current dual-side analysis. To fully address these challenges and evaluate the broader applicability of the YOLOv11 model, additional experimental studies would be required. Such studies should focus on individual defect types, assessing their distinct characteristics and determining the effectiveness of the YOLOv11 model in automating defect detection across a comprehensive range of industrial leather imperfections. These future investigations would provide a more robust understanding of the model’s potential for universal application in the industrial leather inspection process. It has been demonstrated that YOLOv11 is highly successful in detecting and classifying defects, even for those that are difficult to classify due to similar manifestations on the surface. By applying a controlled environment in the form of a digitization chamber and using the YOLO model on data obtained in controlled conditions, we have shown that it is possible to implement computer vision in industrial settings, creating the potential to overcome manual labor and reduce leather waste by optimizing utilization in the pre-cutting process of finished leather hides. Furthermore, the results underscore the importance of considering the physical properties of the leather’s grain and flesh sides when developing automated defect detection systems. The superior detection accuracy on the flesh side highlights how surface texture and defect visibility can significantly impact model performance. This study demonstrates how controlled lighting, optimized imaging conditions, and side-specific analysis can provide scalable solutions for improving quality control in the leather industry. The ability to accurately classify defects in an industrial environment contributes to the possibility of assessing tolerance and meeting different customer standards, which can significantly enhance utilization optimization.
This study highlights the practical implications of applying YOLOv11 in the finished leather industry. The model’s superior defect detection and classification accuracy, combined with its adaptability to industrial conditions, addresses long-standing challenges in quality control. These advancements offer manufacturers a pathway to enhance operational efficiency, reduce waste, and support sustainability, while also emphasizing the importance of adopting advanced AI models for automated quality inspection.

Author Contributions

Conceptualization, N.B., K.T., J.Ž. and M.M.; methodology, N.B., K.T. and J.Ž.; software, N.B., K.T. and J.Ž.; validation, N.B., K.T. and J.Ž. and M.M.; writing—original draft preparation, N.B., K.T. and J.Ž.; writing—review and editing, N.B., K.T., J.Ž. and M.M.; supervision, N.B. and M.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by “EU for Green Agenda in Serbia” grant number 00136377/00127312/2023/26.

Data Availability Statement

Data available on request from the authors.

Acknowledgments

This research was conducted as part of the innovation grant “Improving the Utilization of Industrially Processed Animal Leather through the Application of Artificial Intelligence”, Ref. No.: 00136377/00127312/2023/26, within the project “EU for Green Agenda in Serbia”, with the technical and financial support of the European Union and in partnership with the Ministry of Environmental Protection, and is implemented by UNDP in cooperation with the Embassy of Sweden and the European Investment Bank (EIB), with additional funding from the Governments of Sweden, Switzerland, and Serbia.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Grand View Research. Available online: https://www.grandviewresearch.com/industry-analysis/leather-goods-market (accessed on 25 October 2024).
  2. Cognitive Market Research. Available online: https://www.cognitivemarketresearch.com/genuine-leather-market-report (accessed on 25 October 2024).
  3. Mascianà, P. World statistical compendium for raw hides and skins, leather and leather footwear. In Intergovernmental Group on Meat and Dairy Products Sub-Group on Hides and Skins; Food and Agricultural Organization of the United Nations: Rome, Italy, 2015. [Google Scholar]
  4. Food and Agriculture Organization of the United Nations—FAOSTAT Database. Available online: https://www.fao.org/faostat/en/#data/QCL (accessed on 26 October 2024).
  5. Chen, Z.; Jiehang, D.; Zhu, Q.; Wang, H.; Chen, Y. A systematic review of machine-vision-based leather surface defect inspection. Electronics 2022, 11, 2383. [Google Scholar] [CrossRef]
  6. Chen, Z.; Zhu, Q.; Zhou, X.; Deng, J.; Song, W. Experimental Study on YOLO-Based Leather Surface Defect Detection. IEEE Access 2024, 12, 32830–32848. [Google Scholar] [CrossRef]
  7. Wang, M.; Xie, X.; Qiu, H.; Li, J. GEI-YOLOv9-based leather defect detection algorithm research. In Proceedings of the Fourth International Conference on Image Processing and Intelligent Control, Kuala Lumpur, Malaysia, 10–12 May 2024; Volume 13250, pp. 571–576. [Google Scholar]
  8. Tabernik, D.; Šela, S.; Skvarč, J.; Skočaj, D. Segmentation-based deep-learning approach for surface-defect detection. J. Intell. Manuf. 2020, 31, 759–776. [Google Scholar] [CrossRef]
  9. Wróbel, A.; Szymczyk, P. An Application of a Material Defect Detection System Using Artificial Intelligence. Int. J. Mod. Manuf. Technol. 2023, 15, 221–228. [Google Scholar] [CrossRef]
  10. Thangakumar, J. Revolutionizing leather quality assurance through deep learning powered precision in defect detection and segmentation by a comparative analysis of Mask RCNN and YOLO v8. In Proceedings of the 2024 International Conference on Advances in Data Engineering and Intelligent Computing Systems (ADICS), Chennai, India, 18–19 April 2024. [Google Scholar]
  11. Silva, V.; de Pinho, R.; Allahdad, M.K.; Silva, J.; Ferreira, M.J.; Magalhães, L. A Robust Real-time Leather Defect Segmentation Using YOLO. In Proceedings of the 2023 18th Iberian Conference on Information Systems and Technologies (CISTI), Aveiro, Portugal, 20–23 June 2023. [Google Scholar]
  12. Wang, C.Y.; Yeh, I.H.; Liao, H.Y.M. Yolov9: Learning what you want to learn using programmable gradient information. In Proceedings of the European Conference on Computer Vision, Milan, Italy, 29 September–4 October 2024. [Google Scholar]
  13. Yang, S.; Cao, Z.; Liu, N.; Sun, Y.; Wang, Z. Maritime electro-optical image object matching based on improved YOLOv9. Electronics 2024, 13, 2774. [Google Scholar] [CrossRef]
  14. Huang, X.; Liang, C.; Li, X.; Kang, F. An Underwater Crack Detection System Combining New Underwater Image-Processing Technology and an Improved YOLOv9 Network. Sensors 2024, 24, 5981. [Google Scholar] [CrossRef] [PubMed]
  15. Rizzieri, N.; Dall’Asta, L.; Ozoliņš, M. Diabetic Retinopathy Features Segmentation without Coding Experience with Computer Vision Models YOLOv8 and YOLOv9. Vision 2024, 8, 48. [Google Scholar] [CrossRef] [PubMed]
  16. Wang, X.; Zhang, C.; Qiang, Z.; Liu, C.; Wei, X.; Cheng, F. A Coffee Plant Counting Method Based on Dual-Channel NMS and YOLOv9 Leveraging UAV Multispectral Imaging. Remote Sens. 2024, 16, 3810. [Google Scholar] [CrossRef]
  17. Bustamante, A.; Belmonte, L.M.; Morales, R.; Pereira, A.; Fernández-Caballero, A. Bridging the Appearance Domain Gap in Elderly Posture Recognition with YOLOv9. Appl. Sci. 2024, 14, 9695. [Google Scholar] [CrossRef]
  18. Mi, Z.; Yan, W.Q. Strawberry Ripeness Detection Using Deep Learning Models. Big Data Cogn. Comput. 2024, 8, 92. [Google Scholar] [CrossRef]
  19. Li, J.; Feng, Y.; Shao, Y.; Liu, F. IDP-YOLOV9: Improvement of Object Detection Model in Severe Weather Scenarios from Drone Perspective. Appl. Sci. 2024, 14, 5277. [Google Scholar] [CrossRef]
  20. Wan, L.; Li, Z.; Zhang, C.; Chen, G.; Zhao, P.; Wu, K. Algorithm Improvement for Mobile Event Detection with Intelligent Tunnel Robots. Big Data Cogn. Comput. 2024, 8, 147. [Google Scholar] [CrossRef]
  21. Xu, W.; Zhu, D.; Deng, R.; Yung, K.; Ip, A.W. Violence-YOLO: Enhanced GELAN Algorithm for Violence Detection. Appl. Sci. 2024, 14, 6712. [Google Scholar] [CrossRef]
  22. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J.; Ding, G. Yolov10: Real-time end-to-end object detection. arXiv 2024, arXiv:2405.14458. [Google Scholar]
  23. Hussain, M.; Khanam, R. In-Depth Review of YOLOv1 to YOLOv10 Variants for Enhanced Photovoltaic Defect Detection. Solar 2024, 4, 351–386. [Google Scholar] [CrossRef]
  24. Tan, L.; Liu, S.; Gao, J.; Liu, X.; Chu, L.; Jiang, H. Enhanced self-checkout system for retail based on improved YOLOv10. J. Imaging 2024, 10, 248. [Google Scholar] [CrossRef]
  25. Qiu, X.; Chen, Y.; Cai, W.; Niu, M.; Li, J. LD-YOLOv10: A Lightweight Target Detection Algorithm for Drone Scenarios Based on YOLOv10. Electronics 2024, 13, 3269. [Google Scholar] [CrossRef]
  26. Liu, W.; Wang, S.; Gao, X.; Yang, H. A Tomato Recognition and Rapid Sorting System Based on Improved YOLOv10. Machines 2024, 12, 689. [Google Scholar] [CrossRef]
  27. Zhang, C.; Peng, N.; Yan, J.; Wang, L.; Chen, Y.; Zhou, Z.; Zhu, Y. A Novel YOLOv10-DECA Model for Real-Time Detection of Concrete Cracks. Buildings 2024, 14, 3230. [Google Scholar] [CrossRef]
  28. Ali, M.L.; Zhang, Z. The YOLO Framework: A Comprehensive Review of Evolution, Applications, and Benchmarks in Object Detection. Comput. Sci. Math. 2024, 13, 336. [Google Scholar] [CrossRef]
Figure 1. Architectural overview of the YOLOv11 model (v11 innovations are marked with *).
Figure 1. Architectural overview of the YOLOv11 model (v11 innovations are marked with *).
Processes 12 02892 g001
Figure 2. Digitization environment named Prototype P2-24.
Figure 2. Digitization environment named Prototype P2-24.
Processes 12 02892 g002
Figure 3. Example of an enlarged annotated image of grain (left) and flesh (right) leather sides showing a defect named grubs (larval damage).
Figure 3. Example of an enlarged annotated image of grain (left) and flesh (right) leather sides showing a defect named grubs (larval damage).
Processes 12 02892 g003
Figure 4. Example of an enlarged annotated image of grain (left) and flesh (right) leather sides showing a defect named suckout (cut damage).
Figure 4. Example of an enlarged annotated image of grain (left) and flesh (right) leather sides showing a defect named suckout (cut damage).
Processes 12 02892 g004
Figure 5. Normalized confusion matrices for the grain (left) and flesh (right) side.
Figure 5. Normalized confusion matrices for the grain (left) and flesh (right) side.
Processes 12 02892 g005
Figure 6. Bar chart of prediction confidences.
Figure 6. Bar chart of prediction confidences.
Processes 12 02892 g006
Figure 7. Workflow of prediction confidences.
Figure 7. Workflow of prediction confidences.
Processes 12 02892 g007
Table 1. Evolution of YOLO models in 2024.
Table 1. Evolution of YOLO models in 2024.
ModelBackboneHeadLabel Allocation
YOLOv9CSPDarkNet53conv, anchor boxes, multi-label classification, auto-learning bounding boxProgrammable gradient information (PGI)
YOLOv10CSPDarkNet53 or customconv, decoupled detection head, anchor-free, multi-label classificationConsistent dual assignments for NMS-free training
YOLOv11Ultralytics backboneMulti-task support for detection, segmentation, classification, pose estimation, OBBOptimized task-aligner assigner
Table 2. Algorithms used for calculating the cost function.
Table 2. Algorithms used for calculating the cost function.
ModelConfidence Error LossBox Regression LossClassification Loss
YOLOv9BCEPGIoU (Programmable Gradient IoU)BCE
YOLOv10BCEDIoU (Dual IoU)BCE
YOLOv11BCECIoU + DFLBCE
Table 3. Performance results.
Table 3. Performance results.
Sample Size of LeatherAverage Accuracy Detection (%)Average Accuracy Classification (%)
GrubsSuckoutGrubsSuckoutGrubsSuckout
Grain side30030085.887.166.878.2
Flesh side30030093.591.898.297.6
Table 4. Algorithms used for calculating the cost function.
Table 4. Algorithms used for calculating the cost function.
MetricGrain Side (%)Flesh Side (%)
Accuracy8593
F1-Score8693
Precision8792
Recall8494
Suckout Accuracy8792
Grubs Accuracy8594
Table 5. Performance results.
Table 5. Performance results.
GrubsSuckout
Mean Accuracy (%)STD (%)Mean Accuracy (%)STD (%)
Grain side68.445.8378.285.1
Flesh side97.191.5896.51.44
Table 6. Relevant use of YOLO model series for detection and classification of industrial leather defects on finished leather.
Table 6. Relevant use of YOLO model series for detection and classification of industrial leather defects on finished leather.
AuthorSample SizeDefectsLeather SideYOLO
Model
mAP (%)
Wang, M. et al. [7]6288Bubble, dent, broken glueGrainYOLOv994.7
Andrzej Wróbel and Piotr Szymczyk [9]400General defects (not categorized)GrainYOLOv595
Chen, Z. et al. [6]2855Cavity, pinhole, scratch, rotten surface, growth line, healing wound, crease, bacterial woundGrainYOLOv5, YOLOv6, YOLOv7, YOLOv867.2 (YOLOv5), 61 (YOLOv6), 67.01 (YOLOv7), 62.98 (YOLOv8)
Thangakumar, J., et al. [10]Not specifiedVarious leather defectsNot specifiedYOLOv892
The proposed model in this paper 1200Grubs (larval damage) and suckout (cut damage)GrainYOLOv1173.36
The proposed model in this paper 1200Grubs (larval damage) and suckout (cut damage)FleshYOLOv1196.85
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Banduka, N.; Tomić, K.; Živadinović, J.; Mladineo, M. Automated Dual-Side Leather Defect Detection and Classification Using YOLOv11: A Case Study in the Finished Leather Industry. Processes 2024, 12, 2892. https://doi.org/10.3390/pr12122892

AMA Style

Banduka N, Tomić K, Živadinović J, Mladineo M. Automated Dual-Side Leather Defect Detection and Classification Using YOLOv11: A Case Study in the Finished Leather Industry. Processes. 2024; 12(12):2892. https://doi.org/10.3390/pr12122892

Chicago/Turabian Style

Banduka, Nikola, Katarina Tomić, Jovan Živadinović, and Marko Mladineo. 2024. "Automated Dual-Side Leather Defect Detection and Classification Using YOLOv11: A Case Study in the Finished Leather Industry" Processes 12, no. 12: 2892. https://doi.org/10.3390/pr12122892

APA Style

Banduka, N., Tomić, K., Živadinović, J., & Mladineo, M. (2024). Automated Dual-Side Leather Defect Detection and Classification Using YOLOv11: A Case Study in the Finished Leather Industry. Processes, 12(12), 2892. https://doi.org/10.3390/pr12122892

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop