Next Article in Journal
The Improved Cytotoxic Capacity of Functionalized Nanodiamonds with Metformin in Breast and Ovarian Cancer Cell Lines
Next Article in Special Issue
Application of Intelligent Control in Chromatography Separation Process
Previous Article in Journal
Exploring Bayesian Optimization for Photocatalytic Reduction of CO2
Previous Article in Special Issue
Detection of Large Foreign Objects on Coal Mine Belt Conveyor Based on Improved
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Timestep-Adaptive-Diffusion-Model-Oriented Unsupervised Detection Method for Fabric Surface Defects

College of Communication and Information Engineering, Xi’an University of Science and Technology, Xi’an 710054, China
*
Author to whom correspondence should be addressed.
Processes 2023, 11(9), 2615; https://doi.org/10.3390/pr11092615
Submission received: 3 August 2023 / Revised: 25 August 2023 / Accepted: 30 August 2023 / Published: 1 September 2023

Abstract

:
Defect detection is crucial in quality control for fabric production. Deep-learning-based unsupervised reconstruction methods have been recognized universally to address the scarcity of fabric defect samples, high costs of labeling, and insufficient prior knowledge. However, these methods are subject to several weaknesses in reconstructing defect images into defect-free images with high quality, like image blurring, defect residue, and texture inconsistency, resulting in false detection and missed detection. Therefore, this article proposes an unsupervised detection method for fabric surface defects oriented to the timestep adaptive diffusion model. Firstly, the Simplex Noise–Denoising Diffusion Probabilistic Model (SN-DDPM) is constructed to recursively optimize the distribution of the posterior latent vector, thus gradually approaching the probability distribution of surface features of the defect-free samples through multiple iterative diffusions. Meanwhile, the timestep adaptive module is utilized to dynamically adjust the optimal timestep, enabling the model to flexibly adapt to different data distributions. During the detection, the SN-DDPM is employed to reconstruct the defect images into defect-free images, and image differentiation, frequency-tuned salient detection (FTSD), and threshold binarization are utilized to segment the defects. The results reveal that compared with the other seven unsupervised detection methods, the proposed method exhibits higher F1 and IoU values, which are increased by at least 5.42% and 7.61%, respectively, demonstrating that the proposed method is effective and accurate.

1. Introduction

Fabric defect detection plays a key role in controlling textile quality. Fabric defects may influence the appearance of the product, resulting in performance degradation or even functional failure. Punctual detection and repair of fabric defects can lower the defective rate and scrapped quantity, reduce waste and repeated production costs, ensure the qualification of final products, and improve customer satisfaction and brand reputation [1]. However, manual detection currently still prevails in many enterprises, which not only places high requirements on the technical qualifications of inspectors but also leads to a heavy burden of labor costs. Consequently, promoting a highly precise and efficient fabric defect detection system is extremely significant in improving product quality, ensuring the smooth running of production machines, and effectively lowering labor costs [2].
Machine vision has been highly recognized and is gradually replacing manual vision, becoming an important application in fabric defect detection. Machine vision detection, a traditional processing method for images, can extract low-level features of images by obtaining prior knowledge of defect features, thus identifying and classifying the defects [3]. Under the premise of ensuring accuracy, machine vision detection can realize automation and intelligence while imposing high requirements for camera performance and light source environment [4]. Fabric defects are diversified in form and complicated in texture, as shown in Figure 1. Notably, small defects, as displayed in Figure 1b, occupy fewer pixels and have little impact on the overall structure or pattern of the image, increasing the difficulty of distinguishing them from the surrounding texture or pattern. Owning to the above factors, machine vision fails to recognize defects in specific textured fabrics, which is only one of its application restrictions.
Deep learning technologies break through the deficiencies of traditional machine learning technologies. The deep-learning-based supervised detection method [5,6], which automatically extracts the features of the detected object, has shown significant effectiveness in image classification, which further supports its application in surface defect detection. Notably, the deep-learning-based supervised detection method possesses high performance but still requires lots of annotated data to train the model and a certain number of defect samples as references. Actually, collecting and annotating numerous defect sample data is challenging and even impractical [7]. Therefore, some scholars [8,9,10] have comprehensively and extensively studied deep-learning-based unsupervised detection methods to detect surface defects. A prevailing method [11,12] is to obtain the reconstruction models with positive product features by learning defect-free samples, then reconstructing defect images into defect-free images by utilizing the trained reconstruction models, and positioning defects by comparing the differences before and after reconstruction.
The above method is superior because it does not require obtaining the defect type in advance and free from labeling of the sample defects. Currently, the unsupervised defect detection models primarily include the generative adversarial network (GAN) [13] and the automatic encoder (AE) [14]. Nevertheless, these models face challenges in reconstructing defect images into defect-free images in a high-quality manner [15], causing lower accuracy in defect detection. GAN is composed of manually designed generators and discriminators and focuses on solving potential gradient vanishing or explosions, increasing its training difficulty. AE maps high-dimensional feature images to low-dimensional vector representations, which leads to pixel merging, resulting in blurry reconstructed images. In addition, the fabric surface texture features are distributed in a non-periodic manner, and such decentralized data are extremely likely to generate a highly similar image to the original one, resulting in residual defects in the reconstructed image. During the post-processing, the fixed threshold segmentation makes it hard to distinguish defects from reconstruction differences, especially for small and low-contrast defects, making it difficult to accurately position defects. The denoising diffusion probabilistic model [16], as a new generative model, is of higher stability and controllability, and can effectively solve the saddle point by minimizing the convex cross-entropy loss [17]. In consideration of the above contents, a timestep-adaption-diffusion-model-oriented unsupervised detection method for fabric surface defects is proposed in this article. By recursively optimizing the distribution of the posterior latent vector and fitting a distribution that is closer to the real one, it effectively solves the poor model reconstruction mentioned above. First, regarding the low accuracy of GAN-based and AE-based methods in repairing defect images, the simplex noise [18]–denoising diffusion probabilistic model (SN-DDPM) is proposed to control the diffusion to repair defect images and keep the authenticity and interpretability of the image. Secondly, targeting inefficient high-quality reconstruction and no appropriate timestep for the diffusion model, the structural similarity index (SSIM) [19] and mean squared error (MSE) are employed as the guided timestep adaption modules, aiming at the optimal step size of the SN-DDPM and high-quality reconstruction. Additionally, an effective defect segmentation algorithm that utilizes image difference and FTSD [20] is employed to highlight the morphological features of defects. Furthermore, the adaption threshold binarization and closed operations are adopted to segment the defect precisely and improve the detection accuracy.
In summary, the contributions of this article may be summarized as follows:
Applying the SN-DDPM to repair fabric defect images for precise detection;
Employing SSIM and MSE as the parameterized timestep adaption modules to achieve the optimal timestep of the model (DDPM);
Proposing a post-processing method based on FTSD to achieve pixel-level segmentation of defects.
This article is organized structurally as follows.
In Section 2, the unsupervised detection models and DDPM are introduced. Section 3 elaborates on the proposed timestep-adaption-diffusion-model-oriented unsupervised detection method for fabric surface defects. The applied dataset, training details, and evaluation indicators are described in Section 4. The next section summarizes and discusses the experimental results. The last section is the conclusion highlighting the experimental results and prospects for future research directions.

2. Related Works

Recently, unsupervised detection models have been recognized widely due to their outstanding performance. This section will introduce partial models and explore the application prospects of DDPMs in defect detection.

2.1. Unsupervised Detection Method

The current unsupervised detection methods benefit from the support of image reconstruction technologies, which combine higher reconstruction results accuracy with applying other measurement methods (such as potential vector errors) to identify and discover defects. Thus, the quality of reconstructed images directly influences the final detection effect. There are many new technologies and algorithms with excellent performance in enhancing the quality of reconstructed images. Li et al. [21] initially used a denoising automatic encoder to reconstruct fabric defect images, which can categorize defect and defect-free images and segment defects by fixed thresholds. While this method holds great potential for improvement in small defects with low contrast, Zhang et al. [22] put forward a multi-scale U-shaped denoising convolutional autoencoder model and applied it to defect detection. Their experimental results disclosed that this model has good generalization capability. Li et al. [23] constructed a generative network with an encoder–decoder structure and introduced multi-scale channel attention and pixel attention into the encoder network. Meanwhile, they improved the performance of defect detection by applying consistency loss constraints in the reconstruction of pixels, structure, and gradients of the image. In terms of image generation, GAN outperforms the AE-based method [13], increasingly extending its application in derived models for defect detection. Zhang et al. [24] integrated attention mechanisms based on GAN to enhance its feature representation capability for high-quality information, achieving better reconstruction. Wei et al. [25] conducted multi-stage training based on a deep convolutional generation adversarial network (DCGAN) and reduced the interference of defects in the image reconstruction using the linear weighted integration method. They proved that the constructed method outperformed others in terms of f-score measurement. However, the fabric images exhibit abundant texture details, complicated color changes, and irregular patterns, increasing the difficulty for GAN to capture their true distribution. Thus, the generator learns many subtle differences and local structures, extremely weakening the gradient signal and increasing the likelihood of gradient cancellation [26]. In addition, there are multi-scale and multi-level structures in the texture of fabric images, which not only increases the risk of pattern breakdown but also results in the failure to generate real textures and detail changes in the test image.

2.2. Denoising Diffusion Probabilistic Models

DDPMs have shown excellent performance in various applications, such as image synthesis, video generation, and molecular design [27]. DDPMs improve the training stability by systematically adding the noise to the generated data and real data, and sending them to the discriminator for processing [28]. Moreover, DDPMs are capable of effectively solving the instability caused by a mismatch between the distribution of generated data and real data during GAN training. Müller-Franzes et al. [29] verified that the DDPM showed better precision and recall than the GAN-based models during the generation of medical images. Lugmayr et al. [30] developed a DDPM model based on mask repairing that generates an image of the masked area by reasoning the unmasked image information. Their results were more semantic and authentic in contrast to those of other models. Li et al. [31] reconstructed the super-resolution of the image using a DDPM and obtained simpler and more stable properties in the training process compared with the GAN-based model. With only one loss term, the adopted DDPM could complete the training without an additional discriminator module, enhancing the convenience and efficiency of the model in practical application. Additionally, Gedara Chaminda Bandara et al. [32] pre-trained a DDPM to obtain information on unannotated remote sensing images and then utilized the multi-scale feature representation of the diffusion model decoder to train a lightweight change detection classifier. The method was proved to extract key semantics of remote sensing images and produce better feature representations than VAE-based and GAN-based methods.
Thus, it is evident that DDPMs have demonstrated outstanding performance in image generation, not only better preserving the structure and detailed features of images, but also presenting unique advantages in solving the instability during GAN training. Therefore, SN-DDPM is adopted in this article to repair fabric defect images, reconstruct defect images into defect-free images of higher quality, and position the defect areas more accurately.

3. Proposed Methods

Furthermore, this article proposes a timestep-adaptive-diffusion-model-oriented unsupervised detection method for fabric surface defects. This mainly contributes to the feature extraction of good fabric surface and defect detection with SN-DDPM, as illustrated in Figure 2.
(1)
Surface Feature Extraction of Flawless Fabrics
As demonstrated in Figure 2a, the constructed SN-DDPM gradually adds SN to the training data from a certain target distribution X 0 through the forward diffusion process q ( x t | x t 1 ) to obtain the pure noise X T . The model converts X T into X 0 by learning the reverse process p θ ( x t | x t 1 ) , and iteratively outputs the optimal timestep t k through the timestep adaptive module.
(2)
Defect Detection with SN-DDPM
As explicated in Figure 2b, SN is added to a defective image, with a timestep of t k and reconstructed images are obtained after denoising. The grayscale processing and Gaussian filtering are performed on the defect image and reconstructed image, followed by an absolute difference to obtain a residual image. Finally, FTSD is employed to highlight defects, followed by threshold binarization and closed operation to obtain the detection results.

3.1. Surface Feature Extraction of Flawless Fabrics

SN-DDPM is generative and can produce high-quality images by narrowing the distribution of training data after training, so as to capture the essential characteristics of the fabric surface. Figure 2a reveals the two processes during diffusion, forward and reverse. During the forward diffusion, SN is gradually added to the original image X 0 until the image completely turns to pure noise X T . The reverse diffusion transfers the X T to X 0 gradually through training the denoising Unet, and iteratively outputs the optimal timestep t k using the timestamp adaptive module.

3.1.1. Forward Diffusion

In each step of forward diffusion, an SN with a variance of β t is added to X t 1 to generate a new hidden variable X t , with a distribution of q ( x t | x t 1 ) . The specific diffusion process is expressed in Formula (1) below.
q ( x t | x t 1 ) = N ( x t | x t 1 1 β t , β t I )
where N ( x t | x t 1 1 β t , β t I ) represents the normal distribution of mean x t 1 1 β t and covariance β t I that produces x t ; I is the identity matrix, showing that each dimension exhibits the same standard deviation β t , which satisfies β 1 < β 2 < < β T ; q ( x t | x t 1 ) represents the normal distribution, with a mean value of x t 1 1 β t and a variance of β t I . To sample X t at any timestep t [16], α t = 1 β t and α t ¯ = i = 0 T α i are set herein, and the following two formulas can be obtained:
q ( x t | x 0 ) = N ( x t | x 0 α t ¯ , ( 1 α t ¯ ) Ι )
x t = x 0 α t ¯ + ϵ t 1 α t ¯ , ϵ t N ( 0 , Ι )
where ϵ t serves as a learned gradient of the data density. Using the above methods, X t can be acquired at once without sampling the t 1 times, and thus the noisy image X t can be generated faster, further improving the overall diffusion efficiency.

3.1.2. Simplex Noise

SN possesses a higher frequency than Gaussian noise, due to which it shows the complicated details and textures of the fabric surface better. Regarding the coordinate transformation, the simplex coordinate is transformed into the positive super lattice volume space of the corresponding space by skewing, as follows:
x = x + ( x + y ) × F
y = y + ( x + y ) × F
where x and y are the coordinates of the original super lattice body; x and y are the coordinates of the positive super lattice body; and F can be calculated as follows:
F = n + 1 1 n
where n denotes the spatial dimension, which is assigned as 2 in this article for two-dimensional image processing.
Then, the simplex lattice should be determined. The vectors of pixel points are sequenced from largest to smallest to obtain a new vector, and the largest value in the dimension is taken in sequence until three vertices are obtained. According to the obtained vertices, the vertex gradient vector g r a d can be determined, taking the permutation sequence table as the indexing to obtain the vertex gradient value, the same as Perlin noise.
To obtain the distance vector d i s t between pixel points and vertices, the inverse function G in the skewing formula F is applicable, and G can be expressed as Formula (7):
G = 1 1 n + 1 n
Then, d i s t can be expressed as follows:
d i s t = ( x 1 + 2 G , y 1 + 2 G )
Finally, the radial attenuation function is applied to calculate the contribution value of each vertex (Formula (9)), and the values are summed.
( max ( 0 , r 2 | d i s t | 2 ) ) 4 × d o t ( d i s t , g r a d )
where better visual effects can be obtained at r 2 = 0.6 [18].

3.1.3. Reverse Diffusion

Being opposite to the forward diffusion, the reverse process is to remove noise. It can be realized by learning a model p θ by the denoising Unet to approximately simulate the conditional probability q ( x t 1 | x t ) . By parameterizing the mean and variance values, p θ can be obtained:
p θ ( x t 1 | x t ) = N ( x t 1 ; μ θ ( x t , t ) , θ ( x t , t ) )
x 0 is known, so the following expression can be obtained through the Bayesian formula:
q ( x t 1 | x t , x 0 ) = N ( x t 1 ; μ ˜ ( x t , x 0 ) , β ˜ t I )
where β ˜ t = 1 α ¯ t 1 1 α ¯ t β t and μ ˜ t ( x t , x 0 ) = α t ( 1 α ¯ t 1 ) 1 α ¯ t 1 x t + α ¯ t 1 β t 1 α ¯ t x 0 .
By combining with Formula (3), the below expression can be obtained:
μ ˜ t = 1 α t ( x t 1 α t 1 α ¯ t ϵ t )
Therefore, the training model μ θ ( x t , t ) is applied to estimate μ ˜ t , while x t serves as input during the training, so the model can estimate the noise ϵ t .

3.1.4. Denoising Unet

The structure of the denoising Unet is explicated in Figure 3 below:
The denoising Unet possesses an encoder–decoder structure, where the right half is down-sampling and the left half is up-sampling. During the encoder operation, the resolution of the image can be gradually reduced through continuous down-sampling to obtain image information at different scales. Meanwhile, this process can support the model to extract low-level features, such as points, lines, and gradients from the underlying image information, and gradually transition to high-level features, such as contours and more abstract information. In this way, the network fulfills the feature extraction and combination from details to the whole, making the finally obtained features more comprehensive. In addition, using the denoising Unet with the supplemented skip connection structure, the network integrates various feature diagrams of encoder positioning to the channel while the up-sampling is implemented at each level. Furthermore, with the integration of the underlying and apparent features, the network can maintain more high-resolution details contained in high-level feature diagrams, improving the accuracy of image reconstruction.
To support the model in estimating noise ϵ t , the logarithmic likelihood of the predicted distribution of the model should be maximized and the negative logarithmic likelihood should be optimized by using the lower bound of variation. After that, the following formulas can be obtained:
L V L B = L T + L T 1 + + L 0
L T = D K L ( q ( x T | x 0 ) | | p θ ( x T ) )
L t = D K L ( q ( x t | x t 1 , x 0 ) | | p θ ( x t | x t + 1 ) ) ;    1 t T 1
L 0 = log p θ ( x 0 | x 1 )
Since the forward diffusion consists of no learnable parameters, x T is pure noise and can be ignored as a constant, so the loss function can be simplified [16] and calculated:
L t s i m p l e = E x 0 , t , ϵ [ | | ϵ ϵ θ ( a ¯ t x 0 + 1 a ¯ t ϵ , t ) | | 2 ]
Furthermore, the above formula can be utilized to predict the noise ϵ at each time t to allow the model to accurately predict by measuring the difference between ϵ and the real noise ϵ θ till they are the same.

3.1.5. Timestep Adaptive Module

The timestep largely determines the quality of reconstructed images. Relevant experiments and research [27,33] show the single valley function relationship between timestep and the quality of reconstructed images. In this article, the advance–retreat method is adopted as the core concept of the timestep adaptive module. It is a dominant optimization algorithm, with adjusted search steps to achieve the closest optimal solution in accordance with the change in objective function.
The timestep adaptive module is illustrated in Figure 4.
The training data for the feature extraction of defect-free surfaces are defect-free images only. In consideration of this, SSIM and MSE served as evaluation indicators to ensure the reconstruction result is maximally similar to the original image, which can be defined as follows:
𝕃 = ( 1 α ) M S E x , y + α ( 1 S S I M x , y )
where x and y represent the original and reconstructed data, respectively; and α refers to the weight factor to balance the relative importance of the pixels and SSIM. Herein, α = 0.5 is designated to balance the degree of distortion and SSIM of the images. SSIM can be calculated based on the data of brightness, contrast, and structure:
S S I M x , y = ( 2 μ x μ y + c 1 ) ( 2 σ x y + c 2 ) ( μ x 2 + μ y 2 + c 1 ) ( σ x 2 + σ y 2 + c 2 )
In the formula above, μ x and μ y represent the average values of x and y , respectively; σ x and σ y denote the variance of x and y , respectively; σ x y is the covariance of x and y ; c 1 = ( k 1 L ) 2 and c 2 = ( k 2 L ) 2 are constants to maintain stability; and L stands for the dynamic range of pixel values, with k 1 = 0.01 and k 2 = 0.03 [19].
MSE can be calculated with the following formula:
M S E x , y = 1 n i = 1 n ( x y ) 2
After the initial point T 0 and the initial step h = 100 are set, the next detection point can be written as T 100 = T 0 + h , based on which the 𝕃 ( T 0 ) and 𝕃 ( T 100 ) can be calculated and compared. If 𝕃 ( T 0 ) 𝕃 ( T 100 ) , the forward search can be continued, otherwise, the reverse search should be implemented, and the search step size is defined as h 2 . To prevent the search from falling into an infinite loop, a counter is set herein to control the number of searches. It will stop the search and output T k when the number of cycles n reaches 20.

3.2. Defect Detection with SN-DDPM

The SN-DDPM only containing the features of flawless products is obtained, as mentioned in Section 3.1, in which the corresponding optimal timestep is clarified. Here, the defect image is reconstructed, and the defect is located accurately. As shown in Figure 1b, the defect is segmented into main three steps, namely, image reconstruction, image difference, and FTSD, as specified below (Algorithm 1).
Algorithm 1: Defect Detection with SN-DDPM
Input: RGB image X
Output: Defect detection result X r e s u l t
1:   Step 1: Obtaining the optimal timestep and reconstructing the defect image X ^ .
2:   Step 2: Processing the images as follows:
3:   Converting the RGB image to grayscale: X g r a y = 0.2125 X r + 0.7154 X g + 0.0721 X b
4:   Gaussian filter: X g a u s s i a n = 1 2 π σ x σ y exp ( x 2 + y 2 2 σ x σ y )
5:   Step 3: Absolute difference:
6:    Δ X = | X g a u s s i a n X ^ g a u s s i a n |
7:   Step 4: Performing FTSD:
8:   Applying the Gaussian filter to smooth the residual image
9:   Converting the smoothed image to LAB color space
10:    Calculating the average image feature vector
11:    Calculating the pixel vector value
12:    Calculating the saliency image from normalized Euclidean distance
13:  Step 5: Binarization:
14:    Calculating the threshold value: T = μ + σ
15:    Binarizing the saliency image: O = 0 ,    if   p T 255 ,   o t h e r w i s e
16:  Step 6: Closed operation:
17:     X r e s u l t = ( X s a l i e n c y s ) s
Step 1: The defect image is reconstructed using the previously obtained reconstruction model and the optimal timestep. With the defect image serving as input, the optimal timestep controls the SN to generate a noisy image and input it into the reconstruction model to obtain the reconstructed image, which maximally keeps the features of flawless products, with the defects repaired.
Step 2: The grayscale processing and Gaussian filtering are conducted on defect images and reconstructed images, respectively, as expressed in Formulas (21) and (22), respectively:
X g r a y = 0.2125 X r + 0.7154 X g + 0.0721 X b
G ( x , y ) = 1 2 π σ x σ y exp ( x 2 + y 2 2 σ x σ y )
where, X g r a y represents the grayscale image; and X r , X g , and X b are the pixel values of the red, green, and blue (RGB) channels, respectively. Meanwhile, the convolution kernel with the size of 3 × 3 is selected for Gaussian filtering, and σ x and σ y denote the pixel deviations in the x-axis and y-axis directions of the image, respectively.
Step 3: The absolute difference operation (Formula (23) is performed on the test diagram and the reconstructed grayscale Gaussian image to obtain a residual image.
Δ x ( m , n ) = | x ( m , n ) x ^ ( m , n ) |
In the expression above, x ( m , n ) and x ^ ( m , n ) represent the standard image and operated image with the dimension of m × n , respectively, and Δ x ( m , n ) refers to the residual image.
Step 4: The residual image is subjected to FTSD to obtain a saliency image. As demonstrated in Figure 5, the specific process includes smoothing the residual image using a 7 × 7 Gaussian filter to eliminate noise and preserve the overall structure of the image. The image obtained at this time is called a Gaussian image, which is then converted from the RGB color space to the LAB color space to obtain the Lab image. Subsequently, the average pixel values L μ , a μ , and b μ of the L, A, and B channels of the converted Lab image are calculated to obtain the average image feature vector I μ and pixel vector value I ω h c ( x , y ) . After the calculation and normalization of the Euclidean distance of these two vectors, the saliency image is finally obtained.
Step 5: The random noise in the unreal defect area is filtered for more accurate detection. Noise usually obeys a normal distribution, so binarization is employed to segment a gray residual image, with the threshold being defined as follows:
T = μ + σ
where, T is the adaptive threshold, and μ and σ are the mean and standard deviation of the saliency image, respectively. The binarization and segmentation operations are expressed as follows:
p = 0 ,       p T p = 255 ,   p > T
where p represents the pixel value of the residual image; 0 is defined if p is less than or equal to the threshold, otherwise 255 is designated.
Step 6: Finally, through a closed operation, small holes are eliminated, and cracks in the contour line are filled, thus obtaining a complete defect form. The closed operation is expanded and corroded, with the formula as follows:
X r e s u l t = ( X s a l i e n c y s ) s
where X r e s u l t is the image after the closed operation, s represents the structural element, is the expansion operation, and refers to the corrosion operation.
The spatial complexity of the SN-DDPM model is significant, with 239.6 M parameters, each of which is stored using a 32-bit floating point data type. Such a large set of parameters endows SN-DDPM with powerful characterization capabilities, enabling it to extract more subtle patterns and associations, giving it an advantage in complex conditions, large-scale datasets, or high-dimensional data.

4. Experimental Setup

4.1. Datasets

The colored fabrics from the small lattice (SL) dataset from the Yarn-dyed Fabric Image Dataset Version 1 (YDFID-1) [34] were selected, which consisted of 3245 defect-free samples and 254 defect samples. The fabric pattern is primarily that of small lattices, displayed in RGB images of 512 × 512 pixels. To verify the applicability of SN-DDPM on different color fabrics, eight types of typical fabrics with different textures and colors were selected, which are SL1, SL2, SL5, SL8, SL9, SL10, SL11, and SL13. Images of some defect-free and defect samples were selected for comparison, as given in Figure 6. This dataset contains highly complicated defect categories and fabric textures, providing a sound solution for verifying the performance of deep learning models in detecting complicated defects.
Influenced by insufficient data of defect-free samples in the dataset, the sample size was increased using data enhancement methods, which is beneficial for improving the model invariance. A total of 51,888 high-quality images were obtained by rotating the original defect-free images at 90°, 180°, and 270°, as well as flipping the upper, lower, left, and right mirror surfaces. They served as training sets, while the rest of the samples were test sets.

4.2. Training Process

During the model training, the flawless fabric image was adopted only to train fully extracting the characteristics of flawless samples with the principle of unsupervised learning. The model was trained through flawless images, obtaining the feature distribution of these samples. The detection results of the proposed method were compared with those of DCAE [35], DCGAN [36], Recycle-GAN [37], MSCDAE [38], UDCAE [39], VAE-L2SSIM [40], and AFFGAN [12]. All models were trained with a batch size of 8, an epoch of 5000, and a learning rate of 1 × 10−4. Meanwhile, the model loss should be maintained in a stable state. A system equipped with an Intel i9-12900H CPU and Nvidia RTX3070ti GPU was employed to train and test the models in this article.

4.3. Evaluation Method

4.3.1. Evaluation Indicator of Image Reconstruction Results

The peak signal-to-noise ratio (PSNR) [41] and SSIM, two commonly applied indicators to assess image quality, can quantitatively analyze the reconstruction results and evaluate the model’s capability to retain details during image reconstruction objectively and accurately. SSIM was discussed in Section 3.1.5, and PSNR is introduced herein.
By combining Formula (20), the PSNR can be defined as follows:
P S N R = 10 log 10 ( M A X I 2 M S E )
M A X I 2 in the above expression represents the maximum possible pixel value of the image. The image pixels used in the article are represented by 8-bit binary, so its value is 255.
With large PSNR and SSIM values, the reconstruction model can better preserve the details of the original image, the reconstruction results are closer to the original image, and the image quality is higher. Therefore, the larger the values of PSNR and SSIM, the higher the similarity between the reconstructed image and the original image, and the stronger the model’s capability to reconstruct details.

4.3.2. Evaluation Indicator Defect Detection Results

Precision (P), recall (R), accuracy (Acc), F1 value, and intersection over union (IoU), as defined in Formulas (28)–(32), were employed to quantitatively analyze the defect detection results of different models.
P = T P T P + F P × 100 %
R = T P T P + F N × 100 %
A c c = T P + T N T P + F P + T N + F N × 100 %
F 1 = 2 × T P 2 × T P + F P + F N × 100 %
I o U = T P T P + F N + F P × 100 %
The relationships among the four indicators, true positive ( T P ), false positive ( F P ), false negative ( F N ), and true negative ( T N ), are summarized in Figure 7, with gray representing the test result and brown representing the reference value. Actually, T P represents the number of pixels that are successfully detected and confirmed as defect areas; F P represents the number of pixels that are defect areas but erroneously identified as non-defective areas; F N refers to the number of pixels that are non-defect areas but erroneously identified as defective areas; and T N stands for the number of pixels which are successfully detected and confirmed as defect-free areas. P and R denote the precision of the model in predicting whether it is correct or not. The higher the P and R values, the better the performance of the defect detection method. Nevertheless, it is worth noting that there are contradictions between P and R under certain circumstances, increasing the difficulty in acquiring higher values of both, while the F 1 value can better reflect the overall detection performance. In addition, A c c indicates the model’s accuracy in predicting the correct region, and I o U measures the accuracy of the model in judging the defect position. A c c and I o U can reflect whether the model has detected the defect, instead of unilaterally paying attention to the accuracy of defective pixel detection.

5. Experimental Results and Discussion

5.1. Fabric Images Reconstruction Experiments

The reconstruction capability of the unsupervised detection model has a direct effect on the detection results, which is primarily reflected in the repair of the defect area of the detected fabric image and that of the image details of the defect-free area. For comparing the reconstruction capabilities of DCAE, DCGAN, Recycle-GAN, MSCDAE, UDCAE, VAE-L2SSIM, AFFGAN, and SN-DDPM, the fabric samples with different textures, background colors, and types of defects were selected in this study. Figure 8 shows the repairing results of eight models on the image to be tested, in which the sample SL1 contains a large defect area. The DCGAN and VAE-L2SSIM exhibit no remarkable defect areas but fail to visualize details of the defect-free areas. The reconstructed image of DCGAN displays obvious stitching traces, while that of VAE-L2SSIM possesses blurring and texture disorder. The DCAE, MSCDAE, AFFGAN, and UDCAE all show traces of defect areas, of which the reconstructed results of MSCDAE are quite different from the original images. Recycle-GAN cannot effectively reconstruct defect areas into good products, and there is a big difference between non-defect areas before and after reconstruction. Compared with the above six models, SN-DDPM shows higher capability to repair defect areas and displays the non-defect areas almost the same as the original images, which is beneficial to the subsequent defect positioning. The second line shows the sample SL8 with monofilament stripe defects and the construction and repair results of each model. After repair by the DCAE, DCGAN, and VAE-L2SSIM, traces of defect areas are not observed, but the texture information of non-defect areas is lost. Furthermore, block stitching traces are visible in the reconstructed images after repair by DCGAN. MSCDAE retains the defects on the reconstruction diagram, and the reconstruction results of non-defect areas are not as good as expected. Additionally, UDCAE is unable to effectively reconstruct the images and Recycle-GAN fails in and enlarges the defect area. In addition, AFFGAN is capable of effectively reconstructing a defect image into a defect-free image but exhibits a too regular and unnatural texture of the reconstructed image compared with the original image. The above comparison and discussion reveal that SN-DDPM possesses leading advantages in repairing defect areas and maintaining relevant details of non-defect areas. The results of eight models in reconstructing and removing small defects in sample SL9 containing small defects are listed in the third row. It can be observed that most models exhibit the repaired defect areas. Nevertheless, VAE-L2SSIM and DCGAN fail to effectively reconstruct the image, while Recycle-GAN and MSCDAE suffer from blurred reconstructed images and defect retention. Therefore, SN-DDPM has a better visual reconstruction effect of texture details than that of other models, thus reasonably demonstrating competitive reconstruction performance.
Furthermore, PSNR and SSIM are selected to assess the image quality, based on which the reconstruction capabilities of various models are quantitatively compared. There is no defect-free image corresponding to the defect sample in the data set so the defect-free image is used as the reconstruction object. As listed in Table 1, SN-DDPM obtains the best SSIM values, which verifies that the timestep adaptive module constrained by SSIM can effectively improve the reconstruction capability of the model in defect-free areas. In addition, SN-DDPM also obtains the highest PSNR values, indicating that this model has strong a reconstruction capability. Notably, both AFFGAN and SN-DDPM can capture the structural and textural features of fabrics, but the pixels of the reconstructed image using AFFGAN on the sample SL11 are closer to those of the original image. Thus, AFFGAN and SN-DDPM demonstrate close SSIM but greatly different PSNR.
The above results suggest that DCGAN learns the features of defect-free regions, so that it can effectively reconstruct the defect regions. The block reconstruction of the DCGAN model on the image results in the failed connection of the fabric textures in the adjacent grid boundary areas, causing obvious stitching marks in the reconstruction results. The partially reconstructed images by DCAE, MSCDAE, and UDCAE demonstrate observable defects. Due to fewer network layers and smaller receptive domains, DCAE and MSCDAE are applicable for simple mapping transformations on input images only, making it difficult to extract the essential texture information of flawless areas. In comparison with DCAE and MSCDAE, the UDCAE model presents poor connections among image pixels even though the network structure has been deepened. As a result, the deep model compresses the images by force, so that some details are lost. In addition, the overall color scale is similar, so that it is easy to form the color block kernels, as observed in the sample SL8, intensifying the difficulty of accurately positioning the defects. AFFGAN enhances the feature representation capability of defect-free textures based on the attention mechanism, maintaining good reconstruction results. Notably, some reconstruction results still are subjected to a small number of residual defects. As observed in the reconstructed image of SN-DDPM, the defect areas in the color-patterned fabric image are repaired, and features of the defect-free areas are clearly and intuitively displayed. This indicates that SN-DDPM exhibits the best performance in capturing the essential information of color-patterned fabric images. In addition, detail textures are concerned more with introducing a timestamp adaptive module guided by SSIM and MSE, achieving optimal reconstruction and restoration results.

5.2. Defect Detection Experiments

To verify whether the SN-DDPM can accurately locate the defects at the pixel level, relevant tests were performed on fabric samples containing different textures, background colors, and types of defects. The overall results are shown in Figure 9, where the original images, reconstructed images, heat maps, saliency images, final results, and ground truth are displayed in sequence from top to bottom. As shown in the third lines (heat maps), SN-DDPM performs well in the reconstruction of defect images, so the reconstruction error of defect-free areas exerts a small influencing effect on the detection results. In addition, the saliency images in the fourth line suggest that the salient algorithm can accurately segment defects and highlight defect areas, exhibiting obvious effects on small defects with low contrast in samples SL11 and SL13. The final results are extremely close to the ground truth except for the sample SL5. However, because of the similarity between the defect color and the background color, the sample SL5 suffers significant pixel loss in the defect area during the absolute difference process, resulting in incomplete defect morphology in the saliency image. However, this does not influence the capability to determine the shape and location of the defects based on the detection results. Therefore, applying SN-DDPM for defect image reconstruction and combining saliency algorithms for defect detection can achieve excellent results and high reliability in defect positioning.
To evaluate the performance of SN-DDPM objectively and accurately, Table 2 presents the values of evaluation indicators of DCAE, DCGAN, Recycle GAN, MSCDAE, UDCAE, VAE-L2SSIM, AFFGAN, and SN-DDPM in each test set. In terms of average value, SN-DDPM achieves the optimal results in all indicators, of which the F1 and IoU values typically reflect an increase in model performance by at least 5.42% and 7.61%, respectively. The table reveals that UDCAE is comparable to SN-DDPM in terms of P and Acc values but exhibits a significantly lower F1 value. This is attributed to the fact that compared to SN-DDPM, UDCAE shows poorer detection results in the ground truth, which can be expressed as a larger FN in the confusion matrix diagram. Due to the higher number of pixels in the defect-free region, AFFGAN maintains a higher Acc value but a higher FP rate in recognizing the defects, resulting in lower P and F1 values. The P and R values of SN-DDPM are basically complementary, indicating that when the P value is positive, the R value is negative, and vice versa. Compared to other models, the difference between the lower and higher performance values of SN-DDPM is not significant, which is more intuitive in the F1 value. Such a result suggests that SN-DDPM improves the F1 value significantly, demonstrating its superiority in overall detection performance. Meanwhile, SN-DDPM demonstrates its advantage in Acc value, although it is not so significant in comparison to other models. The primary reason is that the number of pixels in the defect areas used is much smaller than that in the defect-free area, rendering Acc unable to objectively describe the quality of the detection results. Referring to the value of IoU, SN-DDPM is highly competitive, with the highest IoU values, demonstrating its accuracy and reliability in defect prediction.

5.3. Ablation Study

The effectiveness of the timestep adaptive module is based on the condition that the evaluation indicator 𝕃 is a single valley function because it is difficult to solve the optimal timestep in case L is a multi-valley function. The defect-free samples SL1, SL2, SL8, SL9, and SL10 are selected as the experimental subjects to verify whether the evaluation indicator 𝕃 is a single valley function. Specifically, 𝕃 is calculated every 10 steps, with a total timestep of 1000. The experimental results are illustrated in Figure 10.
The figure discloses that the evaluation indicators of various samples show the characteristics of a unimodal function. Some fluctuations in the waveform activate within the error range and are free from significant influence on the overall trend. In addition, a lower value of L reflects a better reconstruction effect. Based on the optimal step sizes corresponding to various types of samples, it can be observed that all results are within the range of 0–1000 and are different in each type of sample. In this case, it is impossible to represent the optimal step sizes with fixed values, which further proves the effectiveness and applicability of the timestep adaptive module.
To further verify the validity of α = 0.5 in Formula (18), α is assigned different values based on the above experiments, and the optimal timestep is obtained by the timestep adaptive module. Meanwhile, the F1 and IoU values of the final result are calculated as the evaluation criteria, as listed in Table 3.
As observed in Table 3, the results of F1 and IoU values are not satisfactory at α = 0.1 and α = 0.9 but are the best at α = 0.5 . These outcomes suggest that α = 0.5 balances the degree of distortion and structural similarity of the image. Meanwhile, the experimental results in this article confirm that α = 0.5 is effective.

5.4. Model Failure Experiment

SN-DDPM exhibits strong reconstruction performance but has insufficient robustness during the experimental process. Specifically, the reconstruction results are chaotic when there is an overwhelming number of sample types during the model training. To ensure the diversity of training samples, all fabric samples (19 types in total) in the dataset are selected as the training set, with equal numbers of each sample type, while some defect images serve as the test set. The experimental results are summarized as follows (Figure 11):
As shown in Figure 11, only the samples SL1 and SL5 can be reconstructed normally, while the reconstruction results of the remaining samples show significant pixel deviations. The texture of the defect-free area in sample SL7 is matched to that of the original image, while its color is closer to that of the sample SL2. By contrast, the samples SL2, SL3, and SL9 not only showed significant pixel deviations but also failed to reconstruct the texture details normally. Therefore, it can be concluded that SN-DDPM exhibits good detection only by training a small number of types of samples, while it has to be retrained or structurally adjusted based on different data distributions and its features adapted to other types of defect detection tasks. Consequently, it is not feasible as a unified model for all defect detection tasks.
Relevant research [42,43] shows that the major challenge faced by the diffusion model is the instability and inconsistency of the output, failing to accurately associate attributes with its objects (e.g., color). In this article, SN-DDPM demonstrates weak reconstruction results in the presence of multiple sample datasets, showing that its generalization capability still needs to be studied and improved. Therefore, the following two methods are proposed: (1) Using the mask-based SN-DDPM, the image can be reconstructed well by masking the suspicious defect areas and utilizing the defect-free features around the mask. Meanwhile, it can preserve the semantic information of the defect-free areas in the original image. Thus, the repaired image will better match the viewer’s understanding of the scene and objects, effectively avoiding the chaotic reconstruction caused by excessive training features in the model. (2) Attention mechanisms can be introduced to help the model better focus on important regions and features in the image, thereby improving the accuracy and consistency of the generated output. Through learning attention weights, the model can better comprehend the attributes and associations of objects, thereby outputting more accurate results.

6. Conclusions

In this article, a timestep-adaptive-diffusion-model-oriented unsupervised detection method is applied to the detection of fabric surface defects. It only employs the fabric defect-free samples to train the model and takes SSIM and MSE as the guided timestep adaptive modules to obtain the optimal timestep. During the detection, SN-DDPM with the optimal timestep is employed to reconstruct the defect image into a defect-free image. After that, the residual images before and after reconstruction are processed through FTSD to highlight the defect area. Finally, a discrimination threshold is utilized to segment the defect. Experimental results based on public datasets reveal that SN-DDPM can more effectively extract the essential characteristics of fabrics in contrast to other unsupervised reconstruction models. Meanwhile, its reconstruction results are closer to the true feature distribution, effectively solving the blurring, defect residue, or texture inconsistency in the reconstruction results obtained by other models. These findings suggest that the SN-DDPM demonstrates superior reconstruction capability and outstanding detection performance. In addition, the instability and inconsistency of SN-DDPM in diverse sample datasets are discussed and feasible effective solutions are recommended to help develop more reliable and powerful models.

Author Contributions

All authors participated in some part of the work for this article. In the investigation, Z.J. and S.T. proposed the idea and conceived the design; Z.J. carried out the simulation and wrote the original draft; Y.Z. and J.Y. analyzed and discussed the results; H.L. and J.L. reviewed and edited the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Key Research and Development Program of China (2018YFC0808300), Shaanxi Science and Technology Plan Key Industry Innovation Chain (Group)—Project in Industrial Field (2020ZDLGY15-07).

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ngan, H.Y.; Pang, G.K.; Yung, N.H.J.I. Automated fabric defect detection—A review. Image Vis. Comput. 2011, 29, 442–458. [Google Scholar] [CrossRef]
  2. Wong, W.; Jiang, J. Computer vision techniques for detecting fabric defects. In Applications of Computer Vision in Fashion and Textiles; Elsevier: Amsterdam, The Netherlands, 2018; pp. 47–60. [Google Scholar]
  3. Rasheed, A.; Zafar, B.; Rasheed, A.; Ali, N.; Sajid, M.; Dar, S.H.; Habib, U.; Shehryar, T.; Mahmood, M.T. Fabric Defect Detection Using Computer Vision Techniques: A Comprehensive Review. Math. Probl. Eng. 2020, 2020, 8189403. [Google Scholar] [CrossRef]
  4. Xiang, J.; Pan, R.; Gao, W. Online Detection of Fabric Defects Based on Improved CenterNet with Deformable Convolution. Sensors 2022, 22, 4718. [Google Scholar] [CrossRef] [PubMed]
  5. Chen, M.; Yu, L.; Zhi, C.; Sun, R.; Zhu, S.; Gao, Z.; Ke, Z.; Zhu, M.; Zhang, Y. Improved faster R-CNN for fabric defect detection based on Gabor filter with Genetic Algorithm optimization. Comput. Ind. 2022, 134, 103551. [Google Scholar] [CrossRef]
  6. Jing, J.F.; Zhuo, D.; Zhang, H.H.; Liang, Y.; Zheng, M. Fabric defect detection using the improved YOLOv3 model. J. Eng. Fiber. Fabr. 2020, 15, 1558925020908268. [Google Scholar] [CrossRef]
  7. Ren, Z.; Fang, F.; Yan, N.; Wu, Y. State of the Art in Defect Detection Based on Machine Vision. Int. J. Precis. Eng. Manuf.-Green Technol. 2021, 9, 661–691. [Google Scholar] [CrossRef]
  8. Szarski, M.; Chauhan, S. An unsupervised defect detection model for a dry carbon fiber textile. J. Intell. Manuf. 2022, 33, 2075–2092. [Google Scholar] [CrossRef]
  9. Wang, Y.Y.; Song, K.C.; Niu, M.H.; Bao, Y.Q.; Dong, H.W.; Yan, Y.H. Unsupervised defect detection with patch-aware mutual reasoning network in image data. Automat. Constr. 2022, 142, 104472. [Google Scholar] [CrossRef]
  10. Zhang, N.; Zhong, Y.; Dian, S.J.O. Rethinking unsupervised texture defect detection using PCA. Opt. Laser. Eng. 2023, 163, 107470. [Google Scholar] [CrossRef]
  11. Zhang, H.W.; Chen, X.W.; Lu, S.; Yao, L.; Chen, X. A contrastive learning-based attention generative adversarial network for defect detection in colour-patterned fabric. Color. Technol. 2023, 139, 248–264. [Google Scholar] [CrossRef]
  12. Zhang, H.; Qiao, G.; Lu, S.; Yao, L.; Chen, X. Attention-based Feature Fusion Generative Adversarial Network for yarn-dyed fabric defect detection. Text. Res. J. 2022, 93, 1178–1195. [Google Scholar] [CrossRef]
  13. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. Commun. ACM 2020, 63, 139–144. [Google Scholar] [CrossRef]
  14. Zhai, J.; Zhang, S.; Chen, J.; He, Q. Autoencoder and its various variants. In Proceedings of the 2018 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Miyazaki, Japan, 7–10 October 2018; pp. 415–419. [Google Scholar]
  15. Kahraman, Y.; Durmusoglu, A. Deep learning-based fabric defect detection: A review. Text. Res. J. 2023, 93, 1485–1503. [Google Scholar] [CrossRef]
  16. Ho, J.; Jain, A.; Abbeel, P. Denoising diffusion probabilistic models. Adv. Neural Inf. Process. Syst. 2020, 33, 6840–6851. [Google Scholar]
  17. Bansal, A.; Borgnia, E.; Chu, H.-M.; Li, J.S.; Kazemi, H.; Huang, F.; Goldblum, M.; Geiping, J.; Goldstein, T.J. Cold diffusion: Inverting arbitrary image transforms without noise. arXiv 2022, arXiv:2208.09392. [Google Scholar]
  18. Perlin, K. Improving noise. In Proceedings of the 29th Annual Conference on Computer Graphics and Interactive Techniques, New York, NY, USA, 23–26 July 2002; pp. 681–682. [Google Scholar]
  19. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
  20. Achanta, R.; Hemami, S.; Estrada, F.; Susstrunk, S. Frequency-tuned salient region detection. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 1597–1604. [Google Scholar]
  21. Li, Y.; Zhao, W.; Pan, J. Deformable Patterned Fabric Defect Detection with Fisher Criterion-Based Deep Learning. IEEE Trans. Autom. Sci. Eng. 2017, 14, 1256–1264. [Google Scholar] [CrossRef]
  22. Zhang, H.W.; Liu, S.T.; Tan, Q.L.; Lu, S.; Yao, L.; Ge, Z.Q. Colour-patterned fabric defect detection based on an unsupervised multi-scale U-shaped denoising convolutional autoencoder model. Color. Technol. 2022, 138, 522–537. [Google Scholar] [CrossRef]
  23. Li, X.; Zheng, Y.; Chen, B.; Zheng, E. Dual Attention-Based Industrial Surface Defect Detection with Consistency Loss. Sensors 2022, 22, 5141. [Google Scholar] [CrossRef]
  24. Zhang, H.W.; Qiao, G.H.; Liu, S.T.; Lyu, Y.T.; Yao, L.; Ge, Z.Q. Attention-based vector quantisation variational autoencoder for colour-patterned fabrics defect detection. Color. Technol. 2023, 139, 223–238. [Google Scholar] [CrossRef]
  25. Wei, C.; Liang, J.; Liu, H.; Hou, Z.; Huan, Z. Multi-stage unsupervised fabric defect detection based on DCGAN. Visual Comput. 2022, 1–17. [Google Scholar] [CrossRef]
  26. Zhang, G.; Cui, K.; Hung, T.-Y.; Lu, S. Defect-GAN: High-fidelity defect synthesis for automated defect inspection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, Hawaii, 3–7 January 2021; pp. 2524–2534. [Google Scholar]
  27. Yang, L.; Zhang, Z.; Song, Y.; Hong, S.; Xu, R.; Zhao, Y.; Shao, Y.; Zhang, W.; Cui, B.; Yang, M.-H. Diffusion models: A comprehensive survey of methods and applications. arXiv 2022, arXiv:2209.00796. [Google Scholar]
  28. Croitoru, F.A.; Hondru, V.; Ionescu, R.T.; Shah, M. Diffusion Models in Vision: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 10850–10869. [Google Scholar] [CrossRef]
  29. Müller-Franzes, G.; Niehues, J.M.; Khader, F.; Arasteh, S.T.; Haarburger, C.; Kuhl, C.; Wang, T.; Han, T.; Nebelung, S.; Kather, J.N.J. Diffusion Probabilistic Models beat GANs on Medical Images. arXiv 2022, arXiv:2212.07501. [Google Scholar]
  30. Lugmayr, A.; Danelljan, M.; Romero, A.; Yu, F.; Timofte, R.; Van Gool, L. Repaint: Inpainting using denoising diffusion probabilistic models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11461–11471. [Google Scholar]
  31. Li, H.; Yang, Y.; Chang, M.; Chen, S.; Feng, H.; Xu, Z.; Li, Q.; Chen, Y. SRDiff: Single image super-resolution with diffusion probabilistic models. Neurocomputing 2022, 479, 47–59. [Google Scholar] [CrossRef]
  32. Gedara Chaminda Bandara, W.; Gopalakrishnan Nair, N.; Patel, V.M.J. Remote Sensing Change Detection (Segmentation) using Denoising Diffusion Probabilistic Models. arXiv 2022, arXiv:2206.11892. [Google Scholar]
  33. Graham, M.S.; Pinaya, W.H.; Tudosiu, P.-D.; Nachev, P.; Ourselin, S.; Cardoso, J. Denoising diffusion models for out-of-distribution detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 2947–2956. [Google Scholar]
  34. Zhang, H. Yarn-Dyed Fabric Image Dataset Version 1. 2021. Available online: http://github.com/ZHW-AI/YDFID-1 (accessed on 2 August 2023).
  35. Zhang, H.; Tang, W.; Zhang, L.; Li, P.; Gu, D. Defect detection of yarn-dyed shirts based on denoising convolutional self-encoder. In Proceedings of the 2019 IEEE 8th Data Driven Control and Learning Systems Conference (DDCLS), Dali, China, 24–27 May 2019; pp. 1263–1268. [Google Scholar]
  36. Hu, G.; Huang, J.; Wang, Q.; Li, J.; Xu, Z.; Huang, X. Unsupervised fabric defect detection based on a deep convolutional generative adversarial network. Text. Res. J. 2019, 90, 247–270. [Google Scholar] [CrossRef]
  37. Bansal, A.; Ma, S.; Ramanan, D.; Sheikh, Y. Recycle-gan: Unsupervised video retargeting. In Proceedings of the European conference on computer vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 119–135. [Google Scholar]
  38. Mei, S.; Yang, H.; Yin, Z. An Unsupervised-Learning-Based Approach for Automated Defect Inspection on Textured Surfaces. IEEE Trans. Instrum. Meas. 2018, 67, 1266–1277. [Google Scholar] [CrossRef]
  39. Zhang, H.; Tan, Q.; Lu, S. Yarn-dyed shirt piece defect detection based on an unsupervised reconstruction model of the U-shaped denoising convolutional auto-encoder. J. Xidian Univ. 2021, 48, 123–130. [Google Scholar]
  40. Wei, W.; Deng, D.; Zeng, L.; Zhang, C. Real-time implementation of fabric defect detection based on variational automatic encoder with structure similarity. J. Real-Time Image Process. 2020, 18, 807–823. [Google Scholar] [CrossRef]
  41. Huynh-Thu, Q.; Ghanbari, M. Scope of validity of PSNR in image/video quality assessment. Electron. Lett. 2008, 44, 800–801. [Google Scholar] [CrossRef]
  42. Du, C.; Li, Y.; Qiu, Z.; Xu, C. Stable Diffusion is Unstable. arXiv 2023, arXiv:2306.02583. [Google Scholar]
  43. Chefer, H.; Alaluf, Y.; Vinker, Y.; Wolf, L.; Cohen-Or, D. Attend-and-excite: Attention-based semantic guidance for text-to-image diffusion models. arXiv 2023, arXiv:2301.13826. [Google Scholar] [CrossRef]
Figure 1. Surface defects of complicated texture fabrics: (a) linear defect; (b) spot defect; and (c) planar defect.
Figure 1. Surface defects of complicated texture fabrics: (a) linear defect; (b) spot defect; and (c) planar defect.
Processes 11 02615 g001
Figure 2. Framework of the timestep-adaptive-diffusion-model-oriented unsupervised detection method for fabric surface defects.
Figure 2. Framework of the timestep-adaptive-diffusion-model-oriented unsupervised detection method for fabric surface defects.
Processes 11 02615 g002
Figure 3. Structure of the denoising Unet.
Figure 3. Structure of the denoising Unet.
Processes 11 02615 g003
Figure 4. Example of a timestep adaptive diagram, with the advance–retreat method applied to solve the closest solution to the optimal solution T k .
Figure 4. Example of a timestep adaptive diagram, with the advance–retreat method applied to solve the closest solution to the optimal solution T k .
Processes 11 02615 g004
Figure 5. Process of the FTSD method to remove noise and highlight defects.
Figure 5. Process of the FTSD method to remove noise and highlight defects.
Processes 11 02615 g005
Figure 6. Images for defect-free and defect samples of colored fabrics from the SL dataset.
Figure 6. Images for defect-free and defect samples of colored fabrics from the SL dataset.
Processes 11 02615 g006
Figure 7. Confusion matrix diagram to show the relationship among TP, FP, TN, and FN.
Figure 7. Confusion matrix diagram to show the relationship among TP, FP, TN, and FN.
Processes 11 02615 g007
Figure 8. Qualitative comparison of the reconstruction results of different models, with the red boxes marking the defect areas.
Figure 8. Qualitative comparison of the reconstruction results of different models, with the red boxes marking the defect areas.
Processes 11 02615 g008
Figure 9. The defect detection effect of SN-DDPM.
Figure 9. The defect detection effect of SN-DDPM.
Processes 11 02615 g009
Figure 10. Relationship between timestep and evaluation score for SL1, SL2, SL8, SL9, and SL10.
Figure 10. Relationship between timestep and evaluation score for SL1, SL2, SL8, SL9, and SL10.
Processes 11 02615 g010
Figure 11. Model failure experiment: 19 types of defect-free fabric samples are trained and partial defect images are tested.
Figure 11. Model failure experiment: 19 types of defect-free fabric samples are trained and partial defect images are tested.
Processes 11 02615 g011
Table 1. PSNR and SSIM values in the reconstructed images of 7 models.
Table 1. PSNR and SSIM values in the reconstructed images of 7 models.
IndexMethodSL1SL2SL5SL8SL9SL10SL11SL13Average Value
SSIMDCAE0.95840.80350.82640.93410.69420.89070.75300.88860.8436
DCGAN0.54770.16820.53920.09860.74620.38400.35680.04600.3608
Recycle-GAN0.01510.27210.13970.36430.07870.34720.14950.03300.1750
MSCDAE0.40840.42380.15860.32860.76450.39880.46720.43850.4236
UDCAE0.95580.79560.82340.82670.85640.84620.78690.70930.8250
VAE-L2SSIM0.57030.16990.32950.38850.46950.46290.54280.39210.4157
AFFGAN0.97480.85420.85940.86930.91350.94910.94460.91360.9098
SN-DDPM0.96460.90290.86970.89380.93960.90770.94810.95910.9232
PSNR (dB)DCAE26.264126.743827.413827.110824.903728.380527.916728.673127.1758
DCGAN14.811612.865714.58698.187614.637913.848913.284612.201013.0530
Recycle-GAN11.767417.564411.756418.372318.916819.008214.673611.037915.3871
MSCDAE19.960421.756412.469214.899024.951323.143721.200425.133320.4392
UDCAE25.349625.643225.389122.867526.320425.626721.976925.059624.7791
VAE-L2SSIM20.848510.034812.867624.673818.785722.976720.647226.123519.6197
AFFGAN28.156728.994727.094727.887728.925429.318930.019227.719128.5146
SN-DDPM28.146429.840025.458927.895629.177130.291928.395030.132928.6672
Note: The optimal result is marked with a bold number.
Table 2. Model detection accuracies on various types of fabrics.
Table 2. Model detection accuracies on various types of fabrics.
Metric (%)MethodSL1SL2SL5SL8SL9SL10SL11SL13Average Value
PDCAE37.9237.7348.8763.4916.2946.5955.8347.5244.28
DCGAN22.2938.1366.4531.9116.238.760.000.0022.97
Recycle-GAN36.2425.3920.3342.7731.2523.8535.7844.1532.47
MSCDAE51.3936.1749.6856.7844.6843.6654.0949.6848.27
UDCAE54.9455.5587.7515.5359.1451.1415.6987.7553.44
VAE-L2SSIM0.0042.6925.0070.1314.2824.482.2824.0625.37
AFFGAN62.0117.0221.8463.2635.8547.6934.6729.8639.02
SN-DDPM61.1058.9733.4857.4560.4751.0461.4446.1053.76
RDCAE72.9265.0451.5781.0813.5162.7460.0365.8059.09
DCGAN20.0835.936.7017.7310.001.000.0099.4423.86
Recycle-GAN79.5660.2256.6873.8767.8083.4674.2875.2771.39
MSCDAE74.4474.1571.1586.5526.0371.2376.1971.1568.86
UDCAE82.1161.6135.668.0878.4544.2015.1235.6645.11
VAE-L2SSIM0.0014.140.9959.6022.502.8111.6634.1018.22
AFFGAN75.8957.4269.0979.3080.5764.4138.1244.7963.70
SN-DDPM83.0770.6187.0184.2064.1183.6580.8976.9278.81
AccDCAE98.3697.8596.9799.2397.9998.5999.2699.3798.45
DCGAN97.6398.9397.0399.1797.8498.8499.150.0086.07
Recycle-GAN99.0997.6997.4799.0598.2699.0199.2599.1098.62
MSCDAE98.7897.5294.9299.2498.2398.5399.2394.9297.67
UDCAE98.9498.7497.8499.0098.6798.7599.2197.8498.62
VAE-L2SSIM98.6898.7296.9099.3698.2798.8599.5399.5398.73
AFFGAN99.1697.3497.9799.2399.8298.6699.1799.5598.86
SN-DDPM99.3697.8797.6199.4099.4299.4299.6298.9098.95
F1DCAE46.5546.4148.0267.2614.7450.1752.6045.3646.39
DCGAN16.3236.2410.4821.725.361.750.000.0011.48
Recycle-GAN46.0524.9727.650.0037.8833.3144.0852.5133.31
MSCDAE58.4047.2957.6466.5925.6851.9059.6657.6453.10
UDCAE63.1753.3646.998.6360.6039.2213.1446.9941.51
VAE-L2SSIM0.0019.341.9063.6815.044.8622.4222.4218.71
AFFGAN65.1516.4131.5366.5749.6252.5732.3529.9343.02
SN-DDPM65.6255.6144.7764.6761.4457.1164.7654.2058.52
IoUDCAE31.4531.8532.9852.2423.8734.9838.0330.4234.48
DCGAN10.1128.706.6915.122.960.990.000.008.07
Recycle-GAN33.2216.8116.5438.2223.8321.5430.6039.1727.49
MSCDAE42.8031.2544.5950.9117.4536.5044.0744.5939.02
UDCAE47.3139.4332.496.3944.0626.378.6232.4929.65
VAE-L2SSIM0.0013.400.9948.659.442.7512.7612.7612.59
AFFGAN50.099.3319.1851.3233.0037.1425.4820.8430.80
SN-DDPM53.2547.2531.5250.9448.9246.8054.0940.2646.63
Note: The optimal result is marked with a bold number.
Table 3. Ablation study for different values of α .
Table 3. Ablation study for different values of α .
Metric (%)αSL1SL2SL8SL9SL10Average Value
F10.132.4529.1729.9430.1730.7430.49
0.358.5155.5656.8347.0052.1354.01
0.565.6255.6164.6761.4457.1160.89
0.749.0046.0547.9543.2641.4645.54
0.924.2725.8023.0826.5124.1224.76
IoU0.119.8117.1117.6217.7718.1618.09
0.341.8938.9340.3530.7135.2537.43
0.553.2547.2550.9448.9246.8049.43
0.726.1033.2235.0129.9528.1930.49
0.914.9816.2414.2816.5314.8915.38
Note: The optimal result is marked with a bold number.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tang, S.; Jin, Z.; Zhang, Y.; Lu, J.; Li, H.; Yang, J. A Timestep-Adaptive-Diffusion-Model-Oriented Unsupervised Detection Method for Fabric Surface Defects. Processes 2023, 11, 2615. https://doi.org/10.3390/pr11092615

AMA Style

Tang S, Jin Z, Zhang Y, Lu J, Li H, Yang J. A Timestep-Adaptive-Diffusion-Model-Oriented Unsupervised Detection Method for Fabric Surface Defects. Processes. 2023; 11(9):2615. https://doi.org/10.3390/pr11092615

Chicago/Turabian Style

Tang, Shancheng, Zicheng Jin, Ying Zhang, Jianhui Lu, Heng Li, and Jiqing Yang. 2023. "A Timestep-Adaptive-Diffusion-Model-Oriented Unsupervised Detection Method for Fabric Surface Defects" Processes 11, no. 9: 2615. https://doi.org/10.3390/pr11092615

APA Style

Tang, S., Jin, Z., Zhang, Y., Lu, J., Li, H., & Yang, J. (2023). A Timestep-Adaptive-Diffusion-Model-Oriented Unsupervised Detection Method for Fabric Surface Defects. Processes, 11(9), 2615. https://doi.org/10.3390/pr11092615

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop