You are currently on the new version of our website. Access the old version .
SensorsSensors
  • Article
  • Open Access

30 July 2020

A Robust Fabric Defect Detection Method Based on Improved RefineDet

and
School of Mathematics and Computer Science, Fuzhou University, Fuzhou 350108, China
*
Author to whom correspondence should be addressed.
This article belongs to the Section Intelligent Sensors

Abstract

This paper proposes a robust fabric defect detection method, based on the improved RefineDet. This is done using the strong object localization ability and good generalization of the object detection model. Firstly, the method uses RefineDet as the base model, inheriting the advantages of the two-stage and one-stage detectors and can efficiently and quickly detect defect objects. Secondly, we design an improved head structure based on the Full Convolutional Channel Attention (FCCA) block and the Bottom-up Path Augmentation Transfer Connection Block (BA-TCB), which can improve the defect localization accuracy of the method. Finally, the proposed method applies many general optimization methods, such as attention mechanism, DIoU-NMS, and cosine annealing scheduler, and verifies the effectiveness of these optimization methods in the fabric defect localization task. Experimental results show that the proposed method is suitable for the defect detection of fabric images with unpattern background, regular patterns, and irregular patterns.

1. Introduction

In the textile industry, defect detection is an important part of the quality control of fabric products. It aims to detect fabric products with defective areas efficiently and timely to reduce additional economic losses caused by these low-quality products. Many textile mills still use the manual inspection method to detect defects [1], which causes inspectors to inadvertently miss some products with defects after a prolonged observation. To this end, it is necessary to develop an automatic defect detection method based on computer vision.
Since the 1980s, there have been a lot of works in the field of automated defect detection. Among them, the traditional works are mainly divided into the following categories: (1) Spectral-based methods, (2) statistical-based methods, (3) structural-based methods, and (4) traditional learning-based methods. For example, Jia et al. [2] proposed a fabric defect detection method based on lattice segmentation and the Gabor filtering. The method mainly used the Gabor filter band, a kind of spectral-based method, to extract features of the image patch (i.e., semantic lattice). Jing et al. [3] proposed two defect detection methods, i.e., Gabor preprocessed Golden Image Subtraction (GGIS), and Gabor preprocessed Golden Image Subtraction Based on a Genetic Algorithm (GAGIS), and the two methods mainly used the Gabor filter to preprocess the input image and used the image subtraction technique to locate the defective area. Pan et al. [4] and Zhang et al. [5] used the mathematical morphology method, a kind of Statistical-based method, to enhance the defective area in unpatterned images. Liu et al. [6] proposed a defect detection method proposed based on unsupervised segmentation (a dictionary learning method based on image patch) and Extreme Learning Machine (ELM), and the method mainly extracted geometric features (using Hu invariant moment method) and texture features (using an optimal wavelet packet technology) for ELM classifier. These traditional defect detection methods can get good results on some specific fabric products (unpatterned or regular patterned background). However, most of them are mainly based on predefined features or hand-craft features [7], including statistical features, structural features, and spectral features of the images. This means that the configuration of model parameters requires some prior knowledge or problem-specific research. When the fabric products with new design patterns appear, these detection methods must be modified or even redesigned [8].
With the development of deep Convolution Neural Networks (CNN) and Graphics Processing Units (GPU) computing power, many researchers pay more attention to studying an efficient defect detection method based on deep learning. Different from traditional methods, these deep learning methods mainly use the CNN, one of the most typical representative of deep neural networks, for feature extraction. Through a series of convolution, activation, and pooling operations, the CNN network can adaptively generate a hierarchy of features of the input image (i.e., feature maps) [9]. With this powerful feature learning capability, deep learning models based on CNN have been widely applied to various sub-fields of defect detection, such as fabric defect detection, lithium battery defect detection. For example, Zhao et al. [7] proposed a fabric defect classification method using a three-parallel-module integrated structure (i.e., Stacked Convolutional Auto-Encoders (SCAE), a shallow CNN and a deep CNN combining non-local block). Gan et al. [9] designed a joint detection CNN architecture, which contains two major parts: The global detection part (used for background classification in image level) and the sub detection part (used for defect classification in image patch level). Mei et al. [10] proposed an unsupervised learning approach based on a convolutional denoising autoencoder at multiple Gaussian pyramid levels (Multi-scale CDAE). Chen et al. [11] proposed a weakly-supervised learning-based surface defect classification and segmentation framework based on CNN with the Spatial Attention Mechanism (SAM). Hu et al. [12] proposed an unsupervised fabric defect detection method based on Deep Convolutional Generative Adversarial Network (DCGAN).
As a branch of deep learning, more and more object detection models based on CNN have been used in recent defect detection research, including the two-stage detector (Faster RCNN [13]) and the one-stage detector (SSD [14], YOLOv2 [15], YOLOv3 [16]). For example, Liu et al. [17] and Liu et al. [18] used Faster RCNN as the base model. [17] applied some data augmentation methods and soft NMS to further improve the performance of the model, and [18] combined the high-level features of the ROI pooling layer output with the low-level features obtained by the Histogram of Oriented Gradient (HOG) in the original Faster RCNN. Wu et al. [19] designed a Composite Interpolating Feature Pyramid Network (CI-FPN) as the neck structure and introduced a guided anchor mechanism and position-sensitive RoI-Align in head structure. Liu et al. [20] introduced SSD to the defect detection for the first time and added the third-level feature conv3_3 to the feature pyramid to achieve the detection of small objects. Zhang et al. [21] verified the performance of three variants of YOLOv2 models and proposed a yarn-dyed fabric defect automatic localization and classification method. Wei et al. [22] designed three models based on YOLOv3 to detect the surface defects on magnetic tiles. Jing et al. [8] introduced the YOLOv3 model to fabric defect detection and added a lower feature layer to the feature pyramid.
The advantages of the fabric defect detection methods based on object detection model are summarized as follows:
  • Powerful feature extraction capability. Object detection uses deep convolutional neural networks as the backbone [23], which can automatically extract the defect features of the input image.
  • Efficient neck structure (feature pyramid and feature fusion) structure. Most object detection models have a neck (i.e., feature pyramid and feature fusion), which can detect defect areas with different sizes in the image. These feature pyramid structures are roughly divided into the following categories [24,25]: (1) SSD-style, (2) FPN-style, (3) STDN-style, (4) M2Det-style, (5) PAN-style. Some recent studies have made improvements on the neck structure and achieved a good detection result [19,26].
  • Flexible selection of the model. Pre-existing object detection models usually can be divided into two categories, the one-stage object detection models and the two-stage object detection models. In general, the two-stage object detection models have higher localization and object classification accuracy and the one-stage object detection models are time-efficient and can be used for real-time detection [27]. According to the needs of different application scenarios, we can choose an appropriate model for training. In the field of fabric defect detection, the one-stage object detector can be selected as the base model to meet the needs of real-time detection.
  • Classification and localization results based on image patch level. The object detection model generates candidate object bounding boxes (i.e., image patches with defect area or background area) from the feature maps and sends them to the classification subnetwork and regression subnetwork, respectively. After Non-Maximum Suppression (NMS), we can directly get the defect categories and location of each predicted defect image patch.
  • Various general optimization methods. In the whole training and testing stages of the object detection model, many researchers proposed various general optimization methods [25,27,28], including data augmentation method, attention mechanism, learning rate scheduling strategy, activation function selection, loss function optimization, and post-processing method improvements. These optimization methods can boost the performance of all popular object detection models without introducing extra computational cost during inference [28].
  • Good generalization ability. The object detection model mainly learns the feature of defect objects, rather than the background. Therefore, when there are enough training images with defect objects, it can be suitable for fabric defect detection under different texture backgrounds. Meanwhile, some state-of-the-art data augment [25,29,30,31], and Weakly Supervised Object Localization (WSOL) [32,33] methods can alleviate the problems of insufficient training samples and expensive manual labeling to a certain extent.
In view of this, a fabric defect detection method based on an improved RefineDet is proposed in this paper. Firstly, we use a one-stage object detection model, RefineDet [34], as the base model, which has the advantages of both one-stage and two-stage object detection models. Secondly, we design an improved head structure combing Fully Convolutional Channel Attention (FCCA) block and Bottom-up path Augmentation Transfer Connection Block (BA-TCB), and this improved head structure can effectively improve the model’s ability to locate defective areas. Finally, we apply and verify the influence of many general optimization methods in the field of fabric defect detection, which can boost the performance of the proposed defect detection model without introducing too much extra computational cost during inference. On the three public defect datasets TILDA, Hong Kong Patterned Textures Database, and DAGM2007, the mAP and F1-score metrics of the proposed method exceed 80%. Experimental results show that the proposed method is suitable for the defect detection of fabric images with unpattern background, regular patterns, and irregular patterns and can detect different fabric defects at high speed.
The main contributions of this work are summarized as follows:
  • We use RefineDet as the base model of defect detection. To the best of our knowledge, this is the first time that RefineDet has been used for fabric defect detection. Using the special two-step classification and regression structure of RefineDet, the proposed method can better detect the defect area compared with other common object detectors.
  • We design an improved head structure. This improved head structure consists of Fully Convolutional Channel Attention-based Anchor Refinement Module (FCCA-ARM), BA-TCB, and Object Detection Module (ODM). By adding the channel attention mechanism (i.e., FCCA block) and designing the bottom-up path augmentation structure (i.e., BA-TCB), the detection accuracy of the proposed method is further improved.
  • We research and verify the influence of many general optimization methods in the field of fabric defect detection. The state-of-the-art general optimization methods, such as attention mechanism, DIoU-NMS, and cosine annealing scheduler, are successfully applied to our detection model, which is an important reference for researchers in the fabric defect detection field.
The remainder of this paper is organized as follows: Related work is introduced in Section 2. The details of our proposed method are included in Section 3. Experiment results and discussion are provided in Section 4. Finally, Section 5 is the conclusion of this work.

3. Methodology

As shown in Figure 2, we propose a robust fabric defect detection method based on improved RefineDet. The improved RefineDet network is consist of the following part: VGG-16 based backbone, FCCA-ARM, BA-TCB, ODM. In addition, many general optimization methods, such as mosaic data augmentation, attention mechanism, DIoU-NMS, and cosine annealing scheduler are applied to the training and testing stage of our model, which effectively improves the defect localization accuracy of the proposed method.
Figure 2. The network architecture of the proposed method.

3.1. Data Augmentation

In general, neural networks need to train millions of parameters. The premise of making these parameters work correctly is enough input data for training. However, in fabric defect detection, the data is not as much as we think. The size of the fabric datasets is dozens to thousands. Insufficient of defect sample is a challenge in the research of fabric defect detection methods.
To address the issue above, we mainly use data augmentation methods in the image preprocessing stage. In addition to using some data augmentation methods based on photometric distortion and geometric distortion, we also use the mosaic method in our work.
As shown in Figure 3, we can perform a combination of various data augmentation methods (i.e., image flip, rotation, scale-up, brightness change, contrast change, and mosaic) to increase the number of training images. In addition, when the generated images are close to the natural image, these data augmentation methods can increase the diversity of the input images, which makes the trained model more robust.
Figure 3. Data augmentation methods.

3.2. VGG-16 Based Backbone

We do not change the backbone of the original RefineDet, i.e., we use VGG-16 as the base feature extract network, remove fc8, replace fc6, fc7 into convolution layers (Conv6, Conv7) and add two extra convolution layers (Conv8_1, Conv8_2) after VGG-16. The configuration of each layer of the backbone network can be seen in Table 1. Between Conv1_1 and Conv7, all convolution layers are followed by the Relu activation function.
Table 1. The configuration of the VGG-16 based backbone network.

3.3. The Improved Head Structure

Inspired by BI-FPN [26], PANet [45], and CBAM [51], we design an improved head structure to improve the defect localization ability of the proposed method. The head in our model is consists of three parts: Fully convolutional Channel Attention-based Anchor Refinement Module (FCCA-ARM), Bottom-up path Augmentation Transfer Connection Block (BA-TCB), and ODM.

3.3.1. FCCA-ARM

The Layers in the feature pyramid come from Conv4_3, Conv5_3, Conv7, and Conv8_2 of the backbone network. By using the attention mechanism, the model can focus on important features and suppress unnecessary ones [51]. Therefore, we design an improved ARM that uses fully convolutional channel attention blocks to process those layers before ARM.
As seen in Figure 4c, given an intermediate feature map X W H C as input, the output Y of fully convolutional channel attention block is computed as Equation (1).
Y = σ 1 ( F C o n v ( A v g P o o l ( X ) ) + F C o n v ( M a x P o o l ( X ) ) )
in which,
F C o n v ( ) = σ 2 ( C o n v 2 ( σ 2 ( C o n v 1 ( ) ) ) )
Figure 4. Different channel attention blocks. (a) SE block; (b) CA block; (c) the Full Convolutional Channel Attention (FCCA) block.
A v g P o o l ( ) is the global average pool operation. M a x P o o l ( ) is the global max pool operation. C o n v 1 ( ) is convolution layer with 1∗1-s1-p0-C/R (i.e., kernel size is 1∗1, the stride is 1, padding is 0, output channel is C/R, R is the reduction ratio). C o n v 2 ( ) is the convolution layer with 1∗1-s1-p0-C. σ 1 ( ) is the Sigmoid activation function. Σ 2 ( ) is the Relu activation function.
Different from the SE block (see Figure 4a), the channel attention mechanism uses both average-pooled and max-pooled features simultaneously. In addition, we use two convolutional layers instead of the multilayer perceptron (MLP) in the original channel attention block (see Figure 4b), which can increase representation power and decrease the parameters of our model.

3.3.2. BA-TCB

Feature maps in deep layers often contain more global context information, while other feature maps in shallower layers contain more local textures and pattern structures. Thus, FPN adds a top-down path to propagate these semantically strong features. PANet adds an additional bottom-up pathway on top of the FPN. This bidirectional (top-down and bottom-up) structure can be used widely in BI-FPN, CI-FPN and have proved effective in practical applications.
Inspired by this bidirectional structure, we design four new blocks, TCBv2, to form a bottom-up pathway and add four fully convolutional channel attention blocks between TCB and TCBv2.
As seen in Figure 5, the four input of TCBv2 is denoted as ( P 3 , P 4 , P 5 , P 6 ) , and the output feature maps ( N 3 , N 4 , N 5 , N 6 ) are computed as Equations (3) and (4).
N 3 = T C B v 2 ( zeros , P 3 )
N i = T C B v 2 ( N i 1 , P I ) ,   i = 4 , 5 , 6
where I refers to a feature level with a resolution of 1 2 i of the input images (e.g., P 3 , N 3 are feature maps with resolution 40∗40 when the input image is 320∗320).
Figure 5. The structure of BA-TCB. (a) Bidirectional structure (Top-down TCB and Bottom-up TCBv2); (b) The structure of TCBv2.

3.3.3. ODM

ODM is consists of four convolutional layers with channel k∗(C+4). Where C is the number of defect classes for the multi-classification task, 4 refers to the four coordinates value (i.e., xmin, ymin, w, h) of predicted boxes for the regression task. K refers to the number of aspect ratios, and in our model k is set to 3, which means that the aspect ratios are 1:1, 1:2, and 2:1.

3.4. The Two-Step Loss Function and DioU-NMS

Similar to original RefineDet, in the training stage, the total two-step loss function is designed as Equation (5).
L ( { p i , X i } , { c i , B i } ) = a r m _ c ( { p i } ) + a r m _ l ( { x i } ) + o d m _ c ( { c i } ) + o d m _ l ( { t i } )
in which,
a r m _ c ( { p i } ) = 1 N a r m ( i [ l i g t 1 ] × C E ( p i , l i g t ) )
a r m _ l ( { X i } ) = 1 N a r m ( i [ l i g t 1 ] × L 1 ( X i , B i g t ) )
o d m _ c ( { c i } ) = 1 N o d m ( i [ l i g t 1 ] × C E ( c i , l i g t ) )
o d m _ l ( { B i } ) = 1 N o d m ( i [ l i g t 1 ] × L 1 ( B i , B i g t ) )
I is the index of anchor boxes in a training batch. p i ,   X i are the predicted defect class confidence being a defect object and localization coordinates of the anchor box i in the CA-ARM structure, respectively. c i ,   B i are the predicted object class confidence and coordinates of the refined box in the ODM, respectively. N a r m , N o d m are the numbers of defect class boxes in the CA-ARM and ODM, respectively. l i g t is the ground truth class label of anchor box i . B i g t is the ground truth localization coordinates. The indicator function [ l i g t 1 ] means that when the box i belongs to the defect class, it outputs 1, otherwise 0. In a r m _ c , CE is the cross-entropy loss over two classes (defect and background), and in o d m _ c , C E ( ) is the cross-entropy function over all defect classes. L 1 ( ) is the Smooth L1 function.
After calculating the total loss, backpropagation is performed to update the model parameters. After the specified number of iterations (in our experiment, we set it 120 k), the training of model finish.
Generally, the IoU-based method can be defined as IoU ( B , B gt ) . The penalty term ( B , B gt ) is designed for minimizing the normalized distance between predicted boxes B and the ground truth boxes B gt and making the regression (i.e., defect area localization) more accurate and faster. The three important geometric factors that affect the accuracy of the regression of the anchor boxes are the overlapping area between B and B gt , the distance between the center points of B and B gt and the aspect ratio of the box. Considering these factors, many improved IoU calculation methods have appeared in recent years. For example, gIoU designs a novel penalty item based on IoU to solve the problem of gradient disappearance when the bounding boxes do not overlap (as seen in Equation (10)). dIoU takes into account the factor of the distance between the center points (as seen in Equation (11)). By adding a penalty term to the original IoU, the normalized distance between the center points of the two bounding boxes is directly minimized, so that it can converge faster than the gIoU loss. Considering all three geometric factors, cIoU adds a penalty item to the aspect ratio based on the dIoU, which making regression more accurate and faster (as seen in Equation (12)).
gIoU ( B , B gt ) = IoU ( B , B gt ) gIoU ( B , B gt ) = IoU ( B , B gt ) | A c A B gt B | | A c |
dIoU ( B , B gt ) = IoU ( B , B gt ) dIoU ( B , B gt ) = IoU ( B , B gt ) ρ 2 ( b , b g t ) l c 2
cIoU ( B , B gt ) = IoU ( B , B gt ) cIoU ( B , B gt ) = IoU ( B , B gt ) ρ 2 ( b , b g t ) l c 2 α υ
in which,
v = 4 π 2 ( a r c t a n w g t h g t a r c t a n w h ) 2
α = v ( 1 IoU ( B , B gt ) ) + v
A c is the area of the smallest closure box of B gt and B. A B gt B is the intersection area of B gt and B. ρ ( ) is the Euclidean distance. b ,   b g t are the central points of B gt and B, respectively. l c is the diagonal length of the smallest closure box of B gt and B. And w g t h g t , w h are the aspect ratio of B gt and B, respectively.
Although cIoU is better than dIoU under normal circumstances, dIoU is better than cIoU in small object detection, i.e., the consistency of the aspect ratio may not affect the regression accuracy of small objects [54]. Therefore, we use dIoU-NMS to post-process the detection results of the proposed methods in the testing stage. For the predicted box M with the highest score, the dIoU-NMS can be formally defined as Equation (15).
s i = { s i , D I o U ( M , B i ) < T s 0 , D I o U ( M , B i ) T s
This means that these predicted defect boxes B i that overlap with M greater than the NMS threshold T s will be filtered.

3.5. Leaning Rate Adjustment Method Based on Cosine Annealing Scheduler

The common learning rate adjustment strategies mainly include the following: (1) StepLR (i.e., the learning rate in the training stage decreases according to regular intervals); (2) MultiStepLR (i.e., the learning rate decreases according to preset intervals); (3) ExponentialLR (i.e., the learning rate decreases according to exponential decay); (4) ReduceLROnPlateau (i.e., the learning rate in the training stage decreases when an indicator no longer changes); (5) CosineAnnealingLR (i.e., the learning rate in the training stage change according to the period of the cosine function). In the training stage, we use the cosine annealing scheduler to adjust the learning rate. The cosine annealing scheduler adjusts the learning rate in an intermediate iteration or epoch η t as Equation (16).
η t = η m i n + 1 2 ( η m a x η m i n ) ( 1 + c o s ( T c u r T m a x π ) )
where η m i n is the valley value of cosine function, and η m i n is set to 0 in our experiment. η m a x is the peak value of cosine function (i.e., initial learning rate), η m a x is set to 10 3 . T m a x is half of the Cosine function period, and T m a x is set to 120 k. T c u r refers to the number of epochs or iterations recorded since the last start.
As seen in Figure 6, the cosine annealing scheduler takes the cosine function as a period and resets the learning rate at the maximum value of each period. Taking the initial learning rate as the maximum learning rate and taking 2∗ T m a x as the period, it decreases first and then rises within a period.
Figure 6. The cosine annealing leaning rate in different Tmax.

4. Results and Discussion

In this section, we will show the robust fabric defect detection ability of the proposed method through a large number of comparison experiments and ablation experiments. All experiments were performed on an LZ-748GT workstation configured with Intel E5-2600 CPU (2200 MHz), 32GB RAM, and a Nvidia 16GB TITAN XP GPU. The code and detection results of the proposed method are available at https://github.com/2089527378/Fabric-defect-detection-based-on-improved-RefineDet.

4.1. The Datasets and Evaluation Metrics

In order to show the robust ability of the proposed method in the fabric defect detection task, we use three representative public defect datasets (i.e., TILDA dataset, Hong Kong Patterned Textures Database, and DAGM2007 dataset) in our experiment. In addition, to better compare the improvements, we also show the performance of our method on the PASCAL VOC dataset, an object detection standard dataset and the details can be seen in Appendix A.
  • TILDA dataset
TILDA is a common public textile texture dataset that was developed within the framework of the working group Texture Analysis of the DFG’s (Deutsche Forschungsgemeinschaft) major research programme “Automatic Visual Inspection of Technical Objects” [35]. A total of eight representative textile kinds (covering four unpatterned fabrics, two regular patterned fabrics, and two complex irregular patterned fabrics) were included in the dataset. Based on the analysis of textile atlases, seven defect classes are defined. In order to evaluate the performance of the proposed method, we use four common defect classes of fabric defects, which are hole (E1), spot (E2), wire (E3), and dark thread (E4). An illustration of defects and texture background can be found in Figure 7. The size of the images is 768 × 512 pixels, and the number of defective images is 1597.
Figure 7. Illustration of defect images on TILDA dataset.
  • Hong Kong Patterned Textures Database
Hong Kong Patterned Textures Database (we call it “Hong Kong dataset”) is a regular patterned database provided by Dr. Ngan, the Research Assistant Professor, Department of Mathematics, Hong Kong Baptist University. It consists of non-defective and defective images in three regular patterned fabrics (i.e., star-patterned, box-patterned, and dot-patterned fabrics). As shown in Figure 8, Six defect classes are defined, which are BrokenEnd (bn), Hole (he), NettingMultiple (nm), Knots (kn), ThickBar (tk), and ThinBar (tb). This dataset is highly challenging because of the small number of samples (only 82 defective images in all six defect classes). The size of images on the Hong Kong dataset is 450 × 450 pixels.
Figure 8. Illustration of images on Hong Kong dataset.
  • DAGM2007 Dataset
DAGM2007 is a benchmark defect dataset that is provided by the International Association for Pattern Recognition (IAPR) and the German Chapter of the European Neural Network Society (GNSS). The defect area is artificially generated, but similar to real-world problems. The texture background is mainly unpatterned or simple irregular patterned, and there are ten defect classes in DAGM2007 dataset which are Class1, Class2… Class10 (see Figure 9). The size of images is 512 × 512 pixels. The number of training defect images is 1046, and the number of testing images is 1054.
Figure 9. Illustration of images on DAGM2007 Dataset.
  • Evaluation Metrics
Six common evaluation metrics (i.e., Precision (P), Recall I, F1-score, mean Average Precision (mAP), Model Parameter (Param.), and detection time) are used. These evaluation metrics utilized in our experiments include two parts: The localization accuracy and the complexity of the model. We use P, R, F1-score, and mAP to evaluate the classification and localization accuracy of the proposed method. We can calculate P, R, F1-score, and mAP in the image-patch level as Equations (17)–(20).
P = k = 1 C P k
R = k = 1 C R k
m A P = k = 1 C A P k
F 1 s c o r e = 2 × P × R P + R
in which,
P k = T P k T P k + F P k
R k = T P k T P k + F N k
A P k = 0 1 P k ( R k )   d R k
C is the number of defect classes on a fabric dataset. T P k refers to the number of real defect class k objects that are correctly detected as defective boxes. F N k refers to the number of real defect class k objects that are falsely detected as non-defective boxes (i.e., texture background) or other class boxes. And F P k refers to the number of real background area that is falsely detected as defective boxes.
The comprehensive metric mAP refers to the area enclosed by the P-R curve and the R axis. The F1-score metric is a comprehensive evaluator that uses both the R and P metrics. In addition, we use Param. Metric and detection time to quantitatively show the space and time complexity of the model.

4.2. Experimental Results and Discussion on TILDA Dataset

In order to test the ability of the method to detect defects on fabric images containing multiple texture backgrounds (including unpatterned background, regular patterned background, and irregular patterned background), we used the TILDA dataset for experiments.

4.2.1. Experimental Settings

On the TILDA dataset, we divide the data set according to the ratio of 7:3, i.e., use 1117 images as the training set and 480 images as the testing set. The training steps of our method is 120 k; the batch size is 32; the optimizer is stochastic gradient descent optimizer (SGD); the momentum and weight decay are respectively set as 0.9 and 0.0005. The η m a x , η m i n , T m a x of cosine annealing scheduler are set as 1e-3, 0, 120 k, respectively. Considering the base models of the existing object detection-based fabric defect detection methods, we use a variety of common object detection models, including Faster RCNN (2015, NIPS), SSD (2016, ECCV), YOLOv3 (2018, CVPR), and FCOS (2019, CVPR) as comparison methods to show the advantages of the proposed method.
The configurations in our experiment are as follows:
  • Faster RCNN: ResNet-50 backbone + FPN + RPN + SGD optimizer + StepLR scheduler +NMS.
  • SSD: VGG-16 backbone + six layer feature pyramid + SGD optimizer + MultiStepLR scheduler + NMS.
  • YOLOv3: DarkNet-53 backbone + three feature pyramid and feature fusion + Adam optimizer + NMS.
  • FCOS (anchor free detector): ResNet-50 backbone + five layer FPN + NMS.
  • Original RefineDet: VGG-16 backbone + four layer feature pyramid and head structure (ARM, TCB and ODM) +SGD optimizer + MultiStepLR scheduler + NMS.
  • Ours: VGG-16 backbone + improved head structure (FCCA-ARM, BA-TCB and ODM) +SGD optimizer + Cosine annealing scheduler + DIoU-NMS.

4.2.2. Results and Discussion

As shown in Figure 10a, on the TILDA dataset, the loss function of the proposed method continuously decreases during the training stage until convergence. After 120 k iterations, the training of model parameters is completed. Compared with five common detection methods, the proposed method has the highest defect localization accuracy, and the detection speed can reach real-time. As shown in Table 2 and Figure 10b, on comprehensive metrics F1-score, the proposed method is 21.7%, 17%, 35.4%, and 3.4% higher than Faster RCNN, SSD, YOLOv3, and FCOS, respectively. On the object detection benchmark metric mAP, the proposed method is 21.3%, 19.8%, 46.9%, and 3.4% higher than Faster RCNN, SSD, YOLOv3, and FCOS, respectively. In terms of detection speed, our method only adds a small amount of space and time cost compared with the original RefineDet. The parameter size of the entire model only reaches 43.1 million (i.e., only occupies about 168MB of disk space), and the detection time reaches 34 frames per second (FPS). These experimental results show that the proposed method achieves a trade-off between detection time and localization accuracy, and can efficiently and quickly detect various defects on the fabric image.
Figure 10. Experiment display of the proposed method on the TILDA dataset. (a) The loss of the proposed method on the TILDA dataset in the training stage; (b) The P-R curve of the proposed method on the TILDA testing set.
Table 2. Performance comparison of different object detection approaches on TILDA dataset.
As shown in Figure 11, the TILDA dataset contains eight different texture backgrounds, including unpatterned textures, regular patterned textures, and complex irregular patterned textures. For these challenging images of fabric defects, our method is effective mainly because of the good generalization ability of the object detection model. The object detection model mainly learns the feature of defect objects, rather than the background. Complex and diverse texture backgrounds may affect the detection accuracy of the model, but when there are enough training images with defect object, it also can be suitable for the fabric defect detection under different texture backgrounds. The results at the TILDA dataset demonstrate the advantages of our method on fabric images with complex texture backgrounds. Meanwhile, in Section 4.4, we will specifically show the results of the ablation experiment and the advantages of each component of the proposed method.
Figure 11. The detection result of the proposed method on the TILDA testing set.

4.3. Experimental Results and Discussion on the Hong Kong Dataset and DAGM2007 Dataset

In order to demonstrate the good generalization ability of the proposed method, we use the Hong Kong Patterned Textures Database to verify the advantages of the proposed method on the small sample dataset and use the DAGM2007 dataset to show the robust defect detection ability of our method on the general texture defect dataset.

4.3.1. Experimental Settings

In this experiment, we did not change the hyperparameters of the proposed method. The training steps is 120 k; the batch size is 32; the optimizer is SGD optimizer; the momentum and weight decay are respectively set as 0.9 and 0.0005. The η m a x , η m i n , T m a x of cosine annealing scheduler are set as 1e-3, 0, 120 k, respectively. On Hong Kong Patterned Textures Database, we divide the dataset according to the ratio of 6:4—i.e., 50 images are used as the training set, and 32 images are used as the testing set. Because there are a few defective images on the Hong Kong dataset, we mainly use the offline data augmentation methods to process the training data and expand the number of training set to 633. At the same time, interference operations (image brightness change, image blur, random flip, and random rotation) are added to images of the testing set to further verify the anti-interference ability of the model. On the DAGM2007 dataset, the number of training set images is 1046, and the number of images in the testing set is 1054. We mainly use online data augmentation methods to process training data.

4.3.2. Results and Discussion

As shown in Figure 12, the loss function continuously decreases during the training stage until convergence. After 120 k iterations, the training of model parameters is completed.
Figure 12. Loss function display of the proposed method on the Hong Kong dataset and DAGM2007 dataset. (a) The loss of the proposed method on the Hong Kong dataset in the training stage; (b) The loss of the proposed method on the DAGM2007 dataset in the training stage.
As shown in Table 3, the proposed method is 1.1% and 0.2% higher than baseline in mAP metric on Hong Kong testing set and DAGM2007 testing set, respectively, and in F1-score, the proposed method is 3.4% and 3.8% higher than the baseline, respectively, which shows our method can get a good trade-off between P and R metrics. We also applied interference operations, such as image brightness change, image blur, random flip, and random rotation, to the testing images of the Hong Kong dataset, increasing the number of testing images to 128. In the case of unchanged model parameters, the proposed method can still effectively detect defect areas on Hong Kong testing set with interference (82.6% mAP and 81.5% F1-score), which shows that the proposed method has a strong anti-interference ability. In particular, as shown in Figure 13, the P-R curve of the proposed method is close to the upper right corner, and the detection results (mAP value) for three testing sets are all higher than 82%. Meanwhile, our method consumes considerable time and space costs. When the model input is 320 ∗ 320, the detection time reaches high detection speed (higher than 18 FPS), and the model parameters only reach about 43 million in space, and only occupy about 168 MB of disk space.
Table 3. Performance results on Hong Kong dataset and DAGM2007 dataset.
Figure 13. P-R curve display of the proposed method on the Hong Kong dataset and DAGM2007 dataset. (a) The P-R curve of the proposed method on Hong Kong testing set; (b) The P-R curve of the proposed method on Hong Kong testing set with interference; (c) The P-R curve of the proposed method on DAGM2007 testing set.
The experimental results on the Hong Kong testing set show that the proposed method obtains better detection results on the Hong Kong dataset. For this small sample data set, the defective samples of the training set are expanded by means of offline data augmentation. Under the premise that the generated images are close to the real images, the method in this paper is also applicable. Meanwhile, as seen in Figure 14a,b. The Hong Kong dataset is mainly periodic texture images of star type, box type, and dot type, which also verifies that our method is suitable for defect detection on periodic texture fabrics, even if there are many types of defects and a small number of samples in each defect class. As shown in Figure 14c, the DAGM2007 dataset mainly contains unpatterned background and simple irregular patterned background images. In this dataset, the proposed method achieves more than 96% in both mAP and F1-score metric. The experimental results on this dataset show that the proposed method is also suitable for general texture defect detection.
Figure 14. The partial detection result of the proposed method. (a) The detection result of the proposed method on Hong Kong testing set; (b) The detection result of the proposed method on Hong Kong testing set with interference; (c) The detection result of the proposed method on DAGM2007 testing set.

4.4. Ablation Experiments

In this part, on the one hand, we show the advantages of each component of the proposed method. On the other hand, we verify the impact of some general state-of-the-art optimization methods on the defect detection results.
As shown in Table 4, it can be seen from experiments 1, 2, and 3 that the FCCA-ARM and BA-TCB structure in our model can effectively improve the localization accuracy of the model. The improved head structure makes the mAP metric increase by 2.2%, and the F1-score increases by 0.9%. Although the parameter amount of the model has increased slightly, it can also reach real-time detection. In addition, experiments 1, 4, and 5 prove that DIoU-NMS and cosine annealing optimization methods can boost the performance of the model without introducing extra computational cost. As shown in Figure 15, when the Tmax of the cosine annealing scheduler is set to 120 k (the maximum number of iterations), the mAP and F1-score of the model are better than the original RefineDet (baseline). The combination of these components and the optimization methods makes our model have higher detection accuracy and achieve a trade-off between localization accuracy and speed.
Table 4. Performance of components and optimization methods in our method.
Figure 15. Detection results with different Tmax of Cosine annealing scheduler.
As shown in Table 5, we also verified the impact of other state-of-the-art general optimization methods, such as Swish activation function, Mish activation function, SE attention mechanism, SAM attention mechanism, GIoU regression loss, and DIoU regression loss on model performance. It can be seen from experiments 1, 2, and 3 that when replacing all the Relu activation functions of the backbone network, the Mish and Swish activation function does not greatly improve the model in the mAP metric. Experiments 1, 4, 5 prove that CAM works better in our model than SE and SAM. We also used GIoU loss and DIoU loss as odm_l loss; however, they made the localization accuracy decrease sharply, we still use Smooth L1 function as regression loss.
Table 5. Performance of different other optimization methods.

4.5. The Shortcomings and Outlook of The Proposed Method

The proposed method can detect the most defective areas efficiently and quickly. However, as shown in Figure 16, it is inevitable that some false predicted boxes and missed defect areas will be generated, mainly due to the highly similar shapes of some defect classes, the similar color of the defect, and the texture background.
Figure 16. Some examples of false detected images in the experimental results.
In order to solve this problem, on the one hand, the defect classification capability of the model should be further improved in future work. On the other hand, the phenomenon of blur and uneven illumination during the image acquisition process should be avoided as much as possible. In order to reduce the cost of manual labeling and better meet the needs of the textile industry, we also consider introducing weakly-supervised object detection components in future work. Meanwhile, considering that Mixup, CutOut, and CutMix can effectively improve the classification ability of the model, we plan to use Mixup, CutOut, CutMix, and Mosaic augmentation methods to obtain more number of data samples for the training in future work.

5. Conclusions

A fabric defect detection method based on improved RefineDet is proposed in this paper. On the one hand, the proposed method uses RefineDet as the base model, and have an improved head based on the FCCA block and BA-TCB structure, which efficiently improve the defect localization accuracy of the model. On the other hand, the proposed method applies many general optimization methods, such as attention mechanism, DIoU-NMS, and cosine annealing scheduler, and verifies the effectiveness of these optimization methods in the fabric defect detection. Experimental results show that the proposed method is suitable for the defect detection of fabric images with unpattern background, regular patterns, and irregular patterns and can detect different fabric defects at high speed.

Author Contributions

Conceptualization, H.X. and Z.W.; methodology, H.X. and Z.W.; software, Z.W.; validation, H.X.; formal analysis, Z.W.; writing—original draft preparation, Z.W.; writing—review and editing, H.X.; supervision, H.X.; funding acquisition, H.X. Both authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number 61801121.

Acknowledgments

The authors would like to thank the financial support received from the National Natural Science Foundation of China (No.61801121). Thanks Ngan for the Hong Kong Patterned Textures Database.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

In order to further demonstrate the localization ability of the proposed method in the general object detection domain, we also use the PASCAL VOC dataset (https://pjreddie.com/projects/pascal-voc-dataset-mirror/)an object detection standard dataset, to show the performance of the proposed method.

A.1. Experimental Settings

The network hyperparameters and data preprocessing methods are consistent with ones in Zhang et al. [34]—i.e., the backbone is VGG-16; the input size of the model is 320 × 320; the training steps is 120 k; the batch size is 32; the optimizer is SGD optimizer; the momentum and weight decay are, respectively, set as 0.9 and 0.0005. The PASCAL VOC dataset contains 20 common object classes that are diningtable, person, bottle, boat, train, bird, dog, cat, tvmonitor, cow, car, sofa, horse, chair, pottedplant, bicycle, motorbike, aeroplane, sheep, and bus. In this experiment, the training set is the VOC2007 trainval set (5011 images and 20 object classes) plus VOC2012 trainval set (11,540 images and 20 object classes). The testing set is the VOC2007 test set (4952 images and 20 object classes).

A.2. Results and Discussion on PASCAL VOC Dataset

As shown in Figure A1b, after 120 k iterations, the training of our model parameters is completed. Meanwhile, as shown in Table A1, the mAP and F1-score metrics of the proposed method on the VOC2007 test set are 0.2% and 1.5% higher than the original RefineDet, respectively. Although the size and detection time of the proposed method increases a little, the localization accuracy of our model has been improved. On the premise of ensuring the real-time detection, our method achieves a trade-off between detection accuracy and detection speed. The results in this experiment show that our method can also be applied in the field of general object detection.
Figure A1. Loss function and P-R curve display on the VOC dataset. (a) The loss function of the original RefineDet; (b) The loss function of the proposed method; (c) The P-R curve of the original RefineDet; (d) The P-R curve of the proposed method.
In addition, it can be seen from Figure A1c,d that compared with other classes, many object detectors have lower detection accuracy for bottle, chair, and pottedplant objects. For example, the APs for bottle, chair, and pottedplant classes are 63.0%, 62.6%, and 49.8%, respectively in our method; they are 61.4%, 62.2%, 49.3%, respectively in original RefineDet; they are 52.1%, 52.0%, 38.8% in original Faster RCNN (data comes from Table 6 in [13]); they are 47.6%, 54.7%, 48.6% in original SSD (data comes from Table 1 in [14], the input size of SSD is 300 × 300). This means that the proposed method is better than these common models in the detection of the bottle, chair, and pottedplant classes.
Table A1. Performance reults on PASCAL VOC2007 test set.
Table A1. Performance reults on PASCAL VOC2007 test set.
MethodP (%)R (%)mAP (%)F1-Score (%)Parm.Detection Time (FPS)
Original RefineDet
(Data come from Table 1 in [34])
--80.0--40.3
Original RefineDet
(Our reproduction in Pytorch)
35.986.079.950.734.4 M37.2
Ours
(RefineDet + FCCA-ARM + BA-TCB + DIoU-NMS + Cosine annealing scheduler)
37.486.380.152.243.6 M29.2

References

  1. Hanbay, K.; Talu, M.F.; Özgüven, Ö.F. Fabric defect detection systems and methods—A systematic literature review. Optik 2016, 127, 11960–11973. [Google Scholar] [CrossRef]
  2. Jia, L.; Chen, C.; Liang, J. Fabric defect inspection based on lattice segmentation and Gabor filtering. Neurocomputing 2017, 238, 84–102. [Google Scholar] [CrossRef]
  3. Jing, J.; Chen, S.; Li, P. Fabric defect detection based on golden image subtraction. Color. Technol. 2017, 133, 26–39. [Google Scholar] [CrossRef]
  4. Pan, Q.; Chen, M.; Zuo, B.; Hu, Y. The inspection of raw- silk defects using image vision. J. Eng. Fiber Fabr. 2018, 13, 78–86. [Google Scholar] [CrossRef]
  5. Zhang, B.; Tang, C. Fabric defect detection based on relative total variation model and adaptive mathematical morphology. Text. Res. J. 2017, 38, 145–149. [Google Scholar]
  6. Liu, L.; Zhang, J.; Liu, L.; Huang, Q.; Fu, X. Unsupervised segmentation and elm for fabric defect image classification. Multimed. Tools Appl. 2019, 78, 12421–12449. [Google Scholar] [CrossRef]
  7. Zhao, Y.; Hao, K.; He, H.; Tang, X.; Wei, B. A visual long-short-term memory based integrated CNN model for fabric defect image classification. Neurocomputing 2020, 380, 259–270. [Google Scholar] [CrossRef]
  8. Jing, J.; Zhuo, D.; Zhang, Y.; Liang, Y.; Zheng, M. Fabric defect detection using the improved YOLOv3 model. J. Eng. Fiber Fabr. 2020, 15, 1–10. [Google Scholar] [CrossRef]
  9. Gan, X.; Lv, R.; Zhu, H.; Ma, L.; Wang, X.; Zhang, Z.; Huang, Z.; Zhu, H.; Ren, W. A fast and robust convolutional neural network-based defect detection model in product quality control. Int. J. Adv. Manuf. Technol. 2018, 94, 3465–3471. [Google Scholar]
  10. Mei, S.; Wang, Y.; Wen, G. Automatic fabric defect detection with a multi-scale convolutional denoising autoencoder network model. Sensors 2018, 18, 1064. [Google Scholar] [CrossRef]
  11. Chen, H.; Hu, Q.; Zhai, B.; Chen, H.; Liu, K. A robust weakly supervised learning of deep Conv-Nets for surface defect inspection. Neural Comput. Appl. 2020. [Google Scholar] [CrossRef]
  12. Hu, G.; Huang, J.; Wang, Q.; Li, J.; Xu, Z.; Huang, X. Unsupervised fabric defect detection based on a deep convolutional generative adversarial network. Text. Res. J. 2020, 90, 247–270. [Google Scholar] [CrossRef]
  13. Ren, S.; He, K.M.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. In Proceedings of the 29th Annual Conference on Neural Information Processing Systems, Montreal, QC, Canada, 7–9 December 2015; pp. 91–99. [Google Scholar]
  14. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.; Berg, A.C. SSD: Single Shot Multibox Detector. In Proceedings of the 14th European Conference on Computer Vision, Amsterdam, The Netherlands, 10–16 October 2016; pp. 21–37. [Google Scholar]
  15. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Hawaii, HI, USA, 25–30 June 2017; pp. 6517–6525. [Google Scholar]
  16. Redmon, J.; Farhadi, A. YOLOv3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  17. Liu, Z.; Liu, X.; Li, C.; Li, B.; Wang, B. Fabric Defect Detection Based on Faster R-CNN. In Proceedings of the 9th International Conference on Graphic and Image Processing, Qingdao, China, 14–16 October 2017; pp. 1–9. [Google Scholar]
  18. Liu, Z.; Guo, Z.; Yang, J. Research on Texture Defect Detection Based on Faster-RCNN and Feature Fusion. In Proceedings of the 11th International Conference on Machine Learning and Computing, Zhuhai, China, 22–23 February 2019; pp. 429–433. [Google Scholar]
  19. Wu, Y.; Zhang, X.; Fang, F. Automatic Fabric Defect Detection Using Cascaded Mixed Feature Pyramid with Guided Localization. Sensors 2020, 20, 871. [Google Scholar] [CrossRef]
  20. Liu, Z.; Liu, S.; Li, C.; Ding, S.; Dong, Y. Fabric Defects Detection Based on SSD. In Proceedings of the 2nd International Conference on Graphics and Signal Processing, Sydney, Australia, 6–8 October 2018; pp. 74–78. [Google Scholar]
  21. Zhang, H.; Zhang, L.; Li, P.; Gu, D. Yarn-Dyed Fabric Defect Detection with YOLOV2 Based on Deep Convolution Neural Networks. In Proceedings of the IEEE 7th Data Driven Control and Learning Systems Conference, Enshi, China, 25–27 May 2018; pp. 170–174. [Google Scholar]
  22. Wei, J.; Zhu, P.; Qian, X.; Zhu, S. One-Stage Object Detection Network for Inspecting the Surface Defects of Magnetic Tiles. In Proceedings of the 2019 IEEE International Conference on Imaging Systems and Techniques, Abu Dhabi, UAE, 9–10 December 2019; pp. 1–6. [Google Scholar]
  23. Jiao, L.; Zhang, F.; Liu, F.; Yang, S.; Li, L.; Feng, Z.; Qu, R. A Survey of Deep Learning-Based Object Detection. IEEE Access 2019, 7, 128837–128868. [Google Scholar] [CrossRef]
  24. Zhao, Q.; Sheng, T.; Wang, Y. M2det: A Single-Shot Object Detector Based on Multi-Level Feature Pyramid Network. In Proceedings of the 33th AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; pp. 9259–9266. [Google Scholar]
  25. Bochkovskiy, A.; Wang, C.; Liao, H.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  26. Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and efficient object detection. arXiv 2020, arXiv:1911.09070. [Google Scholar]
  27. Wu, X.; Sahoo, D.; Hoi, S.C.H. Recent advances in deep learning for object detection. Neurocomputing 2020, 396, 39–64. [Google Scholar] [CrossRef]
  28. Zhang, Z.; He, T.; Zhang, H.; Zhang, Z.; Xie, J.; Li, M. Bag of Freebies for Training Object Detection Neural Networks. arXiv 2019, arXiv:1902.04103. [Google Scholar]
  29. Zhang, H.; Cisse, M.; Dauphin, Y.N.; David, L. MixUp: Beyond empirical risk minimization. arXiv 2017, arXiv:1710.09412. [Google Scholar]
  30. DeVries, T.; Taylor, G.W. Improved regularization of convolutional neural networks with CutOut. arXiv 2017, arXiv:1708.04552. [Google Scholar]
  31. Yun, S.; Han, D.; Chun, S.; Oh, S.J.; Yoo, Y.; Choe, J. CutMix: Regularization Strategy to Train Strong Classifiers with Localizable Features. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 6023–6032. [Google Scholar]
  32. Zhang, X.; Wei, Y.; Feng, J.; Yang, Y.; Huang, T. Adversarial Complementary Learning for Weakly Supervised Object Localization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1325–1334. [Google Scholar]
  33. Zhang, C.; Cao, Y.; Wu, J. Rethinking the Route towards Weakly Supervised Object Localization. arXiv 2020, arXiv:2002.11359. [Google Scholar]
  34. Zhang, S.; Wen, L.; Bian, X. Single-Shot Refinement Neural Network for Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4203–4212. [Google Scholar]
  35. Xie, H.; Zhang, Y.; Wu, Z. Fabric Defect Detection Method Combing Image Pyramid and Direction Template. IEEE Access 2019, 7, 182320–182334. [Google Scholar] [CrossRef]
  36. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  37. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. arXiv 2015, arXiv:1512.03385. [Google Scholar]
  38. He, K.; Georgia, G.; Dollar, P.; Girshick, R. Mask R-CNN. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 42, 386–397. [Google Scholar] [CrossRef]
  39. Xie, S.N.; Girshick, R.; Dollar, P. Aggregated Residual Transformations for Deep Neural Networks. In Proceedings of the IEEE Computer Vision and Pattern Recognition, Honolulu, HI, USA, 22–25 July 2017. [Google Scholar]
  40. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  41. Sandler, M.; Howard, A.; Zhu, M. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the IEEE Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4510–4520. [Google Scholar]
  42. Zhang, X.; Zhou, X.; Lin, M. ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices. In Proceedings of the IEEE Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6848–6856. [Google Scholar]
  43. Lin, T.; Dollar, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. arXiv 2017, arXiv:1612.03144. [Google Scholar]
  44. Chiasi, G.; Lin, T.Y.; Le, Q.V. NAS-FPN: Learning Scalable Feature Pyramid Architecture for Object Detection. In Proceedings of the IEEE Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 7029–7038. [Google Scholar]
  45. Liu, S.; Qi, H.; Shi, J. Path Aggregation Network for Instance Segmentation. In Proceedings of the IEEE Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8759–8768. [Google Scholar]
  46. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the IEEE Computer Vision and Pattern Recognition, Columbus, OH, USA, 24–27 June 2014; pp. 580–587. [Google Scholar]
  47. Bodla, N.; Singh, B.; Chellappa, R. Soft-NMS–Improving Object Detection with One line of Code. In Proceedings of the International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5561–5569. [Google Scholar]
  48. Liu, S.; Huang, D.; Wang, Y. Adaptive NMS: Refining pedestrian detection in a crowd. arXiv 2019, arXiv:1904.03629. [Google Scholar]
  49. Geirhos, R.; Rubisch, P.; Michaelis, C.; Bethge, M.; Wichmann, F.A.; Brendel, W. ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness. arXiv 2019, arXiv:1811.12231. [Google Scholar]
  50. Hu, J.; Shen, L.; Albanie, S.; Wu, E. Squeeze-and-excitation networks. arXiv 2018, arXiv:1709.01507. [Google Scholar]
  51. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional block attention module. arXiv 2018, arXiv:1807.06521. [Google Scholar]
  52. Ramachandran, P.; Zoph, B.; Le, Q.V. Searching for activation functions. arXiv 2017, arXiv:1710.05941. [Google Scholar]
  53. Misra, D. Mish: A self regularized nonmonotonic neural activation function. arXiv 2019, arXiv:1908.08681. [Google Scholar]
  54. Zheng, Z.H.; Wang, P.; Liu, W.; Li, J.; Ye, R.; Ren, D. Distance-IoU Loss: Faster and better learning for bounding box regression. arXiv 2019, arXiv:1911.08287. [Google Scholar]

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.