Next Article in Journal
Resilient Consensus Control for Multi-Agent Systems: A Comparative Survey
Next Article in Special Issue
Waste Detection System Based on Data Augmentation and YOLO_EC
Previous Article in Journal
Data Interworking Model and Analysis for Harmonization of Smart Metering Protocols in IoT-Based AMI System
Previous Article in Special Issue
Underwater Object Detection Using TC-YOLO with Attention Mechanisms
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

ARTD-Net: Anchor-Free Based Recyclable Trash Detection Net Using Edgeless Module

1
Visual Information Processing, Korea University, Seoul 02841, Republic of Korea
2
Department of Electrical Engineering, Korea University, Seoul 02841, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(6), 2907; https://doi.org/10.3390/s23062907
Submission received: 6 February 2023 / Revised: 2 March 2023 / Accepted: 2 March 2023 / Published: 7 March 2023
(This article belongs to the Special Issue Image Processing and Analysis for Object Detection)

Abstract

:
Due to the sharp increase in household waste, its separate collection is essential in order to reduce the huge amount of household waste, since it is difficult to recycle trash without separate collection. However, since it is costly and time-consuming to separate trash manually, it is crucial to develop an automatic system for separate collection using deep learning and computer vision. In this paper, we propose two Anchor-free-based Recyclable Trash Detection Networks (ARTD-Net) which can recognize overlapped multiple wastes of different types efficiently by using edgeless modules: ARTD-Net1 and ARTD-Net2. The former is an anchor-free based one-stage deep learning model which consists of three modules: centralized feature extraction, multiscale feature extraction and prediction. The centralized feature extraction module in backbone architecture focuses on extracting features around the center of the input image to improve detection accuracy. The multiscale feature extraction module provides feature maps of different scales through bottom-up and top-down pathways. The prediction module improves classification accuracy of multiple objects based on edge weights adjustments for each instance. The latter is an anchor-free based multi-stage deep learning model which can efficiently finds each of waste regions by additionally exploiting region proposal network and RoIAlign. It sequentially performs classification and regression to improve accuracy. Therefore, ARTD-Net2 is more accurate than ARTD-Net1, while ARTD-Net1 is faster than ARTD-Net2. We shall show that our proposed ARTD-Net1 and ARTD-Net2 methods achieve competitive performance in mean average precision and F1 score compared to other deep learning models. The existing datasets have several problems that do not deal with the important class of wastes produced commonly in the real world, and they also do not consider the complex arrangement of multiple wastes with different types. Moreover, most of the existing datasets have an insufficient number of images with low resolution. We shall present a new recyclables dataset which is composed of a large number of high-resolution waste images with additional essential classes. We shall show that waste detection performance is improved by providing various images with the complex arrangement of overlapped multiple wastes with different types.

1. Introduction

Recently, the increasing population has been accompanied by a vast amount of resource consumption. Therefore, separate collection is essential to reduce the huge amount of household waste by recycling trash. Since it is costly and time-consuming to separate trash manually, it is very important to develop an automatic system for separate collection using deep learning and computer vision [1,2,3,4]. In previous research on automatic systems for separate collection, many methods have been proposed for trash classification using deep learning and computer vision [5,6,7]. However, they only deal with the simple case for the image with a single trash. In order to overcome the limitations of single garbage classification, previous research has worked on methods for detecting multiple types of trash [8,9]. However, the trash datasets used in their research have several problems which make it difficult for deep learning models to detect trash in the real world. First, the existing dataset does not efficiently deal with various wastes of the real world, since it does not include important types of waste such as cans and PETs. Second, since the existing dataset is used for classifying single waste in one image, it does not consider data with the complex arrangement, i.e., with several overlapped wastes with different types. Third, most of the existing datasets have a small number of data with low resolution. In order to resolve those problems, we are concerned with the design of deep learning based model for recognizing multiple wastes of different types in the complex image which may be overlapped as well as the creation of large-scale high-resolution dataset with various types.
In this paper, we shall propose two Anchor-free based Recyclable Trash Detection Networks (ARTD-Net) for accurately recognizing overlapped multiple wastes of different types by using edgeless module: ARTD-Net1 and ARTD-Net2. The former is an anchor-free-based one-stage deep learning model which consists of three modules: centralized feature extraction, multiscale feature extraction and prediction. The centralized feature extraction module is used as the backbone in the model, focusing on extracting features around the center of the input image. The extracted feature maps are passed on to the multiscale feature extraction module for further processing. The multiscale feature extraction module generates five feature maps of different sizes through the combination of bottom-up and top-down pathways. The prediction module improves detection accuracy by exploiting feature maps of different sizes and efficiently finding small-sized objects using anchor-free detection, and adjusting edge weights for each instance to improve classification accuracy of multiple objects. The latter is an anchor-free based multi-stage deep learning model which can efficiently finds each of waste regions by additionally exploiting Region Proposal Network and RoIAlign. It sequentially performs classification and regression to improve accuracy. Additionally, we build a new recyclables dataset which consists of a large number of high-resolution images of waste. Our dataset provides various images with complex arrangements, such as overlapped wastes of different types, and comprises additional essential classes.
The contributions of our research are summarized as follows:
  • We propose two Anchor-free based Recyclable Trash Detection Networks (ARTD-Net) which can recognize overlapped multiple wastes of different types efficiently by using edgeless module: ARTD-Net1 and ARTD-Net2. ARTD-Net2 is more accurate than ARTD-Net1, while ARTD-Net1 is faster than ARTD-Net2. We shall show that our proposed ARTD-Net1 and ARTD-Net2 achieve competitive performance in mean average precision and F1 score compared to other deep learning models.
  • We propose a background weight adjustment block in the centralized feature extraction module which improves the detection accuracy by focusing on around the center of the input image based on centralized weights adjustments.
  • We propose an object instance separation block in the prediction module which improves the classification accuracy of multiple objects based on edge weights adjustments for each instance.
  • We present a multi-stage model for ARTD-Net which efficiently finds each of waste regions by using Region Proposal Network and RoIAlign.
  • We contribute a new large scale recyclables dataset which comprises various additional essential classes, and include various high resolution images with the complex arrangement, i.e., with several overlapped wastes with different types.
The remainder of this paper comprises four sections. In Section 2, we review studies on waste classification and anchor-free models. In Section 3, we present the overall architecture for our waste detection model. In Section 4, we describe our experimental results, and finally in Section 5 we conclude with our results.

2. Related Work

2.1. Trash Dataset

2.1.1. Classification

TrashNet dataset [10] is a collection of 2527 images of six classes of waste, including trash, metal, plastic, glass, cardboard and paper. A number of subsequent studies have been conducted with the TrashNet dataset [11]. RecycleNet achieved 81% accuracy on the Trashnet dataset. The best result was achieved with the DenseNet121 [12] model. They changed the connection pattern between high-density blocks and reduced the parameters of the 121-layer network from 7 million to approximately 3 million in order to compensate for the slow prediction time [13]. Ruiz et al. achieved an accuracy of 88.6% based on the ResNet model [14] after adding non-waste class to the TrashNet dataset. They experimented with CNN [15], VGG [16] and Inception for automatic garbage classification [17]. The DNN-TC model is based on the ResNext architecture and applies several preprocessing techniques. Two fully connected layers were added to reduce redundancy and improve performance. In addition, they collected inorganic, organic, and medical waste from Vietnam and included them in the dataset. The performance of DNN-TC model was compared with that of the existing TrashNet dataset, and the results were analyzed [18].

2.1.2. Segmentation

The TACO dataset was created to detect waste dumped in the sea. This dataset consists of 1500 images with 4,784 annotations that are labeled in 60 subcategories that belong to 28 super-categories [19]. MJU-Waste dataset was built to compensate for the shortcomings of the TACO dataset. It consists of 2475 RGBD images taken with Microsoft Kinect RGBD camera. They improved detection accuracy by applying the intensity and depth information to multiple levels of spatial granularity to previous deep learning models. The model with ResNet-101 as the backbone achieved an Intersection over Union (IoU) score of 87.84 [20].

2.2. Object Detection Method

Object detection methods can generally be divided into two main methods: one-stage and multi-stage. One-stage method performs classification and regression at the same time. This results in faster processing times compared to the multi-stage method. We analyze YOLO series [21,22,23,24,25,26], SSD [27], FASF [28], FCOS [29], SABL [30], SOLO [31], CornerNet [32], and CentripetalNet [33] and compare their performance with ARTD-Net1. The multi-stage method performs regression and classification sequentially. It is more accurate than one-stage method. We analyze Faster R-CNN [34], Cascade R-CNN [35], Double-Head RCNN [36], Sparse R-CNN [37] and DetectoRS [38] and compare their performance with ARTD-Net2.
YOLO is a real-time object detector designed for high-speed processing. It finds the location and class using a single network pass. YOLO divides the input image into multiple grids, where the grid cell closest to the center of the object is responsible for detecting the object. Each grid cell predicts the corresponding bounding boxes, along with their confidence scores and conditional class probabilities. YOLO has a limitation where the performance is somewhat reduced for small objects due to the small difference in IoU values. To address this limitation, improved models such as YOLOx, YOLOF, YOLOv5 and YOLOv6 have been proposed.
SSD detects objects of various sizes by dividing the input image into grids of different sizes across six feature maps. It is constructed using six additional convolution layers on the fifth convolution layer of VGG-16.
FSAF is a one-stage model based on RetinaNet [39]. They improved the performance by using multi-level anchor-free branches which solved the problems of the existing anchor-based model. During training, anchor-free branches attached to each level of the Feature Pyramid Network (FPN) [40] select the most appropriate feature level for training, which effectively represents the instance. As a result of training, the model achieves better performance than existing one-stage detectors in detecting small objects.
FCOS is a one-stage model which detects objects in a per-pixel prediction fashion. It is based on an anchor-free method and reduces training time by eliminating the need for complex calculations associated with anchor boxes such as calculating overlapping during training. They utilize centerness to decrease the impact of bounding boxes generated far from the center of an object. Centerness reduces the influence of predicted values at locations far from the center of the object using the center of the bounding box, right-bottom, left-top and corner pairs.
CornerNet predicts object bounding boxes using a keypoint pair instead of anchors, where the keypoints represent the top-left and top-right corners of the target object. Instead of using anchor boxes, CornerNet estimates these keypoints based on feature points. CentripetalNet is a keypoint-based object detector which utilizes centripetal shift to generate a keypoint pair representing the corners of the same object.
Faster R-CNN is an object detection model which is developed to improve the slow speed of R-CNN caused by selective search. The selective search is the slowest part of R-CNN. It is computed in the CPU. To improve this, Faster R-CNN introduces RPN that can perform computations on the GPU. RPN takes a feature map as input which is obtained from the previous convolutional neural network layer. A 256-dimension vector is obtained using a sliding window on the received feature map. At this time, the anchor to be used as the window being set in advance. In Faster R-CNN, nine anchors with various width, height, ratio, and size are used. The class and location of the object are calculated through two layers using the 256-dimensional vector obtained in this way.
Cascade R-CNN proposes a method to address two issues that occur with the increase in IoU thresholds. With an increase in IoU thresholds, the number of positive samples decreases exponentially, which can lead to overfitting during training. In addition, if there is a difference between IoU thresholds used during training and those used at inference, it can result in decreased accuracy. Cascade R-CNN is comprised of a sequence of detectors with different IoU thresholds set. Detectors are connected sequentially and use the output from the previous step as input for the next step. As a result, each detector has the positive set of examples of equivalent size to solve the overfitting problem. It shows that performance is improved through the process of gradual training using the proposals of the learned detectors at low IoU.
The Double-Head method utilizes the commonly used two head structures for classification and localization tasks in R-CNN based detectors. They found that the fully connected head is better for classification while the convolution head is better for localization. To leverage the strengths of both structures, they proposed the Double-Head method which combines a fully connected head for classification and a convolution head for bounding box regression.
Sparse R-CNN is a purely sparse method for object detection. It utilizes a fixed sparse set of learned object proposals provided to the object recognition head for classification and location. Sparse R-CNN directly outputs final predictions without a non-maximum suppression post-procedure and demonstrates accuracy performance on par with well-established detector baselines on the COCO dataset.
DetectoRS proposes a backbone design utilizing a see and think mechanism with improvements at two levels: macro and micro. At the macro level, the method involves building Recursive Feature Pyramid by adding feedback connections to the existing FPN. At the micro level, they propose Switchable Atrous Convolution which utilizes a switch function to collect features extracted at various atrous rates.

3. Model Architecture

In this section, we shall present ARTD-Net which is an anchor-free based deep learning model for detecting multiple wastes with various types. We propose two versions of ARTD-Net: ARTD-Net1 and ARTD-Net2. ARTD-Net1 is a one-stage model which performs classification and regression at the same time, while ARTD-Net2 a multi-stage model which sequentially performs classification and regression. ARTD-Net1 is faster than ARTD-Net2, but ARTD-Net2 is more accurate than ARTD-Net1.

3.1. ARTD-Net1

As shown in Figure 1, ARTD-Net1 is an anchor-free-based one-stage model which consists of three modules: centralized feature extraction, multiscale feature extraction and prediction. The centralized feature extraction module is used as backbone in the model. It focuses on extracting features around the center of the input image. It consists of three layers each of which extracts a feature map of different size. The feature map from each layer is sent to multiscale feature extraction module. The multiscale feature extraction module is composed of five layers each of which produces a feature map of different size. The feature map in each layer is sent to prediction module. Each of five prediction modules finds the region and class for multiple wastes of different size based on the feature map from multiscale feature extraction module.

3.1.1. Centralized Feature Extraction Module

The centralized feature extraction module is obtained by adding BWAB (Background Weight Adjustment Block) to RetinaNet. BWAB exploits a new scheme which efficiently extracts features around the center of the input image while reducing the edge weights around the border of the image. BWAB consists of two components: Background Weight Adjustment kernel and the convolution layer. Let F i R H × W × C be the feature maps at layer i of a backbone, and H and W the height and width of the feature map, respectively, and C the number of class. Let ( x , y ) be the coordinates of the feature map, x X = { 0 , 1 , , W } , y Y = { 0 , 1 , , H } .
Then, the function of Background Weight Adjustment kernel is defined as follows:
K b a c k ( x , y ) = e x p ( ( | x σ | + | y σ | ) γ ) ,
where σ is the scale factor for the area of the kernel, and γ is the scale factor for the gradient of the kernel. Following the bottom-up pathway, centralized feature extraction module reduces the size of the feature map through downsampling by half at each layer for the input image, and creates different feature maps with various scales, which are transferred to multiscale feature extraction module. Regarding Figure 2, H and W are the height and width of feature maps, respectively. s is the downsampling ratio of the feature maps.

3.1.2. Multiscale Feature Extraction Module

The multiscale feature extraction module generates feature maps of different scales through convolution layer based on feature map received from the backbone. The multiscale feature extraction module consists of two pathways: bottom-up and top-down. The bottom-up pathway reduces the size of the feature map by half through downsampling, and the top-down pathway doubles the size of the feature map through upsampling. The previous FPN produces three feature maps using only the top-down pathway as shown in Figure 3. Our multiscale feature extraction module provides consistent detection accuracy by adding the bottom-up pathway to FPN and thus generating five feature maps of different sizes with level from layer 1 through layer 5.

3.1.3. Prediction Module

There are five prediction modules each of which performs classification and regression simultaneously for object instance. They efficiently find the small-sized wastes by using anchor-free detection, and improve detection accuracy by exploiting the feature maps of different size. Prediction module consists of two blocks: OISB (Object Instance Separation Block) and ODB (Object Detection Block). OISB reduces the interference between adjacent objects by reducing the edge weight of each object through instance separation kernel. Let b l = [ b x l , b y l , w l , h l ] be the bounding box information, where i denotes a layer level of prediction module, and ( b x l , b y l ) is the center coordinate of bounding box b l , and ( w l , h l ) is the width and height of bounding box, respectively. Let ( x , y ) be the coordinates of bounding box with ( b x l , b y l ) as origin, and x X = { w l / 2 , , w l / 2 } , y Y = { h l / 2 , , h l / 2 } .
Then, Instance Separation Kernel is defined as follows:
K i n s t a n c e ( x , y ) = 1 2 π σ e x p ( x 2 + y 2 2 σ ) , σ = 3
Feature layer receives the feature map from multiscale feature extraction module, and then instance separation layer adjusts the edge weight of each object using the instance separation kernel in order to improve detection accuracy for overlapped multiple wastes of different types. The feature map is sent to ODB after applying the convolution layer. This phase is illustrated in Figure 4.
ODB detects multiple wastes with various sizes through anchor-free based object detection. Since the size of the anchor is fixed in the existing anchor-based models, it is difficult to detect the object smaller than the size of the anchor. ODB in our model is trained using the ground truth bounding box for the object instead of the fixed anchor most approximate to the bounding box in order to efficiently find the small objects whose bounding boxes are much smaller than the anchor candidate. ODB finds the loss value for our model by using Class Loss Function (CLF) and Box Regression Loss Function (BRLF). CLF calculates a class loss for the object instance based on the focal loss function. It decreases the loss for the high probability class while increasing for the low probability class. Let p be the predicted probability of the class for the object instance. Given p, CLF is defined as follows:
C L F ( p ) = ( 1 p ) γ l o g ( p ) , γ = 2 ,
where γ is the adjustable parameter.
BRLF is a bounding box regression function which calculates a loss by finding the IoU between the ground truth bounding box and the prediction bounding box.
Let x ˜ = ( x 1 g , y 1 g , x 2 g , y 2 g ) R 4 be the coordinates of the ground truth bounding box for the object instance, and ( x 1 g , y 1 g ) and ( x 2 g , y 2 g ) the top-left and bottom-right corners of the ground truth bounding box, respectively. Let x = ( x 1 p , y 1 p , x 2 p , y 2 p ) R 4 be the coordinates of the predicted bounding box of our model for the object instance, and ( x 1 p , y 1 p ) and ( x 2 p , y 2 p ) the top-left and bottom-right corners of the predicted bounding box, respectively. Given ( x , x ˜ ) , BRLF is defined as follows:
X = ( x 2 p x 1 p ) ( y 2 p y 1 p ) , X ˜ = ( x 2 g x 1 g ) ( y 2 g y 1 g ) I = ( m i n ( x 2 g , x 2 p ) m a x ( x 1 g , x 1 p ) ) ( m i n ( y 2 g , y 2 p ) m a x ( y 1 g , y 1 p ) ) U = X + X ˜ I IoU = I / U BRLF ( x , x ˜ ) = ln ( IoU )
Then, the loss function L at layer l in the prediction module is defined as follows:
L ( l ) = C L F l ( p ) + B R L F l ( x , x ˜ ) 2
Then, the final total loss L ( l ^ ) is defined as the minimum loss, where
l ^ = arg m i n l L ( l )

3.2. ARTD-Net2

ARTD-Net2 is a multi-stage model which improves the accuracy of waste detection compared to ARTD-Net1 by attaching Region Proposal Network in the prediction module and additionally exploiting recursive feature extraction scheme as shown in Figure 5. ARTD-Net2 is slower to detect than ARTD-Net2, but ARTD-Net2 has higher detection accuracy than ARTD-Net1.
The centralized feature extraction module is used as the backbone in the multi-stage model. The feature map in each layer is sent to recursive multiscale feature extraction module connected to the layer as shown in Figure 6.
Recursive multiscale feature extraction module recursively performs the process of extracting features by multiscale feature extraction module to minimize feature loss. It improves accuracy by extracting important but missing features through the feedback network which sends the result of the first step to the input of the second step. The top-down pathway of multiscale feature extraction module is implemented as the unrolled iteration, and can be repeated n times.
The prediction module consists of two stages: the first stage has Region Proposal Network (RPN) and RoIAlign, and the second stage OISB and ODB. It is designed to combine anchor-based RPN and anchor-free ODB. RPN improves detection accuracy by increasing the number of anchor candidates. RoIAlign converts the results of RPN, which have different sizes, into a fixed-sized feature map in order to input them into the fully connected layer. OISB reduces the interference between nearby objects by reducing the edge weight of each object through instance separation kernel. ODB improves accuracy by detecting small objects which are not detected in RPN through anchor-free-based object detection.

4. Experiments

In this section, we shall describe about the performance evaluation for ARTD-Net1 and ARTD-Net2, respectively. Our experimental environment consists of eight GPUS, each NVIDIA A100 40 GB, and one AMD EPYC Processor with 92 single-cores and 1.7TB RAM. Our model was implemented using the Pytorch framework and MMDetection [41] with Python. We use Stochastic Gradient Descent (SGD) as the optimizer and the initial learning rate was set at 0.01. Weight decay of 0.0001 and momentum of 0.9 were used. The batch size was set at 32 (four images per GPU). The parameters σ and γ of Background Weight Adjustment kernel were set to 0.85 and 10, respectively. We use official codes downloaded from MMdetection for performance comparison of our models, and fine-tuned the parameters following the rules presented in the respective papers to achieve the best performance. In the case of YOLOv5 and YOLOv6, we downloaded the models from the official GitHub of authors. All models presented in this paper were trained on the recyclables dataset.
We used F1 score and mean Average Precision (mAP) as evaluation metrics to measure the performance of model. The F1 score is the harmonic mean of precision and recall, and is particularly useful for evaluating model performance on datasets with class imbalance. mAP is evaluation metric in object detection tasks. It is calculated as the mean of the Average Precision (AP) scores for each class.
This section is organized as follows: Section 4.1 describes the recyclables dataset used in the experiments in detail. Section 4.2 evaluates the performance of ARTD-Net1 and ARTD-Net2, respectively.

4.1. Recyclables Dataset

We build our recyclables dataset by focusing on the overlapped multiple wastes of different types with the complex arrangement. It comprises a total of 50,183 images of various resolutions and 110,759 annotation data, and has the largest number of annotations among trash datasets. The resolution of the image is 1280 × 720 on average and 4032 × 2268 at maximum. Several types of waste, up to eight, are placed in one image in order to increase the batch complexity over the previous dataset. Recyclable waste objects are classified into ten classes: paper, paper pack, paper cup, can, bottle, PET, plastic, vinyl, cap, label. Our dataset format follows COCO. Table 1 shows the number of annotations for each class of training and validation dataset.
The cap class has the maximum number of 8462 data and the paper pack class has the second largest number of 7705 data. Cap and label classes have the largest number of annotations as they are obtained from PET or bottles.
As shown in Figure 7, we collect overlapped multiple wastes of different types in various backgrounds and lighting conditions to make the deep learning model robust.

4.2. Performance Evaluation

4.2.1. ARTD-Net1

We make use of our dataset in the previous section for the experiments. Experiments are conducted for F1 Score and mAP using the same backbone, ResNet-50. As shown in Table 2, we compare mAP according to the size of the object: small, medium and large.
First, we investigate the performance of centralized feature extraction module with BWAB in the backbone. It shows that adding BWAB into backbone improves the accuracy by reducing the edge weights around the border for the image. As shown in Table 2, our model with BWAB has higher F1 score and mAP than that without BWAB, showing better performance. Our model has F1 score of 82.043 and mAP of 0.495. It has higher F1 score of 2.12 and mAP of 0.027 than the model without both blocks.
Second, we check the performance of prediction module. It shows that adding OISB into prediction module improves the accuracy by reducing the edge weight of each object. our model with OISB has higher F1 score and mAP than that without OISB, showing better performance. Our model has F1 score of 81.688 and mAP of 0.506. It has higher F1 score of 1.765 and mAP of 0.038 than the model without both blocks.
Our model with both of BWAB and OISB has the highest score for F1 score and mAP.
We compare the F1 score and mAP of ARTD-Net1 with the other one-stage models. As shown in Table 3, we compare F1 score and mAP according to the size of the object: small, medium and large. For F1 score and mAP, ARTD-Net1 with ResNext-101 [42] achieves the highest performance. Among the models excluding ARTD-Net1, YOLOv6 with CSPStackRep as the backbone achieved the highest F1 score, and YOLOX-X with Modified CSPDarknet v5 as the backbone showed the highest mAP. ARTD-Net1 with ResNext-101 has higher F1 score and mAP of 0.275 and 0.004, respectively, than those of YOLOX-X with Modified CSPDarknet v5 as backbone. ARTD-Net1 with ResNext-101 has higher F1 score and mAP of 0.231 and 0.007, respectively than those of YOLOv6 with CSPStackRep as backbone. Among models excluding YOLO series, CornerNet with HourGlass-104 [43] as a backbone has the highest F1 score of 84.724 and CentripetalNet with HourGlass-104 as a backbone has the highest mAP of 0.586. ARTD-Net1 shows the best performance among the one-stage models.

4.2.2. ARTD-Net2

We make use of our dataset for the experiments. Experiments are conducted for F1 Score and mAP using the same backbone, ResNet-101. As shown in Table 4, we compare mAP according to the size of the object: small, medium and large.
First, we investigate the performance of centralized feature extraction module with BWAB in the backbone of ARTD-Net2. It shows that adding BWAB into backbone improves the accuracy by reducing the edge weights around the border for the image. As shown in Table 4, our model with BWAB has higher F1 score and mAP than that without BWAB, showing better performance. Our model has F1 score of 85.31 and mAP of 0.57. It has higher F1 score of 1.44 and mAP of 0.03 than the model without both blocks.
Second, we check the performance of prediction module of ARTD-Net2. It shows that adding OISB into prediction module improves the accuracy by reducing the edge weight of each object. Our model with OISB has a higher F1 score and mAP than that without OISB, showing better performance. Our model has an F1 score of 85.14 and mAP of 0.581. It has a higher F1 score of 1.27 and mAP of 0.041 than the model without both blocks.
For all the cases, F1 Score and mAP achieve the highest score when using both BWAB and OISB.
We compare the F1 score and mAP of ARTD-Net2 with the other multi-stage models. As shown in Table 5, we compare F1 score and mAP according to the size of the object: small, medium and large. For all the cases, ARTD-Net2 with ResNeXt-101 achieves the highest performance. Among the models excluding ARTD-Net2, DetectoRS with ResNeXt-101 as the backbone has the highest performance. ARTD-Net2 with ResNeXt-101 has higher F1 score and mAP of 2.105 and 0.005, respectively than those of DetectoRS with ResNeXt-101 as backbone.

4.2.3. Performance Comparison between ARTD-Net1 and ARTD-Net2

We analyze the two experimental results to compare the performance of ARTD-Net1 and ARTD-Net2: Confusion Matrix, Processing Time and Accuracy.
First, we compare the accuracy of each class by using the confusion matrix after training our model with dataset. Figure 8 shows the accuracy of each class on ARTD-Net1 and ARTD-Net2 when using ResNet-101 as a backbone. According to the confusion matrix, ARTD-Net has significantly higher accuracy for bottle and vinyl. Those two components are not difficult to detect, since their features are clear. Large-sized recyclables are generally well detected. However, in the case of cap, false positives occur frequently, since they are small and similar to the background. In particular, the paper cup class is one of the most difficult to detect among the recyclable classes. The accuracy for paper cup class on ARTD-Net1 is 20%, while the accuracy for paper cup class on ARTD-Net2 is 37%, resulting in an improvement in performance. The inference results of ARTD-Net1 and ARTD-Net2 are shown in Figure 9.
Second, we compare the performance of ARTD-Net1 and ARTD-Net2 through accuracy and processing time of our model. As shown in Figure 10, ARTD-Net1 with ResNet-50 achieves the fastest processing time. ARTD-Net2 with ResNeXt-101 as the backbone has the highest performance. ARTD-Net2 with ResNet-101 has higher accuracy than ARTD-Net1 with ResNeXt-101. However, ARTD-Net1 with ResNeXt-101 has faster processing speed than ARTD-Net2 with ResNet-101.
Finally, we show that ARTD-Net1 is faster than ARTD-Net2 but ARTD-Net2 is more accurate than ARTD-Net1.

5. Conclusions

In this paper, we have proposed two anchor-free-based Recyclable Trash Detection Networks (ARTD-Net) which can recognize overlapped multiple wastes of different types efficiently by using edgeless module: ARTD-Net1 and ARTD-Net2. The former is an anchor-free based one-stage deep learning model which consists of three modules: centralized feature extraction, multiscale feature extraction and prediction. The centralized feature extraction module in the backbone architecture is obtained by adding BWAB to RetinaNet. BWAB exploits a new scheme which efficiently extracts features around the center of the input image while reducing the edge weights around the border of the image. The multiscale feature extraction module provides feature maps of different scales through bottom-up and top-down pathways. The prediction module consists of two blocks: OISB and ODB. OISB reduces the interference between nearby objects by reducing the edge weight of each object through instance separation kernel. ODB improves accuracy by detecting small objects which are not detected in RPN through anchor-free-based object detection. The latter is an anchor-free based multi-stage deep learning model which can efficiently finds each of waste regions by additionally exploiting Region Proposal Network and RoIAlign. It sequentially performs classification and regression to improve accuracy. We have presented a new recyclables dataset which is composed of a large number of high-resolution waste images with additional essential classes. We demonstrated that waste detection performance is improved by providing various images with the complex arrangement of overlapped multiple wastes with different types. The performance of ARTD-Net is confirmed through various experiments. ARTD-Net1 with BWAB and OISB achieves F1 score and mAP of 82.849 and 0.518, respectively, which is 2.926 and 0.05 higher than ARTD-Net1 without BWAB and OISB. Among the one-stage models, ARTD-Net1 with ResNext-101 achieves the highest F1 score and mAP of 85.594 and 0.596, respectively. ARTD-Net2 with BWAB and OISB achieves F1 score and mAP of 86.089 and 0.603, respectively, which is 2.219 and 0.063 higher than ARTD-Net2 without BWAB and OISB. Among the multi-stage models, ARTD-Net2 with ResNext-101 achieves the highest F1 score and mAP of 88.351 and 0.617, respectively. Finally, we demonstrated that ARTD-Net1 is faster than ARTD-Net2 but ARTD-Net2 is more accurate than ARTD-Net1.
For future works, we shall work on the details of ARTD-Net such as the number of blocks for each module and hyperparameters, and continue to collect data under more various lighting and background conditions to use our dataset for various works such as field diagnostics.

Author Contributions

Conceptualization, B.K. and C.-S.J.; methodology, B.K. and C.-S.J.; software, B.K.; validation, B.K. and C.-S.J.; formal analysis, B.K. and C.-S.J.; investigation, B.K.; resources, B.K.; data curation, B.K.; writing—original draft preparation, B.K.; writing—review and editing, B.K. and C.-S.J.; visualization, B.K.; supervision, C.-S.J.; project administration, B.K. and C.-S.J.; funding acquisition, C.-S.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

This work was supported by Artificial intelligence industrial convergence cluster development project funded by the Ministry of Science and ICT(MSIT, Korea)&Gwangju Metropolitan City.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, C.; Qin, J.; Qu, C.; Ran, X.; Liu, C.; Chen, B. A smart municipal waste management system based on deep-learning and Internet of Things. Waste Manag. 2021, 135, 20–29. [Google Scholar] [CrossRef] [PubMed]
  2. Cubillos, M. Multi-site household waste generation forecasting using a deep learning approach. Waste Manag. 2020, 115, 8–14. [Google Scholar] [CrossRef] [PubMed]
  3. Zhang, Q.; Yang, Q.; Zhang, X.; Bao, Q.; Su, J.; Liu, X. Waste image classification based on transfer learning and convolutional neural network. Waste Manag. 2021, 135, 150–157. [Google Scholar] [CrossRef] [PubMed]
  4. Liang, S.; Gu, Y. A deep convolutional neural network to simultaneously localize and recognize waste types in images. Waste Manag. 2021, 126, 247–257. [Google Scholar] [CrossRef] [PubMed]
  5. Abdu, H.; Noor, M.H.M. Domestic Trash Classification with Transfer Learning Using VGG16. In Proceedings of the 2022 IEEE 12th International Conference on Control System, Computing and Engineering (ICCSCE), Penang, Malaysia, 21–22 October 2022; pp. 137–141. [Google Scholar]
  6. Fan, M.; Xiao, L.; He, X.; Chen, Y. Trash Classification Network Based on Attention Mechanism. In Proceedings of the 2022 15th International Conference on Advanced Computer Theory and Engineering (ICACTE), Hangzhou, China, 23–25 September 2022; pp. 19–22. [Google Scholar]
  7. Ahmad, K.; Khan, K.; Al-Fuqaha, A. Intelligent fusion of deep features for improved waste classification. IEEE Access 2020, 8, 96495–96504. [Google Scholar] [CrossRef]
  8. Kumsetty, N.V.; Nekkare, A.B.; Kamath, S.S.; Kumar, M.A. Trashbox: Trash detection and classification using quantum transfer learning. In Proceedings of the 2022 31st Conference of Open Innovations Association (FRUCT), Helsinki, Finland, 27–29 April 2022; pp. 125–130. [Google Scholar]
  9. Q. Tomas, J.P.; D. Celis, M.N.; B. Chan, T.K.; A. Flores, J. Trash detection for computer vision using scaled-yolov4 on water surface. In Proceedings of the 11th International Conference on Informatics, Environment, Energy and Applications, Xi’an, China, 4–6 March 2022; pp. 1–8. [Google Scholar]
  10. Aral, R.A.; Keskin, Ş.R.; Kaya, M.; Hacıömeroğlu, M. Classification of trashnet dataset based on deep learning models. In Proceedings of the 2018 IEEE International Conference on Big Data (Big Data), Seattle, WA, USA, 10–13 December 2018; pp. 2058–2062. [Google Scholar]
  11. Yang, M.; Thung, G. Classification of trash for recyclability status. CS229 Proj. Rep. 2016, 2016, 3. [Google Scholar]
  12. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  13. Bircanoğlu, C.; Atay, M.; Beşer, F.; Genç, Ö.; Kızrak, M.A. RecycleNet: Intelligent waste sorting using deep neural networks. In Proceedings of the 2018 Innovations in Intelligent Systems and Applications (INISTA), Thessaloniki, Greece, 3–5 July 2018; pp. 1–7. [Google Scholar]
  14. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  15. Yamashita, R.; Nishio, M.; Do, R.K.G.; Togashi, K. Convolutional neural networks: An overview and application in radiology. Insights Imaging 2018, 9, 611–629. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  17. Ruiz, V.; Sánchez, Á.; Vélez, J.F.; Raducanu, B. Automatic image-based waste classification. In From Bioinspired Systems and Biomedical Applications to Machine Learning: Proceedings of the 8th International Work-Conference on the Interplay between Natural and Artificial Computation, IWINAC 2019, Almería, Spain, 3–7 June 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 422–431. [Google Scholar]
  18. Vo, A.H.; Vo, M.T.; Le, T. A novel framework for trash classification using deep transfer learning. IEEE Access 2019, 7, 178631–178639. [Google Scholar] [CrossRef]
  19. Proença, P.F.; Simoes, P. Taco: Trash annotations in context for litter detection. arXiv 2020, arXiv:2003.06975. [Google Scholar]
  20. Wang, T.; Cai, Y.; Liang, L.; Ye, D. A multi-level approach to waste object segmentation. Sensors 2020, 20, 3816. [Google Scholar] [CrossRef] [PubMed]
  21. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  22. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  23. Jocher, G.; Chaurasia, A.; Stoken, A.; Borovec, J.; Kwon, Y.; Michael, K.; Fang, J. ultralytics/yolov5: v6.2-yolov5 classification models, apple m1, reproducibility, clearml and deci. ai integrations. Zenodo.org 2022. [Google Scholar] [CrossRef]
  24. Chen, Q.; Wang, Y.; Yang, T.; Zhang, X.; Cheng, J.; Sun, J. You only look one-level feature. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13039–13048. [Google Scholar]
  25. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. Yolox: Exceeding yolo series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
  26. Li, C.; Li, L.; Jiang, H.; Weng, K.; Geng, Y.; Li, L.; Ke, Z.; Li, Q.; Cheng, M.; Nie, W.; et al. YOLOv6: A single-stage object detection framework for industrial applications. arXiv 2022, arXiv:2209.02976. [Google Scholar]
  27. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part I 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 21–37. [Google Scholar]
  28. Zhu, C.; He, Y.; Savvides, M. Feature selective anchor-free module for single-shot object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019; pp. 840–849. [Google Scholar]
  29. Tian, Z.; Shen, C.; Chen, H.; He, T. Fcos: Fully convolutional one-stage object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9627–9636. [Google Scholar]
  30. Wang, J.; Zhang, W.; Cao, Y.; Chen, K.; Pang, J.; Gong, T.; Shi, J.; Loy, C.C.; Lin, D. Side-aware boundary localization for more precise object detection. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part IV 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 403–419. [Google Scholar]
  31. Wang, X.; Zhang, R.; Kong, T.; Li, L.; Shen, C. Solov2: Dynamic and fast instance segmentation. Adv. Neural Inf. Process. Syst. 2020, 33, 17721–17732. [Google Scholar]
  32. Law, H.; Deng, J. Cornernet: Detecting objects as paired keypoints. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 734–750. [Google Scholar]
  33. Dong, Z.; Li, G.; Liao, Y.; Wang, F.; Ren, P.; Qian, C. Centripetalnet: Pursuing high-quality keypoint pairs for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 10519–10528. [Google Scholar]
  34. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Cai, Z.; Vasconcelos, N. Cascade r-cnn: Delving into high quality object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6154–6162. [Google Scholar]
  36. Wu, Y.; Chen, Y.; Yuan, L.; Liu, Z.; Wang, L.; Li, H.; Fu, Y. Rethinking classification and localization for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 10186–10195. [Google Scholar]
  37. Sun, P.; Zhang, R.; Jiang, Y.; Kong, T.; Xu, C.; Zhan, W.; Tomizuka, M.; Li, L.; Yuan, Z.; Wang, C.; et al. Sparse r-cnn: End-to-end object detection with learnable proposals. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Online, 19–25 June 2021; pp. 14454–14463. [Google Scholar]
  38. Qiao, S.; Chen, L.C.; Yuille, A. Detectors: Detecting objects with recursive feature pyramid and switchable atrous convolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 10213–10224. [Google Scholar]
  39. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  40. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  41. Chen, K.; Wang, J.; Pang, J.; Cao, Y.; Xiong, Y.; Li, X.; Sun, S.; Feng, W.; Liu, Z.; Xu, J.; et al. MMDetection: Open mmlab detection toolbox and benchmark. arXiv 2019, arXiv:1906.07155. [Google Scholar]
  42. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated residual transformations for deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1492–1500. [Google Scholar]
  43. Newell, A.; Yang, K.; Deng, J. Stacked hourglass networks for human pose estimation. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part VIII 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 483–499. [Google Scholar]
Figure 1. ARTD-Net1.
Figure 1. ARTD-Net1.
Sensors 23 02907 g001
Figure 2. Centralized feature extraction module.
Figure 2. Centralized feature extraction module.
Sensors 23 02907 g002
Figure 3. Multiscale feature extraction module and feature pyramid network.
Figure 3. Multiscale feature extraction module and feature pyramid network.
Sensors 23 02907 g003
Figure 4. Prediction module.
Figure 4. Prediction module.
Sensors 23 02907 g004
Figure 5. ARTD-Net2.
Figure 5. ARTD-Net2.
Sensors 23 02907 g005
Figure 6. Recursive multiscale feature extraction module.
Figure 6. Recursive multiscale feature extraction module.
Sensors 23 02907 g006
Figure 7. Sample images of dataset.
Figure 7. Sample images of dataset.
Sensors 23 02907 g007
Figure 8. Confusion matrix of ARTD-Net1 and ARTD-Net2.
Figure 8. Confusion matrix of ARTD-Net1 and ARTD-Net2.
Sensors 23 02907 g008
Figure 9. Inference result of ARTD-Net1, ARTD-Net2 and existing deep learning model.
Figure 9. Inference result of ARTD-Net1, ARTD-Net2 and existing deep learning model.
Sensors 23 02907 g009
Figure 10. Processing time (ms/img) and accuracy (mAP) using recyclables dataset on ARTD-Net1 and ARTD-Net2.
Figure 10. Processing time (ms/img) and accuracy (mAP) using recyclables dataset on ARTD-Net1 and ARTD-Net2.
Sensors 23 02907 g010
Table 1. Annotation data for each category and class.
Table 1. Annotation data for each category and class.
CategoryPaperPaper PackPaper CupCanBottlePETPlasticVinylCapLabel
Train920815,364786786619075559312,759946818,04313,419
Validation1837665139145122137152112170
Total939115,440793288009220571512,896962018,15513,589
Table 2. Performance of ARTD-Net1, S: small, M: medium, L: large.
Table 2. Performance of ARTD-Net1, S: small, M: medium, L: large.
ModelF1 ScoremAPmAP 50mAP 75
mAP SmAP MmAP L
Without BWAB and OISB79.9230.4680.6610.541
0.1450.3970.561
With BWAB Without OISB82.0430.4950.6770.564
0.1590.4020.573
Without BWAB With OISB81.6880.5060.6630.565
0.160.3980.593
With BWAB and OISB82.8490.5180.6780.585
0.1590.4260.625
Table 3. Comparison between ARTD-Net1 and other one-stage deep learning models.
Table 3. Comparison between ARTD-Net1 and other one-stage deep learning models.
ModelBackboneF1 ScoremAPmAP50mAP75mAPsmAPmmAPl
SSD300VGG-1675.5730.3070.5110.3260.0450.2550.377
SSD512VGG-1675.9520.3380.5430.3570.0510.290.401
RetinaNetResNet-5077.1610.3530.5670.3840.0310.3080.481
RetinaNetResNet-10178.2080.3730.580.40.0480.3250.507
FCOSResNet-5081.460.520.6980.5770.1440.4670.612
FCOSResNet-10182.6230.5250.6950.5820.170.4350.629
FSAFResNet-5080.2120.4790.6570.5390.1470.4060.564
FSAFResNet-10180.630.4960.6830.5730.1650.4320.595
SABLResNet-5079.8050.4750.6240.5310.1550.3860.592
SABLResNet-10180.0830.5130.6420.570.1750.3970.605
CornerNetHourGlass-10484.7240.5750.7110.6310.1810.5170.636
CentripetalNetHourGlass-10479.3260.5860.7420.6390.1980.5020.657
SOLOv2ResNet-5080.710.4850.6730.5330.1090.3920.59
SOLOv2ResNet-10181.6340.5250.7050.5720.1270.4240.619
YOLOv3DarkNet-5371.3790.2390.4110.250.0460.1750.367
YOLOv4CSP-Darkent5381.2720.4990.6520.5440.1480.3960.655
YOLOv5Modified CSP v584.7620.5870.7370.6190.1980.4710.663
YOLOFResNet-5078.2030.3550.5130.370.1690.3120.467
YOLOX-XModified CSP v585.3190.5920.7780.6690.190.5080.699
YOLOv6CSPStackRep85.3630.5890.7610.6810.1970.490.716
ARTD-Net1ResNet-5082.8490.5180.6780.5850.1590.4260.625
ARTD-Net1ResNet-10184.8990.5450.7110.6210.1930.4390.666
ARTD-Net1ResNeXt-10185.5940.5960.7750.6770.1780.4830.733
Table 4. Performance of ARTD-Net2, S: small, M: medium, L: large.
Table 4. Performance of ARTD-Net2, S: small, M: medium, L: large.
ModelF1 ScoremAPmAP 50mAP 75
mAP SmAP MmAP L
Without BWAB and OISB83.870.540.6970.61
0.1630.4340.624
With BWAB Without OISB85.310.570.7420.635
0.1820.4450.638
Without BWAB With OISB85.140.5810.7560.62
0.1940.4620.645
With BWAB and OISB86.0890.6030.7610.646
0.2110.4790.652
Table 5. Comparison between ARTD-Net2 and other multi-stage deep learning models.
Table 5. Comparison between ARTD-Net2 and other multi-stage deep learning models.
ModelBackboneF1 ScoremAPmAP50mAP75mAPsmAPmmAPl
Faster R-CNNResNet-10178.1490.4880.6380.5340.1360.3860.643
Faster R-CNNResNeXt-10180.7330.5210.6770.5880.1670.4180.66
Cascade R-CNNResNet-10181.3730.5290.6860.5750.1520.410.653
Cascade R-CNNResNeXt-10183.5680.5460.6990.5960.1740.4330.68
Double-Head R-CNNResNet-10182.8410.570.710.6110.1720.4420.649
Double-Head R-CNNResNeXt-10184.7790.5830.7530.6430.1950.4860.664
Sparse R-CNNResNet-10183.7170.5780.7150.6380.1760.4760.655
Sparse R-CNNResNeXt-10185.6350.5930.7660.6530.2080.4910.673
DetectoRSResNet-10184.90.580.7440.6470.1940.4890.676
DetectoRSResNeXt-10186.2460.6120.7820.6590.2240.5070.686
ARTD-Net2ResNet-10186.0890.6030.7610.6460.2110.4790.652
ARTD-Net2ResNeXt-10188.3510.6170.8060.6650.2330.5110.711
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kang, B.; Jeong, C.-S. ARTD-Net: Anchor-Free Based Recyclable Trash Detection Net Using Edgeless Module. Sensors 2023, 23, 2907. https://doi.org/10.3390/s23062907

AMA Style

Kang B, Jeong C-S. ARTD-Net: Anchor-Free Based Recyclable Trash Detection Net Using Edgeless Module. Sensors. 2023; 23(6):2907. https://doi.org/10.3390/s23062907

Chicago/Turabian Style

Kang, BoSeon, and Chang-Sung Jeong. 2023. "ARTD-Net: Anchor-Free Based Recyclable Trash Detection Net Using Edgeless Module" Sensors 23, no. 6: 2907. https://doi.org/10.3390/s23062907

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop