Next Article in Journal
Medical Residents’ Behaviours toward Compulsory COVID-19 Vaccination in a Tertiary Hospital in Italy
Previous Article in Journal
The Indirect Role of Passive-Avoidant and Transformational Leadership through Job and Team Level Stressors on Workplace Cyberbullying
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Intelligent Waste-Sorting and Recycling Device Based on Improved EfficientNet

1
Department of Electrical Engineering and Automation, Jiangxi University of Science and Technology, Ganzhou 341000, China
2
Jiangxi Provincial Key Laboratory of Maglev Technology, Ganzhou 341000, China
3
Ganjiang Innovation Academy, Chinese Academy of Sciences, Ganzhou 341000, China
4
Department of Science, Jiangxi University of Science and Technology, Ganzhou 341000, China
*
Author to whom correspondence should be addressed.
Int. J. Environ. Res. Public Health 2022, 19(23), 15987; https://doi.org/10.3390/ijerph192315987
Submission received: 5 November 2022 / Revised: 24 November 2022 / Accepted: 28 November 2022 / Published: 30 November 2022
(This article belongs to the Section Environmental Science and Engineering)

Abstract

:
The main source of urban waste is the daily life activities of residents, and the waste sorting of residents’ waste is important for promoting economic recycling, reducing labor costs, and protecting the environment. However, most residents are unable to make accurate judgments about the categories of household waste, which severely limits the efficiency of waste sorting. We have designed an intelligent waste bin that enables automatic waste sorting and recycling, avoiding the extensive knowledge required for waste sorting. To ensure that the waste-classification model is high accuracy and works in real time, GECM-EfficientNet is proposed based on EfficientNet by streamlining the mobile inverted bottleneck convolution (MBConv) module, introducing the efficient channel attention (ECA) module and coordinate attention (CA) module, and transfer learning. The accuracy of GECM-EfficientNet reaches 94.54% and 94.23% on the self-built household waste dataset and TrashNet dataset, with parameters of only 1.23 M. The time of one recognition on the intelligent waste bin is only 146 ms, which satisfies the real-time classification requirement. Our method improves the computational efficiency of the waste-classification model and simplifies the hardware requirements, which contributes to the residents’ waste classification based on intelligent devices.

1. Introduction

In recent years, as urbanization and living standards have increased, the variety and quantity of waste has increased dramatically [1], putting enormous pressure on resource use, environmental safety, and economical recycling. Urban residents are the main producers of household waste, and they participate in household waste sorting. Recycling is an effective way to utilize waste resources, reduce the quantity of waste, and contribute to sustainable development [2]. However, due to many different categories of waste, extensive sorting knowledge is required. It is difficult to translate residents’ willingness for waste sorting into actual action. As a result, many countries have started researching intelligent waste sorting and recycling devices, which were applied in engineering practice [3,4]. Intelligent recognition of waste categories is a prerequisite for sorting and recycling. Computer vision technology and deep learning technology can automatically detect and classify waste categories [5,6], providing technical support for waste sorting and recycling.
The convolutional neural network (CNN) is one of the main branches of deep learning, and it is the mainstream image-recognition method nowadays. With the rapid development of deep learning technology, CNN has made significant achievements in image classification [7]. Numerous researchers have used the CNN to solve waste image classification tasks [8,9], and have achieved a series of achievements. Ref. [10] improved ResNet18 with a self-monitoring module to enhance the feature map representation, achieving 95.87% accuracy on the TrashNet dataset. Ref. [11] optimized DenseNet121 with data augmentation and genetic algorithms, achieving 99.6% accuracy on the TrashNet dataset. Ref. [12] implemented waste image classification by ResNet50, achieving 95.3% accuracy on the self-built dataset. It is feasible to achieve accurate and reliable waste classification by CNN. However, the above research articles use the large CNN model, which performs well on accuracy. However, they have many parameters, and the inference of the models requires many floating point operations [13,14]. It is impractical to have high-performance chips in every waste-sorting device, which makes their application difficult. Therefore, aiming at embedded devices or platforms with limited resources, numerous researchers have started to explore lightweight CNN models for waste image classification, which are light, efficient, and have acceptable performance. Ref. [15] implemented waste image classification by MobileNetv3 and achieved 94.26% accuracy on the self-built dataset, with a single infer time of only 261.7 ms. Ref. [16] proposed the GCNet based on ShuffleNetv2, which was improved by parallel space and channel attention module, FRelu, and transfer learning, the model parameters were 1.3 M, and the single infer time on the Raspberry Pi 4B was only 105 ms, with an accuracy of 97.9% on the self-built dataset.
The scholars mentioned above have researched waste image classification and verified the effectiveness of CNN, which has practical implications for automatic waste classification and recycling. However, most of these research efforts are based on large models to implement waste image classification, with less focus on lightweight and actual applications, which is not conducive to them being used in real life. Therefore, the aim and novelty of this paper is the proposition of a lightweight and efficient model for waste image classification, which is applied to the intelligent waste bin that we designed. Our proposed method can automatically sort and collect the input household waste, including recyclable waste, hazardous waste, kitchen waste, and other waste, which provides a feasible solution to residents for waste sorting.
Overall, the main contributions of this work include the following.
(1)
An intelligent waste bin has been designed, which can automatically collect the waste put in, improve the efficiency of waste sorting by residents, and reduce the separation work of collection facilities.
(2)
We propose an improved EfficientNet, named GECM-EfficientNet, which accurately classifies different categories of waste by fewer parameters.
(3)
We use transfer learning [17] to initialize the model parameters during training, optimizing the performance of the model without adding extra computation.
(4)
Our waste-classification model balances speed and accuracy with good real-time performance on edge devices, which can reduce hardware costs.
The paper is structured as follows: Section 2 reviews the work related to the large model and lightweight model. Section 3 describes the working process of the intelligent waste bin and the detailed design of the proposed model. Section 4 describes the experimental results and analysis of different datasets. Finally, Section 5 concludes the work of this paper.

2. Related Work

Waste-sorting and recycling devices require frequent forward inference of models, which can be computationally expensive. Unfortunately, it is impractical to equip each device with a high-performance graphics processing unit (GPU), which results in high hardware costs. Cloud deployment [18,19] can free models from reliance on local computation, but this is heavily dependent on the Internet. When there is no network or poor network connectivity, cloud deployment is not possible. Compared with large models, lightweight models tend to be slightly less accurate, but smaller and faster. Therefore, for waste sorting and recycling devices, local deployment of lightweight models would be an effective measure.
In 2012, AlexNet [20] won the ImageNet image classification competition, triggering the boom in CNN for image classification, giving rise to a series of models, such as VGG [21], GoogleNet [22], and ResNet [23], with superior performance. Ref. [24] implements three VGGs in tandem for electrocardiogram classification, achieving high accuracy of 97.23% on the PTB-XL dataset. Ref. [25] used transfer learning to improve the VGG, achieving 98.4% and 95.71% accuracy on the self-built grape and tomato pest datasets. Refs. [26,27] achieve classification of plants and botrytis by ResNet, with an accuracy of over 98%. Ref. [28] initialized GoogleNet through transfer learning, which achieved up to 99.94% accuracy on the self-built northern maize leaf blight dataset. These papers implemented image classification through large models. They excelled in accuracy, all achieving over 95% accuracy. However, large models require large memory and hardware resources, which hinders their usefulness on resource-limited embedded devices. For this, lightweight models would be a viable solution.
In 2016, ref. [29] first used lightweight ideas to design models and proposed SqueezeNet, with a model size of 0.5 M. After that, numerous developers continued to explore lightweight models, proposing MobileNet [30,31], ShuffleNet [32,33], and EfficientNet [34] (EfficientNetB0- EfficientNetB7). Among the mainstream models, EfficientNet achieves the most excellent ImageNet accuracy and has a highly efficient execution; thus, it is widely used in image classification. Ref. [35] improves EfficientNetB0 by adjusting the number of MBConv modules and using the residual structure and LeakRelu, with the parameter of 1.03 M, achieving 99.69% accuracy on the self-built human behavioral point cloud dataset. Ref. [36] initialized the weights of EfficientNetB4 by transfer learning, achieving the plant nutrient deficiency diagnosis with an accuracy of 98.52% on the DND-SB dataset. Ref. [37] embeds the spatial attention module into EfficientNetB4, which improves the accuracy by about 1% on the RFMID dataset (fundus disease dataset). Ref. [38] improves EfficientNetB0 based on the convolutional block attention module (CBAM) and coordinate attention (CA) module, improving the classification accuracy by 3.5% on the self-built cervical cancer dataset. Ref. [39] implements the printed circuit board (PCB) classification recovery model based on EfficientNetB3, improved by transfer learning, and achieves the accuracy of 94.37% on the PCB DSLR dataset.
Inspired by the many areas of EfficientNet, we chose EfficeientNet as the baseline model. Among several versions of EfficientNet, we prioritized real-time and chose EfficientNetB0, which has the fewest parameters. Based on this, we focused on exploring the application of EfficientNetB0 to waste-sorting tasks and making improvements.

3. Materials and Methodology

3.1. Waste Sorting Device

3.1.1. Hardware Structure

In this research, we first used SolidWorks to build the mechanical structure of the intelligent waste bin. Then, the actual intelligent waste bin was built through hardware devices, Figure 1 shows its mechanical modelling, simulation modelling, and physical construction. The intelligent waste bin consists of three parts: waste recognition, actuator device, and collection device. In the waste-recognition process, the camera captures image frames, which are then passed into the classification model to identify the waste category. The actuator device consists of two servos and the attached paddle plate and baffle plate, which sort the waste into the bins by rotating at different angles. The collection device is four fan-shaped waste bins, which are set up for recyclable waste, hazardous waste, kitchen waste, and other waste according to the standards in the literature [40]. Note that the camera is HF867 with the following parameters: maximum resolution 1280 × 720; frame rate 30/s; sensitivity 39 db.
The intelligent waste bin works as follows. First, an infrared sensor senses the waste input. Then, the camera takes the image of the waste, which is passed into the waste-classification model for recognition. Finally, based on the recognition result, the servo controls the paddle plate and baffle plate to sort the waste into the corresponding bins. After the sorting is complete, the paddle plate and baffle plate return to their original position.

3.1.2. Control Circuit

The Raspberry Pi 4B [41,42] is widely used in the intelligence field. We use it as the main control device in this research. The Raspberry Pi 4B processor is BROADCOM BCM2711 with 4-core CORTEX-A72, frequency is 1.5 GHZ, memory is 8 GB, and 40 expandable pins. This paper uses the Raspberry Pi 4B to deploy the waste classification model and control the infrared sensor, camera, and servo. The control circuit is shown in Figure 2.

3.2. Dataset Used

3.2.1. Self-Built Dataset

For the task of waste classification, there is no large dedicated dataset yet. According to the literature [40] on waste classification standards, this paper establishes a scene-rich household waste dataset, and Table 1 shows the details of the dataset. Our dataset contains 18 categories of household waste, with a total of 7361 images, which are classified as recyclable waste, hazardous waste, kitchen waste, and other waste. The dataset is collected from the Internet and photography, and it is our baseline dataset. We will construct the waste-sorting device based on the dataset.

3.2.2. Trashnet Dataset

This paper selects the TrashNet dataset [43] to validate the model performance. The TrashNet dataset is a small public dataset with only 2527 images, widely used in waste image classification tasks. It has six categories of waste images, namely cardboard, glass, metal, paper, plastic, and trash.

3.3. The Basics of Efficientnet

Typically, CNN is developed with a fixed resource budget. If resources are increased, the network can be extended to improve performance. The common method is to scale the model depth, width, and resolution. For example, by deepening the number of layers, ResNet can construct a model with 200 layers. MobileNet v2 sets the scaling factor and resolution factor, which adjusts the number of channels and the input size. Therefore, Google proposed a composite scaling method to improve the model performance, simultaneously scaling the depth, width, and input image resolution. Google designed the baseline model EfficientNetB0 and then scaled it to obtain EfficientNetB1-EfficientNetB7. On the ImageNet dataset, EfficientNet achieved contemporaneous advanced accuracy and speed.
EfficientNet adopts the mobile inverted bottleneck convolution (MBConv [31]) as the basic module, and uses the squeeze-and-excitation (SE) module [44] to calibrate the feature map by the importance of the channels. Figure 3 shows the structure of MBConv. C i and C o are input channel and output channel, respectively. H and W are height and width of the feature map, respectively. DWConv is a deepwise convolution, the kernel size of which is K. BN denotes batch normalization. Swish and Sigmoid are the activation function. In the SE module, C represents the channel of the feature map, and r is the parameter used for dimensionality reduction, which is set to 4. First, the input channels are augmented by 1 × 1 convolution (pointwise convolution, PW), then achieved feature extraction by 3 × 3 or 5 × 5 deepwise convolution. In the SE module, the global features are extracted through global average pooling, and then channel weights are obtained with the two fully connected layers and sigmoid. The channel weights and the feature map are multiplied channel by channel, which implements the channel weighting operation. Finally, the channels of the feature map are adjusted by PW. Shortcut and dropout are used only when the input channel and output channel are equal, and the stride is 1.

3.4. Improved Efficientnet

The intelligent waste bin is used daily, requiring high accuracy and real-time waste classification models. EfficientNetB0 is a simple and elegant model, which combines the advantages of MobileNetv2, but with more efficient feature-extraction capabilities. This paper designed a lightweight and efficient waste image classification model based on EfficientNetB0, named GECM-EfficientNet. The network structure is shown in Figure 4. First, the number of MBConv modules is adjusted, which reduces the model parameters. Secondly, we use the effiicient channel attention (ECA) module [45] to replace the SE module, which solves the shortcomings of the dimensionality reduction operation. Subsequently, parallel connections are made between the coordinate attention (CA) module [46] and the ECA module, enabling spatial weighting operation. Finally, transfer learning is used to initialize the model parameters during training.

3.4.1. Optimising the Network Structure

EfficientNetB0 repeatedly stacks the MBConv module to obtain excellent feature-extraction capabilities. As the network deepens, more channels in the convolutional layer are used to generate more detailed filters, which results in larger parameters. With the structure of the MBConv modules unchanged, the number of MBConv modules was reduced in the deeper network layers, which better balanced the accuracy and parameters of the model. The adapted model was named G-EfficientNet. Table 2 shows a comparison in network structure between EfficientNet and G-EfficientNet.

3.4.2. Efficient Channel Attention

Based on SENet, ECANet proposes the ECA module. ECANet shows that the dimensionality reduction operation in the SE module carries side effects. The ECA module implements a local cross-channel interaction strategy without dimensionality reduction by one-dimensional convolution, effectively improving performance with lower parameters. In this paper, the SE module is replaced by an ECA module, which avoids the side effects of the dimensionality reduction operation and with fewer parameters.
Figure 5 shows the working process of the ECA module. First, to obtain the global features without dimensionality reduction, the global average pooling (GAP) is performed on the input feature map. Then, channel weights are generated by one-dimensional convolution and sigmoid function. Finally, the channel weights and the input feature map are multiplied channel by channel, which implements the channel weighting operation.
In the operation of GAP, the input feature map is compressed by global average pooling. After completing GAP, we can obtain global features of dimension 1 × 1 × C . Equation (1) shows the calculation process:
z c = 1 H × W i = 1 H j = 1 W u c ( i , j ) , z R c .
In Equation (3), z c denotes the output features on each channel that uses global average pools. H and W denote the height and width of the input feature map. i and j denote the feature value coordinates of the input feature map.
After implementing GAP, the cross-channel interaction strategy is implemented by one-dimensional convolution with size k. The parameter k is generated through an adaptive function, representing the coverage of the local cross-channel interaction. Equation (2) demonstrates the calculation principle. In Equation (2), C is the total number of channels, and | x | odd denotes the nearest odd number to x. Finally, channel weights are generated by sigmoid and multiplied with the input feature map by channel. We have
k = φ ( C ) = log 2 C + 1 2 o d d .

3.4.3. Coordinate Attention

The ECA module only considers channel weight assignment, which ignores the feature map’s spatial information. CA module embeds location information into the channel weighting operation, allowing the feature map to be weighted by spatial and channel information. Figure 6 shows the structure of CA module. In this paper, we connect the CA module and the ECA module in parallel not only to calibrate the feature maps based on channel information, but also to introduce spatial information.
First, the CA module implements location information embedding for the input feature map X. The dimension of X is H × W × C . Average pooling is used for each channel along the horizontal and vertical directions, with pooling kernels the size of (H,1) and (1,W). In the cth channel, Equations (3) and (4) show the output of the hth row and wth column:
z c h ( h ) = 1 W 0 i < W x c ( h , i )
z c w ( w ) = 1 H 0 j < H x c ( j , w ) .
The components of the input feature map are x c (h,i) and x c (j,w), and the coordinates of the components are (h,i) and (j,w), and the channel is c. z h and z w denote the average pooled output along the horizontal and vertical directions. z c h (h) and z c w (w) denote the output component of the cth channel in row h and column w.
Equation (5) shows the next steps. First, the feature map is concatenated, obtained by the pooling operation. Next, the channels are compressed through a standard convolution F 1 with 1 × 1 . Finally, the intermediate output m is obtained through a nonlinear activation layer δ , choosing h-swish as the activation function:
m = δ ( F 1 ( [ z h , z w ] ) ) .
Then, the intermediate output m is sliced into two feature maps along the spatial dimension. The feature maps are represented as m h and m w , and two 1 × 1 standard convolutions F h and F w transform them to the same channels with the input feature map X. The activation is then performed by the sigmoid function ( σ ). The calculation is shown in Equations (6) and (7):
g h = σ ( F h ( m h ) )
g w = σ ( F w ( m w ) ) .
In the above equation, g h and g w represent the coordinate attention weights along the horizontal and vertical directions. The final output of the CA module is shown in Equation (8), where x c (i,j) and y c (i,j) correspond to the values in the input and output feature maps with the coordinate (i,j) and the channel c:
y c ( i , j ) = x c ( i , j ) × g c h ( i ) × g c w ( j ) .

3.4.4. Transfer Learning

Transfer learning [47,48] allows knowledge learned in different domains or tasks to be transferred, which can reduce training time and improve performance. In transfer learning, domain D is the subject of learning. The domain is divided into the source domain D s and the target domain D t . They consist of the data X and the probability distribution P(X) that generates X, which can be expressed as D = X , P ( X ) . Task T is the goal of learning, divided into the source task T s and the target task T t . The task consists of the label space Y and the prediction function f(·), which can be expressed as T = Y , f ( · ) .
Given the source domain D s and the source task T s , the target domain D t and the target task T t . With D s D t or T s T t , transfer learning solves the target task T t in the target domain D t , through knowledge learned in the source domain D s and the source task T s . This paper implements transfer learning, through the weights of EfficientNetB0 on the ImageNet dataset.

3.5. Experimental Settings

During the training, the process is accelerated by the Tesla P100. The dataset is divided into the training dataset and the test dataset according to 8:2. During model training, the generalization capability of the model is enhanced by data augmentation, with such measures as random size cropping, flipping, and luminance transformations. Adam was chosen as the model optimizer. The learning rate prevented overfitting with cosine annealing [49], the initial learning rate was 0.001, and the cosine annealing parameter was 0.01. The loss function was chosen as cross-entropy. The training period is 200, and one batch is trained with 16 images. This paper sets up the following experiments for analysis and discussion.
(1)
Ablation experiments of the improved model, verifying each improvement’s contribution to the model performance.
(2)
Comparison experiments between the improved model and the mainstream model. All models were trained and tested on both the self-built dataset and the TrashNet dataset, verifying the level of advancement of the improved models.
(3)
Model classification accuracy and inference time test. The model was deployed on a Raspberry Pi 4B for testing, verifying the accuracy and real-time performance of the model.

4. Results and Discussion

4.1. Ablation Experiments

To validate the contribution of each improvement. We selected the top accuracy and parameters as metrics, and experiments were conducted on the self-built dataset. In order to demonstrate that the network structure can be optimized by streamlining the number of MBConv modules, we compare G-EfficientNet with EfficientNetB0. To prove that the ECA module is lighter and more efficient, the SE module was replaced with an ECA module for the experiments. To prove that the CA module can introduce spatial information, the SE and CA modules are connected in parallel. To prove that transfer learning can optimize the model parameters. We initialized the model parameters by EfficientNetB0 weights, trained on the ImageNet dataset. Figure 7 shows the training loss and test accuracy curves of the model. L is the training loss, which is calculated on the training dataset. A is the test accuracy, which is calculated on the test dataset. E is training epoch. Obviously, GECM-EfficientNet achieves the best test accuracy and converges quickly.
Table 3 shows the performance parameters of the above model. A is the top1 accuracy of the model, and P represents the number of parameters of the mode. As can be seen, the parameters of G-EfficientNet are reduced by 72.2% compared to EfficientNetB0, but the accuracy is only reduced by 1.27%. The SE module is replaced with the ECA module, which reduces parameters by 0.1 M and improves accuracy by 1.48%. There is a parallel connection of the CA module and the SE module, with only 0.2 M additional parameters, but with a 1.80% increase in accuracy. Optimizing the model parameters by transfer learning, the parameters remain the same with a 4.53% improvement in accuracy.
First, the number of MBConv modules in EfficientNetB0 is adjusted, which acquires the lighter model G-EfficientNet, Next, we improve the MBConv module. We replaced the SE module with the ECA module, connecting the CA module in parallel. During model training, the model parameters are initialized by transfer learning. Ultimately, GECM-EfficientNet was designed. Compared to EfficientNetB0, the model accuracy was improved by 5.7% on the self-built dataset, with 69.73% fewer parameters.

4.2. Comparison and Analysis of Models

To verify the level of advancement of the improved model, GECM-EfficientNet was compared with the mainstream model, with experiments completed on both the self-built dataset and TrashNet dataset. Finally, the models were deployed on the Raspberry Pi 4B for classification and inference time test. We selected lightweight models such as EfficientNetB0, MobileNetv2, MobileNetv3 [50], and ShuffleNetv2, and large models such as GoogleNet, DenseNet121, ResNet50, Inceptionv3, and VGG16.

4.2.1. On the Self-Built Dataset

Figure 8 shows the training loss and test accuracy curves on the self-built dataset. As can be seen, GECM-EfficientNet is in the lead, achieving an accuracy of approximately 90% in only 20 epochs.
The test accuracy, parameters, and single inference time of models are shown in Table 4, where T represents the single inference time of model on the Raspberry Pi 4B. As can be seen, the lightweight model achieves similar or higher accuracy than the large network, with fewer parameters and high real-time performance. Among the mainstream models, EfficientNetB0 achieved the highest accuracy (88.81%) and low parameters (4.03 M), with single inference taking 0.2 s, second only to MobileNetv3, ShuffleNetv2 1 × , and MobileNetv2. Our proposed GECM-EfficientNet, with only 1.23 M parameters, achieves 94.54% accuracy with a single inference time of 146 ms. The reasons for this analysis are as follows: (1) We adapt the number of MBConv modules to obtain G-EfficientNet, which is a lightweight and excellent model. (2) The ECA module is lighter than the SE module, eliminating dimensionality reduction operations’ side effects. (3) Embedding of location information through the CA attention module, which allows for spatially weighted operations. (4) Optimization of model parameters through EfficientNetB0 weights on the ImageNet dataset, which speeds up convergence and improves accuracy.

4.2.2. On the Trashnet Dataset

To further validate the advances of the improved model, comparison experiments were set up on the TrashNet dataset. Figure 9 shows the training loss and test accuracy curves of models. It is evident that GECM-EfficientNet achieves the highest accuracy and converges quickly.
The experimental results are shown in Table 5. Among the mainstream models, the proposed GECM-EfficientNet is in the leading position. It achieved the highest accuracy (94.23%) and the lowest parameters (1.23 M). As can be verified, GECM-EfficientNet is a lightweight and excellent model. In order to further validate the superiority of the improved model, GECM-EfficientNet was compared with other related studies. Among them, the literature [4] proposes RecycleNet with an accuracy of 81%. Ref. [51] implements a dual fusion approach by PSO and GA, which achieves 94.11% and 94.58% accuracy. Among these related studies, GECM-EfficientNet is also in the lead, with an accuracy similar to the literature [10], the literature [11], and the GA (Ahmad et al. [51]). However, they focus on improving the accuracy of large models, but with large parameters and poor real-time performance. The GECM-EfficientNet parameters are few, but achieve high accuracy. On resource-constrained edge devices, GECM-EfficientNet offers good prospects for application.

4.3. System Testing

4.3.1. Speed Test

In order to verify the real-time performance of GECM-EfficientNet, the above model was deployed to the Raspberry Pi 4B for testing. Figure 10 shows the single inference time. N is the number of inferences. T / s represents the single inference time, which is in seconds. Among the mainstream models, GECM-EfficientNet has a significant advantage on real-time performance. The average single inference time of GECM-EfficientNet is 146 ms, meeting the real-time performance requirements of waste classification.

4.3.2. Classification Test

The confusion matrix is plotted through the test dataset of the self-built dataset, which shows the prediction results for the different categories. Figure 11 shows the results. The rows and columns of the matrix indicate the true and predicted values for waste categories. The values on the diagonal in Figure 11a indicate the number of correctly sorted waste items, whereas the values outside the diagonal indicate the number of incorrectly sorted waste items. Figure 11a is normalized to give Figure 11b, the diagonal values indicate the accuracy of the classification. As can be seen, single category accuracy remains mostly above 90% and up to 99%. Category 8 (waste dry batteries) is the least accurate (85%) because the waste dry battery’s small, cylindrical shape resembles a cigarette butt. The confusion matrix shows that the GECM-EfficientNet can accurately distinguish between different categories of waste.
Selected waste images from the test dataset for testing. Figure 12 shows the category and accuracy of the waste tested. The four types of waste are framed by different colors. The waste type indicates the type of waste containing other waste, kitchen waste, recyclable waste, and hazardous waste. The name indicates the category of waste and accuracy indicates the classification accuracy, which is expressed as a percentage. As can be seen, the model can achieve an accurate classification of different waste. It can be proven that the GECM-EfficientNet has a high classification accuracy, which satisfies the requirements of waste-sorting devices.

4.4. Discussion of Intelligent Waste Bin

Most of the existing normal waste recycling devices are set up with different bins, which then wait for residents to sort and put out waste manually. However, residents often do not have enough knowledge of sorting, and there is difficulty in translating willingness into action. As a result, this paper designs an intelligent waste bin that automatically sorts and recycles household waste. The intelligent waste bin costs just 2500 CNY. The waste-classification model is trained with 18 categories of waste images. However, for other categories of waste, if the model is trained with enough images of the waste, it can also correctly identify the category of waste. The intelligent waste bin recycles waste to different bins through the combined movement of the paddle plate and baffle plate. This mechanical structure can effectively recycle solid waste, but it may not be pleasant for liquid waste, which may be partially left behind. Therefore, the recycling of liquid waste will be one of future highlights.
Nowadays, some academics are also researching intelligent waste-sorting devices. Ref. [9] designed a smart waste bin based on ResNet34, which can dichotomize waste with a single inference time of 950 ms. Ref. [52] constructed a smart bin based on inceptionv3, which can recycle waste into two bins. These smart devices are based on large models for waste classification, which has significance for automatic waste sorting and recycling. However, although they achieve high accuracy, they perform weakly in real time, which can cause unhappiness for users. Ref. [15] build an intelligent waste-sorting system through a lightweight model (MobileNetv3), in which the waste classification model is deployed in the cloud. Cloud deployment avoids the expense of local computing resources, but this is heavily dependent on the Internet. The highlight in this paper is that a lightweight waste-classification model is proposed, which is directly deployed on the embedded device, avoiding Internet dependencies. The model is high in real-time accuracy, with a single inference time of 146 ms. At the same time, this paper proposes the intelligent waste bin that can sort waste more carefully, and the waste will be recycled into four bins. It can be placed in airports, schools, and shopping malls, which contributes to environmental protection and resource recycling.

5. Conclusions

With the increasing focus on environmental safety and resource recycling, society is calling on residents to sort their waste. This requires residents to be knowledgeable about the different categories of waste, which makes it very difficult to sort waste. For this, intelligent waste-sorting devices would be an effective solution. This paper introduces computer vision technology to waste classification, proposing a lightweight and efficient waste-classification model (GECM-EfficientNet), and an intelligent waste bin is designed based on GECM-EfficientNet. On the self-built household waste dataset, GECM-EfficientNet achieved the accuracy of 94.54%, with a single inference time of 146 ms. The intelligent waste bin enables the automatic sorting and recycling of waste, improving the efficiency of waste sorting. It is relevant for environmental protection and resource recycling, but also beneficial for the country’s sustainable development. The main work and contributions in this paper are as follows.
(1)
We chose the lightweight EfficientNetB0 as the baseline model. The MBConv module is first streamlined, optimizing the model structure and reducing complexity. Then, the ECA module and CA module are connected in parallel, replacing the SE module in the MBConv module, which implements the feature map’s spatial and channel weighting operations.
(2)
In the training strategy, the model parameters are initialized by transfer learning, which improves the model performance and convergence speed.
(3)
We verify the superiority of the GECM-EfficientNet performance with the self-built dataset and the TrashNet dataset. Among the many mainstream models and related research, GECM-EfficientNet is in the lead, with outstanding performance in accuracy and real-time performance.
(4)
We design an intelligent waste bin and implement waste classification through GECM-EfficientNet.The model first identifies the input waste, then sorts and recycles into the corresponding bins by the execution structure. This provides a new solution for alleviating the environmental crisis and achieving a circular economy.
In waste sorting and recycling, some research results have been made in this paper, but the following limitations still exist. (1) Waste-classification models can only identify 18 categories of household waste. In reality, there are many categories of waste, and the dataset needs to be expanded later. (2) Consider the use of semi-supervised learning, which makes use of the vast amount of unlabelled image data for learning, facilitating the performance of the classification model. (3) The current mechanical structure is unable to recycle mixed waste. In future work, it may be effective to design a loading device that can separate mixed waste to single waste. In addition, image segmentation and object-detection techniques can identify the different components of mixed waste.

Author Contributions

Conceptualization, J.Y.; Software, Z.C.; Formal analysis, L.L.; Resources, L.C.; Writing—original draft, Z.F. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by a grant from the Research Projects of Ganjiang Innovation Academy, Chinese Academy of Sciences (No. E255J001), the National Natural Science Foundation of China (No. 62063009), and the Jiangxi Postgraduate Innovation Special Fund Project (YC2022-S648).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable. This study did not involve humans or animals.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Tang, D.; Shi, L.; Huang, X.; Zhao, Z.; Zhou, B.; Bethel, B.J. Influencing Factors on the Household-Waste-Classification Behavior of Urban Residents: A Case Study in Shanghai. Int. J. Environ. Res. Public Health 2022, 19, 6528. [Google Scholar] [CrossRef]
  2. Yang, Q.; Fu, L.; Liu, X.; Cheng, M. Evaluating the Efficiency of Municipal Solid Waste Management in China. Int. J. Environ. Res. Public Health 2018, 15, 2448. [Google Scholar] [CrossRef] [Green Version]
  3. Cheah, C.G.; Chia, W.Y.; Lai, S.F.; Chew, K.W.; Chia, S.R.; Show, P.L. Innovation designs of industry 4.0 based solid waste management: Machinery and digital circular economy. Environ. Res. 2022, 213, 113619. [Google Scholar] [CrossRef]
  4. Bircanoğlu, C.; Atay, M.; Beşer, F.; Genç, Ö.; Kızrak, M.A. RecycleNet: Intelligent Waste Sorting Using Deep Neural Networks. In Proceedings of the 2018 Innovations in Intelligent Systems and Applications (INISTA), Thessaloniki, Greece, 3–5 July 2018; pp. 1–7. [Google Scholar] [CrossRef]
  5. Vo, A.H.; Le, H.S.; Vo, M.T.; Le, T. A Novel Framework for Trash Classification Using Deep Transfer Learning. IEEE Access 2019, 7, 178631–178639. [Google Scholar] [CrossRef]
  6. Lin, K.; Zhao, Y.; Kuo, J.H.; Deng, H.; Cui, F.; Zhang, Z.; Zhang, M.; Zhao, C.; Gao, X.; Zhou, T.; et al. Toward smarter management and recovery of municipal solid waste: A critical review on deep learning approaches. J. Clean. Prod. 2022, 346, 130943. [Google Scholar] [CrossRef]
  7. Chen, Z.; Yang, J.; Feng, Z.; Chen, L. RSCNet: An Efficient Remote Sensing Scene Classification Model Based on Lightweight Convolution Neural Networks. Electronics 2022, 11, 3727. [Google Scholar] [CrossRef]
  8. Azhaguramyaa, V.R.; Janet, J.; Narayanan, V.; Sabari, R.S.; Santhosh, K.K. An Intelligent System for Waste Materials Segregation Using IoT and Deep Learning. J. Phys. Conf. Ser. 2021, 1916, 012028. [Google Scholar] [CrossRef]
  9. Kang, Z.; Yang, J.; Li, G.; Zhang, Z. An Automatic Garbage Classification System Based on Deep Learning. IEEE Access 2020, 8, 140019–140029. [Google Scholar] [CrossRef]
  10. Zhang, Q.; Zhang, X.; Mu, X.; Wang, Z.; Liu, X. Recyclable waste image recognition based on deep learning. Resour. Conserv. Recycl. 2021, 171, 105636. [Google Scholar] [CrossRef]
  11. Mao, W.L.; Chen, W.C.; Wang, C.T.; Lin, Y.H. Recycling waste classification using optimized convolutional neural network. Resour. Conserv. Recycl. 2021, 164, 105132. [Google Scholar] [CrossRef]
  12. Adedeji, O.; Wang, Z. Intelligent Waste Classification System Using Deep Learning Convolutional Neural Network. Procedia Manuf. 2019, 35, 607–612. [Google Scholar] [CrossRef]
  13. Gaba, S.; Budhiraja, I.; Kumar, V.; Garg, S.; Kaddoum, G.; Hassan, M.M. A federated calibration scheme for convolutional neural networks: Models, applications and challenges. Comput. Commun. 2022, 192, 144–162. [Google Scholar] [CrossRef]
  14. Chen, Z.; Guo, H.; Yang, J.; Jiao, H.; Feng, Z.; Chen, L.; Gao, T. Fast vehicle detection algorithm in traffic scene based on improved SSD. Measurement 2022, 201, 111655. [Google Scholar] [CrossRef]
  15. Wang, C.; Qin, J.; Qu, C.; Ran, X.; Liu, C.; Chen, B. A smart municipal waste management system based on deep-learning and Internet of Things. Waste Manag. 2021, 135, 20–29. [Google Scholar] [CrossRef] [PubMed]
  16. Chen, Z.; Yang, J.; Chen, L.; Jiao, H. Garbage classification system based on improved ShuffleNet v2. Resour. Conserv. Recycl. 2022, 178, 106090. [Google Scholar] [CrossRef]
  17. Pan, S.J.; Qiang, Y. A Survey on Transfer Learning. IEEE Trans. Knowl. Data Eng. 2010, 22, 1345–1359. [Google Scholar] [CrossRef]
  18. Hemalatha, M. A Hybrid Random Forest Deep learning Classifier Empowered Edge Cloud Architecture for COVID-19 and Pneumonia Detection. Expert Syst. Appl. 2022, 210, 118227. [Google Scholar] [CrossRef]
  19. Yang, C.; Wang, Y.; Lan, S.; Wang, L.; Shen, W.; Huang, G.Q. Cloud-edge-device collaboration mechanisms of deep learning models for smart robots in mass personalization. Robot. Comput.-Integr. Manuf. 2022, 77, 102351. [Google Scholar] [CrossRef]
  20. Krizhevsky, A.; Sutskever, I.; Hinton, G. ImageNet Classification with Deep Convolutional Neural Networks. Adv. Neural Inf. Process. Syst. 2012, 25, 84–90. [Google Scholar] [CrossRef] [Green Version]
  21. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar] [CrossRef]
  22. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar] [CrossRef] [Green Version]
  23. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  24. Fang, R.; Lu, C.C.; Chuang, C.T.; Chang, W.H. A visually interpretable detection method combines 3-D ECG with a multi-VGG neural network for myocardial infarction identification. Comput. Methods Progr. Biomed. 2022, 219, 106762. [Google Scholar] [CrossRef] [PubMed]
  25. Paymode, A.S.; Malode, V.B. Transfer Learning for Multi-Crop Leaf Disease Image Classification using Convolutional Neural Network VGG. Artif. Intell. Agric. 2022, 6, 23–33. [Google Scholar] [CrossRef]
  26. Ganguly, S.; Bhowal, P.; Oliva, D.; Sarkar, R. BLeafNet: A Bonferroni mean operator based fusion of CNN models for plant identification using leaf image classification. Ecol. Inform. 2022, 69, 101585. [Google Scholar] [CrossRef]
  27. Yan, Z.; Liu, H.; Li, T.; Li, J.; Wang, Y. Two dimensional correlation spectroscopy combined with ResNet: Efficient method to identify bolete species compared to traditional machine learning. LWT 2022, 162, 113490. [Google Scholar] [CrossRef]
  28. Pan, S.Q.; Qiao, J.F.; Wang, R.; Yu, H.L.; Wang, C.; Taylor, K.; Pan, H.Y. Intelligent diagnosis of northern corn leaf blight with deep learning model. J. Integr. Agric. 2022, 21, 1094–1105. [Google Scholar] [CrossRef]
  29. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size. arXiv 2016, arXiv:1602.07360. [Google Scholar] [CrossRef]
  30. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. arXiv 2017, arXiv:1704.04861. [Google Scholar] [CrossRef]
  31. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar] [CrossRef] [Green Version]
  32. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6848–6856. [Google Scholar] [CrossRef] [Green Version]
  33. Ma, N.; Zhang, X.; Zheng, H.T.; Sun, J. ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design. In Proceedings of the Computer Vision—ECCV 201, Munich, Germany, 8–14 September 2018; pp. 122–138. [Google Scholar] [CrossRef] [Green Version]
  34. Tan, M.; Le, Q. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. In Proceedings of the 36th International Conference on Machine Learning, PMLR, Vancouver, BC, Canada, 13 December 2019; Volume 97, pp. 6105–6114. [Google Scholar] [CrossRef]
  35. Luo, C.Y.; Cheng, S.Y.; Xu, H.; Li, P. Human behavior recognition model based on improved EfficientNet. Procedia Comput. Sci. 2022, 199, 369–376. [Google Scholar] [CrossRef]
  36. Espejo-Garcia, B.; Malounas, I.; Mylonas, N.; Kasimati, A.; Fountas, S. Using EfficientNet and transfer learning for image-based diagnosis of nutrient deficiencies. Comput. Electron. Agric. 2022, 196, 106868. [Google Scholar] [CrossRef]
  37. Sun, K.; He, M.; He, Z.; Liu, H.; Pi, X. EfficientNet embedded with spatial attention for recognition of multi-label fundus disease from color fundus photographs. Biomed. Signal Process. Control 2022, 77, 103768. [Google Scholar] [CrossRef]
  38. Guo, Y.; Wang, Y.; Yang, H.; Zhang, J.; Sun, Q. Dual-attention EfficientNet based on multi-view feature fusion for cervical squamous intraepithelial lesions diagnosis. Biocybern. Biomed. Eng. 2022, 42, 529–542. [Google Scholar] [CrossRef]
  39. Soomro, I.A.; Ahmad, A.; Raza, R.H. Printed Circuit Board identification using Deep Convolutional Neural Networks to facilitate recycling. Resour. Conserv. Recycl. 2022, 177, 105963. [Google Scholar] [CrossRef]
  40. Peng, H.; Shen, N.; Ying, H.; Wang, Q. Factor analysis and policy simulation of domestic waste classification behavior based on a multiagent study—Taking Shanghai’s garbage classification as an example. Environ. Impact Assess. Rev. 2021, 89, 106598. [Google Scholar] [CrossRef]
  41. Sobota, J.; Goubej, M.; Königsmarková, J.; Čech, M. Raspberry Pi-based HIL simulators for control education. IFAC-PapersOnLine 2019, 52, 68–73. [Google Scholar] [CrossRef]
  42. Zhou, K.; Yuan, Y. A Smart Ammunition Library Management System Based on Raspberry Pie. Procedia Comput. Sci. 2020, 166, 165–169. [Google Scholar] [CrossRef]
  43. Thung, G.; Yang, M. Classification of Trash for Recyclability Status. CS229 Proj.Report2016. 2016, pp. 940–945. Available online: http://cs229.stanford.edu/proj2016/report/ThungYang-ClassificationOfTrashForRecyclabilityStatus-report.pdf (accessed on 14 January 2022).
  44. Jie, H.; Li, S.; Gang, S.; Albanie, S. Squeeze-and-Excitation Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 42, 2011–2023. [Google Scholar] [CrossRef] [Green Version]
  45. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 11531–11539. [Google Scholar] [CrossRef]
  46. Hou, Q.; Zhou, D.; Feng, J. Coordinate Attention for Efficient Mobile Network Design. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 13708–13717. [Google Scholar] [CrossRef]
  47. Orenstein, E.C.; Beijbom, O. Transfer Learning and Deep Feature Extraction for Planktonic Image Data Sets. In Proceedings of the 2017 IEEE Winter Conference on Applications of Computer Vision (WACV), Santa Rosa, CA, USA, 24–31 March 2017; pp. 1082–1088. [Google Scholar] [CrossRef]
  48. Guo, Y.; Shi, H.; Kumar, A.; Grauman, K.; Rosing, T.; Feris, R. SpotTune: Transfer Learning Through Adaptive Fine-Tuning. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 4800–4809. [Google Scholar] [CrossRef] [Green Version]
  49. He, T.; Zhang, Z.; Zhang, H.; Zhang, Z.; Xie, J.; Li, M. Bag of Tricks for Image Classification with Convolutional Neural Networks. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 558–567. [Google Scholar] [CrossRef] [Green Version]
  50. Howard, A.; Sandler, M.; Chen, B.; Wang, W.; Chen, L.C.; Tan, M.; Chu, G.; Vasudevan, V.; Zhu, Y.; Pang, R.; et al. Searching for MobileNetV3. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1314–1324. [Google Scholar] [CrossRef]
  51. Ahmad, K.; Khan, K.; Al-Fuqaha, A. Intelligent Fusion of Deep Features for Improved Waste Classification. IEEE Access 2020, 8, 96495–96504. [Google Scholar] [CrossRef]
  52. Gupta, T.; Joshi, R.; Mukhopadhyay, D.; Sachdeva, K.; Jain, N.; Virmani, D.; Garcia-Hernandez, L. A deep learning approach based hardware solution to categorise garbage in environment. Complex Intell. Syst. 2022, 8, 1129–1152. [Google Scholar] [CrossRef]
Figure 1. The intelligent waste bin mechanical modeling and simulation modeling are shown in (a) and (b) respectively, where they were created by SolidWorks. Subfigure (c) shows the physical construction of the intelligent waste bin, which is constructed by the camera, servos, bins, infrared sensor, baffle plate, and paddle plate.
Figure 1. The intelligent waste bin mechanical modeling and simulation modeling are shown in (a) and (b) respectively, where they were created by SolidWorks. Subfigure (c) shows the physical construction of the intelligent waste bin, which is constructed by the camera, servos, bins, infrared sensor, baffle plate, and paddle plate.
Ijerph 19 15987 g001
Figure 2. Control circuit. It is built through the Raspberry Pi 4B, camera, servos, and infrared sensor.
Figure 2. Control circuit. It is built through the Raspberry Pi 4B, camera, servos, and infrared sensor.
Ijerph 19 15987 g002
Figure 3. MBConv module structure.
Figure 3. MBConv module structure.
Ijerph 19 15987 g003
Figure 4. Improved EfficientNet structure.
Figure 4. Improved EfficientNet structure.
Ijerph 19 15987 g004
Figure 5. Efficient channel attention module.
Figure 5. Efficient channel attention module.
Ijerph 19 15987 g005
Figure 6. Coordinate attention module.
Figure 6. Coordinate attention module.
Ijerph 19 15987 g006
Figure 7. The results of the ablation experiment on the self-built dataset. Subfigure (a) shows the loss curve, where the loss value is calculated on the training dataset. Subfigure (b) shows the accuracy curve, where the accuracy is calculated on the test dataset.
Figure 7. The results of the ablation experiment on the self-built dataset. Subfigure (a) shows the loss curve, where the loss value is calculated on the training dataset. Subfigure (b) shows the accuracy curve, where the accuracy is calculated on the test dataset.
Ijerph 19 15987 g007
Figure 8. The results of the model performance comparison on the self-built dataset. Subfigure (a) shows the loss curve, where the loss value is calculated on the training dataset. Subfigure (b) shows the accuracy curve, where the accuracy is calculated on the test dataset.
Figure 8. The results of the model performance comparison on the self-built dataset. Subfigure (a) shows the loss curve, where the loss value is calculated on the training dataset. Subfigure (b) shows the accuracy curve, where the accuracy is calculated on the test dataset.
Ijerph 19 15987 g008
Figure 9. The results of the model performance comparison on the TrashNet dataset. Subfigure (a) shows the loss curve, where the loss value is calculated on the training dataset. Subfigure (b) shows the accuracy curve, where the accuracy is calculated on the test dataset.
Figure 9. The results of the model performance comparison on the TrashNet dataset. Subfigure (a) shows the loss curve, where the loss value is calculated on the training dataset. Subfigure (b) shows the accuracy curve, where the accuracy is calculated on the test dataset.
Ijerph 19 15987 g009
Figure 10. Inference time on the Raspberry Pi 4B.
Figure 10. Inference time on the Raspberry Pi 4B.
Ijerph 19 15987 g010
Figure 11. Confusion matrix of GECM-EfficientNet on the test dataset of the self-built dataset. The x-axis and y-axis represent the true and predicted categories of waste respectively, and the different categories are represented by numbers. Subfigure (a) shows the number of predictions for each waste category. Subfigure (b) is obtained by normalizing subfigure (a).
Figure 11. Confusion matrix of GECM-EfficientNet on the test dataset of the self-built dataset. The x-axis and y-axis represent the true and predicted categories of waste respectively, and the different categories are represented by numbers. Subfigure (a) shows the number of predictions for each waste category. Subfigure (b) is obtained by normalizing subfigure (a).
Ijerph 19 15987 g011
Figure 12. Test results for partial waste images, where the waste type, name and classification accuracy are shown.
Figure 12. Test results for partial waste images, where the waste type, name and classification accuracy are shown.
Ijerph 19 15987 g012
Table 1. Self-built household waste dataset.
Table 1. Self-built household waste dataset.
CategoryNameNumberCategoryNameNumber
Kitchen wasteFruit flee425Recyclable wastePlastic bowl417
Waste bone379Can374
Biscuits484Waste paper510
Tea426Cardboard348
Eggshell401Plastic bottle511
Other wasteSoiled plastic422Hazardous wasteDry battery342
Cigarette395Accumulator307
Cotton swab595Circuit board299
Chopsticks371Insecticide355
Table 2. G-EfficientNet model structure.
Table 2. G-EfficientNet model structure.
StageOperatorResolutionChannelRepeats
(Orignal)
Repeats
(Adapt)
1Conv 3 × 3 224 × 224 3211
2MBConv1, k 3 × 3 112 × 112 1611
3MBConv6, k 3 × 3 112 × 112 2422
4MBConv6, k 5 × 5 56 × 56 4022
5MBConv6, k 3 × 3 28 × 28 8032
6MBConv6, k 5 × 5 14 × 14 11231
7MBConv6, k 5 × 5 14 × 14 19240
8MBConv6, k 3 × 3 7 × 7 32011
9Conv 1 × 1 &Pooling&FC 7 × 7 128011
Table 3. Ablation experiments on the self-built dataset.
Table 3. Ablation experiments on the self-built dataset.
ModelECA ModuleCA ModuleTransfer LearningA/%P/M
EfficientNetB0---88.814.03
G-EfficientNet---87.541.12
--89.021.02
--89.351.32
--91.761.12
94.541.22
Table 4. Results of experiments with the self-built dataset.
Table 4. Results of experiments with the self-built dataset.
ModelA/%P/MT/msModelA/%P/MT/ms
GECM-EfficienNet94.541.23146.82ShuffleNetv2 1×86.731.27101.51
G-EfficientNet87.541.12127.78ShuffleNetv2 2×88.135.38225.42
EfficientNetB088.814.03211.21GoogleNet85.925.61278.73
MobileNetv287.722.2593.26DenseNet12188.026.97367.51
MobileNetv3 (small)87.751.5495.89ResNet5084.723.55524.35
MobileNetv3 (large)88.474.23207.78Inceptionv386.421.82451.49
VGG1683.88134.33837.95----
Table 5. Experimental results on the TrashNet dataset.
Table 5. Experimental results on the TrashNet dataset.
ModelA/%ModelA/%ModelA/%
GECM-EfficienNet94.23ShuffleNetv2 1×87.27VGG1684.09
G-EfficientNet89.26ShuffleNetv2 2×88.67[11]94.02
EfficientNetB091.65GoogleNet84.69PSO( Ahmad et al. [51])94.11
MobileNetv288.86DenseNet12188.46[10]95.87
MobileNetv3 (small)88.27ResNet5088.86[4]81
MobileNetv3 (large)90.45Inceptionv388.27GA( Ahmad et al. [51])94.58
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Feng, Z.; Yang, J.; Chen, L.; Chen, Z.; Li, L. An Intelligent Waste-Sorting and Recycling Device Based on Improved EfficientNet. Int. J. Environ. Res. Public Health 2022, 19, 15987. https://doi.org/10.3390/ijerph192315987

AMA Style

Feng Z, Yang J, Chen L, Chen Z, Li L. An Intelligent Waste-Sorting and Recycling Device Based on Improved EfficientNet. International Journal of Environmental Research and Public Health. 2022; 19(23):15987. https://doi.org/10.3390/ijerph192315987

Chicago/Turabian Style

Feng, Zhicheng, Jie Yang, Lifang Chen, Zhichao Chen, and Linhong Li. 2022. "An Intelligent Waste-Sorting and Recycling Device Based on Improved EfficientNet" International Journal of Environmental Research and Public Health 19, no. 23: 15987. https://doi.org/10.3390/ijerph192315987

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop