Next Article in Journal
Resilience Indicators for a Road Transport Network to Access Emergency Health Services
Previous Article in Journal
Title Walking the Soundscape: Creative Learning Pathways to Environmental Education in Chilean Schools
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Ensemble Learning Model for Waste Classification Systems

by
Ahmet Alkılınç
1,*,
Feyza Yıldırım Okay
2,
İbrahim Kök
3 and
Suat Özdemir
1
1
Department of Computer Engineering, Hacettepe University, 06800 Ankara, Türkiye
2
Department of Computer Engineering, Gazi University, 06570 Ankara, Türkiye
3
Department of Artificial Intelligence and Data Engineering, Ankara University, 06100 Ankara, Türkiye
*
Author to whom correspondence should be addressed.
Sustainability 2026, 18(1), 24; https://doi.org/10.3390/su18010024
Submission received: 21 November 2025 / Revised: 13 December 2025 / Accepted: 16 December 2025 / Published: 19 December 2025

Abstract

Waste classification is a critical aspect of sustainable waste management systems. Traditional methods for waste classification are often inadequate to handle the complexity and diversity of materials encountered in real-world scenarios. This paper proposes novel deep ensemble learning models that combine pre-trained models with ensemble methods to improve waste classification performance. The proposed model leverages transfer and ensemble learning techniques, employing both averaging and weighted averaging methods to enhance waste classification accuracy. The proposed model is evaluated comprehensively on four publicly available waste image datasets containing various waste categories: TrashNet, TrashBox, Waste Pictures and Garbage Classification. The obtained results show that the averaging and weighted averaging ensemble methods improved classification accuracy by 1% to 3% over the strongest individual models. The weighted ensemble method achieves 96% accuracy, 94% precision, 97% recall and 95% F1 score on the TrashNet dataset. Statistical significance is verified using 5-fold cross-validation and paired t-tests (p < 0.05). To ensure model explainability, the localization of important object regions is demonstrated with Grad-CAM visualizations. Overall, this study validates the potential of integrating deep image classification models with ensemble methodologies to improve the accuracy and efficiency of waste classification. The main contributions of this study can be summarized as follows: we design an efficient deep ensemble method that leverages multiple pre-trained models and ensemble techniques; we employ averaging and weighted averaging techniques to improve classification accuracy and model robustness; and lastly, we evaluate the model using multiple datasets to demonstrate generalizability, scalability and robustness.

1. Introduction

Rapid population growth and urbanization in cities are leading to an unprecedented increase in waste generation, posing significant challenges that necessitate a paradigm shift in conventional methods in resource management, environmental sustainability, and community resilience. According to recent studies, a 70% rise is expected in the global volume of waste by 2050 [1,2].
In this context, waste classification systems emerge as a critical subject, playing a significant role in mitigating environmental pollution, preserving resources, and fostering efforts towards the waste recycling process. Traditional methods such as physical separation, chemical analysis, and visual inspection that rely on predefined rules and heuristic approaches are commonly used in waste classification [3]. However, the utilization of these methods can be challenging due to reasons such as the diversity of waste, human error, cost, time and identification of hazardous waste [4]. For these reasons, the development of automated waste classification systems integrated with modern technologies is crucial in overcoming these challenges and enhancing the efficiency of waste management processes. Recently, digital technologies such as the Internet of Things (IoT), Artificial Intelligence (AI) and sensor networks have started to play a major role in improving waste collection, sorting and management operations [5]. In addition, these technologies are widely used to improve operational efficiency, reduce collection route optimization and provide real-time decision-making.
The recent advances in Deep Learning (DL) and computer vision have made significant progress in the field of image classification. Many different methods are presented to enhance the accuracy and effectiveness of image classification. AlexNet [6], GoogLeNet [7] and ResNet [8] models have obtained remarkable achievements in large-scale image classification. Ioffe et al. [9] proposed Batch Normalization (BN) which has successfully alleviated the problem of gradient vanishing. Xie et al. [10] introduced ResNeXt that both reduced the number of parameters and further improved overall performance. DenseNet [11] addressed the problem of model performance degradation by implementing jump links. This resulted in significant improvements in both accuracy and speed compared to previous network architectures. EfficientNet [12] developed by Google achieved significant performance improvements.
DL has demonstrated outstanding achievements in various challenging computer vision domains and it has proven as a powerful paradigm for extracting robust and generalizable feature representations. Especially in the field of medical image analysis, DL architectures have performed successfully in critical tasks like disease detection and diagnosis [13,14,15,16]. In addition, DL methods have made significant advances in autonomous navigation systems in which robust feature extraction and decision-making processes are critical under dynamic and uncertain conditions [17,18]. These studies demonstrate the ability of DL models to process complex and highly variable data and provide a strong foundation for the application of DL models in waste classification where varied material types, inconsistent imaging conditions, and complex visual patterns are significant issues.
The advancements in DL techniques in recent decades have encouraged extensive research into various algorithms for waste classification. Sakr et al. [19] proposed an algorithm utilizing the AlexNet framework for waste image classification. A study by Yang et al. [20] conducted a comparison between Convolutional Neural Network (CNN)-based models and Support Vector Machines (SVMs) using the TrashNet dataset. Bircanoglu et al. [21] proposed the RecycleNet algorithm which is created by modifying the connection models of skip connections inside dense blocks. Kang et al. [2] presented a ResNet-34 based algorithm, incorporating three specific modifications: multiple feature fusion, reuse of features in the residual unit, and optimization of the activation function. Zeng et al. [22] proposed PublicGarbageNet, which is a multi-task classification algorithm where one task identifies four main categories of household waste and another task recognizes ten sub-classes of waste. Yang et al. [4] introduced GarbageNet, a new incremental learning framework for waste classification. Chen et al. [23] designed GCNet, which is a lightweight and effective waste classification model based on ShuffleNet v2.
Although DL significantly improves the waste classification, our literature review shows that most existing waste classification studies focus either on traditional Machine Learning (ML) models or approaches that utilize a single or individual pre-trained DL model. However, these models suffer from low accuracy and poor generalization issues when trained on small-sized and imbalanced datasets. The main objective of this study is to address these issues in waste classification systems; while many pre-trained CNN models are applied separately for waste classification, there is a significant gap in evaluating how ensemble learning strategies can effectively leverage the complementary strengths of multiple models to improve the performance of waste classification. Therefore, we take advantage of average and weighted average ensemble methods by combining high-performance pre-trained CNN models. Ensemble methods [24] have the potential to improve the accuracy of DL models. In the context of waste classification, employing an ensemble method can effectively boost classification accuracy by leveraging the strengths of multiple models. This collaborative approach enables the development of a more robust model by utilizing each model’s individual strengths while mitigating their weaknesses. Accordingly, first, the waste classification models are trained using transfer learning with various pre-trained models such as DenseNet121, DenseNet201, MobileNetV2, MobileNetV3Large, InceptionV3, ResNet50, ResNet50V2, ResNet101, ResNet101V2, Xception, ConvNeXtTiny, ConvNeXtLarge, EfficientNetB0, EfficientNetB7, EfficientNetV2B0, EfficientNetV2L. Then, the three models are selected from these trained models to implement averaging and weighted averaging ensemble learning methods. In this study, we propose the combination of pre-trained deep models with ensemble methods to improve accuracy while preserving robustness. We provide a comprehensive benchmark analysis on four publicly available waste datasets and present an optimized weighted ensemble strategy which improves the prediction performance while maintaining practical applicability. The main contributions of the paper are as follows:
  • We present efficient deep ensemble learning models, integrating pre-trained models with ensemble methods to provide more accurate results for waste management systems.
  • We perform a comparative evaluation of the waste classification problem using sixteen different pre-trained DL models on four waste datasets.
  • We provide a detailed overview of the existing studies on waste classification.
  • We implement Grad-CAM method to ensure the explainability of models within the waste classification task.
This paper is structured as follows. Section 2 presents a review of the literature on waste classification. Section 3 describes the pre-trained models, ensemble learning methods and also describes the proposed waste classification method. Section 4 presents the dataset, performance metrics and experimental study and results. Finally, Section 5 provides general concluding remarks.

2. Related Work

Enhancing waste management practices can facilitate waste recycling and ensure efficient and economical use of resources. Research in this area covers a wide range of studies investigating various methodologies. Traditional approaches are largely based on manual sorting techniques which are labor-intensive and error-prone. However, recent developments in computer vision, ML and DL have significantly contributed to the waste classification process. These developments have resulted in extensive research targeting waste classification. The related studies in waste classification field are summarized and grouped by their dataset and methodology in Table 1.
Yang et al.’s [20] hand-collected, publicly available dataset with approximately 400–500 images per class is created to classify waste into recycling category by providing an effective waste processing method. Utilizing SVM with scale-invariant feature transformation (SIFT) features and CNN, their study compares their performance. Experimental results indicate that SVM reaches 63% accuracy. Zhihong et al. [25] introduced a robotic grasping system designed for automatic waste sorting using ML methods. To achieve precise grasping, they used VGG-16 model and Region Proposal Generation (RPN) for pose estimation and object recognition. Bai et al. [26] introduced an intelligent waste collection robot capable of detecting and operating autonomously. The robot utilizes a deep neural network for both waste recognition and ground segmentation. They achieved 95% accuracy in waste recognition using ResNet-34. Chu et al. [27] suggested a multi-layer hybrid DL system (MHS) that combines high-resolution cameras and sensors to classify urban waste automatically. With this approach, they achieved over 90% accuracy in distinguishing recyclable items from non-recyclable items and outperformed the traditional CNN method based on images alone. Rabano et al. [28] introduced a model for waste classification using the MobileNet model on the Trashnet dataset. They utilized transfer learning from a pre-trained model on ImageNet. The suggested model achieved a test accuracy of 87.2%.
Satvilkar et al. [29] addressed the important topic of waste classification, highlighting its importance in waste management and the unique role of image processing through analytics. The paper explores multiple methods to determine the most efficient approach or combination for image classification using methods such as CNN, SVM, XGB, RF, KNN. CNN achieved the best classification accuracy of 89.81%. Bircanoglu et al. [21] proposed RecycleNet which is a deep CNN architecture optimized to classify recyclable objects by modifying skip connections within dense blocks. The proposed model decreased the number of parameters in a 121-layer network to about 3 million and achieved 81% accuracy on the TrashNet dataset. Zeng et al. [30] presented a multiscale convolutional neural network (MSCNN) method that uses aerial hyperspectral data for large-area waste detection. The proposed algorithm shows strong performance in litter detection and outperforms existing hyperspectral image (HSI) classification methods on publicly available datasets such as Indian Pines and Pavia University. Adeeji et al. [31] suggested a method for waste classification utilizing the ResNet-50 CNN model as a feature extractor and SVM for classification. This approach obtained an accuracy of 87% on the TrashNet dataset. Ruiz et al. [32] trained and evaluated different DL architectures with the TrashNet dataset to classify waste. Specifically, they compared CNN models including ResNet, Inception and VGG. The highest accuracy is achieved from the combined Inception-ResNet model, which achieves 88.6% accuracy.
Endah et al. [33] suggested a method for waste classification using transfer learning on the TrashNet dataset. They applied Xception, VGG16 and ResNet-50 models. The results demonstrate that the Xception model attains the best accuracy of 88%. Toğaçar et al. [34] proposed a model that reconstructs the used dataset with AutoEncoder network and extracts feature sets using CNN architectures. In the proposed model, these extracted features are subsequently merged and Ridge Regression (RR) is applied to decrease the number of features. SVM is used as the classifier and the classification accuracy is 99.95% in experiments on a two-class open-access dataset. Azis et al. [35] presented a DL and computer vision-based method for waste classification using Inception-v3. The proposed model achieved 92.5% accuracy on TrashNet. Shi et al. [36] addressed challenges in waste classification by proposing a technique which is using a multilayer hybrid convolution neural network (MLH-CNN). This method, resembling VggNet yet with a simpler design and fewer parameters, aims to enhance accuracy and reduce computational time. Experimental findings on the TrashNet dataset demonstrate a classification accuracy of up to 92.6%. Zhang et al. [37] proposed a DL-based classification model for recyclable waste images, which integrates a self-monitoring module into the residual network model to improve feature representation and extract features from various waste images. The proposed model achieved an accuracy of 95.87%. Majchrowska et al. [38] introduced new datasets, classify-waste and detect-waste, which combine open-source datasets with merged annotations covering various waste categories. They also introduced a two-stage detector that uses EfficientDet-D2 for waste detection and EfficientNet-B2 for its classification. Kumsetty et al. [39] introduced TrashBox, which is a new waste dataset consisting of seven classes. In addition, a new DL framework utilizing quantum transfer learning is explored and promising results are obtained in experimental evaluations on different datasets. Zhou et al. [40] presented a DL method for the classification of medical waste. The suggested classification method uses ResNeXt as a deep neural network and achieves remarkable results using transfer learning methods. In the experiments, it correctly classifies eight types of medical waste in the medical dataset with an accuracy of 97.2%.
Table 1. Summary of the literature on waste classification.
Table 1. Summary of the literature on waste classification.
RefYearTaskML/DL Model(s)Dataset#ClassPerformance Measure
[20]2016Waste classificationSIFT + SVMTrashNet6Accuracy
[25]2017Waste sortingVGG-16Own data1Miss rate
False rate
[26]2018Waste recognitionResNet-34Own data6Accuracy
[27]2018Waste classificationCNN, MLP, AlexNetOwn data2Accuracy, Precision, Recall
[28]2018Waste classificationMobileNetTrashNet6Accuracy
[29]2018Waste classificationCNN, SVM, XGB,
RF, KNN
TrashNet6Accuracy, Precision, Recall,
F1-score
[21]2018Waste classificationResNet50TrashNet6Accuracy
[30]2019Waste detection,
HSI classification
Multi-Scale CNN,Indian Pines dataset,
Pavia University dataset
16
9
Overall accuracy, Average
accuracy, Kappa coefficient
[31]2019Waste classificationResNet-50, SVMTrashNet6Accuracy
[41]2019Waste classificationVGG, Inception, ResNetTrashNet6Accuracy
[34]2020Waste classificationAutoEncoder, AlexNet,
GoogLeNet, ResNet-50,
SVM
Waste Classification data2Accuracy, Precision,
F1-score
[33]2020Waste classificationVGG16, ResNet-50,
Xception
TrashNet6Accuracy, Precision, Recall
[35]2020Waste classificationInception-v3TrashNet6Accuracy
[42]2021Waste classificationInceptionV3Garbage Classification12Accuracy, Precision,
Recall, F0.5-score
[36]2021Waste classificationMLH-CNN, VGG16,
AlexNet, ResNet50
TrashNet6Accuracy, Precision,
Recall, F1-score
[37]2021Waste classificationResNet18TrashNet6Accuracy, Precision,
Recall, F1-score
[38]2022Waste detection
Waste classification
EfficientDet-D2,
EfficientNet-B2
Detect-waste
Cassify-waste
8mAP, Precision,
Recall, F1-score
[40]2022Waste classificationResNeXt-50Medical waste dataset8Precision, Recall, F1-score
[39]2022Waste detection
Waste classification
ResNet-34, ResNet-50,
ResNet-101, VGG-19,
DenseNet-121
TrashNet
TACO
TrashBox
6
28
7
Accuracy
[43]2023Waste classificationGoogleNet, ResNet, DenseNet,
ResNeXt, EfficientNet
Garbage Classification12Accuracy, Precision,

Recall, F1-score
[44]2023Waste classificationVGGNet16, Resent50,
MobileNetV2, InceptionV3, CNN
Garbage Classification8Accuracy, Precision,
Recall, F1-score
[45]2023Waste classificationResNet18TrashNet6Accuracy, Precision,
Recall, F1-score
[46]2023Waste detection
Waste classification
MobileNet-v2HUAWEI-404Accuracy, Precision
[47]2023Waste classificationResNet-34, ResNet-101,
VGG-16
TrashNet,
TACO
6Accuracy
[48]2024Waste classificationGoogleNet, ResNet50,
Inception-v3, MobileNet-v2,
DenseNet201.
TrashNet6Accuracy, Precision,
Recall, Specificity, F1-score
[49]2024Waste classificationVGG-16, ResNet-34, ResNet-50,
AlexNet, LSTM
Recycle Waste
image dataset
2Accuracy, Precision,
Recall, F1-score
[50]2024Waste classificationVGG19TrashNet,
GarClass
6
6
Accuracy, Precision,
Recall, F1-score
[51]2024Waste classificationYOLO 5, YOLO 7e-waste dataset5mAP, Accuracy, Precision,
Recall, F1-score
[52]2024Waste classificationVGG16TrashNet,
Own data
6Accuracy
[53]2025Waste classificationCNN, YOLOOwn data4Accuracy, Precision,
Recall, F1-score
In another study, Shukurov [43] proposed an algorithm using fine-tuned CNN (GoogleNet, DenseNet, ResNet, ResNeXt, EfficientNet). Its experiments are conducted on an open-source dataset of 12 categories using SGD and Adam optimization algorithms. The experimental results show that fine-tuned ResNeXt model reaches 95% accuracy. Chen et al. [42] presented an algorithm based on InceptionV3 networks for waste classification. They conducted their experiments on the open-source Garbage Classification (12 classes) dataset on Kaggle using transfer learning. The suggested model achieved 93.1% accuracy in waste classification. Dey et al. [44] presented a custom CNN model designed for waste classification. In the study, pre-trained models (ResNet50, VGGNet16, InceptionV3, MobileNetV2 ) are compared with a proposed custom CNN model. The suggested model achieved 97.58% accuracy. Huang et al. [45] proposed ResMsCapsule network, a capsule network-based waste classification model. This model improves the performance of the basic capsule network by combining the residual network and the multiscale module. Experiments with the TrashNet dataset showed that ResMsCapsule has a classification accuracy of 91.41%. Jin et al. [46] developed a new machine vision system based on DL and transfer learning for waste classification. The proposed model is based on MobileNetV2. The model uses an attention mechanism and transfer learning to improve the accuracy. Dimension reduction with PCA was applied to enable the model to run in real time on edge devices. In the experiments, 90.7% accuracy was achieved on the Huawei Cloud dataset. Kumsetty et al. [47] proposed a computer-vision-based automatic waste classification method. In the study, the quality of the datasets was improved by data augmentation and image processing techniques and transfer learning-based models such as ResNet and VGG were used. In the experiments on TrashNet and TACO datasets, 93.13% and 91.34% accuracy was achieved, respectively.
Hossen et al. [48] introduced RWC-Net, which is based on a combination of DenseNet201 and MobileNet-V2 for the classification of recyclable waste. This model uses the feature extraction and learning capabilities from both pre-trained models. The model accomplishes an overall accuracy of 95.01% on the TrashNet dataset. Lilhore et al. [49] proposed a waste classification model that combines CNN and long short-term memory (LSTM). The suggested model also uses a transfer learning method that includes the benefits offered by ImageNet. This model’s performance was evaluated on a two-class dataset and the proposed model achieved a precision of 95.45%. Quan et al. [50] proposed a novel method for waste classification in IoT environments that prioritizes privacy preservation. The method combines differential privacy, federated learning and transfer learning techniques to enable collaborative model training while preserving sensitive data. Sarswat et al. [51] proposed a new method of electronic waste classification using DL and computer vision. Researchers have developed real-time object recognition systems for e-waste classification using YOLO 7 and 5 algorithms. At the end of experimental studies, the YOLO 7 model achieved 94% prediction accuracy. Lin et al. [52] presented a multiscale feature fusion strategy (MFFS) for the classification of real-world waste with similar shape, texture or contaminated appearance. This method improves classification accuracy by combining local fine-grained details and global coarse-grained features. The presented model achieved 94.1% accuracy on the TrashNet dataset and 95.5% accuracy on the custom dataset. Kumar et al. [53] presented an innovative framework integrating AI and IoT technologies to enable the safe and efficient management of hazardous waste in hospitals. In this study, IoT-enabled smart waste bins ensure a secure waste collection process, while CNN and Adaptive Neuro Fuzzy Inference System (ANFIS) increase waste detection and classification accuracy. The preliminary experimental results obtained show a reduction in manual intervention and a significant increase in classification performance.
While existing studies in the literature have demonstrated the potential of DL models for automatic waste classification, there are critical gaps in the field that remain unaddressed. First, most studies are performed on a single dataset or limited data conditions which limits the generalizability of the proposed models to various waste environments. Second, the existing studies focus mainly on traditional ML models or a single pre-trained DL approach. However, these models demonstrate low accuracy and poor generalization performance when trained on small and imbalanced datasets. Although many pre-trained CNN models for waste classification are applied individually in the literature, there is a significant gap in evaluating how ensemble learning strategies can effectively leverage the complementary strengths of multiple models to improve waste classification performance. This study aims to address the aforementioned methodological and practical gaps in the waste classification by integrating optimized ensemble learning methods with experiments on multiple datasets.

3. Proposed Deep Ensemble Model for Waste Classification

This section overviews pre-trained models trained on large image datasets and explores ensemble learning methods. Furthermore, we introduce our proposed deep ensemble model architecture. The methodological workflow diagram is shown in Figure 1.

3.1. Pre-Trained Models

CNNs are widely used for image-based feature extraction and spatiotemporal forecasting [54,55]. Traditional ML techniques require manually prepared feature extraction process and usually creating difficulties in generalizing on different image conditions such as lighting changes, texture variations and background complexity. On the other hand, CNNs automatically learn discriminative features directly from the raw images, enhancing the classification robustness and scalability. Residual learning-based architectures such as ResNet help prevent vanishing gradients through skip connections and enable deeper networks to be trained effectively [8]. ConvNeXt [56] and EfficientNet [12] architectures further improve performance by utilizing modern design techniques such as compound scaling and improved convolutional blocks. However, CNN models still suffer from challenges caused by dataset variability, class imbalance and overfitting, particularly in waste classification where object contamination and deformation are common. To address these challenges, we introduce an ensemble-based approach by combining the predictions of multiple models with the goal of achieving better classification accuracy.
When selecting pre-trained models for our method, we prioritized architectures that have shown strong performance on large-scale benchmarks such as ImageNet and proven effectiveness in waste classification. The selected models represent a variety of architectural designs, including CNNs with different depths, residual connections, and initialization modules. This diversity enables the proposed ensemble model to capture a wide range of feature representations and this is crucial given the complex and heterogeneous nature of waste images which differ in texture, shape and background complexity. Utilizing such complementary models allows our model to better generalize across different waste types and environmental conditions.
InceptionV3 [57] is a CNN architecture that uses inception modules with various kernel sizes to capture features of different scales. It efficiently learns fine and complex patterns by factorizing convolutions into smaller operations.
ResNet-50 and ResNet-101 [8] are deep residual networks with 50 and 101 layers. Their skip connections help the model learn residual functions instead of full mappings, allowing deeper architectures to be trained while mitigating vanishing gradients. These models strike a balance between model complexity and computational efficiency, making them popular choices for various computer vision tasks, including waste classification.
DenseNet-121 [11] is a CNN architecture characterized by its dense connectivity, in which every layer is directly linked to each other layer in a feed-forward. This design produces compact and expressive models. There are different variants like DenseNet-121, DenseNet-169 and DenseNet-201, which offer different levels of depth and accuracy.
Xception [58] is a deep CNN architecture that emphasizes depthwise separable convolutions. It reduces computation by separating spatial and channel operations while capturing complex structures. Furthermore, it performs well on challenging tasks and is suitable for environments with limited processing power.
MobileNetV2 [59] is a lightweight architecture designed for mobile and edge devices. It incorporates separable convolutions in the depth direction and uses residuals and linear bottlenecks to achieve superior performance on resource-constrained devices. Its small size and low latency make it ideal for real-time, resource-constrained applications.
EfficientNet [12] is a CNN architecture that systematically scales the model’s depth, width and resolution to optimize efficiency and accuracy. The models range from EfficientNetB0 to EfficientNetB7, allowing users to choose architectures based on their computational needs.
ConvNeXt [56] is based on ResNet-style architectures and incorporates modern design principles such as improved convolutional operations and scaling strategies. These enhancements make ConvNeXt models both accurate and computationally efficient, providing flexibility in model size, performance and resource usage. A summary of the pre-trained models is shown in Table 2.

3.2. Ensemble Learning Methods

Ensemble learning [24] is a ML technique aimed at enhancing the accuracy and robustness of predictions by aggregating the outputs of multiple models. By leveraging the collective intelligence of these models, ensemble learning reduces the impact of errors or biases that might be present in individual models. This method improves both accuracy and robustness to uncertainties in the data. The effectiveness of ensemble learning in combining predictions from multiple models has made it a powerful tool across various fields. Ensemble learning is used in different areas, such as image classification [60,61], text classification [62,63], intrusion detection [64], and activity recognition [65].
The averaging ensemble method [24] simply averages the predictions of individual models to obtain the final prediction. If you have N models, each providing a prediction y i for a given input, the ensemble prediction y ensemble is calculated as the average:
y ensemble = 1 N i = 1 N y i
Averaging is commonly employed when all models are assumed to contribute equally to the final prediction.
The weighted averaging ensemble method [24] assigns different weights to the predictions of individual models before averaging them. The weights reflect the importance or confidence assigned to each model. If w i represents the weight assigned to the prediction of model i, the weighted ensemble prediction y ensemble is calculated as follows:
y ensemble = i = 1 N w i · y i i = 1 N w i
The weighted averaging ensemble method is beneficial when some models are considered more reliable or accurate than others. Assigning higher weights to more trustworthy models allows them to have a more significant influence on the final prediction.

3.3. Proposed Deep Ensemble Model

In this section, we elaborate on the experimental methodology employed in our study, which involves the utilization of pre-trained DL models and ensemble learning methods to enhance the overall model performance.
Firstly, we conduct experiments using sixteen different pre-trained DL models. These models are selected based on their built-in capabilities in various computer vision tasks. Each model is applied individually to the dataset to extract features and make predictions. Then, a comprehensive evaluation is performed to assess their respective performances. For the evaluation, we use performance measures including loss value, accuracy, precision, and recall. Then, we select three models and use ensemble learning techniques to leverage the effectiveness of the selected models and improve the overall performance. We apply two ensemble methods, averaging and weighted average, for combining the predictions from the selected models. The proposed models are shown in Figure 2.
In the averaging ensemble method, the predictions of the three models are combined and averaged to obtain a prediction. This method assumes that each model contributing to the final prediction is of equal importance. In the weighted averaging ensemble method, it is calculated by assigning different weights to the predictions of each model. The weights are determined according to the performance and reliability of each model, thus more effective models have a greater effect on the prediction. In order to determine the optimal weights for the weighted averaging ensemble, we performed a grid search optimization over the possible combinations of weights assigned to each model. Let m 1 , m 2 , m 3 be the three pre-trained models utilized in the ensemble. The final prediction P is calculated as follows:
P = i = 1 3 w i · p i
where p i is the probability vector output of model m i and w i is the corresponding weight. In order to identify the best-performing weight configuration, a grid search was conducted by iterating w 1 , w 2 , w 3 { 1 , 2 , , 9 } . To ensure that the weights form a valid probability distribution, they were normalized such that:
w ˜ i = w i w 1 + w 2 + w 3 , for i { 1 , 2 , 3 } ,
where w ˜ i denotes the normalized weight assigned to model m i . After evaluating all valid weight combinations, the weights that produce the highest classification accuracy on the test set are selected for the final test. This tuning is performed separately for each dataset to ensure the best performance. The pseudo-code of the proposed weighted averaging ensemble method is given in Algorithm 1.
Algorithm 1 weighted averaging ensemble method
1:
Generate predictions from selected three models: p r e d s 1 ,   p r e d s 2 ,   p r e d s 3 and true labels l a b e l s
2:
Initialize m a x _ a c c u r a c y 0
3:
Initialize b e s t _ w e i g h t s [ 0 , 0 , 0 ]
4:
for  w 1 from 1 to 9 do
5:
    for  w 2 from 1 to 9 do
6:
       for  w 3 from 1 to 9 do
7:
          Compute total weight:
8:
           t o t a l w 1 + w 2 + w 3
9:
          Normalize the weights:
10:
           w w 1 / t o t a l ,   w 2 / t o t a l ,   w 3 / t o t a l
11:
          Compute weighted predictions:
12:
           w t e d _ p r e d s i = 1 3 w i · p r e d s i
13:
          Determine final predictions:
14:
           f i n a l _ p r e d s argmax ( w t e d _ p r e d s ,   a x i s = 1 )
15:
          Compute accuracy:
16:
           a c c u r a c y AccuracyScore ( l a b e l s ,   f i n a l _ p r e d s )
17:
          if  a c c u r a c y > m a x _ a c c u r a c y  then
18:
              m a x _ a c c u r a c y a c c u r a c y
19:
              b e s t _ w e i g h t s ( w 1 ,   w 2 ,   w 3 )
20:
          end if
21:
       end for
22:
     end for
23:
end for
24:
return  b e s t _ w e i g h t s

4. Experimental Results and Evaluation

4.1. Datasets

We evaluated the proposed method using four distinct datasets, namely TrashNet [66], TrashBox [39], Waste Pictures [41] and Garbage Classification [67]. The statistical information of the datasets is shown in Table 3. The data sets are divided into 80% training, 10% validation and 10% testing. In the pre-processing step, input images are resized to 224 × 224 pixels to match the pre-trained models’ input size. We apply the preprocess_input function provided by the framework of each pre-trained model (e.g., Keras) for proper normalization and scaling of input data. We used several data augmentation techniques during training to make the model more generalizable and reduce the overfitting. These included random rotations by at most 30 , horizontal and vertical shifts in the range 0.1 , zoom transformations in the range 0.1 , horizontal flipping and brightness changes in the range [ 0.8 , 1.2 ] .
We analyzed the class distributions of all datasets to address class imbalance issues and used oversampling and class weighting methods to handle this problem. In the TrashNet and TrashBox datasets, there is no significant class imbalance, so we applied oversampling to increase the number of samples from the minority classes. On the other hand, the Garbage Classification and Waste Pictures datasets have higher class imbalance due to the significant differences in the number of categories and the number of images per class. In order to mitigate this issue without excessively increasing the dataset size, we used class weighting during model training. To maintain the integrity of the evaluation and prevent any train-test contamination, all data augmentation and oversampling operations are applied after the dataset is split into training, validation and test subsets and only to the training data. Table 4 shows the number of training, test and validation samples before oversampling, while Table 5 shows the number of training, test and validation samples after oversampling.

4.1.1. TrashNet

The TrashNet [66] dataset consists of six common waste categories, namely cardboard, glass, plastic, paper, metal, and trash. The images were taken by positioning the object on a white poster board and utilizing natural sunlight and ambient room lighting. They were subsequently resized to 512 × 384. The dataset contains a total of 2527 waste images in jpg format. Example images from the dataset are shown in Figure 3.

4.1.2. TrashBox

The TrashBox [39] dataset is a comprehensive collection of images specifically organized for waste classification and detection research. It contains a total of 17,853 waste images in seven classes (plastic, glass, metal, cardboard, e-waste, paper, medical waste). It is a valuable resource for researchers interested in developing and evaluating ML and DL models for automatic waste detection and classification.

4.1.3. Waste Pictures

The Waste Pictures [41] dataset consists of approximately 23,087 waste images from Google search results. The dataset contains 34 different image types, including organic, recyclable, and hazardous waste, which gives the dataset significant diversity.

4.1.4. Garbage Classification

The Garbage Classification [67] dataset is an open-source dataset created on Kaggle using web scraping. The dataset contains 15,515 images from 12 different classes of household waste: cardboard, paper, biological, plastic, metal, brown glass, green glass, white glass, shoes, clothes, trash, and batteries.

4.2. Performance Metrics

The model performance is evaluated by measures like accuracy, precision, recall, and F1 score.
Accuracy refers to how close a measurement is to the true or accepted value. It is calculated by dividing the number of correctly predicted samples by the total number of samples.
Accuracy = True Positives + True Negatives Total Instances
Precision quantifies the accuracy of the model’s positive predictions. It is calculated as the ratio of correctly predicted positive observations to the total number of positive predictions.
Precision = True Positives True Positives + False Positives
Recall, also known as sensitivity or true positive rate, measures the capability of the model to capture all of the relevant positive samples. It is the ratio of accurately predicted positive observations to the total actual positives.
Recall = True Positives True Positives + False Negatives
F1-score is the harmonic average of precision and recall. It gives a balanced evaluation of a model’s performance by taking into account both false positives and false negatives.
F 1 - Score = 2 × Precision × Recall Precision + Recall

4.3. Experimental Results

The experiments are conducted to compare the performance of various pre-trained models in waste classification and to develop a better model with ensemble learning methods. Table 6 summarizes the main findings of the pre-trained models. Furthermore, Figure 4 graphically shows the results. The experiments are performed on Google Colab using the Keras library with the TensorFlow backend. We performed hyperparameter tuning using the KerasTuner framework to optimally configure the base models. A grid search was performed to choose the best optimizer from {Adam, RMSprop, SGD} and to find the learning rate from {0.01, 0.001, 0.0001} for each model. The epoch number was set to 50, the batch size to 32, and early stopping (patience = 10) was applied to prevent overfitting. The selected optimizers and learning rates for each pre-trained model after hyperparameter tuning are shown in Table 7.
Among the pre-trained models on the TrashNet dataset, ResNet-50 and ConvNeXtLarge achieve the highest accuracy with 93% accuracy. According to the loss values, ConvNeXtLarge has shown outstanding performance, recording the lowest loss of 18%. This shows that ConvNeXtLarge not only achieves high accuracy but also remains effective in minimizing prediction errors. To exploit the collective power of different models, we evaluated various model combinations based on their robust balance between accuracy and loss.
The accuracy values of the average and weighted average ensemble models for different model combinations are shown in Table 8. The averaging ensemble model achieved 94.9% accuracy with the DenseNet121, ResNet50, ConvNeXtLarge model combination, while the weighted average ensemble model further improved performance, increasing accuracy to 96%. A summary of the results on the TrashNet dataset is presented in Table 9 and all experimental results are provided in Table A1.
In the experiments on the TrashBox dataset, ConvNeXtLarge is the best performing model with 94%. However, InceptionV3 shows the lowest accuracy among the models, with 80%. According to the loss values, ConvNeXtLarge shows the lowest loss with 17%. This shows that ConvNeXtLarge is the most effective model on the TrashBox dataset with high accuracy and low loss value. The evaluation of ensemble methods for different model combinations in the TrashBox dataset is presented in Table 10. The ResNet50, ResNet101, ConvNeXtLarge model combination obtained the highest accuracy on the TrashBox dataset. The averaging ensemble model achieved 93.5% accuracy and the weighted average ensemble model achieved 95.8% accuracy. A summary of the results on the TrashBox dataset is given in Table 11 and all experimental results are presented in Table A2.
In the experiments on the Waste Pictures dataset, ConvNeXtLarge is the top-performing model with an accuracy of 97%. It also demonstrates the lowest loss value at 10%, indicating its effectiveness on this dataset. The evaluation of ensemble methods for different model combinations in the Waste Pictures dataset is given in Table 12. The average ensemble model achieved 97.2% accuracy with the ResNet50, EfficientNetB0 and ConvNeXtLarge model combination, while the weighted average ensemble model reached 98% accuracy. A summary of the results on the Waste Pictures dataset is provided in Table 13 and all experimental results are shown in Table A3.
On the Garbage Classification dataset, ConvNeXtLarge is the most effective model, with 99% accuracy, and EfficientNetV2B0 follows close behind with 98% accuracy. On this dataset, we combine the predictions of EfficientNetB7, Xception and ConvNeXtLarge to create ensemble models. The optimal weights for maximum accuracy are 0.2, 0.3, 0.6, respectively. The average ensemble model shows an accuracy of 98.6% and the weighted average ensemble model shows an accuracy of 99.1%. The evaluation of ensemble methods for different model combinations is given in Table 14. A summary of the results for the Garbage Classification dataset is presented in Table 15, and all experimental results are given in Table A4.
Figure 5 shows the confusion matrix result of the selected base models and the weighted average ensemble on the TrashNet dataset. The weighted average ensemble demonstrates better performance in terms of classification accuracy and reduces misclassifications between classes with similar visual appearances. For instance, the baseline models confused paper with cardboard, whereas the ensemble model provided a clearer distinction between these classes. Furthermore, the ensemble model reduced misclassification errors between glass and plastic that commonly occurred due to varying light. The results demonstrate the ensemble approach’s effectiveness in combining the complementary strengths of individual models and correcting their systematic errors.
To enhance the explainability of our models, we apply Gradient-Weighted Class Activation Mapping (Grad-CAM) [68] to visualize the regions of input images that contribute most to classification decisions. Figure 6 shows the Grad-CAM visualizations of waste images from the TrashNet dataset for the ConvNeXtLarge model. The results indicate that the model consistently focuses on the most relevant object regions such as the central parts of recyclable materials (e.g., plastic bottles or paper surfaces), rather than irrelevant background areas. This provides that the model has learned meaningful and discriminative features and, by providing insight into the model’s decision-making process, it effectively selects the relevant features of the waste items, thus ensuring that the waste is correctly classified. These findings validate the accuracy of the learned representations and also provide practical insights for the potential use of the system in real-world waste-sorting scenarios where explainability is critical for user trust and system transparency.
The paired t-test is applied using the results obtained from 5 to fold cross-validation to compare the proposed weighted ensemble model with the pre-trained models that performed best on the TrashNet dataset. Table 16 gives the p-values of the t-tests performed and the statistical significance of the differences in the values of the performance metrics. All the p-values are less than the 0.05 significance level, this confirms that the improvements achieved by the weighted ensemble are statistically significant. These results demonstrate that the ensemble model consistently outperforms pre-trained models, thereby validating its robustness and generalization ability.
Experimental results obtained on four waste classification datasets illustrate the effectiveness of the proposed ensemble learning approach. Although the individual DL models, including ResNet101, ConvNeXtLarge, and EfficientNetB0, perform well, the ensemble methods achieve superior accuracy, precision, recall and F1 scores. In particular, the weighted average ensemble method that is optimized through an extensive grid search proves its effectiveness, yielding the best classification accuracy across all datasets. Overall, this study provides a more robust and efficient solution for the automated waste sorting problem by integrating model diversity and optimized ensemble weighting. In addition, it supports future applications in practical smart recycling environments.

4.4. Time Complexity Analysis

The deep learning models evaluated in this study are trained under the same experimental conditions to ensure a fair comparison. In particular, the same input image resolution, batch size and data augmentation configuration settings are used for all models. Furthermore, GPU acceleration is enabled to increase computational efficiency. Early stopping is applied during the training, and the reported times represent the convergence time of the model rather than the maximum epoch limits. Therefore, large and optimized models such as EfficientNet and ConvNeXt achieved faster convergence than smaller models. This demonstrates that modern architectures usually contain advanced optimization mechanisms and more efficient computational blocks, reaching convergence in less time with fewer epochs. Moreover, the training time analysis refers solely to the duration of individual model training and ensemble inference time is not included in this training time to avoid confusing the efficiency of model training with ensemble-level performance. The training time for all neural network models across four waste classification datasets is shown in Table 17.
The proposed ensemble method has a longer training time and higher computations on four datasets. This shows that training multiple models together increases the time cost of the processes. Although this strategic combination to improve the performance of the model improves accuracy, it has a significant disadvantage in terms of time and resource cost. This limits the application of this method, especially in systems with limited hardware resources or in real-time applications. Due to the dramatic increase in training time, maintaining, updating and retraining the model becomes more costly. Therefore, the applicability of the ensemble method should be carefully evaluated depending on the usage scenario. This cost may be acceptable if the target is an academic or industrial project requiring high accuracy, but for resource-constrained systems, simpler and optimized models should be preferred.

4.5. Discussion

The combination of transfer learning and ensemble learning allows us to overcome the challenges of standalone ML/DL models. Transfer learning uses knowledge from a wider domain to improve the generalization capability of the model. This becomes specifically effective in using the proposed model in areas that have sparse or very inconsistent training data.
The proposed model makes use of pre-trained deep learning models to generalize from large datasets. This offers high accuracy and efficiency in addition to reduced training time. Moreover, because here the model’s feature space depends upon diverse pre-trained data and not just the subtleties of the target data, this generalization in transfer learning will reduce overfitting on the target dataset and yield more reliable predictions. Ensemble learning combines several individual classifier predictions to improve the classification performance. Thus, it minimizes the weaknesses of individual models, increases the stability of the prediction and provides higher accuracy and robustness compared to using a single model. This approach is particularly useful in waste classification, where variability in waste types can create significant challenges. The experimental results clearly demonstrate that the presented ensemble methods obtain superior accuracy by better generalizing across different waste categories. Experimental results from existing studies in the literature are shown in Table 18 and Table 19. Our results show improvement compared to previous studies; however, differences in data set splits, pre-processing steps and evaluation protocols between studies may affect the reported results. Therefore, the performance comparisons should be interpreted with these differences in mind.
The proposed weighted ensemble method outperforms state-of-the-art DL methods in waste classification. Most of the previous studies which are usually based on a single CNN architecture reported classification accuracies ranging from 80% to 95%, and often struggled with visually similar waste types, such as glass and plastic, or paper and cardboard. For instance, Endah et al. [33] proposed a method for waste classification using the Xception, VGG16 and ResNet-50 models on the TrashNet dataset. They obtained 88% accuracy with the Xception model. Azis et al. [35] introduced a DL-based method for waste classification using Inception-v3. The proposed model achieved an accuracy rate of 92.5% on TrashNet. Kumsetty et al. [47] achieved 93.1% accuracy on the TrashNet datasets using the ResNet34 model. Hossen et al. [48] presented RWC-Net which is based on a combination of DenseNet201 and MobileNet-V2 for waste classification. This model utilizes the feature extraction and learning capabilities of both pre-trained models. The proposed model achieved an accuracy rate of 95.01% on the TrashNet dataset. On the other hand, our ensemble approach leverages the complementary strengths of several architectures to achieve up to 96% accuracy. The proposed model provides better accuracy and efficiency for waste classification. The correct classification of waste is crucial for environmental sustainability and practical applications in waste management systems. Accurately and reliably classifying waste enables more efficient material recovery by minimizing the risks and costs of manual sorting. The proposed ensemble method is scalable and adaptable for real-world applications in waste management. The scalability of the proposed method allows for a number of integrations, such as AI-enabled collection bins, automated facility conveyors and smart city waste management infrastructure. Moreover, it can be deployable in distributed computing environments like cloud-based platforms or edge devices, enabling large-scale, real-time processing of waste images collected from multiple sources. It also contributes to global sustainability goals and environmental protection initiatives.
The results of this study emphasize the effectiveness of the ensemble methods, which show a significant improvement over individual models. Ensemble methods increase the robustness and accuracy of the prediction model by combining different architectures, while individual models are capable on their own, they often suffer from limitations inherent to their specific architecture. For instance, a model can be quite effective in recognizing specific patterns but cannot generalize to a larger dataset due to overfitting. In contrast, ensemble methods alleviate these limitations by leveraging the strengths of multiple models. Ensemble methods reduce the variance in the predictions, thus resulting in a more reliable and accurate model. This research shows that the combination of different models within an ensemble framework can significantly improve overall performance in waste classification. Achieving high accuracy in waste classification is essential for effective waste management and recycling processes. The accurate classification of various waste types facilitates more efficient sorting processes, reduces pollution, and improves the quality of recycled materials.
While the proposed ensemble model provides an increase in accuracy compared to a single model, it comes with a significant computational cost. Training an ensemble model consists of running multiple models simultaneously, which not only increases training time but also increases inference time. This is an important practical consideration in real-world waste management systems where both efficiency and accuracy are important. In resource-constrained waste sorting environments, such as low-power IoT devices installed in recycling facilities, the inference latency is a major issue. For example, in conveyor belt-based waste sorting platforms, materials are transported at high speeds, and the classifier must provide real-time predictions to ensure accurate sorting. In this case, accuracy improvement cannot compensate for the increase in computational requirements, because system efficiency and computational costs are prioritized. Hence, utilizing a single fine-tuned model having optimal hyperparameters might be the most cost-effective solution in such systems. On the other hand, ensemble models can be useful in scenarios where the decisions are less time-sensitive but require high reliability. For instance, in offline applications like municipal waste management or environmental research, a little longer processing time is acceptable if the resulting classifications are more robust and less error-prone. Similarly, in high-risk applications such as hazardous or medical waste detection, the ensemble method can be favored since its more reliable predictions.
Although ensemble learning and transfer learning employing pre-trained models are well-known methods in computer vision, our work advances the field in meaningful ways, especially adapted to a challenging domain like waste classification. Our work specifically applies ensemble methods to waste classification which is a domain characterized by high intra-class variability, and noisy and real-world image conditions. We validate the proposed models on four different datasets, each with a different number of classes, image sources and environmental complexity. This detailed and comprehensive evaluation in our study has not been addressed in previous research. By evaluating multiple datasets of different complexity, our study provides a strong benchmark for future research in waste classification. Furthermore, our study compares two ensemble methods (averaging and weighted averaging) when combining pre-trained models on waste datasets, while previous studies generally use a single pre-trained model or simple CNN architectures, our work uniquely explores ensemble learning with many pre-trained models.

4.6. Limitations and Future Work

While the proposed model has been evaluated on various datasets, there are still some limitations in their full applicability to real-world waste management environments. Although datasets such as TrashBox, Garbage Classification and Waste Pictures offer a wide variety of waste types and visual conditions due to being collected from different sources, they may lack a consistent representation of real-world waste. Furthermore, these datasets contain images derived from online searches, which may introduce biases in composition, lighting and background that do not accurately reflect the unpredictable conditions encountered in waste bins, public spaces or industrial sorting facilities. The TrashNet dataset, while useful for controlled benchmarking, it is limited in both size and environmental variability. Although the datasets utilized in the study provide a baseline for benchmarking, they still lack some significant characteristics of real-world waste. In particular, the waste items in practical waste sorting scenarios are often overlapped or partially covered and this may mask key features and increase intra-class variability. Since most objects, like plastics, contaminated with food residues or fluids are dirty or soiled, they can distort the color, texture and shape information on which the model relies. Furthermore, inconsistent lighting conditions in waste sorting facilities and the mixed materials in the background can further complicate the image, thus making it difficult for models to generalize. These missing features in the datasets can lead to decreased accuracy, increased misclassification rates and reduced robustness when models are used in real-world waste management systems. Therefore, in future research we will evaluate our model with real waste datasets to make it more reliable and accurate in real-world applications. In order to further validate the robustness of our proposed ensemble learning approach, in future research work we will evaluate model performance on unseen, real-world data obtained from waste management systems. Real-world data will include all variability and uncertainty inherent in waste and allow for more intense testing of generalizability. Furthermore, methods such as data augmentation and synthetic data production can be used to address overfitting, especially where there are small or skewed datasets.
Real-world waste images commonly contain various types of noise, such as lighting variations, motion blur and background complexity. This can reduce the performance of DL models. Although the dataset used in this study includes clean images, dealing with noise is crucial to improve the robustness and scalability of waste classification systems used in uncontrolled environments. To address these difficulties, noise reduction techniques such as Gaussian filtering, median filtering and contrast normalization can be applied during the pre-processing stages. Recent studies [69,70] have shown the effect of noise on model accuracy and the potential for integrating noise reduction mechanisms into the classification process. These methods can be included in ensemble classification methods to increase classification accuracy in noisy environments. Future work will focus on integrating noise-sensitive pre-processing methods and noise reduction modules to improve the efficiency of the proposed ensemble method in naturally noisy environments.
A key limitation of our proposed deep ensemble learning method is the lack of interpretability common to most DL and ensemble models, while the methods generally obtain high accuracy, their complex architecture makes it challenging to provide clear, intuitive explanations for individual predictions. This “black box” nature can prevent confidence and transparency which are important factors for practical application in waste classification systems, especially when decisions can impact operational workflows or environmental compliance. As a result, users and other stakeholders may struggle to understand the rationale behind classification results and this can potentially limit adoption in environments where explainability is critical. In order to address this limitation, we used Grad-CAM [68] in this study. Moreover, post hoc explainability methods such as SHAP [71] and LIME [72] can also be used to help interpret model decisions.
Another limitation of our study is that while the proposed method has shown improved accuracy on various datasets, it is designed on established architectures rather than novel architectural designs or new learning algorithms that can better extract characteristic features of waste images. In our future work, we intend to create novel architectures or custom feature extraction approaches specifically taking into account the challenges faced in waste classification.
We used basic averaging and weighted averaging techniques to combine the outputs of multiple pre-trained models. The reason for choosing these techniques was computational efficiency and ease of implementation which are essential properties for practical deployment in waste management systems that have limited computational resources, while these methods effectively improve the classification accuracy by exploiting the complementary strengths of individual models, there is still a need to develop advanced and potentially novel ensemble approaches for waste classification. Therefore, in our future work, we aim to explore some of the techniques such as dynamic weighting schemes that learn to adaptively assign weights based on model confidence or input features, stacking or meta-learning methods and designing custom ensemble architectures for waste classification. Future work exploring such novel approaches is valuable for improving the accuracy and efficiency of waste classification and improving the practical utility of these systems in various real-world scenarios.

5. Conclusions

This paper presents an effective deep ensemble learning model-based waste classification for sustainable waste management systems. The study utilizes high-performance, state-of-the-art CNN architectures and implements average and weighted average ensemble techniques to improve classification accuracy and model robustness. Furthermore, model explainability is ensured by using Grad-CAM for the waste classification task and the robustness and reliability of the findings are verified through five-fold cross-validation and paired t-tests. Experimental results on four different datasets have shown that the weighted average ensemble method outperforms the individual models. The weighted ensemble method performs at 96% accuracy, 94% precision, 97% recall, and 95% F1 score on the TrashNet dataset. These results clearly demonstrate that utilizing the complementary strengths of multiple CNN’s improves the classification accuracy and also reduces the error rate in visually similar classes.
Although the study has achieved promising results, it has several limitations. First, the datasets used are of relatively small size and representation, hence may lack generalization to broader waste categories and real-world scenarios, including variable lighting conditions, surface contamination and deformation. Second, the performance of models is evaluated in a controlled setting rather than a real waste-sorting environment and is not evaluated in terms of hardware constraints like energy efficiency. In future research, large and diverse datasets will be included to enhance the robustness of the model. Moreover, real-time application studies using edge devices or conveyor-based sorting systems will assist in evaluating operational feasibility.
On the whole, the proposed ensemble model improves resource recovery, reduces landfill usage and protects the environment by providing an effective solution for improving automated recycling operations. With appropriate optimization and system integration, this approach can be widely applied in smart waste management infrastructures and also it can support circular economy initiatives and global sustainability goals.

Author Contributions

Conceptualization, A.A., F.Y.O. and İ.K.; methodology, A.A., F.Y.O. and İ.K.; software, A.A.; writing—original draft preparation, A.A.; writing—review and editing, A.A., F.Y.O., İ.K. and S.Ö; visualization, A.A.; supervision, S.Ö. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Datasets used in this study are publicly available on Internet.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1. Experimental results on the TrashNet dataset.
Table A1. Experimental results on the TrashNet dataset.
Macro AverageWeighted Average
ModelAccuracy (%)Precision (%)Recall (%)F1-Score (%)Precision (%)Recall (%)F1-Score (%)
DenseNet12190.588.191.989.491.290.590.6
DenseNet20187.486.386.786.288.187.487.4
MobileNetV284.683.485.083.985.284.684.7
MobileNetV3L89.387.189.988.289.889.389.4
InceptionV383.480.980.480.583.783.483.4
ResNet5092.990.692.691.493.292.993.0
ResNet50V283.081.080.780.783.683.083.2
ResNet10192.590.092.690.993.192.592.6
ResNet101V287.085.287.785.788.287.087.0
Xception83.881.284.481.985.383.884.1
ConvNeXtTiny84.682.484.682.786.184.684.8
ConvNeXtLarge92.990.692.991.593.292.993.0
EfficientNetB786.683.286.584.087.986.686.8
EfficientNetB088.585.389.286.390.088.588.9
EfficientNetV2B089.386.389.987.390.689.389.6
EfficientNetV2L85.881.683.681.987.785.886.3
Averaging ensemble94.993.094.094.095.095.095.0
Weighted average
ensemble
96.094.097.095.096.096.096.0
Table A2. Experimental results on the TrashBox dataset.
Table A2. Experimental results on the TrashBox dataset.
Macro AverageWeighted Average
ModelAccuracy (%)Precision (%)Recall (%)F1-Score (%)Precision (%)Recall (%)F1-Score (%)
DenseNet12183.783.583.883.584.083.783.7
DenseNet20185.185.185.384.985.885.185.2
MobileNetV280.680.580.780.281.180.680.5
MobileNetV3L85.885.785.785.685.985.885.7
InceptionV379.679.779.779.380.379.679.6
ResNet5085.985.885.985.886.085.985.8
ResNet50V282.482.882.082.382.782.482.4
ResNet10187.087.187.087.087.287.087.0
ResNet101V284.984.884.884.784.984.984.8
Xception82.182.382.082.082.582.182.2
ConvNeXt87.287.187.287.187.387.287.2
ConvNeXtLarge94.995.094.894.994.994.994.9
EfficientNetB787.387.187.487.287.487.387.3
EfficientNetB087.187.087.287.087.387.187.1
EfficientNetV2B088.388.088.388.188.488.388.3
EfficientNetV2L87.587.387.587.487.587.587.5
Averaging ensemble93.593.093.093.093.093.093.0
Weighted average
ensemble
95.896.096.096.096.096.096.0
Table A3. Experimental results on the Waste Pictures dataset.
Table A3. Experimental results on the Waste Pictures dataset.
Macro AverageWeighted Average
ModelAccuracy (%)Precision (%)Recall (%)F1-Score (%)Precision (%)Recall (%)F1-Score (%)
DenseNet12189.890.687.688.690.189.889.6
DenseNet20190.390.888.989.590.690.390.3
MobileNetV289.389.988.789.089.789.389.3
MobileNetV3L91.191.389.890.491.291.191.0
InceptionV386.587.384.885.687.086.586.5
ResNet5090.891.089.890.290.990.890.7
ResNet50V289.790.888.489.290.289.789.7
ResNet10191.891.990.390.992.091.891.8
ResNet101V288.087.886.786.988.588.087.9
Xception87.888.485.586.588.187.887.7
ConvNeXt90.591.189.189.890.890.590.5
ConvNeXtLarge97.297.496.997.197.297.297.2
EfficientNetB791.491.190.290.591.591.491.3
EfficientNetB093.893.993.593.693.893.893.8
EfficientNetV2B093.893.993.193.393.993.893.7
EfficientNetV2L90.490.989.490.090.590.490.4
Averaging ensemble97.297.096.097.097.097.097.0
Weighted average
ensemble
98.098.098.098.098.098.098.0
Table A4. Experimental results on the Garbage Classification dataset.
Table A4. Experimental results on the Garbage Classification dataset.
Macro AverageWeighted Average
ModelAccuracy (%)Precision (%)Recall (%)F1-Score (%)Precision (%)Recall (%)F1-Score (%)
DenseNet12195.493.993.293.595.595.495.4
DenseNet20196.394.994.794.796.596.396.3
MobileNetV295.393.893.393.495.595.395.3
MobileNetV3L95.694.293.793.995.795.695.6
InceptionV393.090.189.589.793.193.093.0
ResNet5096.194.894.894.796.296.196.1
ResNet50V294.191.090.890.794.294.194.1
ResNet10195.994.594.094.096.295.995.9
ResNet101V294.892.392.492.394.994.894.8
Xception94.592.191.791.794.794.594.5
ConvNeXt96.895.695.495.496.996.896.8
ConvNeXtLarge98.898.698.798.698.998.898.8
EfficientNetB795.994.694.094.395.995.995.9
EfficientNetB096.395.094.994.996.496.396.3
EfficientNetV2B097.696.596.696.597.797.697.6
EfficientNetV2L96.595.495.195.296.696.596.5
Averaging ensemble98.698.098.098.099.099.099.0
Weighted average
ensemble
99.199.099.099.099.099.099.0

References

  1. Kaza, S.; Yao, L.; Bhada-Tata, P.; Van Woerden, F. What a Waste 2.0: A Global Snapshot of Solid Waste Management to 2050; World Bank Publications: Washington, DC, USA, 2018. [Google Scholar]
  2. Kang, Z.; Yang, J.; Li, G.; Zhang, Z. An Automatic Garbage Classification System Based on Deep Learning. IEEE Access 2020, 8, 140019–140029. [Google Scholar] [CrossRef]
  3. Huang, G.L.; He, J.; Xu, Z.; Huang, G. A combination model based on transfer learning for waste classification. Concurr. Comput. Pract. Exp. 2020, 32, e5751. [Google Scholar] [CrossRef]
  4. Yang, J.; Zeng, Z.; Wang, K.; Zou, H.; Xie, L. GarbageNet: A Unified Learning Framework for Robust Garbage Classification. IEEE Trans. Artif. Intell. 2021, 2, 372–380. [Google Scholar] [CrossRef]
  5. Okoya, S.A.; Oyinlola, M.; Schröder, P.; Kolade, O.; Abolfathi, S. Enhancing decentralised recycling solutions with digital technologies. In Digital Innovations for a Circular Plastic Economy in Africa; Taylor & Francis: Abingdon, UK, 2023; Volume 208. [Google Scholar]
  6. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; Pereira, F., Burges, C., Bottou, L., Weinberger, K., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2012; Volume 25. [Google Scholar]
  7. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going Deeper with Convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  8. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  9. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015; PMLR: Cambridge MA, USA, 2015; pp. 448–456. [Google Scholar]
  10. Xie, S.; Girshick, R.; Dollar, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  11. Huang, G.; Liu, Z.; van der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26, July 2017. [Google Scholar]
  12. Tan, M.; Le, Q. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; Chaudhuri, K., Salakhutdinov, R., Eds.; PMLR: Cambridge MA, USA, 2019; Volume 97, pp. 6105–6114. [Google Scholar]
  13. Pacal, I.; Kilicarslan, S.; Ozdemir, B.; Deveci, M.; Kadry, S. Efficient and autonomous detection of olive leaf diseases using AI-enhanced MetaFormer. Artif. Intell. Rev. 2025, 58, 303. [Google Scholar] [CrossRef]
  14. Pacal, I.; Akhan, O.; Deveci, R.T.; Deveci, M. NeXtBrain: Combining local and global feature learning for brain tumor classification. Brain Res. 2025, 1863, 149762. [Google Scholar] [CrossRef] [PubMed]
  15. Pacal, I.; Attallah, O. InceptionNeXt-Transformer: A novel multi-scale deep feature learning architecture for multimodal breast cancer diagnosis. Biomed. Signal Process. Control 2025, 110, 108116. [Google Scholar] [CrossRef]
  16. Pacal, I.; Attallah, O. Hybrid deep learning model for automated colorectal cancer detection using local and global feature extraction. Knowl.-Based Syst. 2025, 319, 113625. [Google Scholar] [CrossRef]
  17. Nahavandi, S.; Alizadehsani, R.; Nahavandi, D.; Mohamed, S.; Mohajer, N.; Rokonuzzaman, M.; Hossain, I. A comprehensive review on autonomous navigation. ACM Comput. Surv. 2025, 57, 1–67. [Google Scholar] [CrossRef]
  18. Cerrato, S.; Mazzia, V.; Salvetti, F.; Martini, M.; Angarano, S.; Navone, A.; Chiaberge, M. A deep learning driven algorithmic pipeline for autonomous navigation in row-based crops. IEEE Access 2024, 12, 138306–138318. [Google Scholar] [CrossRef]
  19. Sakr, G.E.; Mokbel, M.; Darwich, A.; Khneisser, M.N.; Hadi, A. Comparing deep learning and support vector machines for autonomous waste sorting. In Proceedings of the 2016 IEEE International Multidisciplinary Conference on Engineering Technology (IMCET), Beirut, Lebanon, 2–4 November 2016; pp. 207–212. [Google Scholar] [CrossRef]
  20. Yang, M.; Thung, G. Classification of Trash for Recyclability Status; Technical Report; CS229 Project Report; Stanford University: Stanford, CA, USA, 2016. [Google Scholar]
  21. Bircanoğlu, C.; Atay, M.; Beşer, F.; Genc, O.; Kızrak, M.A. RecycleNet: Intelligent Waste Sorting Using Deep Neural Networks. In Proceedings of the 2018 Innovations in Intelligent Systems and Applications (INISTA), Thessaloniki, Greece, 3–5 July 2018; pp. 1–7. [Google Scholar] [CrossRef]
  22. Zeng, M.; Lu, X.; Xu, W.; Zhou, T.; Liu, Y. PublicGarbageNet: A Deep Learning Framework for Public Garbage Classification. In Proceedings of the 2020 39th Chinese Control Conference (CCC), Shenyang, China, 27–29 July 2020; pp. 7200–7205. [Google Scholar] [CrossRef]
  23. Chen, Z.; Yang, J.; Chen, L.; Jiao, H. Garbage classification system based on improved ShuffleNet v2. Resour. Conserv. Recycl. 2022, 178, 106090. [Google Scholar] [CrossRef]
  24. Sagi, O.; Rokach, L. Ensemble learning: A survey. WIREs Data Min. Knowl. Discov. 2018, 8, e1249. [Google Scholar] [CrossRef]
  25. Zhihong, C.; Hebin, Z.; Yanbo, W.; Binyan, L.; Yu, L. A vision-based robotic grasping system using deep learning for garbage sorting. In Proceedings of the 2017 36th Chinese Control Conference (CCC), Dalian, China, 26–28 July 2017; pp. 11223–11226. [Google Scholar] [CrossRef]
  26. Bai, J.; Lian, S.; Liu, Z.; Wang, K.; Liu, D. Deep Learning Based Robot for Automatically Picking Up Garbage on the Grass. IEEE Trans. Consum. Electron. 2018, 64, 382–389. [Google Scholar] [CrossRef]
  27. Chu, Y.; Huang, C.; Xie, X.; Tan, B.; Kamal, S.; Xiong, X. Multilayer Hybrid Deep-Learning Method for Waste Classification and Recycling. Comput. Intell. Neurosci. 2018, 2018, 5060857. [Google Scholar] [CrossRef]
  28. Rabano, S.L.; Cabatuan, M.K.; Sybingco, E.; Dadios, E.P.; Calilung, E.J. Common Garbage Classification Using MobileNet. In Proceedings of the 2018 IEEE 10th International Conference on Humanoid, Nanotechnology, Information Technology, Communication and Control, Environment and Management (HNICEM), Baguio City, Philippines, 29 November–2 December 2018; pp. 1–4. [Google Scholar] [CrossRef]
  29. Satvilkar, M. Image Based Trash Classification Using Machine Learning Algorithms for Recyclability Status. Master’s Thesis, National College of Ireland, Dublin, Ireland, 2018. [Google Scholar]
  30. Zeng, D.; Zhang, S.; Chen, F.; Wang, Y. Multi-Scale CNN Based Garbage Detection of Airborne Hyperspectral Data. IEEE Access 2019, 7, 104514–104527. [Google Scholar] [CrossRef]
  31. Adedeji, O.; Wang, Z. Intelligent Waste Classification System Using Deep Learning Convolutional Neural Network. Procedia Manuf. 2019, 35, 607–612. [Google Scholar] [CrossRef]
  32. Ruiz, V.; Sánchez, Á.; Vélez, J.F.; Raducanu, B. Automatic Image-Based Waste Classification. In Proceedings of the From Bioinspired Systems and Biomedical Applications to Machine Learning, Almería, Spain, 3–7 June 2019; Springer International Publishing: Cham, Switzerland, 2019; pp. 422–431. [Google Scholar]
  33. Rismiyati; Endah, S.N.; Khadijah; Shiddiq, I.N. Xception Architecture Transfer Learning for Garbage Classification. In Proceedings of the 2020 4th International Conference on Informatics and Computational Sciences (ICICoS), Semarang, Indonesia, 10–11 November 2020; pp. 1–4. [Google Scholar] [CrossRef]
  34. Toğaçar, M.; Ergen, B.; Cömert, Z. Waste classification using AutoEncoder network with integrated feature selection method in convolutional neural network models. Measurement 2020, 153, 107459. [Google Scholar] [CrossRef]
  35. Azis, F.A.; Suhaimi, H.; Abas, E. Waste Classification using Convolutional Neural Network. In Proceedings of the 2020 2nd International Conference on Information Technology and Computer Communications, Kuala Lumpur, Malaysia, 12–14 August 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 9–13. [Google Scholar] [CrossRef]
  36. Shi, C.; Tan, C.; Wang, T.; Wang, L. A Waste Classification Method Based on a Multilayer Hybrid Convolution Neural Network. Appl. Sci. 2021, 11, 8572. [Google Scholar] [CrossRef]
  37. Zhang, Q.; Zhang, X.; Mu, X.; Wang, Z.; Tian, R.; Wang, X.; Liu, X. Recyclable waste image recognition based on deep learning. Resour. Conserv. Recycl. 2021, 171, 105636. [Google Scholar] [CrossRef]
  38. Majchrowska, S.; Mikołajczyk, A.; Ferlin, M.; Klawikowska, Z.; Plantykow, M.A.; Kwasigroch, A.; Majek, K. Deep learning-based waste detection in natural and urban environments. Waste Manag. 2022, 138, 274–284. [Google Scholar] [CrossRef] [PubMed]
  39. Kumsetty, N.V.; Bhat Nekkare, A.; Sowmya, S.K.; Kumar, A.M. TrashBox: Trash Detection and Classification using Quantum Transfer Learning. In Proceedings of the 2022 31st Conference of Open Innovations Association (FRUCT), Helsinki, Finland, 27–29 April 2022; pp. 125–130. [Google Scholar] [CrossRef]
  40. Zhou, H.; Yu, X.; Alhaskawi, A.; Dong, Y.; Wang, Z.; Jin, Q.; Hu, X.; Liu, Z.; Kota, V.G.; Abdulla, M.H.A.H.; et al. A deep learning approach for medical waste classification. Sci. Rep. 2022, 12, 2159. [Google Scholar] [CrossRef]
  41. Waste Pictures. 2019. Available online: https://www.kaggle.com/wangziang/waste-pictures (accessed on 1 May 2024).
  42. Chen, Y.; Han, W.; Jin, J.; Wang, H.; Xing, Q.; Zhang, Y. Clean Our City: An Automatic Urban Garbage Classification Algorithm Using Computer Vision and Transfer Learning Technologies. J. Phys. Conf. Ser. 2021, 1994, 012022. [Google Scholar] [CrossRef]
  43. Shukurov, R. Garbage classification based on fine-tuned state-of-the-art models. In Proceedings of the 2023 9th International Conference on Control, Decision and Information Technologies (CoDIT), Rome, Italy, 3–6 July 2023; pp. 841–846. [Google Scholar] [CrossRef]
  44. Dey, D.; Shama, U.S.; Akash, M.; Karim, D.Z. Automatic Waste Classification System using Deep Leaning Techniques. In Proceedings of the 2023 3rd International Conference on Electrical, Computer, Communications and Mechatronics Engineering (ICECCME), Tenerife, Canary Islands, Spain, 19–21 July 2023; pp. 1–6. [Google Scholar] [CrossRef]
  45. Huang, L.; Li, M.; Xu, T.; Dong, S.Q. A waste classification method based on a capsule network. Environ. Sci. Pollut. Res. 2023, 30, 86454–86462. [Google Scholar] [CrossRef]
  46. Jin, S.; Yang, Z.; Królczykg, G.; Liu, X.; Gardoni, P.; Li, Z. Garbage detection and classification using a new deep learning-based machine vision system as a tool for sustainable waste recycling. Waste Manag. 2023, 162, 123–130. [Google Scholar] [CrossRef]
  47. Kumsetty, N.V.; Nekkare, A.B.; Sowmya Kamath, S.; Anand Kumar, M. An Approach for Waste Classification Using Data Augmentation and Transfer Learning Models. In Machine Vision and Augmented Intelligence: Select Proceedings of MAI 2022; Springer Nature: Singapore, 2023; pp. 357–368. [Google Scholar] [CrossRef]
  48. Hossen, M.M.; Majid, M.E.; Kashem, S.B.A.; Khandakar, A.; Nashbat, M.; Ashraf, A.; Hasan-Zia, M.; Kunju, A.K.A.; Kabir, S.; Chowdhury, M.E.H. A Reliable and Robust Deep Learning Model for Effective Recyclable Waste Classification. IEEE Access 2024, 12, 13809–13821. [Google Scholar] [CrossRef]
  49. Lilhore, U.K.; Simaiya, S.; Dalal, S.; Damaševičius, R. A smart waste classification model using hybrid CNN-LSTM with transfer learning for sustainable environment. Multimed. Tools Appl. 2024, 83, 29505–29529. [Google Scholar] [CrossRef]
  50. Quan, M.K.; Nguyen, D.C.; Nguyen, V.D.; Wijayasundara, M.; Setunge, S.; Pathirana, P.N. Toward Privacy-Preserving Waste Classification in the Internet of Things. IEEE Internet Things J. 2024, 11, 24814–24830. [Google Scholar] [CrossRef]
  51. Sarswat, P.K.; Singh, R.S.; Pathapati, S.V.S.H. Real time electronic-waste classification algorithms using the computer vision based on Convolutional Neural Network (CNN): Enhanced environmental incentives. Resour. Conserv. Recycl. 2024, 207, 107651. [Google Scholar] [CrossRef]
  52. Lin, Z.; Xu, H.; Zhou, M.; Wang, B.; Qin, H. Waste classification strategy based on multi-scale feature fusion for intelligent waste recycling in office buildings. Waste Manag. 2024, 190, 443–454. [Google Scholar] [CrossRef]
  53. Kumar, A.K.; Ali, Y.; Kumar, R.R.; Assaf, M.H.; Ilyas, S. Artificial Intelligent and Internet of Things framework for sustainable hazardous waste management in hospitals. Waste Manag. 2025, 203, 114816. [Google Scholar] [CrossRef]
  54. Alzubaidi, L.; Zhang, J.; Humaidi, A.J.; Al-Dujaili, A.; Duan, Y.; Al-Shamma, O.; Santamaría, J.; Fadhel, M.A.; Al-Amidie, M.; Farhan, L. Review of deep learning: Concepts, CNN architectures, challenges, applications, future directions. J. Big Data 2021, 8, 53. [Google Scholar] [CrossRef]
  55. Donnelly, J.; Daneshkhah, A.; Abolfathi, S. Forecasting global climate drivers using Gaussian processes and convolutional autoencoders. Eng. Appl. Artif. Intell. 2024, 128, 107536. [Google Scholar] [CrossRef]
  56. Liu, Z.; Mao, H.; Wu, C.Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A ConvNet for the 2020s. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 11976–11986. [Google Scholar]
  57. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  58. Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  59. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  60. Bhuiyan, M.; Islam, M.S. A new ensemble learning approach to detect malaria from microscopic red blood cell images. Sensors Int. 2023, 4, 100209. [Google Scholar] [CrossRef]
  61. Vij, R.; Arora, S. A hybrid evolutionary weighted ensemble of deep transfer learning models for retinal vessel segmentation and diabetic retinopathy detection. Comput. Electr. Eng. 2024, 115, 109107. [Google Scholar] [CrossRef]
  62. Kang, M.; Ahn, J.; Lee, K. Opinion mining using ensemble text hidden Markov models for text classification. Expert Syst. Appl. 2018, 94, 218–227. [Google Scholar] [CrossRef]
  63. Chen, H.; Zhang, Z.; Huang, S.; Hu, J.; Ni, W.; Liu, J. TextCNN-based ensemble learning model for Japanese Text Multi-classification. Comput. Electr. Eng. 2023, 109, 108751. [Google Scholar] [CrossRef]
  64. Cao, Y.; Wang, Z.; Ding, H.; Zhang, J.; Li, B. An intrusion detection system based on stacked ensemble learning for IoT network. Comput. Electr. Eng. 2023, 110, 108836. [Google Scholar] [CrossRef]
  65. Jethanandani, M.; Sharma, A.; Perumal, T.; Chang, J.R. Multi-label classification based ensemble learning for human activity recognition in smart home. Internet Things 2020, 12, 100324. [Google Scholar] [CrossRef]
  66. Thung, G.; Yang, M. Trashnet. 2016. Available online: https://github.com/garythung/trashnet (accessed on 1 May 2024).
  67. Garbage Classification. 2021. Available online: https://www.kaggle.com/datasets/mostafaabla/garbage-classification (accessed on 1 May 2024).
  68. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  69. Attallah, O.; Ibrahim, R.A.; Zakzouk, N.E. A lightweight deep learning framework for transformer fault diagnosis in smart grids using multiple scale CNN features. Sci. Rep. 2025, 15, 14505. [Google Scholar] [CrossRef] [PubMed]
  70. Chen, C.; Liang, R.; Song, M.; Zhang, Z.; Tao, J.; Yan, B.; Cheng, Z.; Chen, G. Noise-assisted data enhancement promoting image classification of municipal solid waste. Resour. Conserv. Recycl. 2024, 209, 107790. [Google Scholar] [CrossRef]
  71. Lundberg, S.M.; Lee, S.I. A Unified Approach to Interpreting Model Predictions. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
  72. Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why should i trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1135–1144. [Google Scholar]
Figure 1. Workflow of the proposed methodology.
Figure 1. Workflow of the proposed methodology.
Sustainability 18 00024 g001
Figure 2. Proposed model.
Figure 2. Proposed model.
Sustainability 18 00024 g002
Figure 3. Sample data images in the TrashNet dataset.
Figure 3. Sample data images in the TrashNet dataset.
Sustainability 18 00024 g003
Figure 4. Loss and accuracy of the models on different datasets. (a) TrashNet, (b) TrashBox, (c) Garbage Classification, and (d) Waste Pictures dataset.
Figure 4. Loss and accuracy of the models on different datasets. (a) TrashNet, (b) TrashBox, (c) Garbage Classification, and (d) Waste Pictures dataset.
Sustainability 18 00024 g004
Figure 5. Confusion matrices for TrashNet dataset. (a) DenseNet121, (b) ResNet50, (c) ConvNeXtLarge and (d) weighted average ensemble.
Figure 5. Confusion matrices for TrashNet dataset. (a) DenseNet121, (b) ResNet50, (c) ConvNeXtLarge and (d) weighted average ensemble.
Sustainability 18 00024 g005
Figure 6. Grad-CAM visualizations of the ConvNeXtLarge model on the TrashNet dataset.
Figure 6. Grad-CAM visualizations of the ConvNeXtLarge model on the TrashNet dataset.
Sustainability 18 00024 g006
Table 2. The features of pre-trained models.
Table 2. The features of pre-trained models.
ModelYearParameterDepthComplexity
InceptionV3201523.9 MDeepComplex
ResNet50201525.6 MDeepModerate
ResNet50V2201625.6 MDeepModerate
ResNet101201544.7 MDeepComplex
ResNet101V2201644.7 MDeepComplex
DenseNet12120168.1 MDeepComplex
DenseNet201201720.2 MDeepComplex
Xception201722.9 MDeepComplex
MobileNetV220183.5 MShallowLightweight
MobileNetV3L20193.9 MShallowLightweight
EfficientNetB020195.3 MModerateEfficient
EfficientNetB7201966.7 MModerateEfficient
EfficientNetV2B020217.2 MModerateEfficient
EfficientNetV2L2021119.0 MModerateEfficient
ConvNeXtTiny202228.6 MModerateComplex
ConvNeXtLarge2022197.7 MModerateComplex
Table 3. Statistics of waste datasets.
Table 3. Statistics of waste datasets.
Dataset# Images# ClassesClass ImbalanceTypeAnnotation
TrashNet25276LowClassificationClear background
TrashBox17,8537LowClassificationScraped from web
Garbage Classification15,51512HighClassificationScraped from web
Waste Pictures23,08734HighClassificationScraped from Google search
Table 4. Number of images in the training, validation and test sets before oversampling.
Table 4. Number of images in the training, validation and test sets before oversampling.
DatasetTrainValidationTestTotal
TrashNet20212532532527
TrashBox14,2791781179317,853
Waste Pictures14,2683567525223,087
Garbage Classification12,4121551155215,515
Table 5. Number of images in the training, validation and test sets after oversampling.
Table 5. Number of images in the training, validation and test sets after oversampling.
DatasetTrainValidationTestTotal
TrashNet35642532534070
TrashBox16,8351781179320,409
Waste Pictures14,2683567525223,087
Garbage Classification12,4121551155215,515
Table 6. Loss and accuracy of the pre-trained models.
Table 6. Loss and accuracy of the pre-trained models.
DatasetTrashNetTrashBoxWaste PicturesGarbage Classification
ModelLossAccuracyLossAccuracyLossAccuracyLossAccuracy
DenseNet1210.290.910.490.840.340.900.130.95
DenseNet2010.320.870.440.850.330.900.120.96
MobileNetV20.450.850.570.810.360.890.170.95
MobileNetV3L0.340.890.450.860.310.910.120.96
InceptionV30.480.830.600.800.440.870.220.93
ResNet500.270.930.490.860.290.910.140.96
ResNet50V20.480.830.580.820.350.900.180.94
ResNet1010.270.920.450.870.300.920.130.96
ResNet101V20.440.870.520.850.420.880.180.95
Xception0.450.840.520.820.400.880.160.94
ConvNeXtTiny0.440.850.390.870.310.910.120.97
ConvNeXtLarge0.180.930.170.950.100.970.050.99
EfficientNetB00.330.890.390.870.210.940.120.96
EfficientNetB70.380.870.390.870.320.910.130.96
EfficientNetV2B00.350.890.360.880.220.940.110.98
EfficientNetV2L0.390.860.410.870.340.900.130.97
Table 7. Summary of the CNN models that are trained on all datasets.
Table 7. Summary of the CNN models that are trained on all datasets.
ModelOptimizerLR# Parameters
DenseNet121RMSprop1 × 10−37,044,679
DenseNet201RMSprop1 × 10−318,335,431
MobileNetV2RMSprop1 × 10−32,266,951
MobileNetV3LRMSprop1 × 10−33,003,079
InceptionV3Adam1 × 10−321,817,127
ResNet50Adam1 × 10−323,602,055
ResNet50V2SGD1 × 10−223,579,143
ResNet101RMSprop1 × 10−342,672,519
ResNet101V2SGD1 × 10−242,640,903
XceptionRMSprop1 × 10−320,875,823
ConvNeXtTinyAdam1 × 10−327,825,511
ConvNeXtLargeAdam1 × 10−3196,241,095
EfficientNetB0Adam1 × 10−34,058,538
EfficientNetV2B0Adam1 × 10−35,928,279
EfficientNetV2LAdam1 × 10−3117,755,815
EfficientNetB7Adam1 × 10−364,115,614
Table 8. Evaluation of the averaging and weighted average ensemble models’ accuracy on the TrashNet dataset.
Table 8. Evaluation of the averaging and weighted average ensemble models’ accuracy on the TrashNet dataset.
ModelsAveraging
Ensemble (%)
Weighted Average
Ensemble (%)
Weights
DenseNet121, ResNet50, ConvNeXtLarge94.996.00.33, 0.22, 0.44
DenseNet121, ResNet101, ConvNeXtLarge93.796.00.35, 0.18, 0.47
ResNet50, ResNet101, ConvNeXtLarge94.995.30.17, 0.33, 0.5
DenseNet121, MobileNetV3, ConvNeXtLarge94.195.30.17, 0.33, 0.5
ResNet50, MobileNetV3, ConvNeXtLarge94.195.30.34, 0.06, 0.6
ResNet101, EfficientNetB0, ConvNeXtLarge93.394.50.25, 0.25, 0.5
EfficientNetB7, Xception, ConvNeXtLarge91.394.10.17, 0.16, 0.67
ResNet50, EfficientNetB0, ConvNeXtTiny91.794.10.62, 0.31, 0.07
Table 9. Summary of experimental results on the TrashNet dataset.
Table 9. Summary of experimental results on the TrashNet dataset.
ModelAccuracy (%)Precision (%)Recall (%)F1-Score (%)
DenseNet12190.588.191.989.4
ResNet5092.990.692.691.4
ConvNeXtLarge92.990.692.991.5
Averaging ensemble94.993.094.094.0
Weighted average
ensemble
96.094.097.095.0
Table 10. Evaluation of the averaging and weighted average ensemble models’ accuracy on the TrashBox dataset.
Table 10. Evaluation of the averaging and weighted average ensemble models’ accuracy on the TrashBox dataset.
ModelsAveraging
Ensemble (%)
Weighted Average
Ensemble (%)
Weights
ResNet50, ResNet101, ConvNeXtLarge93.595.80.07, 0.33, 0.6
ResNet101, EfficientNetB0, ConvNeXtLarge94.195.60.33, 0.09, 0.58
DenseNet121, ResNet101, ConvNeXtLarge93.895.60.09, 0.33, 0.58
DenseNet121, MobileNetV3, ConvNeXtLarge93.495.50.1, 0.2, 0.7
DenseNet121, ResNet50, ConvNeXtLarge93.095.40.18, 0.18, 0.64
EfficientNetB7, Xception, ConvNeXtLarge93.995.30.22, 0.22, 0.56
ResNet50, EfficientNetV2B0, ConvNeXtTiny91.591.90.3, 0.35, 0.35
DenseNet121, DenseNet201, EfficientNetB088.689.70.11, 0.33, 0.56
Table 11. Summary of experimental results on the TrashBox dataset.
Table 11. Summary of experimental results on the TrashBox dataset.
ModelAccuracy (%)Precision (%)Recall (%)F1-Score (%)
ResNet5085.985.885.985.8
ResNet10187.087.187.087.0
ConvNeXtLarge94.995.094.894.9
Averaging ensemble93.593.093.093.0
Weighted average
ensemble
95.896.096.096.0
Table 12. Evaluation of the averaging and weighted average ensemble models’ accuracy on the Waste Pictures dataset.
Table 12. Evaluation of the averaging and weighted average ensemble models’ accuracy on the Waste Pictures dataset.
ModelsAveraging
Ensemble (%)
Weighted Average
Ensemble (%)
Weights
ResNet50, EfficientNetB0, ConvNeXtLarge97.298.00.22, 0.22, 0.56
ResNet101, EfficientNetB0, ConvNeXtLarge96.898.00.19, 0.25, 0.56
EfficientNetB0, EfficientNetV2B0, ConvNeXtLarge97.397.90.25, 0.19, 0.56
EfficientNetV2B0, ResNet50, ConvNeXtLarge97.097.90.27, 0.18, 0.54
EfficientNetB0, InceptionV3, ConvNeXtLarge97.097.90.31, 0.07, 0.62
DenseNet121, ResNet101, ConvNeXtLarge96.097.80.1, 0.2, 0.7
ResNet50, ResNet101, ConvNeXtLarge96.197.70.15, 0.23, 0.62
ResNet101V2, InceptionV3, ConvNeXtLarge95.597.70.13, 0.12, 0.75
Table 13. Summary of experimental results on the Waste Pictures dataset.
Table 13. Summary of experimental results on the Waste Pictures dataset.
ModelAccuracy (%)Precision (%)Recall (%)F1-Score (%)
ResNet5090.891.089.890.2
ConvNeXtLarge97.297.496.997.1
EfficientNetB093.893.993.593.6
Averaging ensemble97.297.096.097.0
Weighted average
ensemble
98.098.098.098.0
Table 14. Evaluation of the averaging and weighted average ensemble models’ accuracy on the Garbage Classification dataset.
Table 14. Evaluation of the averaging and weighted average ensemble models’ accuracy on the Garbage Classification dataset.
ModelsAveraging
Ensemble (%)
Weighted Average
Ensemble (%)
Weights
EfficientNetB7, Xception, ConvNeXtLarge98.699.10.2, 0.2, 0.6
DenseNet121, MobileNetV3, ConvNeXtLarge98.099.00.06, 0.41, 0.53
ResNet50, ResNet101, ConvNeXtLarge98.199.00.08, 0.25, 0.67
ResNet101, EfficientNetB0, ConvNeXtLarge98.399.00.25, 0.25, 0.5
DenseNet121, ResNet101, ConvNeXtLarge98.199.00.14, 0.14, 0.72
DenseNet121, ResNet50, ConvNeXtLarge97.998.90.14, 0.14, 0.72
ResNet50, MobileNetV3, ConvNeXtLarge98.398.90.2, 0.2, 0.6
ResNet50, EfficientNetV2B0, ConvNeXtTiny97.697.70.1, 0.8, 0.1
Table 15. Summary of experimental results on the Garbage Classification dataset.
Table 15. Summary of experimental results on the Garbage Classification dataset.
ModelAccuracy (%)Precision (%)Recall (%)F1-Score (%)
Xception94.592.191.791.7
ConvNeXtLarge98.898.698.798.6
EfficientNetB795.994.694.094.3
Averaging ensemble98.698.098.098.0
Weighted average
ensemble
99.199.099.099.0
Table 16. Statistical significance results on the TrashNet dataset.
Table 16. Statistical significance results on the TrashNet dataset.
ModelAccuracy p-ValueF1-Score p-Value
DenseNet1210.0010.001
ResNet500.0020.002
MobileNetV3Large0.0010.001
EfficientNetV2B00.0020.001
ConvNeXtLarge0.020.03
Table 17. Training time of models.
Table 17. Training time of models.
DatasetTrashNetTrashBoxWaste PicturesGarbage Classification
ModelTraining Time(s)
DenseNet121770359628422173
DenseNet201674361512542063
MobileNetV2515358219161997
MobileNetV3L1001468523771400
InceptionV3746390221652607
ResNet50688460016483408
ResNet50V2737424521221685
ResNet101624458521211399
ResNet101V2842352716701401
Xception596355623851702
ConvNeXtTiny1034535628403821
ConvNeXtLarge429467214832667
EfficientNetB0497286428451869
EfficientNetB71066868918092585
EfficientNetV2B0858603330661860
EfficientNetV2L882795637503405
Proposed Method188713,85659756953
Table 18. Performance comparison of different methods on TrashNet dataset.
Table 18. Performance comparison of different methods on TrashNet dataset.
ReferenceMethodTrain / Val / Test Ratio (%)Accuracy (%)
(Yang et al., 2016) [20]SIFT + SVM70/13/1763
(Kumsetty et al., 2022) [39]Quantum ResNet-50Not specified80.5
(Bircanoğlu et al., 2018) [21]RecycleNet70/13/1781
(Adedeji et al., 2019) [31]ResNet-50 + SVMNot specified87
(Rabano et al., 2018) [28]MobileNetNot specified87.2
(Endah et al., 2020) [33]Xception80/2088
(Ruiz et al., 2019) [32]Inception-ResNet80/10/1088.6
(Satvilkar et al., 2018) [29]CNN75/2589.8
(Quan et al., 2024) [50]VGG1980/2090.0
(Huang et al., 2023) [45]ResNet1880/2091.4
(Azis et al., 2020) [35]Inception-v380/10/1092.5
(Shi et al., 2021) [36]MLH-CNN,80/2092.6
(Kumsetty et al., 2023) [47]ResNet-3480/10/1093.1
(Lin et al., 2024) [52]VGG1680/2094.1
(Hossen et al., 2024) [48]DenseNet201,MobileNet-v270/20/1095.0
Proposed model (average ensemble)DenseNet121, ResNet50, ConvNeXtLarge80/10/1094.9
Proposed model (weighted average ensemble)DenseNet121, ResNet50, ConvNeXtLarge80/10/1096.0
Table 19. Performance comparison of different methods on Garbage Classification dataset.
Table 19. Performance comparison of different methods on Garbage Classification dataset.
ReferenceMethod#ClassTrain/Val/Test Ratio (%)Accuracy (%)
(Chen et al., 2021) [42]InceptionV31280/10/1093.1
(Shukurov, 2023) [43]ResNeXt1280/10/1095
(Dey et al., 2023) [44]custom CNN880/2097.58
Proposed model (average ensemble)EfficientNetB7, Xception, ConvNeXtLarge1280/10/1098.6
Proposed model (weighted average ensemble)EfficientNetB7, Xception, ConvNeXtLarge1280/10/1099.1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alkılınç, A.; Okay, F.Y.; Kök, İ.; Özdemir, S. Deep Ensemble Learning Model for Waste Classification Systems. Sustainability 2026, 18, 24. https://doi.org/10.3390/su18010024

AMA Style

Alkılınç A, Okay FY, Kök İ, Özdemir S. Deep Ensemble Learning Model for Waste Classification Systems. Sustainability. 2026; 18(1):24. https://doi.org/10.3390/su18010024

Chicago/Turabian Style

Alkılınç, Ahmet, Feyza Yıldırım Okay, İbrahim Kök, and Suat Özdemir. 2026. "Deep Ensemble Learning Model for Waste Classification Systems" Sustainability 18, no. 1: 24. https://doi.org/10.3390/su18010024

APA Style

Alkılınç, A., Okay, F. Y., Kök, İ., & Özdemir, S. (2026). Deep Ensemble Learning Model for Waste Classification Systems. Sustainability, 18(1), 24. https://doi.org/10.3390/su18010024

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop