Next Article in Journal
The Assessment of the Spatiotemporal Characteristics of the Eco-Environmental Quality in the Chishui River Basin from 2000 to 2020
Previous Article in Journal
Critical Conditions for Wellbore Failure during CO2-ECBM Considering Sorption Stress
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Performance Improvement of Machine Learning Model Using Autoencoder to Predict Demolition Waste Generation Rate

1
School of Science and Technology Acceleration Engineering, Kyungpook National University, Daegu 41566, Republic of Korea
2
School of Architectural, Civil, Environmental and Energy Engineering, Kyungpook National University, Daegu 41566, Republic of Korea
3
Division of Smart Safety Engineering, Dongguk University Wise Campus, 123 Dongdae-ro, Gyeongju 38066, Republic of Korea
*
Author to whom correspondence should be addressed.
Sustainability 2023, 15(4), 3691; https://doi.org/10.3390/su15043691
Submission received: 31 January 2023 / Revised: 12 February 2023 / Accepted: 15 February 2023 / Published: 16 February 2023

Abstract

:
Owing to the rapid increase in construction and demolition (C&D) waste, the information of waste generation (WG) has been advantageously utilized as a strategy for C&D waste management. Recently, artificial intelligence (AI) has been strategically employed to obtain accurate WG information. Thus, this study aimed to manage demolition waste (DW) by combining three algorithms: artificial neural network (multilayer perceptron) (ANN-MLP), support vector regression (SVR), and random forest (RF) with an autoencoder (AE) to develop and test hybrid machine learning (ML) models. As a result of this study, AE technology significantly improved the performance of the ANN model. Especially, the performance of AE (25 features)–ANN model was superior to that of other non-hybrid and hybrid models. Compared to the non-hybrid ANN model, the performance of AE (25 features)–ANN model improved by 49%, 27%, 49%, and 22% in terms of the MAE, RMSE, R2, and R, respectively. The hybrid model using ANN and AE proposed in this study showed useful results to improve the performance of the DWGR ML model. Therefore, this method is considered a novel and advantageous approach for developing a DWGR ML model. Furthermore, it can be used to develop AI models for improving performance in various fields.

1. Introduction

The World Bank (2018) predicted that the annual volume of municipal solid waste (MSW) generated in cities will increase to 3.4 billion tons in 2050 [1]. In addition to the deterioration of the urban environment, the increase in waste poses various environmental and health risks, such as groundwater pollution, land degradation, increased cancer incidence, child mortality, and birth defects [2]. Therefore, governments worldwide and researchers in related industries have been striving to realize effective waste management (WM) strategies, and in recent years, advanced technologies and intelligent systems have been introduced in an attempt to manage waste.
Accounting for 35–40% of global waste, construction and demolition waste (C&DW) is increasing at an alarming rate and has emerged as a concern for the national economy and sustainable development goals [3,4,5]. In addition, the generation of C&DW is steadily increasing [6,7,8], wherein demolition waste (DW) accounts for 70–90% of C&DW [9,10]. As such, copious amounts of generated C&DW create considerable problems and increase the social and environmental burden through their deleterious effects on the environment. Therefore, pertinent management of C&DW is essential, and as a potential WM strategy, several researchers have utilized the information on waste generation (WG) as an advantageous tool for C&DW management. However, WM processes involve complex systems influenced by several factors, and obtaining adequate and satisfactory information for WM is a challenging task.
Consequently, the application of artificial intelligence (AI) technology has been actively considered in recent years to derive useful and satisfactory information for WM from multidimensional and noisy data. In this regard, several studies have employed AI technology for predicting the volume of C&DW or MSW generated. For instance, an artificial neural network (ANN) algorithm has been applied to predict WG [11,12,13,14,15,16,17,18,19]. In addition, a support vector machine (SVM) algorithm has been utilized to develop a WG prediction model [12,13,18,20,21,22,23,24,25,26,27,28], and several studies have developed a WG prediction model using linear regression (LR) [11,13,29,30,31,32,33,34,35,36]. Moreover, a WG prediction model has been developed using a decision tree (DT) algorithm [3,37,38,39,40]. In particular, scholars prefer using standalone algorithms, e.g., ANN, SVM, LR, and DT algorithms, as AI models for WG prediction. Overall, the ANN and SVM algorithms are the most representative and frequently applied algorithms.
After Abbasi and Hanandeh (2016) developed a WG prediction model based on the k-nearest neighbor algorithm, several scholars have attempted to develop an AI model for WG prediction by applying a random forest (RF) algorithm [27,41,42,43,44]. The performance of the AI models rendered by the aforementioned standalone algorithms vary depending on the characteristics of the data, e.g., data size and input variable type, data processing method, and employed hyperparameters [45]. Generally, ANN and SVM algorithms are unsuitable for datasets in which the input-variable type is categorical data instead of numerical data [39,46]. In contrast, algorithms such as DT can be applied regardless of categorical data or numerical data, thereby yielding an acceptable performance. Furthermore, various performance results can be obtained for each study based on the selection of hyperparameters. Thus, the predictive performance of AI models developed using a standalone algorithm depends on the algorithm type, dataset characteristics (e.g., size or input-variable type), and selected hyperparameters. Accordingly, the predictive performance of AI models can be determined based on these influencing factors (e.g., algorithm type, data characteristics, and selected hyperparameters).
Although certain methods can be used to develop superior prediction models with improved prediction performance, these approaches pose considerable limitations. Recent research have focused on the development of hybrid AI models [13,14,18,19,20,21,24,47] to overcome the limitations associated with the existing standalone algorithms and augment the predictive performance of AI models. To this end, Abbasi et al. (2013, 2014) [20,21], Cai et al. (2020) [24], Dai et al. (2020) [47], Golbaz et al. (2019) [13], and Song et al. (2017) [18] developed hybrid models with improved predictive performance to predict C&DW and MSW generation by applying the following algorithms: the SVM algorithm, wavelet denoising method (WT), partial least-squares (PLS), long- and short-term memory (LSTM), fuzzy information granulation-genetic algorithm (FIG-GA), fuzzy, and gray model (GM). Furthermore, Liang et al. (2021) [14] and Soni et al. (2019) [19] improved the ANN model performance using Archimedes’ optimization algorithm (AOA)–ANN and GA–ANN hybrid models, which enhanced the performance of the MSW prediction model as well. More importantly, the performance of existing hybrid models have improved in terms of their statistical metrics: root–mean–square (RMSE) values improved from 21% [21] to 48% [14], and coefficient of determination (R2) increased from −6% [13] to 21% [19]. Similarly, other performance indicators, such as MSE, MAE, and MAPE, have demonstrated notable performance improvements in AI models. Overall, the characteristics common across most research on existing hybrid models include the application of the ANN and SVM algorithms with numerical data for the input-variable type of the dataset. Notably, ANN and SVM algorithms yield a superior performance in case of handling numerical data, which supports the stated trend.
This study aims to develop a hybrid DW model for predicting the volume of DW generation. Although existing studies primarily use numerical input data, this study proposes a method for developing a novel, hybrid predictive model that can deliver a superior prediction performance in comparison to the DW prediction model applied with standalone algorithms for categorical input data. The following procedures were performed to fulfill the present research objective:
(1)
Two distinct data-preprocessing methods were applied to construct the dataset. The first data-preprocessing method constructed a dataset by eliminating outliers, standardizing, and label-encoding categorical variables. The second method developed a dataset by reconstructing the numerical data using AE technology on the dataset obtained from the first method. After the application of the AE, various feature groups (i.e., number of features: 3, 6, 9, 15, 25, and 30) were created according to the representation size;
(2)
The development of DW predictive models with standalone algorithms (i.e., ANN, SVR, and RF) and hybrid DW predictive models (i.e., AE–ANN, AE–SVR, and AE–RF) with respective algorithms and AE technology;
(3)
The leave-one-out cross-validation (LOOCV) technique was utilized for model validation. The performances of various developed models were evaluated based on the statistical metrics of R, RMSE, R2, and MAE;
(4)
The performance results of DW predictive models and hybrid DW predictive models via standalone algorithms were compared and discussed. Finally, the optimal hybrid predictive model yielding the greatest performance improvement was proposed for DW generation, and the corresponding application method was discussed.
The remainder of the paper is organized as follows. The data and data processing methods used in this study, including the applied algorithms and the underlying reasons for their application along with the verification and evaluation method of the model, are introduced in Section 2. Thereafter, the performance results of the DW predictive models developed with standalone algorithms and the hybrid AE-based DW predictive models are comparatively analyzed in Section 3. In Section 4, the performance and applicability of the proposed hybrid model were compared with those of the existing models developed for predicting C&DW and MSW generation. Lastly, the major findings and limitations of this study along with the future scope of research are summarized in Section 5.

2. Methods and Materials

The data used in this study, data processing method, characteristics of the employed algorithms, and the validation and evaluation methods of the developed predictive models for predicting the demolition waste generation rate (DWGR) are described herein. In particular, the size and characteristics of the acquired data are detailed in Section 2.1; the data-preprocessing methods, including the categorical variables, are elaborated in Section 2.2, after which the application of the unsupervised and supervised machine learning (ML) algorithms is detailed in Section 2.3. Subsequently, the setup of the hyperparameters prior to the application of the algorithm is discussed in Section 2.4, and lastly, the verification and evaluation methods of the ML models developed for DWGR prediction are presented in Section 2.5. A schematic of the current research flow is illustrated in Figure 1.

2.1. Data Source

Building characteristics (e.g., floor area, usage, structure, and region) are the core influencing factors of DW generation [48]. Additionally, Banias et al. (2011) [49] and Cha et al. (2017) [37] considered the building usage (e.g., residential or commercial) as an important influencing factor for DW generation. Furthermore, Wang et al. (2021) [50] and Wu et al. (2021) [51] showed that the region in which the building is located influences the DW generation. Cha et al. (2017) [37] showed that wall type and roof type are also important factors influencing DW generation. Thus, to acquire information on the amount of decommissioning waste generated, the building characteristics (e.g., location, structure, usage, wall type, roof type, and gross floor area (GFA)) were obtained from a direct survey of 782 buildings prior to the demolition of buildings. The data included the records of dismantling waste discharge (kg/m2) collected from demolition sites in redevelopment areas of two South Korean cities—Daegu and Busan. Thereafter, the dataset was constructed based on the information acquired from trucks disposing the demolition waste after building demolition, and the details of the demolition WG (kg) were obtained from the demolition company. A segment of the raw constructed dataset is presented in Table 1, wherein the size of the entire dataset was 782 rows × 7 columns. In Table 1, the building characteristics, e.g., the location, structure, usage, GFA, wall type, and roof type, represent the major factors affecting the DWGR. Therefore, we leveraged these six building features (i.e., location, structure, usage, GFA, wall type, and roof type) as variables of DWGR prediction. The relationship between the DWGR and these six building characteristics is expressed in Equation (1), and the DWGR is defined in Equation (2).
DWGR = f (location, structure, usage, GFA, wall type, roof type),
DWGR i =   A   of   building i   GFA   of   building i ,
where DWGRi denotes the demolition waste generation rate (kg·m2) of building i, A indicates the WG of building i (quantity; kg), and GFA symbolizes the gross floor area (m2) of building i.
The basic statistical analysis of DWGR in terms of the building status and characteristics derived from the collected data is presented in Table 2, wherein the mean values of the DWGR characteristics observably varied with the location, usage, structure, wall type, and roof type of the buildings. In particular, the GFA and DWGR of most buildings were ≤300 (m2) and ≤3000 (kg·m−2), respectively, which can be attributed to the acquirement of data from redevelopment areas that predominantly housed old low-rise buildings.

2.2. Data Preprocessing and Dataset Size

As the construction of a stable dataset is a prerequisite for improving the performance of ML predictive models, we performed data preprocessing on the acquired dataset. The categorical input variables were converted into numerical variables by preprocessing the data as follows: encoding, outlier elimination, and standardization. First, encoding was performed to convert the data into vectors of real numbers, which would act as the input variables for the AE. As AEs contain neural network structures that operate on real vectors [52], the categorical data must be converted into a real vector to enable the application of a neural network on qualitative data [53]. This study implemented label encoding to convert the categorical variables into real vectors; the label encoding regime for each categorical variable is listed in Table 3. The data-encoding process was followed by outlier elimination according to Equation (3), after which the size of the processed dataset was reduced to 690 rows × 7 columns. In particular, this encoded and processed dataset was employed for developing the ML model. Thereafter, data standardization was performed according to Equation (4) to construct a dataset of the same scale unit.
Q1 − 1.5 × IQR < selecting data < Q3 + 1.5 × IQR,
where IQR denotes the interquartile range, derived as Q3 − Q1; Q denotes the quartile, Q1 indicates the 25th percentile, and Q3 denotes the 75th percentile.
x standardization = x   x ¯ σ ,
where x,   x ¯ , and σ represent the element, mean, and standard deviation of the data, respectively.

2.3. Application of ML Algorithms

The inherent properties of the AE algorithm, the reasons for its implementation in this study, and its overall applications are described herein. In addition, we examined the characteristics of the supervised learning algorithms adopted for developing the DWGR prediction model and discussed the reasons for its adoption.

2.3.1. Autoencoder (Unsupervised Learning)

AE is an ANN-based unsupervised ML algorithm [54] that utilizes a neural network to reconstruct an output value equal to an arbitrary input value [55]. In particular, the basic AE contains a symmetrical structure comprising two functional segments—encoder and decoder—and three layers: input layer, hidden layer, and output layer. The encoder transforms the original input data (X) into a lower-dimensional layer, called the compressed representation (also known as feature or latent vector). Additionally, the decoder decompresses the representation into new input data (X′) reconstructed according to the relationship between the input variables [56]. Thus, the features of the input values regenerated by the AE exhibit numerical differences. This AE characteristic is beneficial for supplementing the characteristics of numerical values simply converted via label encoding into ordinal variables. This is because the categorical variables converted via label encoding are arbitrarily assigned integer values, and information cannot be used because of the numerical difference between the encoded values [53]. As the model performance can be expectedly improved by considering the numerical difference characteristics to these categorical variables, this study implemented an AE to convert the categorical input variables into input variables with numerical information.
In the AE, the size of the hidden layer determines the size of the representation (latent vector or feature) [57]. This algorithmic property of the AE can be utilized for feature engineering. Specifically, the number of features depends on the size of the hidden layer. Although AEs can improve the performance of regression and classification tasks through lower-dimensional representations [52], the dimensionality in this case should be reduced to accommodate the substantially high number of features. Conversely, in the case of fewer features, the number of features can be increased using an overcomplete AE on the original dataset, i.e., an AE architecture with a hidden layer dimensionally larger than the input layer [58]. Nonetheless, an overcomplete representation is advantageous because it increases the stability of the representation [59]. However, as reported by Fan et al. (2019) [60] and Meyer (2015) [61], if the hidden layer size is larger than the input layer size in a basic AE, a problem may occur in the generalization of the mode. As observed in stacked AEs, the generalization issue can be resolved by incorporating a hidden layer [62]. In a stacked AE, the pre-training is performed by considering a single layer at a given instant. The layers of the stacked AE are trained to individually minimize the errors arising from the input reconstruction of the layers [63]. In particular, the pre-training aids generalization by ensuring that the learned information is extracted from the input [62,63].
Therefore, the stacked AE architecture was applied in this study to implement a hidden layer that was larger than the input layer, as illustrated in Figure 2. Moreover, we constructed a dataset in which the number of features was adjusted according to the undercomplete and overcomplete AE architectures by altering the representation size of the hidden layer. As indicated in Figure 2, the architectures of an undercomplete AE (feature representation size = 3) and six overcomplete AEs (feature representation sizes = 6, 9, 15, 20, 25, and 30) were applied to construct the feature sets with data sizes of 3, 6, 9, 15, 20, 25, and 30 columns.
In general, the AE minimizes the distance between the input and output by maximally recovering the information from the original input [64]. To this end, the AE model uses loss functions to recreate the features and updates the weight parameters to obtain more efficient results, as well as reduce the likelihood of errors. In addition to minimizing the loss, the AE converges the value of X′ to that of X. Thus, the appropriate transformation of the input variable via the AE can be determined from the loss result. In particular, the loss function based on the reconstruction error is expressed in Equation (5). According to the loss result, the data conversion conducted in this study was assessed to be appropriate.
L ( x ,   x ) = 1 n   i = 1 n ( x x ) 2 ,
where x and x′ indicate denotes the output and input of the decoder, respectively.

2.3.2. Supervised Learning Techniques Used for DWGR Estimation

This study applied supervised learning algorithms to various sets of features constructed using the AE, i.e., subsets comprising 3, 6, 9, 15, 20, 25, and 30 features. In particular, the SVR, ANN, and RF were selected as the learning algorithms. The SVR and ANN algorithms have been widely applied in the field of waste prediction [45,65,66]. Moreover, RF is an ML algorithm with excellent predictive performance and is considered one of the ten best classifiers [67]. In particular, RF can utilize both categorical and continuous variables and is useful for comparing performance variations with the application of AE on SVR and ANN. The details of the supervised learning techniques employed in this study are described below.

Random Forest

Initially proposed by Breiman (2001) [68], RF is a representative ensemble technique based on bagging, which is used to perform bootstrap sampling. Overall, RF is considered one of the most powerful ML algorithms. In principle, RF creates a tree (called weak learner) for each subset by extracting multiple subsets (bootstrap sampling) from the original dataset. The final prediction determines the strong learner based on the majority of votes obtained from the results of each tree. Through this process, RF can prevent overfitting as the number of trees increases and is minimally influenced by outliers. Even for an unbalanced class, this classification algorithm offers superior predictive performance compared to other machine learning algorithms [67]. Recently, certain researchers used RF to predict the amount of waste generated in the fields of C&D and solid WM; Cha et al. (2020, 2021) [41,42] applied RF to estimate the DWGR for diverse types of waste generated during the building–dismantling process, whereas Kumar et al. (2018) [27] leveraged RF to predict the plastic WG rate. Additionally, Dissanayaka and Vasanthapriyan (2019) [43] employed RF to predict MSW generation, and Namoun et al. (2022) [69] used RF to predict household solid WG. Furthermore, this algorithm was implemented by Rosecký et al. (2021) [39] to predict MSW generation at the regional level.

Support Vector Regression

The working principle of SVR is stated as follows: a linear decision function is constructed in the feature space by mapping the input data to the feature space using a nonlinear map. The principal aim of SVR is to determine the optimal decision function. Thereafter, using kernels, SVR nonlinearly maps a linear decision function of the feature space to the original space [28]. The SVR model intends to overcome the fundamental drawback of parametric regression. Overall, it is a novel and powerful ML technique based on statistical learning theory and adheres to the principle of structural risk minimization, i.e., it aims to minimize the upper bound of the generalization error instead of minimizing the training error [25,28,70]. As reported, the SVR model performs reliably in solving problems with small samples, nonlinearities, and high-dimensional characteristics [71]. Coupled with ANN, the SVR model has been effectively utilized for AI models in various fields, and several researchers have applied it in WG-related fields as well. In this regard, Abbasi et al. (2014, 2013) [20,21], Abbasi and Hanandeh (2016) [22], Abunama et al. (2019) [23], Dai et al. (2020) [47], Golbaz et al. (2019) [13], Graus et al. (2018) [26], Kumar et al. (2018) [27], and Song et al. (2017) [18] used SVR to predict MSW generation, whereas Cai et al. (2020) [24] and Cha et al. (2022) [12] employed SVR to predict C&DW and DW generation.

Artificial Neural Networks

The ANN theory was initially proposed by McCulloch and Pitts (1943). According to the signal transmission modes, ANNs are classified as feedforward and feedback neural networks. Owing to their simpler and superior performance over feedback neural networks, feedforward neural networks have been widely used in the field of WM [68]. As reported in the literature, the multilayer perceptron (MLP) is a prominent feedforward ANN architecture that is employed for forecasting problems [72,73]. The basic structure of an MLP contains three layers: input, hidden, and output layers, and the nonlinear transfer function comprising multiple layers of neurons enables the learning of nonlinear and linear relationships between input and output neurons. These neurons are connected to the adjacent layer and hidden layer, and the interconnections of all neurons contain weights indicating the strength. The number of input neurons—identical to the number of input variables—is responsible for receiving external information. In the hidden layer, the sum of weights is passed through an activation function that determines the relationship between the input and output. Subsequently, the output of the hidden layer is inputted to the output layer, and the weights are calculated and transformed via the linear activation function of the output layer [29,74]. The ANN algorithm is the most frequently used algorithm for AI models in the field of WM [45]. In relation to WG, ANNs have been utilized by Golbaz et al. (2019) [13] to predict solid WG from hospitals, by Liang et al. (2021) [14], Shamshiry et al. (2014) [17], and Soni et al. (2019) [19] to predict MSW generation, and by Cha et al. (2022) [12] and Song et al. (2017) [18] to predict C&DW generation.

2.4. Application of Algorithms and Hyperparameter-Tuning

Hyperparameters significantly influence the predictive performance, robustness, and generalization ability of a model. Therefore, prior to the application of AE, we tuned the hyperparameters to derive the optimal performance model for each algorithm (i.e., RF, ANN-MLP, and SVR). For instance, we adjusted the ANN models according to the number of hidden layers and neurons, and the ANN (MLP) models were tested with 1–4 hidden layers and 20, 30, and 40 neurons for each hidden layer. Based on the test results, we selected the ANN-MLP model with the structure of input layer—first hidden layer (10 neurons); second layer (30 neurons)—output layer and two hidden layers, featuring 10 neurons in the first hidden layer and 30 neurons in the second hidden layer.
For the SVR model, hyperparameters such as the kernel, kernel coefficient (γ), and penalty parameter of the error term (C) were adjusted in this study. In addition, a radial basis function kernel, widely used in regression problems, was considered for the kernel. This strategy was selected considering that it can nonlinearly map samples to a high-dimensional space and can easily handle nonlinear relationships between class labels and properties [75]. As for the kernel coefficient, γ = 1/6 was applied according to γ = 1/K (where K denotes the number of features), as proposed by Chang and Lin (2011) [76]. For the penalty parameter of the error term (C), the optimal value was searched for values ranging from 10−5 to 101, and the value of C = 10−3 yielded the optimal performance results.
For the RF model, the optimal predictive performance was derived by adjusting the number of trees and features. The number of trees was varied in increments of 50 units ranging from 100 to 500 and tested with submodels containing 3, 4, 5, and 6 variables. The test results revealed that the RF model delivered the optimal results with 450 trees and 6 variables.

2.5. Model Evaluation

2.5.1. Model Validation

In this study, the proposed model was validated using LOOCV, which is a special case of k-fold cross-validation technique. LOOCV is regarded a suitable validation method for small sample sizes [77,78]. Accordingly, several studies have adopted LOOCV to evaluate the performance of algorithms handling a small number of instances in the dataset [8,79]. In principle, LOOCV utilizes all samples as the test-and-training data to secure adequate training and validation sets. Compared with the 10-fold or k-fold cross-validation, LOOCV is advantageous for obtaining stable results with small target datasets [41,42,80,81]. Therefore, considering the size of the dataset in this study, LOOCV was applied as a model validation method.

2.5.2. Performance Measures

The performance of the DWGR predictive models developed in this study were evaluated based on statistical metrics such as the MAE (Equation (6)), RMSE (Equation (7)), R2 (Equation (8)), and R (Equation (9)). Generally, a satisfied model yields high R2 and R values and low MAE and RMSE values.
MAE = i = 1 n | y i x i | n
RMSE = i = 1 n ( y i x i ) 2 n
R 2 = 1 i = 1 n ( y i x i ) 2 i = 1 n ( y i   x ¯ i ) 2
R = i = 1 n ( x i x ¯ i ) ( y i y ¯ i ) i = 1 n ( x i x ¯ i ) 2 i = 1 n ( y i y ¯ i ) 2
where x i denotes the observed quantity of generated DW, y i represents the predicted quantity of the generated DW,   x ¯ i denotes the mean observed quantity of generated DW, y ¯ i indicates the average predicted quantity of generated DW, and n denotes the number of samples.

3. Results

3.1. Learning Validity Assessment of the Stacked AE Utilized in this Study

The loss function values derived during the learning process are plotted in Figure 3. In the current case of the representation applied with six features, the validation loss approximated the training loss at nearly epoch 20, and thereafter, converged to 0. The validation loss results of the six-feature representation revealed that overfitting did not occur in the applied AE model. For the remaining feature representations (i.e., 3, 9, 12, 15, 20, 25 features; refer to Supplementary Material, Figure S1), the validation loss was larger than the train loss, ranging from approximately epochs 15–20. However, in all feature representations beyond this range, it yielded stable results converging to 0. As observed from these loss results (refer to Figure 3 and Figure S1), overfitting did not occur in all feature representations and the validation loss exhibited a pronounced declivity. In addition, the stacked AE architecture evidently facilitated the stable conversion of the categorical input variable data by considering the numerical information relationship between the input variables, as it maintained the existing data characteristics. The distribution of data values converted from categorical variables into numerical variables through the stacked AE is depicted in Figure S2 (refer to Supplementary Material).

3.2. Comparison of Performance Results and Improvement of Models

The performance results of the predictive models based on standalone algorithms (i.e., ANN-MLP, SVR, and RF) and hybrid models combined with AE for the DWGR prediction are comparatively presented in Figure 4. First, as demonstrated by the results of ANN-MLP (Figure 4 and Figure 5a), the performance of the model with AE implementation yielded significantly improved results for all performance indicators (i.e., MAE, RMSE, R2, and R). In terms of model stability, the MAE and RMSE of all AE–ANN (MLP) models were significantly stabilized below 212.133 and 274.371 (i.e., MAE and RMSE values of worst-performing AE (three features)–ANN (MLP) model), respectively, compared to the ANN (MLP) model (MAE: 356.697; RMSE: 316.186). In particular, the AE (25 features)–ANN (MLP) model yielded the lowest values of the MAE and RMSE at 182.105 and 230.819, respectively. In addition, the R2 and R values were significantly improved in comparison to those of the ANN (R2: 0.458; R: 0.676), and the R2 and R of the AE (25 features)–ANN (MLP) model were 0.680 and 0.825, respectively, which were superior to those of the remaining AEs–ANN (MLP) models.
In the SVR results depicted in Figure 4 and Figure 5b, the MAE and RMSE of all AE–SVR models were significantly more stable in comparison to those of the SVR models (MAE: 1020.304; RMSE: 1116.266). Specifically, the MAE and RMSE of the AE (three features)–SVR model were 302.244 and 384.347, respectively, thereby demonstrating optimal results with no significant variation from the MAE and RMSE results of remaining AE–SVR models. Moreover, as noted from the R values, the performance of AE–SVR models was significantly improved compared to that of the SVR model, wherein the AE (15)–SVR model (R: 0.568) yielded the optimal R value. However, in terms of R2 values, all AE–SVR models delivered improved their performance over the existing SVR model (R2: 0.007). In contrast, the R of the AEs–SVR models ranged from 0.071–0.110 and suggested the inferior accuracy of the AEs–SVR models, which signifies the inapplicability of these models for prediction purposes.
The performance results of the RF and AE–RF models are plotted in Figure 4 and Figure 5c, which indicate distinct patterns compared to those of the AE–ANN (MLP) and AE–SVR models. For the prediction model applying solely the RF algorithm, the MAE and RMSE results were 197.004 and 249.597, respectively, whereas those of the AE–RF models were 191.867–206.940 and 253.159–275.385, respectively. In terms of stability, the MAE of certain AE–RF models was superior to that of RF models. However, their RMSE results were inferior to those of RF models in all cases. In terms of R2 and R values, the performance of all the AE–RF models was marginally worse than the RF model. These results indicated negligible improvement in the performance of AE–RF models. Essentially, as the RF algorithm can handle all types of categorical and numerical variables, the application of variable conversion technology such as AEs can be evidently ineffective. However, as depicted in Figure 4 and Figure 5, a significant performance improvement was observed for the ANN and SVR algorithms. The ANN model yielded a significant improvement in all the performance indicators (MAE, RMSE, R2, and R), and in particular, the predictive performance of the AE (25)–ANN (MLP) model was superior to that of the RF model. Furthermore, the SVR model displayed a significant improvement in the performance indicators (MAE, RMSE, and R), and in terms of R2, the resulting performance improvement was inadequate.
As indicated by the aforementioned performance index results, the AE (25)–ANN (MLP) model delivered the most outstanding performance for predicting DW generation. The degree of performance improvement obtained by applying AE can be clearly confirmed by comparing the correlation results between the observed and predictive values of the ANN (MLP) model and AE (25 features)–ANN (MLP) model (Figure 6a,b), and the results of the predictive model (Figure 7). As indicated in Figure 6a,b, the observed and predictive values of the AE (25 features)–ANN (MLP) model, compared to the ANN (MLP) model, were more closely distributed along the line with a correlation coefficient of 1. Moreover, as depicted in Figure 7, the predicted values of the AE (25 features)–ANN (MLP) model more closely approximated the trend of the observed values than the ANN (MLP) model. The mean of the observed values was 1165.04 kg·m−2, and the mean of the predictive values of the ANN (MLP) and AE (25 features)–ANN (MLP) models were 1157.656 kg·m−2 and 1162.437 kg·m−2, respectively. These results confirmed the significant improvement of the predictive performance of the AE (25 features)–ANN (MLP) model. Moreover, the utilization of AE can be considered an advantageous new method for the development of predictive models with excellent performance on datasets containing categorical variables.

4. Discussion and Recommendations

To more accurately estimate the generation of C&DW during building demolition, this study aimed to develop a hybrid predictive model that can improve the predictive performance of a dataset comprising categorical data. To date, numerous studies on hybrid model development have attempted to improve the performance of the WG prediction model in the fields of C&DW and MSW [12,13,14,18,19,20,21,24,47]. As listed in Table 4, the existing hybrid ML models aimed to predict C&DW and MSW generation at various estimation levels such as building, district, city, and region. Commonly, these studies applied the ANN and SVR techniques as ML algorithms, primarily using numerical data as the input variables.
For instance, Abbasi et al. (2013) [20] developed a PLS–SVM hybrid model to improve the performance of the SVM model based on daily solid WG-related time-series data from 2004 to 2005 of Tehran city. Moreover, Abbasi et al. (2014) [21] attempted to enhance the performance of the SVM model using the WT method on the MSW time-series data corresponding to Tehran city. Furthermore, Song et al. (2017) [18] developed a GM–SVR hybrid model to predict C&DW generation in 31 provinces of China, and reported a significant improvement in its performance.
Additionally, Golbaz et al. (2019) [13] leveraged multiple ML models, such as multilinear regression, ANN, SVM, least squares–SVM, and fuzzy logic (F)–SVM, to predict solid WG from hospitals, among which the F–SVM hybrid model yielded the best performance. Furthermore, Soni et al. (2019) [19] developed a GA–ANN model to predict MSW generation. The GA–ANN model yielded performance improvements of 42% and 21% compared to the ANN model, respectively. More recently, Cai et al. (2020) [24] developed hybrid models using the statistical analysis models of autoregressive integrated moving average, SVR, back-propagation neural network, and LSTM algorithms, among which the LSTM–SVR model delivered the best performance.
Dai et al. (2020) [47] developed a FIG–GA–SVR model to predict district-level MSW generation (Huangshi city, Hubei Province, China). The FIG–GA–SVR model produced performance improvements of 84% and 86% in terms of MSE and MAPE compared to the SVR model, respectively. Additionally, Liang et al. conducted a study on hybrid models, combining algorithms such as ANN, GA, particle swarm optimization, a sine–cosine algorithm, and AOA to predict MSW generation in major cities of Iran. Among the hybrid models, the AOA–ANN model yielded the best predictive performance.
The degree of performance improvement of the hybrid AI models developed for predicting C&D and solid WG was 21%–48% in terms of RMSE (derived from the results of [14,19,20,21,24]), 40% in relation to MAE (deduced from the results of [14]), and 0–21% in R2 values (based on the results of [13,14,19,20,21,24]). On the other hand, the AE (25 features)–ANN(MLP) model developed in this study yielded performance improvements of 49%, 27%, 49%, and 22% in terms of MAE, RMSE, R2, and R, respectively (listed in Table 4), compared to the non-hybrid ANN model. Moreover, compared to the non-hybrid SVR model, the AE (three features)–SVR model yielded an improvement of 70% and 66% in terms of MAE and RMSE, respectively (refer to Table 4). Therefore, the application of the AE technique to the categorical variables in this study significantly influenced the stabilization and accuracy of the model, which resulted in superior performance improvements compared to the existing hybrid models. As summarized in Table 4, unlike previous studies that developed hybrid models for numerical data, this study explored a new direction with the proposed hybrid model to significantly improve the prediction performance for categorical data.
As depicted in Figure 4 and Figure 5, the prediction performance of the RF algorithm for categorical data is superior to those of the ANN and SVR algorithms. In general, the RF algorithms are considered adequately robust to ensure an optimal prediction performance [66]. Accordingly, these algorithms have been widely used in several fields owing to their excellent performance, as well as fast and efficient training process [82]. Therefore, developing an ML regression model with a predictive performance superior to that of the RF model is a challenging task. However, the predictive model proposed in this study combined the ANN and SVR algorithms with AE technology and used categorical variables to deliver a superior performance compared to the RF model. Therefore, this study presents a new strategy to overcome the limitations emerging from the characteristics of the variable type. As demonstrated, the development of hybrid model via AEs can be advantageous toward achieving excellent predictive models with superior performance.
This strategy is beneficial in the case of applying ML algorithms that require numerical data, such as ANN and SVR, or in case (1) The dataset contains more categorical data than numerical data; or (2) The implemented ML algorithm is applicable regardless of the input variable type, such as RF (e.g., decision tree, extra classifier tree, Xgboosting, and gradient-boosting machine). In particular, the proposed strategy will be of great interest to demolition companies because it can facilitate the development of a DW waste-prediction model in a limited-data environment (e.g., lack of numerical input information for buildings subjected to demolition and building characteristics data comprising simple nominal variables). Moreover, the proposed method can be combined with various ML algorithms to improve the predictive performance of models for C&DW and MSW management. Furthermore, the application of the current methodology in a similar data environment will be advantageous for developing models with excellent predictive performance in several other fields as well.

5. Conclusions

This study developed a novel hybrid AI model using standalone algorithms with AE technology to predict DWG from the demolition of buildings in redevelopment areas in South Korea. This study reports novel research findings relevant to the field of C&DW and MSW management. The performance of the ANN-MLP and SVR models was improved using categorical data, and a hybrid DW predictive model was developed by applying AE. In particular, compared to the ANN (MLP) model, the AE (25)–ANN (MLP) model improved performance in terms of MAE, RMSE, R2, and R, thereby significantly influencing the model’s stability and accuracy. The mean of the observed values was 1165.04 kg·m−2, and those of the predictive values obtained by the ANN (MLP) and AE (25)–ANN (MLP) models were 1157.656 kg·m−2 and 1162.437 kg·m−2, respectively. In addition, the application of AE technology to the SVR algorithm considerably enhanced the prediction stability of the AI model based on the significantly improved results in terms of MAE and RMSE.
Notably, the potential of performance improvement with AE will be insignificant for algorithms (e.g., RF) that can handle both categorical and numerical data as an input variable type. However, as the AE (25)–ANN (MLP) predictive model demonstrated a superior performance compared to the RF predictive model, this method is considered a novel and advantageous approach for developing a DW predictive model because it produced an excellent predictive performance for datasets comprising categorical data. More importantly, the present findings are crucial for surpassing the limitations of the feasible ML algorithms that rely on data characteristics and developing various AI models with superior predictive performance. Therefore, the results of this study can help develop ML models with better performance in various fields. In addition, the results of this study are expected to be useful for decision-making by related industry officials or policymakers by providing more accurate DW generation prediction information.
However, as this study was limited to the use of ANN, SVR, and RF algorithms, more diverse algorithms should be considered in the future, and further research should be conducted on various datasets. For example, in the future, it is necessary to conduct research on the development of ML models using AE for datasets other than the categorical input variables used in this study. In addition, efforts are required to find an optimal DWGR ML model by applying additional algorithms.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/su15043691/s1, Figure S1: Loss results according to feature representation size of AE: 3, 9, 15, 20, 25, and 30 features arranged from (a–g), respectively; Figure S2: Distribution of data values by feature representation size: 3, 6, 9, 15, 20, 25, and 30 features arranged from (a–g), respectively.

Author Contributions

G.-W.C. and Y.-C.K.: Conceptualization, methodology, validation, and supervision. G.-W.C. and Y.-C.K.: Writing—original draft preparation. G.-W.C.: Formal analysis. G.-W.C. and Y.-C.K.: Resources. G.-W.C., Y.-C.K. and W.-H.H.: Writing—review and editing and funding acquisition. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Research Foundation of Korea (NRF), grant funded by the South Korean government (MSIT) (NRF-2019R1A2C1088446). This work was supported by the National Research Foundation of Korea (NRF), grant funded by the South Korean government (MSIT) (No. NRF-2020R1C1C1009061).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All data included in this study are available upon request by contact with the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

AIArtificial Intelligence
AEAutoencoder
ANNArtificial Neural Network
C&DWWaste, Construction, and Demolition Waste
CVCross Validation
DWDemolition Waste
DWGRDemolition Waste Generation Rate
GFAGross Floor Area
LOOCVLeave-One-Out Cross Validation
MLMachine Learning
MSWMunicipal Solid Waste
MAEMean Absolute Error
MLPMultilayer Perceptron
MSEMean Squared Error
RPearson’s Correlation Coefficient
R2Coefficient of Determination
RMSERoot–Mean–Square Error
RFRandom Forest
SVRSupport Vector Regression
WGWaste Generation

References

  1. Kaza, S.; Yao, L.; Bhada-Tata, P.; Van Woerden, F. What a Waste 2.0: A Global Snapshot of Solid Waste Management to 2050. World Bank Publications: Washington, DC, USA, 2018. [Google Scholar]
  2. Triassi, M.; Alfano, R.; Illario, M.; Nardone, A.; Caporale, O.; Montuori, P. Environmental pollution from illegal waste disposal and health effects: A review on the ‘‘Triangle of Death”. Int. J. Environ. Res. Public Health 2015, 12, 1216–1236. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Huang, X.; Xu, X. Legal regulation perspective of eco-efficiency construction waste reduction and utilization. Urban Dev. Stud. 2011, 9, 90–94. [Google Scholar]
  4. Rani, M.; Gupta, A. Construction waste management in India. Int. J. Sci. Technol. Manag. 2016, 2016. [Google Scholar]
  5. Wu, H.; Zuo, J.; Zillante, G.; Wang, J.; Yuan, H. Status quo and future directions of construction and demolition waste research: A critical review. J. Clean. Prod. 2019, 240, 118163. [Google Scholar] [CrossRef]
  6. Li, J.; Ding, Z.; Mi, X.; Wang, J.A. model for estimating construction waste generation index for building project in China. Resour. Conserv. Recycl. 2013, 74, 20–26. [Google Scholar] [CrossRef]
  7. Llatas, C.A. Model for quantifying construction waste in projects according to the European waste list. Waste Manag. 2011, 31, 1261–1276. [Google Scholar] [CrossRef]
  8. Wang, J.; Li, Z.; Tam, V.W. Identifying best design strategies for construction waste minimization. J. Clean. Prod. 2015, 92, 237–247. [Google Scholar] [CrossRef]
  9. Butera, S.; Christensen, T.H.; Astrup, T.F. Composition and leaching of construction and demolition waste: Inorganic elements and organic compounds. J. Hazard. Mater. 2014, 276, 302–311. [Google Scholar] [CrossRef]
  10. Lu, W.; Yuan, H.; Li, J.; Hao, J.J.; Mi, X.; Ding, Z. An empirical investigation of construction and demolition waste generation rates in Shenzhen city, South China. Waste Manag. 2011, 31, 680–687. [Google Scholar] [CrossRef] [Green Version]
  11. Ali Abdoli, M.; Falah Nezhad, M.; Salehi Sede, R.; Behboudian, S. Long-term forecasting of solid waste generation by the artificial neural networks. Environ. Prog. Sustain. Energy 2012, 31, 628–636. [Google Scholar] [CrossRef]
  12. Cha, G.W.; Moon, H.J.; Kim, Y.C. A hybrid machine-learning model for predicting the waste generation rate of building demolition projects. J. Clean. Prod. 2022, 375, 134096. [Google Scholar] [CrossRef]
  13. Golbaz, S.; Nabizadeh, R.; Sajadi, H.S. Comparative study of predicting hospital solid waste generation using multiple linear regression and artificial intelligence. J. Environ. Health Sci. Eng. 2019, 17, 41–51. [Google Scholar] [CrossRef] [PubMed]
  14. Liang, G.; Panahi, F.; Ahmed, A.N.; Ehteram, M.; Band, S.S.; Elshafie, A. Predicting municipal solid waste using a coupled artificial neural network with Archimedes optimisation algorithm and socioeconomic components. J. Clean. Prod. 2021, 315, 128039. [Google Scholar] [CrossRef]
  15. Milojkovic, J.; Litovski, V. Comparison of some ANN based forecasting methods implemented on short time series. In Proceedings of the 2008 9th Symposium on Neural Network Applications in Electrical Engineering, Belgrade, Serbia, 25–27 September 2008; pp. 175–178. [Google Scholar]
  16. Noori, R.; Abdoli, M.A.; Ghazizade, M.J.; Samieifard, R. Comparison of neural network and principal component-regression analysis to predict the solid waste generation in Tehran, Iran. J. Public Health 2009, 38, 74–84. [Google Scholar]
  17. Shamshiry, E.; Mokhtar, M.; Abdulai, A.M.; Komoo, I.; Yahaya, N. Combining artificial neural network- genetic algorithm and response surface method to predict waste generation and optimize cost of solid waste collection and transportation process in Langkawi island, Malaysia. Malays. J. Sci. 2014, 33, 118–140. [Google Scholar] [CrossRef] [Green Version]
  18. Song, Y.; Wang, Y.; Liu, F.; Zhang, Y. Development of a hybrid model to predict construction and demolition waste: China as a case study. Waste Manag. 2017, 59, 350–361. [Google Scholar] [CrossRef]
  19. Soni, U.; Roy, A.; Verma, A.; Jain, V. Forecasting municipal solid waste generation using artificial intelligence models—A case study in India. SN Appl. Sci. 2019, 1, 1–10. [Google Scholar] [CrossRef] [Green Version]
  20. Abbasi, M.; Abduli, M.A.; Omidvar, B.; Baghvand, A. Forecasting municipal solid waste generation by hybrid support vector machine and partial least square model. Int. J. Environ. Res. 2013, 7, 27–38. [Google Scholar]
  21. Abbasi, M.; Abduli, M.A.; Omidvar, B.; Baghvand, A. Results uncertainty of support vector machine and hybrid of wavelet transform-support vector machine models for solid waste generation forecasting. Environ. Prog. Sustain. Energy 2014, 33, 220–228. [Google Scholar] [CrossRef]
  22. Abbasi, M.; El Hanandeh, A. Forecasting municipal solid waste generation using artificial intelligence modelling approaches. Waste Manag. 2016, 56, 13–22. [Google Scholar] [CrossRef]
  23. Abunama, T.; Othman, F.; Ansari, M.; El-Shafie, A. Leachate generation rate modeling using artificial intelligence algorithms aided by input optimization method for an MSW landfill. Environ. Sci. Pollut. Res. Int. 2019, 26, 3368–3381. [Google Scholar] [CrossRef] [PubMed]
  24. Cai, T.; Wang, G.; Guo, Z. Construction and demolition waste generation forecasting using a hybrid intelligent method. In Proceedings of the 2020 9th International Conference on Industrial Technology and Management (ICITM), Oxford, UK, 1–13 February 2020; pp. 312–316. [Google Scholar]
  25. Dai, C.; Li, Y.P.; Huang, G.H. A two-stage support-vector-regression optimization model for municipal solid waste management—A case study of Beijing, China. J. Environ. Manag. 2011, 92, 3023–3037. [Google Scholar] [CrossRef] [PubMed]
  26. Graus, M.; Niemietz, P.; Rahman, M.T.; Hiller, M.; Pahlenkemper, M. Machine learning approach to integrate waste management companies in micro grids. In Proceedings of the 2018 19th International Scientific Conference on Electric Power Engineering (EPE), Brno, Czech Republic, 16–18 May 2018; pp. 1–6. [Google Scholar]
  27. Kumar, A.; Samadder, S.R.; Kumar, N.; Singh, C. Estimation of the generation rate of different types of plastic wastes and possible revenue recovery from informal recycling. Waste Manag. 2018, 79, 781–790. [Google Scholar] [CrossRef] [PubMed]
  28. Noori, R.; Abdoli, M.A.; Ghasrodashti, A.A.; Jalili Ghazizade, M. Prediction of municipal solid waste generation with combination of support vector machine and principal component analysis: A case study of Mashhad. Environ. Prog. Sustain. Energy. 2009, 28, 249–258. [Google Scholar] [CrossRef]
  29. Azadi, S.; Karimi-Jashni, A. Verifying the performance of artificial neural network and multiple linear regression in predicting the mean seasonal municipal solid waste generation rate: A case study of Fars province, Iran. Waste Manag. 2016, 48, 14–23. [Google Scholar] [CrossRef]
  30. Chhay, L.; Reyad, M.A.H.; Suy, R.; Islam, M.R.; Mian, M.M. Municipal solid waste generation in China: Influencing factor analysis and multi-model forecasting. J. Mater. Cycles Waste Manag. 2018, 20, 1761–1770. [Google Scholar] [CrossRef]
  31. Fu, H.Z.; Li, Z.S.; Wang, R.H. Estimating municipal solid waste generation by different activities and various resident groups in five provinces of China. Waste Manag. 2015, 41, 3–11. [Google Scholar] [CrossRef]
  32. Jahandideh, S.; Jahandideh, S.; Asadabadi, E.B.; Askarian, M.; Movahedi, M.M.; Hosseini, S.; Jahandideh, M. The use of artificial neural networks and multiple linear regression to predict rate of medical waste generation. Waste Manag. 2009, 29, 2874–2879. [Google Scholar] [CrossRef]
  33. Kumar, A.; Samadder, S.R. An empirical model for prediction of household solid waste generation rate—A case study of Dhanbad, India. Waste Manag. 2017, 68, 3–15. [Google Scholar] [CrossRef]
  34. Montecinos, J.; Ouhimmou, M.; Chauhan, S.; Paquet, M. Forecasting multiple waste collecting sites for the agro-food industry. J. Clean. Prod. 2018, 187, 932–939. [Google Scholar] [CrossRef] [Green Version]
  35. Wei, Y.; Xue, Y.; Yin, J.; Ni, W. Prediction of municipal solid waste generation in China by multiple linear regression method. Int. J. Comp. Appl. 2013, 35, 136–140. [Google Scholar]
  36. Wu, Z.; Fan, H.; Liu, G. Forecasting construction and demolition waste using gene expression programming. J. Comp. Civ. Eng. 2015, 29, 04014059. [Google Scholar] [CrossRef]
  37. Cha, G.W.; Kim, Y.C.; Moon, H.J.; Hong, W.H. New approach for forecasting demolition waste generation using chi-squared automatic interaction detection (CHAID) method. J. Clean. Prod. 2017, 168, 375–385. [Google Scholar] [CrossRef]
  38. Kannangara, M.; Dua, R.; Ahmadi, L.; Bensebaa, F. Modeling and prediction of regional municipal solid waste generation and diversion in Canada using machine learning approaches. Waste Manag. 2018, 74, 3–15. [Google Scholar] [CrossRef] [PubMed]
  39. Rosecký, M.; Šomplák, R.; Slavík, J.; Kalina, J.; Bulková, G.; Bednář, J. Predictive modelling as a tool for effective municipal waste management policy at different territorial levels. J. Environ. Manag. 2021, 291, 112584. [Google Scholar] [CrossRef] [PubMed]
  40. Márquez, M.Y.; Ojeda, S.; Hidalgo, H. Identification of behavior patterns in household solid waste generation in Mexicali’s city: Study case. Resour. Conserv. Recy. 2008, 52, 1299–1306. [Google Scholar] [CrossRef]
  41. Cha, G.W.; Moon, H.J.; Kim, Y.M.; Hong, W.H.; Hwang, J.H.; Park, W.J.; Kim, Y.C. Development of a prediction model for demolition waste generation using a random forest algorithm based on small datasets. Int. J. Environ. Res. Public Health 2020, 17, 6997. [Google Scholar] [CrossRef]
  42. Cha, G.W.; Moon, H.J.; Kim, Y.C. Comparison of random forest and gradient boosting machine models for predicting demolition waste based on small datasets and categorical variables. Int. J. Environ. Res. Public Health 2021, 18, 8530. [Google Scholar] [CrossRef]
  43. Dissanayaka, D.M.S.H.; Vasanthapriyan, S. Forecast municipal solid waste generation in Sri Lanka. In Proceedings of the 2019 International Conference on Advancements in Computing (ICAC), Malabe, Sri Lanka, 5–7 December 2019; pp. 210–215. [Google Scholar]
  44. Nguyen, X.C.; Nguyen, T.T.H.; La, D.D.; Kumar, G.; Rene, E.R.; Nguyen, D.D.; Chang, S.W.; Chung, W.J.; Nguyen, X.H.; Nguyen, V.K. Development of machine learning-based models to forecast solid waste generation in residential areas: A case study from Vietnam. Resour. Conserv. Recy. 2021, 167, 105381. [Google Scholar] [CrossRef]
  45. Abdallah, M.; Talib, M.A.; Feroz, S.; Nasir, Q.; Abdalla, H.; Mahfood, B. Artificial intelligence applications in solid waste management: A systematic research review. Waste Manag. 2020, 109, 231–246. [Google Scholar] [CrossRef]
  46. Osisanwo, F.Y.; Akinsola, J.E.T.; Awodele, O.; Hinmikaiye, J.O.; Olakanmi, O.; Akinjobi, J. Supervised machine learning algorithms: Classification and comparison. Int. J. Comp. Trends Technol. 2017, 48, 128–138. [Google Scholar]
  47. Dai, F.; Nie, G.H.; Chen, Y. The municipal solid waste generation distribution prediction system based on FIG–GA-SVR model. J. Mater. Cycles Waste Manag. 2020, 22, 1352–1369. [Google Scholar] [CrossRef]
  48. Chen, X.; Lu, W. Identifying factors influencing demolition waste generation in Hong Kong. J. Clean. Prod. 2017, 141, 799–811. [Google Scholar] [CrossRef] [Green Version]
  49. Banias, G.; Achillas, C.; Vlachokostas, C.; Moussiopoulos, N.; Papaioannou, I. A web-based Decision Support System for the optimal management of construction and demolition waste. Waste Manag. 2011, 31, 2497–2502. [Google Scholar] [CrossRef]
  50. Wang, Z.S.; Xie, W.C.; Liu, J.K. Regional differences and driving factors of construction and demolition waste generation in China. Eng. Constr. Arch. Manag. 2021, 29, 2300–2327. [Google Scholar] [CrossRef]
  51. Wu, H.; Zuo, J.; Zillante, G.; Wang, J.; Duan, H. Environmental impacts of cross-regional mobility of construction and demolition waste: An Australia Study. Resour. Conserv. Recycl. 2021, 174, 105805. [Google Scholar] [CrossRef]
  52. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  53. Hancock, J.T.; Khoshgoftaar, T.M. Survey on categorical data for neural networks. J. Big Data 2020, 7, 1–41. [Google Scholar] [CrossRef] [Green Version]
  54. Baldi, P. Autoencoders, unsupervised learning, and deep architectures. In Proceedings of the ICML Workshop on Unsupervised and Transfer Learning, Bellevue, WA, USA, 2 July 2011; pp. 37–49. [Google Scholar]
  55. Najafabadi, M.M.; Villanustre, F.; Khoshgoftaar, T.M.; Seliya, N.; Wald, R.; Muharemagic, E. Deep learning applications and challenges in big data analytics. J. Big Data 2015, 2, 1–21. [Google Scholar] [CrossRef] [Green Version]
  56. Delgado, J.M.D.; Oyedele, L. Deep learning with small datasets: Using autoencoders to address limited datasets in construction management. Appl. Soft Comput. 2021, 112, 107836. [Google Scholar] [CrossRef]
  57. Kingma, D.P.; Welling, M. An Introduction to Variational Autoencoders; Foundations and Trends® in Machine Learning · Now Publishers Inc.: Hanover, MA, USA, 2019; Volume 12, pp. 307–392. [Google Scholar]
  58. Ranzato, M.A.; Poultney, C.; Chopra, S.; Cun, Y. Efficient learning of sparse representations with an energy-based model. In Advances in Neural Information Processing Systems 19 (NIPS 2006); MIT Press: Cambridge, MA, USA, 2006; Volume 19. [Google Scholar]
  59. Lewicki, M.S.; Sejnowski, T.J. Learning. overcomplete representations. Neural Comput. 2000, 12, 337–365. [Google Scholar] [CrossRef]
  60. Fan, C.; Sun, Y.; Zhao, Y.; Song, M.; Wang, J. Deep learning-based feature engineering methods for improved building energy prediction. Appl. Energy 2019, 240, 35–45. [Google Scholar] [CrossRef]
  61. Meyer, D. Introduction to Autoencoders. 2015. Available online: https://davidmeyer.github.io/ml/ae.pdf (accessed on 30 January 2023).
  62. Bengio, Y.; Lamblin, P.; Popovici, D.; Larochelle, H. Greedy layer-wise training of deep networks. In Advances in Neural Information Processing Systems 19 (NIPS 2006); MIT Press: Cambridge, MA, USA, 2006; Volume 19. [Google Scholar]
  63. Saha, M.; Santara, A.; Mitra, P.; Chakraborty, A.; Nanjundiah, R.S. Prediction. of the Indian summer monsoon using a stacked autoencoder and ensemble regression model. Int. J. Forecast. 2021, 37, 58–71. [Google Scholar] [CrossRef]
  64. Charte, D.; Charte, F.; del Jesus, M.J.; Herrera, F. An analysis on the use of autoencoders for representation learning: Fundamentals, learning task case studies, explainability and challenges. Neurocomputing 2020, 404, 93–107. [Google Scholar] [CrossRef]
  65. Guo, H.N.; Wu, S.B.; Tian, Y.J.; Zhang, J.; Liu, H.T. Application of machine learning methods for the prediction of organic solid waste treatment and recycling processes: A review. Bioresour. Technol. 2021, 319, 124114. [Google Scholar] [CrossRef] [PubMed]
  66. Xia, W.; Jiang, Y.; Chen, X.; Zhao, R. Application of machine learning algorithms in municipal solid waste management: A mini review. Waste Manag. Res. 2022, 40, 609–624. [Google Scholar] [CrossRef]
  67. Fernández-Delgado, M.; Cernadas, E.; Barro, S.; Amorim, D. Do we need hundreds of classifiers to solve real world classification problems? J. Mach. Learn. Res. 2014, 15, 3133–3181. [Google Scholar]
  68. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  69. Namoun, A.; Hussein, B.R.; Tufail, A.; Alrehaili, A.; Syed, T.A.; BenRhouma, O. An ensemble learning based classification approach for the prediction of household solid waste generation. Sensors 2022, 22, 3506. [Google Scholar] [CrossRef] [PubMed]
  70. Chen, K.Y.; Wang, C.H. Support vector regression with genetic algorithms in forecasting tourism demand. Tour. Manag. 2007, 28, 215–226. [Google Scholar] [CrossRef]
  71. Wang, J.; Liu, Z.; Lu, P. Electricity load forecasting using rough set attribute reduction algorithm based on immune genetic algorithm and support vector machines. In Proceedings of the 2008 International Conference on Risk Management & Engineering Management, Beijing, China, 4–6 November 2008; pp. 239–244. [Google Scholar]
  72. Xu, A.; Chang, H.; Xu, Y.; Li, R.; Li, X.; Zhao, Y. Applying artificial neural networks (ANNs) to solve solid waste-related issues: A critical review. Waste Manag. 2021, 124, 385–402. [Google Scholar] [CrossRef]
  73. You, H.; Ma, Z.; Tang, Y.; Wang, Y.; Yan, J.; Ni, M.; Cen, K.; Huang, Q. Comparison of ANN (MLP), ANFIS, SVM, and RF models for the online classification of heating value of burning municipal solid waste in circulating fluidized bed incinerators. Waste Manag. 2017, 68, 186–197. [Google Scholar] [CrossRef] [PubMed]
  74. Wang, Y.; Li, J.; Gu, J.; Zhou, Z.; Wang, Z. Artificial neural networks for infectious diarrhea prediction using meteorological factors in Shanghai (China). Appl. Soft Comput. 2015, 35, 280–290. [Google Scholar] [CrossRef]
  75. Dong, B.; Cao, C.; Lee, S.E. Applying support vector machines to predict building energy consumption in tropical region. Energy Build. 2005, 37, 545–553. [Google Scholar] [CrossRef]
  76. Chang, C.C.; Lin, C.J. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 1–27. [Google Scholar] [CrossRef]
  77. Cheng, J.; Dekkers, J.C.; Fernando, R.L. Cross-validation of best linear unbiased predictions of breeding values using an efficient leave-one-out strategy. J. Anim. Breed. Genet. 2021, 138, 519–527. [Google Scholar] [CrossRef]
  78. Wong, T.T. Performance evaluation of classification algorithms by k-fold and leave-one-out cross validation. Pattern Recognit. 2015, 48, 2839–2846. [Google Scholar] [CrossRef]
  79. Witten, I.H.; Hall, M.A. Data Mining: Practical Machine Learning Tools and Techniques, 3rd ed.; Morgan Kaufmann: Massachusetts, MA, USA, 2011. [Google Scholar]
  80. Cheng, H.; Garrick, D.J.; Fernando, R.L. Efficient. strategies for leave-one-out cross validation for genomic best linear unbiased prediction. J. Anim. Sci. Biotechnol. 2017, 8, 1–5. [Google Scholar] [CrossRef] [Green Version]
  81. Shao, Z.; Er, M.J. Efficient. leave-one-out cross-validation-based regularized extreme learning machine. Neurocomputing 2016, 194, 260–270. [Google Scholar] [CrossRef]
  82. Bai, J.; Li, Y.; Li, J.; Yang, X.; Jiang, Y.; Xia, S.T. Multinomial random forest. Pattern Recognit. 2022, 122, 108331. [Google Scholar] [CrossRef]
Figure 1. Flowchart of research steps performed in this study (undercomplete AE and overcomplete AE are detailed in Section 2.3).
Figure 1. Flowchart of research steps performed in this study (undercomplete AE and overcomplete AE are detailed in Section 2.3).
Sustainability 15 03691 g001
Figure 2. Stacked AE architecture based on applied representation size.
Figure 2. Stacked AE architecture based on applied representation size.
Sustainability 15 03691 g002
Figure 3. Training and validation loss on six-feature representation by AE.
Figure 3. Training and validation loss on six-feature representation by AE.
Sustainability 15 03691 g003
Figure 4. Performance comparison of RMSE and MAE of hybrid models and non-hybrid models. (a) ANN (MLP), (b) SVR, and (c) RF.
Figure 4. Performance comparison of RMSE and MAE of hybrid models and non-hybrid models. (a) ANN (MLP), (b) SVR, and (c) RF.
Sustainability 15 03691 g004
Figure 5. Performance comparison of R2 and R of hybrid models and non-hybrid models. (a) ANN (MLP), (b) SVR, and (c) RF.
Figure 5. Performance comparison of R2 and R of hybrid models and non-hybrid models. (a) ANN (MLP), (b) SVR, and (c) RF.
Sustainability 15 03691 g005
Figure 6. Correlation comparison between predicted and observed values. (a) ANN (MLP) and (b) AE (25)−ANN (MLP).
Figure 6. Correlation comparison between predicted and observed values. (a) ANN (MLP) and (b) AE (25)−ANN (MLP).
Sustainability 15 03691 g006
Figure 7. Comparison results of observed and predicted values.
Figure 7. Comparison results of observed and predicted values.
Sustainability 15 03691 g007
Table 1. Sample building input and output of raw data.
Table 1. Sample building input and output of raw data.
Building FeaturesOutput
Bldg IDLocationStructureUsageWall TypeRoof TypeGFA (m2)Demolition Waste Generation (kg·m−2)
Bldg 1Project BRCResidentialConcreteSlab289.501279.71
Bldg 2Project ARCResidentialBrickSlab114.202060.07
Bldg 3Project ARCResidentialBrickSlab100.453875.76
Bldg 4Project ARCResidentialBrickSlab100.451644.75
Bldg 5Project ARCResidentialBrickSlab197.681458.22
Bldg 6Project ARCResidentialBrickSlab190.362519.33
Bldg 7Project ARCResidentialBrickSlab114.802494.95
Bldg 8Project ARCResidentialBrickSlab118.413398.11
Bldg 9Project ARCResidentialBrickSlab47.111849.38
Bldg 10Project ARCResidentialBrickSlab106.452665.72
Bldg 11Project ARCResidentialBrickSlab87.532805.48
Bldg 12Project ARCResidentialBrickSlab82.403024.04
Bldg 13Project ARCResidentialBrickSlab95.156033.67
Bldg 14Project ARCResidentialBrickSlab51.112426.00
Bldg 15Project ARCResidentialBrickSlab149.511990.99
Bldg 781Project CMasonryResidentialBlockSlate85.66823.51
Bldg 782Project BMasonryCommercialBlockSlab and roofing tile94.441166.09
Table 2. Building status and statistical analysis of raw data.
Table 2. Building status and statistical analysis of raw data.
CategoryNumbersDWGR (kg·m2)
TotalMinMeanMax
LocationProject A343450,31029813136034
Project B356485,0378313628574
Project C83101,53173612231808
UsageResidential595767,5788312908574
Residential and commercial172251,38141814625718
Commercial1519,51060713012474
StructureRC87169,53841819496034
Masonry604788,0428313058574
Wood9180,8892988892237
Wall typeConcrete910,35787111514696
Brick236391,25925216586034
Block500596,7998311948574
Mud plastered and mortar3740,05651710832591
Roof typeSlab289479,35625216596034
Slab and roofing tile3338,87725211781808
Slate178227,92330612808574
Roofing tile282292,3148310372527
Table 3. Label encoding for numerical data conversion of categorical variables in this study.
Table 3. Label encoding for numerical data conversion of categorical variables in this study.
Categorical Variable Numerical Value Assigned by Label Encoding
LocationLocation_project A0
Location_project B1
Location_project C2
StructureStructure_RC0
Structure_masonry1
Structure_wood2
UsageUsage_ residential0
Usage_residential & commercial1
Usage_ commercial2
Wall typeWall type_concrete0
Wall type_brick1
Wall type_block2
Wall type_ mud plastered and mortar3
Roof typeRoof type_slab0
Roof type_slab and roofing tile1
Roof type_slate2
Roof type_roofing tile3
Table 4. Comparison of performance results of non-hybrid and hybrid models for prediction of C&D and solid WG.
Table 4. Comparison of performance results of non-hybrid and hybrid models for prediction of C&D and solid WG.
StudyEstimation LevelWaste TypeInput Variable Data Composition (Number of Input Variables or Characteristics of Data)Model TypePerformance of Predictive Model
RMSEMSEMAEMAPE R2 R Error Rate (%)
This studyBuildingDWNumerical (1); categorical (5)ANN (MLP)316.186 356.697 0.4580.676
AE (25 features)–ANN (MLP)230.819 128.105 0.6800.825
SVR1020.304 1116.266 0.0070.083
AE (3 features)–SVR302.224 384.347 0.1100.494
Abbasi et al., 2013 [20]City MSWNumerical (time series data)SVM2070 0.761
PL–SVM1541 0.869
Abbasi et al., 2014 [21]City MSWNumerical (time series data)SVM814–3268 0.702–0.756
WT-SVM639–2283 0.813–0.887
Song et al., 2017 [18]RegionalC&D wasteNumerical (time series data)GM 21
GM -SVR 4.6
Golbaz et al., 2019 [13]BuildingMSW (hospital solid waste)Numerical (7); categorical (1)SVM 0.001–0.003 0.79–0.98
F–SVM 0.001–0.002 0.79–0.92
Soni et al., 2019 [19]CityMSWNumerical (4)ANN165.5 0.72
GA–ANN95.7 0.87
Cai et al., 2020 [24]RegionalC&D wasteNumerical (time series data)SVR50.19 17.29
LSTM–SVR29.04 10.02
Dai et al., 2020 [47]DistrictMSWNumerical (3; time series data)SVR 34.725 14.4340.8376
FIG–GA–SVM 5.703 2.0120.9845
Liang et al., 2021 [14]CityMSWNumerical (8)ANN11.23 10.29 0.76
AOA–ANN5.89 6.21 0.88
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cha, G.-W.; Hong, W.-H.; Kim, Y.-C. Performance Improvement of Machine Learning Model Using Autoencoder to Predict Demolition Waste Generation Rate. Sustainability 2023, 15, 3691. https://doi.org/10.3390/su15043691

AMA Style

Cha G-W, Hong W-H, Kim Y-C. Performance Improvement of Machine Learning Model Using Autoencoder to Predict Demolition Waste Generation Rate. Sustainability. 2023; 15(4):3691. https://doi.org/10.3390/su15043691

Chicago/Turabian Style

Cha, Gi-Wook, Won-Hwa Hong, and Young-Chan Kim. 2023. "Performance Improvement of Machine Learning Model Using Autoencoder to Predict Demolition Waste Generation Rate" Sustainability 15, no. 4: 3691. https://doi.org/10.3390/su15043691

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop