Prediction of Cooling Load of Tropical Buildings with Machine Learning

: Cooling load refers to the amount of energy to be removed from a space (or consumed) to bring that space to an acceptable temperature or to maintain the temperature of a space at an acceptable range. The study aimed to develop a series of models and determine the most accurate ones in the prediction of the cooling load of low-rise tropical buildings based on their basic architectural and structural characteristics. In this context, a series of machine learning (regression) algorithms were tested during the research to determine the most accurate/efﬁcient prediction model. In this regard, a data set consisting of ten features indicating the basic characteristics of the building (ﬂoor area, aspect ratio, ceiling height, window material, external wall material, roof material, window wall ratio north faced, window wall ratio south faced, horizontal shading, orientation) were used to predict the cooling load of a low-rise tropical building. The dataset was generated utilizing a set of generative and algorithmic design tools. Following the dataset generation, a series of regression models were tested to ﬁnd the most accurate model to predict the cooling load. The results of the tests with different algorithms revealed that the relationship between the predictor variables and cooling load could be efﬁciently modeled through Histogram Gradient Boosting and Stacking models


Introduction
The famous Brundtland Commission Report of the UN defines sustainability as meeting the needs of the present without compromising the ability of future generations to meet their own needs [1]. To make life on earth sustainable by consuming less and polluting less is one of the most important responsibilities of humans. The majority of the energy sources used today come from fossils. In fact, in recent years, more environmentally friendly and sustainable methods have been developed to produce green energy or to consume less. New energy sources are sustainable to a certain extent. However, the most important issue is to reduce energy consumption and to ensure that the same work is carried out with less energy and, thus, energy efficiency is achieved. In collective living areas, most of the energy is used for the heating and cooling needs of indoor spaces.
Global warming refers to the impact of human activities on the climate, especially the burning of fossil fuels and large-scale deforestation, which causes the emission of greenhouse gases, such as carbon dioxide, into the atmosphere [2]. As a result of the increase in energy demand with the recent global warming, it has become important to re-evaluate the old methods and the measures taken in terms of energy efficiency. In this context, it is seen that the implementation of concepts such as green buildings, green industry, and passive houses has increased. Governments have made serious regulations on these issues [3].
The envelope of the buildings has been constantly developed due to the factors such as temperature, humidity, winds and solar radiation of the outside air. These factors are evaluated separately in hot weather and cold weather conditions. Especially the radiation effect of the sun and air temperature are important factors in terms of indoor thermal comfort in both weather conditions. In hot climate regions, the hot period lasts longer than the cold period. For this reason, the hot periods of these regions are important in terms of cooling load of buildings. Climatic changes due to environmental pollution in recent years (greenhouse effect, global warming, . . . ) bring energy expenditures and cooling to the forefront [4]. The use of climatic data in house design is very important in ensuring energy efficiency in the building. The design of buildings according to different climatic characteristics is effective in the formation of suitable indoor conditions and energy conservation. For example, a courtyard with an external part of the building reduces the cooling load by providing abundant air flow [5].
The calculation of the cooling load in every building has become necessary due to the increase in energy costs and to prevent climate changes caused by fuel energy consumption in buildings. Especially for buildings located in hot and tropical climate regions, it becomes especially important to determine and evaluate cooling loads in the design stage. On the other hand, the gradual increase in outside temperatures due to global warming leads to an increase in cooling load, especially in buildings located in hot-humid climates. This high cooling demand leads to more greenhouse gas emissions and thus supports the global warming process [6]. This situation generates a vicious circle which worsens the situation by causing more and more greenhouse gas emissions every day. Therefore, the cooling load should be calculated carefully. Energy efficiency should estimate and limit the energy consumption for the cooling of buildings.
As indicated in Table 1, the production industry was the sector with the highest energy consumption globally. Buildings were the second in terms of global energy consumption. Only this indicator can point out that serious energy savings can be achieved by enabling energy efficiency in the design and use of buildings. To contribute to the energy savings in the building industry, our study mainly focuses on facilitating the design phase of buildings through the prediction of energy consumption parameters, such as cooling load, with a simple but efficient approach. The amount of energy used in buildings at the urban scale varies regionally but accounts for approximately 40% of total consumption [8,9]. This issue clearly shows that buildings are responsible for a large part of energy consumption that cannot be ignored. Therefore, efficient building design with energy-saving features and auxiliary tools that support these practices can be used to improve the energy efficiency of buildings and may be helpful in alleviating its use.
The main purpose of this study is to determine the cooling load in buildings located in tropical climates. Because when the cooling load of the buildings is known, a design can be made according to the required cooling load, the energy to be consumed by the cooling devices can be reduced, and a sustainable design can be realized. Malaysia is a country located in Southeast Asia. Located between one and four degrees north latitude, Malaysia's climate is equatorial. Tropical forests cover 70% of the country. Because of the effect of monsoon winds and continuous rains, especially between January and May, the country's humidity increases during this season. The daily temperature is between 21 • C and 32 • C in the lowlands, while it is lower in the higher regions [10]. Malaysia is known for its hot and humid climate. As a solution to these climatic conditions, air conditioners (AC)s and AC systems, in general, are widely used in all regions of the country. The wide use of AC systems results in high energy consumption in buildings. In countries such as Malaysia, buildings should be designed to minimize the energy consumption related to the cooling of indoor spaces. Establishing the functional link between the architectural properties of the building and its energy consumption is crucial to facilitate energy-efficient building designs. In this study, the following aspect of architectural design is utilized for the prediction framework we have implemented and tested.
There are various types of houses available in the market in Malaysia, but the terraced house type typically accounts for 41% of the total residential property stock in 2018, which is available in one and multi-story types [11,12]. It is followed by low-cost houses (including low-cost house, low-cost flat and flat) with a 30 percent portion, which is purposely constructed for the low-income group. Terraced houses are the most common residential type, classified under low-to medium-cost housing [13], which are preferred by developers due to the speedy construction methods [14]. As the design is based on the British terraced house, it did not take into account local climatic conditions and cultures when it was brought to Malaysia [15]. The scope of this study was double-story terrace houses. The double-story terrace houses are the most common type of low-rise buildings and the biggest fraction of both existing supplies and newly planned residential units in Malaysia. The case study was carried out on an intermediate double-story terrace house to represent the simulation of conventional low-rise residentials in Malaysia. Figure 1 is the representative building used in this study. The building was located in Skudai, Johor Bahru (latitude 1 • 32 N, longitude 103 • 40 E), with a total floor area of 200 m 2 . The building had a rectangular shape floor plan and aspect ratio of (width/length) 1/2 and a ceiling height of 3.8 m. This model is a typical terrace house with an indoor space layout of a living-cum-dining area (DL), kitchen (K), a guest room with one bathroom (WC) on the ground floor as well as one master bedroom (MB), two smaller bedrooms together with a hall area (corridor) along with two bathrooms (WC) on the first floor. The house was facing the South-North direction, with large windows in the front facade (South). The building structure is a reinforced concrete structure with a brick-infilled frame. The roof is pitched, covered with clay tiles and non-insulated walls and roofs. Cement sand renders covered the façade. The walls are built from brick, and the surface is plastered. A single-glazed window frame was made of aluminum, and Table 2 describes the base model materials and  thermal properties. country's humidity increases during this season. The daily temperature is betwe and 32 °C in the lowlands, while it is lower in the higher regions [10]. Malaysia is for its hot and humid climate. As a solution to these climatic conditions, air cond (AC)s and AC systems, in general, are widely used in all regions of the country. T use of AC systems results in high energy consumption in buildings. In countries Malaysia, buildings should be designed to minimize the energy consumption re the cooling of indoor spaces. Establishing the functional link between the arch properties of the building and its energy consumption is crucial to facilitate ene cient building designs. In this study, the following aspect of architectural design is for the prediction framework we have implemented and tested.
There are various types of houses available in the market in Malaysia, but the house type typically accounts for 41% of the total residential property stock in 201 is available in one and multi-story types [11,12]. It is followed by low-cost houses ing low-cost house, low-cost flat and flat) with a 30 percent portion, which is pu constructed for the low-income group. Terraced houses are the most common res type, classified under low-to medium-cost housing [13], which are preferred by d ers due to the speedy construction methods [14]. As the design is based on th terraced house, it did not take into account local climatic conditions and cultures was brought to Malaysia [15]. The scope of this study was double-story terrace The double-story terrace houses are the most common type of low-rise buildings biggest fraction of both existing supplies and newly planned residential units in M The case study was carried out on an intermediate double-story terrace hous resent the simulation of conventional low-rise residentials in Malaysia. Figure 1 is resentative building used in this study. The building was located in Skudai, Joho (latitude 1°32′ N, longitude 103°40′ E), with a total floor area of 200 m 2 . The build a rectangular shape floor plan and aspect ratio of (width/length) 1/2 and a ceilin of 3.8 m. This model is a typical terrace house with an indoor space layout of a livi dining area (DL), kitchen (K), a guest room with one bathroom (WC) on the grou as well as one master bedroom (MB), two smaller bedrooms together with a hall a ridor) along with two bathrooms (WC) on the first floor. The house was facing the North direction, with large windows in the front facade (South). The building stru a reinforced concrete structure with a brick-infilled frame. The roof is pitched, with clay tiles and non-insulated walls and roofs. Cement sand renders covered çade. The walls are built from brick, and the surface is plastered. A single-glazed frame was made of aluminum, and Table 2 describes the base model materials and properties. Figure 1. Representative building used in the study [16].  These aspects were the total floor area, aspect ratio, ceiling height, window material, external wall material, roof material, window wall ratio north faced, window wall ratio south faced, horizontal shading, and orientation of the building. The literature defines the Cooling Load as the building's energy consumption, or the amount of energy required to keep the environment at a constant temperature [17]. The number of variables affecting cooling load calculations is very high. This study focuses on the following subset of these variables that appear as the key ones based on findings in our previous research [16] ( Figure 2).  Representative building used in the study [16]. These aspects were the total floor area, aspect ratio, ceiling height, window material, external wall material, roof material, window wall ratio north faced, window wall ratio south faced, horizontal shading, and orientation of the building. The literature defines the Cooling Load as the building's energy consumption, or the amount of energy required to keep the environment at a constant temperature [17]. The number of variables affecting cooling load calculations is very high. This study focuses on the following subset of these variables that appear as the key ones based on findings in our previous research [16]   Floor area: The floor area is the floor area of the region where the cooling load will be calculated [18]. As it increases, the cooling load will increase too.
Aspect ratio: Aspect ratio is the ratio of the width to the height of the building. Thanks to the optimum aspect ratio, the building is shaded in hot weather, and the energy consumption required for cooling is reduced [19].
Ceiling height: As the height of the ceiling increases, the air volume in the room, which directly affects the dynamics that affect the cooling load, also increases. This also affects cooling efficiency [20].
Window material: Appropriate thermal comfort conditions can be achieved with glass selection according to the characteristic features of climate zones [21].
External wall material: One of the major factors that make up the cooling load is the total heat gained from the external walls [22]. Floor area: The floor area is the floor area of the region where the cooling load will be calculated [18]. As it increases, the cooling load will increase too.
Aspect ratio: Aspect ratio is the ratio of the width to the height of the building. Thanks to the optimum aspect ratio, the building is shaded in hot weather, and the energy consumption required for cooling is reduced [19].
Ceiling height: As the height of the ceiling increases, the air volume in the room, which directly affects the dynamics that affect the cooling load, also increases. This also affects cooling efficiency [20].
Window material: Appropriate thermal comfort conditions can be achieved with glass selection according to the characteristic features of climate zones [21].
External wall material: One of the major factors that make up the cooling load is the total heat gained from the external walls [22]. Roof material: The roof material used is an important factor that has an impact on the cooling load. For example, when comparing a traditional roof with a green roof, it is known that the green roof application saves energy and has less negative impact on the environment [23].
Window wall ratio: The window-to-wall ratio is the ratio of the window area to the entire façade surface area. In regions with climatic conditions where heating energy demand is high, solar energy gain increases as the window/wall ratio increases [24].
Horizontal shading: With horizontal shading installation, the cooling load is reduced compared to the case without shading [25]. Elements such as balconies, overhangs, etc., are horizontal shading elements [26].
Orientation: Energy efficiency can be achieved with the right building orientation. Among the building orientation types, the cooling load increases in the perimeter zones orientated towards the west façade [27].
Heat loads have long been calculated manually and using the instantaneous calculation method, which assumes that heat gains are converted into instantaneous cooling loads. Although this method is simple and fast, it neglects processes such as heat storage and radiation transfer and, therefore, has little reliability [28]. There are many methods for cooling load calculations. Figure 3 shows the relationship between the American Society of Heating, Refrigerating, and Air-conditioning Engineers (ASHRAE) cooling load calculation methods in terms of complexity and accuracy. Roof material: The roof material used is an important factor that has an impact on the cooling load. For example, when comparing a traditional roof with a green roof, it is known that the green roof application saves energy and has less negative impact on the environment [23].
Window wall ratio: The window-to-wall ratio is the ratio of the window area to the entire façade surface area. In regions with climatic conditions where heating energy demand is high, solar energy gain increases as the window/wall ratio increases [24].
Horizontal shading: With horizontal shading installation, the cooling load is reduced compared to the case without shading [25]. Elements such as balconies, overhangs, etc., are horizontal shading elements [26].
Orientation: Energy efficiency can be achieved with the right building orientation. Among the building orientation types, the cooling load increases in the perimeter zones orientated towards the west façade [27].
Heat loads have long been calculated manually and using the instantaneous calculation method, which assumes that heat gains are converted into instantaneous cooling loads. Although this method is simple and fast, it neglects processes such as heat storage and radiation transfer and, therefore, has little reliability [28]. There are many methods for cooling load calculations. Figure 3 shows the relationship between the American Society of Heating, Refrigerating, and Air-conditioning Engineers (ASHRAE) cooling load calculation methods in terms of complexity and accuracy. From Figure 3, it can be seen that ASHRAEs Heat Balance Method has the highest complexity and accuracy. The Heat Balance Method, using the finite differences approach, calculates the inner surface temperatures of each surface in detail, as well as the solar gains, and makes the closest estimation of the heating and cooling load with the inclusion of natural ventilation, shading, HVAC equipment, and thermal mass [29]. The fact that accuracy increases as the complexity increases has led to the search for methods that are less time-consuming and complex. The calculation of the cooling load is more complex than the heating load due to the presence of dynamic responses and thermal mass [30].
In recent years, studies carried out utilizing machine learning methods have achieved very accurate results in the estimation of cooling loads. For example, Xuemei et al. [31] developed the Least Square Support Vector Machine (LS-SVM) for cooling load prediction, and when compared with Back Propagation Neural Network (BPNN), LS-SVM provided a higher accuracy with less error. Similarly, Li et al. [32] utilized a support vector machine (SVM) to predict the hourly building cooling load and achieved effective results. Gao et al. [33] used extreme learning machine (ELM) and random forest (RF) together to From Figure 3, it can be seen that ASHRAEs Heat Balance Method has the highest complexity and accuracy. The Heat Balance Method, using the finite differences approach, calculates the inner surface temperatures of each surface in detail, as well as the solar gains, and makes the closest estimation of the heating and cooling load with the inclusion of natural ventilation, shading, HVAC equipment, and thermal mass [29]. The fact that accuracy increases as the complexity increases has led to the search for methods that are less time-consuming and complex. The calculation of the cooling load is more complex than the heating load due to the presence of dynamic responses and thermal mass [30].
In recent years, studies carried out utilizing machine learning methods have achieved very accurate results in the estimation of cooling loads. For example, Xuemei et al. [31] developed the Least Square Support Vector Machine (LS-SVM) for cooling load prediction, and when compared with Back Propagation Neural Network (BPNN), LS-SVM provided a higher accuracy with less error. Similarly, Li et al. [32] utilized a support vector machine (SVM) to predict the hourly building cooling load and achieved effective results. Gao et al. [33] used extreme learning machine (ELM) and random forest (RF) together to predict the cooling load of large commercial buildings. Sha et al. [34] compared the performances of different ML algorithms in predicting cooling load and showed that gradient tree boosting (GTB) achieves the most accuracy with fewer errors. Ngo [35] applied an ML method for the prediction of cooling loads of buildings, based on data from 243 buildings and observed high accuracy in predictions. Rana et al. [36] proposed a data-driven approach that has shown greater accuracy than gradient tree boosting (GTB). Xuan et al. [37] used the Chaos approach and Wavelet Decomposition (WD) with the Support Vector Regression separately to predict the cooling load, and the results showed that the hybrid forecasting models perform better than the single ones. Zingre et al. [38] applied long short-term memory (LSTM) to estimate cooling load and demonstrated the predictive potential of this method when the data are in the form of a time series.
In this study, we implemented and tested several foundational machine learning models (Linear Regression, Decision Tree, Elastic Net, K Nearest Neighbor, Support Vector Machines) and ensemble machine learning models (Random Forest, Gradient Boosting, Histogram Gradient Boosting, Voting, Stacking) to determine the model with the best performance in the prediction of the cooling load based on architectural aspects of a tropical building. Python [39] programming language v3.6 was used for the cooling load estimation experiments. Anaconda 3 [40] was preferred as the development environment. Numpy [41] and Pandas [42] libraries were used to prepare the data for the training, the sci-kit-learn library was used to develop machine learning models, and the Matplotlib [43] library was utilized for data visualization.

Materials and Methods
The dataset utilized in this study is a 10,000-row subset of the dataset generated in [16]. K-Fold Cross Validation (KFCV) method was used as the training/validation strategy. A set of machine learning algorithms consisting of foundational algorithms (linear regression, decision tree regression, elastic network regression, K nearest neighbor regression, and support vector regression) and ensemble learning algorithms (random forest regression, gradient boosting regression, histogram-based gradient boosting regression and voting and stacking) is implemented in the training and validation stage. In the final stage, performance metrics were used to explore the effectiveness of the trained models. At this stage, different evaluation metrics such as coefficient of determination (R 2 ), mean square error (MSE), root mean square error (RMSE), and mean absolute error (MAE) were checked to discover the most accurate/efficient machine learning model. Finally, the trained models are stored. A schematic representation of the model evaluation process is provided in Figure 4. predict the cooling load of large commercial buildings. Sha et al. [34] compared the performances of different ML algorithms in predicting cooling load and showed that gradient tree boosting (GTB) achieves the most accuracy with fewer errors. Ngo [35] applied an ML method for the prediction of cooling loads of buildings, based on data from 243 buildings and observed high accuracy in predictions. Rana et al. [36] proposed a data-driven approach that has shown greater accuracy than gradient tree boosting (GTB). Xuan et al. [37] used the Chaos approach and Wavelet Decomposition (WD) with the Support Vector Regression separately to predict the cooling load, and the results showed that the hybrid forecasting models perform better than the single ones. Zingre et al. [38] applied long short-term memory (LSTM) to estimate cooling load and demonstrated the predictive potential of this method when the data are in the form of a time series.
In this study, we implemented and tested several foundational machine learning models (Linear Regression, Decision Tree, Elastic Net, K Nearest Neighbor, Support Vector Machines) and ensemble machine learning models (Random Forest, Gradient Boosting, Histogram Gradient Boosting, Voting, Stacking) to determine the model with the best performance in the prediction of the cooling load based on architectural aspects of a tropical building. Python [39] programming language v3.6 was used for the cooling load estimation experiments. Anaconda 3 [40] was preferred as the development environment. Numpy [41] and Pandas [42] libraries were used to prepare the data for the training, the sci-kit-learn library was used to develop machine learning models, and the Matplotlib [43] library was utilized for data visualization.

Materials and Methods
The dataset utilized in this study is a 10,000-row subset of the dataset generated in [16]. K-Fold Cross Validation (KFCV) method was used as the training/validation strategy. A set of machine learning algorithms consisting of foundational algorithms (linear regression, decision tree regression, elastic network regression, K nearest neighbor regression, and support vector regression) and ensemble learning algorithms (random forest regression, gradient boosting regression, histogram-based gradient boosting regression and voting and stacking) is implemented in the training and validation stage. In the final stage, performance metrics were used to explore the effectiveness of the trained models. At this stage, different evaluation metrics such as coefficient of determination (R 2 ), mean square error (MSE), root mean square error (RMSE), and mean absolute error (MAE) were checked to discover the most accurate/efficient machine learning model. Finally, the trained models are stored. A schematic representation of the model evaluation process is provided in Figure 4.

Dataset Generation and Exploratory Data Analysis
Generating large and reliable datasets is pertinent in developing the prediction model as such datasets lead to better accuracy. In this study, two simulation methods namely the BIM application and Monte Carlo simulation are used in order to generate a large dataset. Simulations are carried out to collect reliable and verified data with an emphasis on covering the whole range of available values. The four steps in generating a large and reliable dataset involve: (1) preparing a 3D base model for energy consumption (cooling load) simulation using the BIM application; (2) developing the dataset using the BIM application; (3) generating more data using the Monte Carlo method, and (4) synthesizing the data derived from the BIM application and Monte Carlo simulation into one large dataset [16]. Once the variable ranges were determined, a Monte Simulation was run to select random values from the pre-determined value ranges stochastically. The EnergyPlus software (2010 version) was used to perform energy simulations and calculate annual building energy consumption (Cooling Load) for each design scenario that was generated in the Monte Carlo simulation. The dataset in this research is a randomly selected subset of the original dataset of 90,000+ rows and is composed of 10,000 rows. All variables were continuous. Measures of central tendency and dispersion for predictor variables are given in Table 3, and the histogram and boxplot of all variables in the dataset are provided in Figures 5 and 6. An explanation of all the variables was provided in the previous section. As the dataset was developed with a Monte Carlo Simulation based on discrete value ranges determined earlier based on standards and interviews, the values of all variables have discrete distributions. Although the values of independent variables are discrete, the dependent variable, Cooling Load, has a continuous distribution enabled and supported by a large number of samples generated as a result of the data generation with simulation and generative design tools.

Machine Learning Models
Many years ago, researchers began to focus on the prediction of building cooling load in buildings by using BIM tools such as Energyplus and DOE 2. The operational energy prediction can be expedited using Building Information Modeling (BIM). Nevertheless, energy simulation tools have their own challenges such as the size and number of parameters that need to be considered. If the number of parameters and their variables are large, then this method takes a long time and requires more intention, and it will be very challenging. A small mistake due to carelessness would cause the wrong results, thus the wrong judgment of cooling load prediction. Furthermore, the problems with the dataset that is generated through the simulation are erroneous design assumptions and errors within modeling tools and their application. Due to the reasons above, researchers have to put a direct and significant focus on data-based energy consumption prediction for buildings taking into account data properties and artificial intelligence (machine learning) algorithms.
After the generation of the dataset, several machine learning (ML) models were implemented and tested. Machine learning is a sub-branch of artificial intelligence and refers to the artificial learning process through a set of algorithms that can continuously learn structurally and make meaningful inferences from large and complex data. Machine learning methods have gained importance in today's world [44]. Machine learning algorithms are used in many fields today [45][46][47][48][49][50][51][52]. Supervised learning, which is a type of machine learning, refers to the implementation of classification and regression algorithms

Machine Learning Models
Many years ago, researchers began to focus on the prediction of building cooling load in buildings by using BIM tools such as Energyplus and DOE 2. The operational energy prediction can be expedited using Building Information Modeling (BIM). Nevertheless, energy simulation tools have their own challenges such as the size and number of parameters that need to be considered. If the number of parameters and their variables are large, then this method takes a long time and requires more intention, and it will be very challenging. A small mistake due to carelessness would cause the wrong results, thus the wrong judgment of cooling load prediction. Furthermore, the problems with the dataset that is generated through the simulation are erroneous design assumptions and errors within modeling tools and their application. Due to the reasons above, researchers have to put a direct and significant focus on data-based energy consumption prediction for buildings taking into account data properties and artificial intelligence (machine learning) algorithms.
After the generation of the dataset, several machine learning (ML) models were implemented and tested. Machine learning is a sub-branch of artificial intelligence and refers to the artificial learning process through a set of algorithms that can continuously learn structurally and make meaningful inferences from large and complex data. Machine learning methods have gained importance in today's world [44]. Machine learning algorithms are used in many fields today [45][46][47][48][49][50][51][52]. Supervised learning, which is a type of machine learning, refers to the implementation of classification and regression algorithms

Machine Learning Models
Many years ago, researchers began to focus on the prediction of building cooling load in buildings by using BIM tools such as Energyplus and DOE 2. The operational energy prediction can be expedited using Building Information Modeling (BIM). Nevertheless, energy simulation tools have their own challenges such as the size and number of parameters that need to be considered. If the number of parameters and their variables are large, then this method takes a long time and requires more intention, and it will be very challenging. A small mistake due to carelessness would cause the wrong results, thus the wrong judgment of cooling load prediction. Furthermore, the problems with the dataset that is generated through the simulation are erroneous design assumptions and errors within modeling tools and their application. Due to the reasons above, researchers have to put a direct and significant focus on data-based energy consumption prediction for buildings taking into account data properties and artificial intelligence (machine learning) algorithms.
After the generation of the dataset, several machine learning (ML) models were implemented and tested. Machine learning is a sub-branch of artificial intelligence and refers to the artificial learning process through a set of algorithms that can continuously learn structurally and make meaningful inferences from large and complex data. Machine learning methods have gained importance in today's world [44]. Machine learning algorithms are used in many fields today [45][46][47][48][49][50][51][52]. Supervised learning, which is a type of machine learning, refers to the implementation of classification and regression algorithms where a dependent variable is known in advance. Classification algorithms are implemented when the dependent variable is binary or categorical. Regression algorithms are utilized when the type of the dependent variable is continuous. Different validation strategies can be used in training the models. Two well-known strategies in this regard are train-test split and KFCV. In the train/test split strategy, the data are divided into 2 or 3 subsets (train/test or train/validation/test), trained with the former ones, and tested with the test dataset which the model has not seen before. Another validation strategy is K-Fold Validation where the dataset is trained and tested by dividing into k different parts.
In this study, the dataset is trained with different regression algorithms as the dependent variable (Cooling Load) is in a continuous nature/scale. The algorithms implemented are summarized in the following section. Additionally, the general machine learning process is shown in Figure 7. where a dependent variable is known in advance. Classification algorithms are implemented when the dependent variable is binary or categorical. Regression algorithms are utilized when the type of the dependent variable is continuous. Different validation strategies can be used in training the models. Two well-known strategies in this regard are train-test split and KFCV. In the train/test split strategy, the data are divided into 2 or 3 subsets (train/test or train/validation/test), trained with the former ones, and tested with the test dataset which the model has not seen before. Another validation strategy is K-Fold Validation where the dataset is trained and tested by dividing into k different parts. In this study, the dataset is trained with different regression algorithms as the dependent variable (Cooling Load) is in a continuous nature/scale. The algorithms implemented are summarized in the following section. Additionally, the general machine learning process is shown in Figure 7.

Foundational Algorithms
Linear Regression (LR): Linear regression is used to reveal the cause-and-effect relationship between a dependent variable and one or more independent variables [53]. The regression model associates the dependent variable with the independent variable or variables through a linear function [54]. Multiple linear regression (MLR) is used when there is more than one independent variable [55]. Decision Tree Regression (DTR): A decision tree is a nonparametric prediction model that can be used to represent both classifiers and regression models [56]. The method is widely used because the decision rules used in the creation of the mentioned tree structures are understandable. The decision tree performs a simple decision-making process by transforming complex data into a gradual state with a multi-stage and sequential approach to solve the classification and regression problem [57]. Elastic Network Regression (Elastic Net): Regularization in regression is used to avoid the overfitting of the data, especially when there is a large variance between train and test set performances. Well-known regularization methods in ML include LASSO(L1) and Ridge(L2). The Elastic Net method has emerged to tackle some of the shortcomings of the LASSO (L1) regularization method. Elastic Net uses LASSO (L1) regression and Ridge (L2) regression together [58]. K Nearest Neighbor Regression (k-NNR): The k Nearest Neighbor algorithm can be used for both classification and regression. Briefly defined, the data of unknown class is compared with other data in the training set, and a distance measurement is made. According to this calculated distance, the most optimal class is found for the data that has not yet been assigned to a class [59]. Support Vector Regression (SVR): Support vector regression is the use of support vector machines (SVM) for regression. In SVM, the optimum separation hyperplane is found to separate the classes from each other, and the distance between the support vectors of different classes is maximized [60].

Ensemble Algorithms
Ensemble learning is the joint decision-making of more than one learning algorithm. Unlike previous machine learning models, more than one learning model is run on the same data set and a joint decision is made according to certain rules [61]. Random Forest Regression (RF): In the RF method, a random forest is generated with multiple decision

Foundational Algorithms
Linear Regression (LR): Linear regression is used to reveal the cause-and-effect relationship between a dependent variable and one or more independent variables [53]. The regression model associates the dependent variable with the independent variable or variables through a linear function [54]. Multiple linear regression (MLR) is used when there is more than one independent variable [55]. Decision Tree Regression (DTR): A decision tree is a nonparametric prediction model that can be used to represent both classifiers and regression models [56]. The method is widely used because the decision rules used in the creation of the mentioned tree structures are understandable. The decision tree performs a simple decision-making process by transforming complex data into a gradual state with a multi-stage and sequential approach to solve the classification and regression problem [57]. Elastic Network Regression (Elastic Net): Regularization in regression is used to avoid the overfitting of the data, especially when there is a large variance between train and test set performances. Well-known regularization methods in ML include LASSO(L1) and Ridge(L2). The Elastic Net method has emerged to tackle some of the shortcomings of the LASSO (L1) regularization method. Elastic Net uses LASSO (L1) regression and Ridge (L2) regression together [58]. K Nearest Neighbor Regression (k-NNR): The k Nearest Neighbor algorithm can be used for both classification and regression. Briefly defined, the data of unknown class is compared with other data in the training set, and a distance measurement is made. According to this calculated distance, the most optimal class is found for the data that has not yet been assigned to a class [59]. Support Vector Regression (SVR): Support vector regression is the use of support vector machines (SVM) for regression. In SVM, the optimum separation hyperplane is found to separate the classes from each other, and the distance between the support vectors of different classes is maximized [60].

Ensemble Algorithms
Ensemble learning is the joint decision-making of more than one learning algorithm. Unlike previous machine learning models, more than one learning model is run on the same data set and a joint decision is made according to certain rules [61]. Random Forest Regression (RF): In the RF method, a random forest is generated with multiple decision trees. In the algorithm, the decision trees are combined to obtain a more accurate and stable forecast [62]. Gradient Boosting Regression (GBR): In the gradient boosting algorithm, the goal is to iteratively predict and improve the error rate. In this model, we tried to fit the loss function found in the previous step to a negative gradient vector. To reduce the error, the incremental reduction results are added to the function [63]. Histogrambased Gradient Boosting Regression (HGBR): Histogram-based gradient boosting is a combination of a gradient boosting machine and a histogram-based algorithm [64]. The histogram-based gradient boosting algorithm provides faster training of decision trees [65]. Voting: In the voting method, the class predictions made by various classifiers are voted on, and the class that obtains the most votes as a result of the voting becomes the class prediction of the majority [66]. Voting in regression is to combine different machine learning regressor models to return predicted average values. This is useful for balancing the individual weaknesses of the models [67]. In voting classification, more than one classification algorithm can be trained with the same training set, or a single algorithm can be trained with the same data set using different parameter values, and different classification models are created, and the final output value is produced by putting the outputs obtained from the models into the voting mechanism [68].
Stacking: The stacking ensemble learning method is based on the principle of generating a higher-performing prediction from these predictions by accepting the predictions of different types of classifiers as input for the metaclassifier [69]. The stacking ensemble learning method accepts the predictions of different types of classifiers as input for the metaclassifier, producing a higher-performing prediction from these predictions [70]. In stacking in classification, the output from different classifiers is passed as input to a metaclassifier for the final classification task [71].

K-Fold Cross Validation
The validation strategy for this study was selected as KFCV. KFCV helps determine a more accurate estimate of model prediction when compared with the train/test set validation strategy. This strategy helps in reducing the variance of the performance estimate and allows the use of more data for training. KFCV also avoids overfitting, as it exposes the model to different subsets of the data. In our implementation of KFCV, the number of folds (k) chosen is 10. In the 10-fold cross validation method, the dataset is divided into 10 equal parts. In each iteration, 9/10 of the dataset is used for training and the reserved 1/10 is used for testing. In each iteration, a different part of the data is reserved for testing. The basic structure of 10-fold cross validation is shown in Figure 8. trees. In the algorithm, the decision trees are combined to obtain a more accurate and stable forecast [62]. Gradient Boosting Regression (GBR): In the gradient boosting algorithm, the goal is to iteratively predict and improve the error rate. In this model, we tried to fit the loss function found in the previous step to a negative gradient vector. To reduce the error, the incremental reduction results are added to the function [63]. Histogram-based Gradient Boosting Regression (HGBR): Histogram-based gradient boosting is a combination of a gradient boosting machine and a histogram-based algorithm [64]. The histogrambased gradient boosting algorithm provides faster training of decision trees [65]. Voting: In the voting method, the class predictions made by various classifiers are voted on, and the class that obtains the most votes as a result of the voting becomes the class prediction of the majority [66]. Voting in regression is to combine different machine learning regressor models to return predicted average values. This is useful for balancing the individual weaknesses of the models [67]. In voting classification, more than one classification algorithm can be trained with the same training set, or a single algorithm can be trained with the same data set using different parameter values, and different classification models are created, and the final output value is produced by putting the outputs obtained from the models into the voting mechanism [68].
Stacking: The stacking ensemble learning method is based on the principle of generating a higher-performing prediction from these predictions by accepting the predictions of different types of classifiers as input for the metaclassifier [69]. The stacking ensemble learning method accepts the predictions of different types of classifiers as input for the metaclassifier, producing a higher-performing prediction from these predictions [70]. In stacking in classification, the output from different classifiers is passed as input to a metaclassifier for the final classification task [71].

K-Fold Cross Validation
The validation strategy for this study was selected as KFCV. KFCV helps determine a more accurate estimate of model prediction when compared with the train/test set validation strategy. This strategy helps in reducing the variance of the performance estimate and allows the use of more data for training. KFCV also avoids overfitting, as it exposes the model to different subsets of the data. In our implementation of KFCV, the number of folds (k) chosen is 10. In the 10-fold cross validation method, the dataset is divided into 10 equal parts. In each iteration, 9/10 of the dataset is used for training and the reserved 1/10 is used for testing. In each iteration, a different part of the data is reserved for testing. The basic structure of 10-fold cross validation is shown in Figure 8. The most important advantage of 10-fold cross validation is that all samples in the dataset are used in the training and testing phases. Thus, the positive or negative effects that some examples may have had during the education process are eliminated. Each sample is guaranteed to be used 9 times for training and 1 time for testing. As a result, 10-fold cross validation tests whether the accuracy value of the success of the model is random (or by chance) and provides a verified success rate. The most important advantage of 10-fold cross validation is that all samples in the dataset are used in the training and testing phases. Thus, the positive or negative effects that some examples may have had during the education process are eliminated. Each sample is guaranteed to be used 9 times for training and 1 time for testing. As a result, 10-fold cross validation tests whether the accuracy value of the success of the model is random (or by chance) and provides a verified success rate.

Model Performance Metrics
It is important to know the performance of the established models to understand which model performs better than the others. There are several metrics defined in the literature to evaluate the performance of regression algorithms/models. Some of these metrics are given in Table 4. In the formulas presented in Table 4, n represents the number of predictions; y i represents the actual value of the ith observation, x i represents the predicted value of the ith observation, and − x i represents the average of the predicted values. Table 4. Performance metrics used in this study.

Performance Metrics Description
Coefficient of determination The following metrics are used to evaluate the models in this study. Coefficient of Determination (R 2 ): This coefficient is a measure that measures how well a statistical model predicts an outcome. It measures the percentage of variability within the values of the dependent variable that can be explained by the regression model. A high R 2 indicates a good regression model fit. As the number of data increases, the reliability of R 2 also increases.
Mean Squared Error (MSE): MSE is a metric that measures the error rate of the prediction model as the square of the error. A small difference between the actual and predicted value indicates a good prediction. The closer the MSE value is to 0, the better the forecast is [73].
Root Mean Squared Error (RMSE): The RMSE can be found by taking the square root of MSE. Root mean squared error is the standard deviation of the difference between actual and predicted values. The closer the RMSE value obtained by the model is to zero, the closer the values predicted by the model will be to the real values. Since RMSE penalizes large errors more, it may be more appropriate in some situations. RMSE avoids the undesirable use of absolute values in many mathematical calculations [74].
Mean Absolute Error (MAE): Mean absolute error (MAE) is frequently used in regression problems because it is easy to interpret. The MAE is the measure of the difference between two continuous variables, i.e., the sum of the absolute values of the differences between the actual and predicted values [74]. An MAE value close to 0 indicates a successful analysis.

Results
In this study, predictions were made for cooling load. The inputs are the total floor area, aspect ratio, ceiling height, window material, external wall material, roof material, window wall ratio north faced, window wall ratio south faced, horizontal shading, and orientation of the building. The output is cooling load (Figure 9).
In this study, five foundational regression algorithms and five ensemble algorithms were used for the generation, training, and validation of the models. The Voting algorithm has been implemented with different combinations of base learners (lr: linear regression, knr: k-neighbor regression, ent: elastic net regression, dtr: decision tree regression, svr: support vector regression, rfr: random forest regression, gbr: gradient boosting regression, hgbr: a histogram-based gradient boosting regression). The Stacking algorithm has been implemented by utilizing different combinations of base learners and Gradient Boosting Regressor as the final estimator. The accurate metric values obtained as a result of the predictions made using each algorithm and each implementation of Voting and Stacking algorithm are provided in Table 5. In this study, five foundational regression algorithms and five ensemble algorithms were used for the generation, training, and validation of the models. The Voting algorithm has been implemented with different combinations of base learners (lr: linear regression, knr: k-neighbor regression, ent: elastic net regression, dtr: decision tree regression, svr: support vector regression, rfr: random forest regression, gbr: gradient boosting regression, hgbr: a histogram-based gradient boosting regression). The Stacking algorithm has been implemented by utilizing different combinations of base learners and Gradient Boosting Regressor as the final estimator. The accurate metric values obtained as a result of the predictions made using each algorithm and each implementation of Voting and Stacking algorithm are provided in Table 5.   As mentioned previously, the R 2 value approaching one indicates that the model has a high success rate. When the results of all models are compared, the lowest R 2 value (0.7341) was obtained in the SVR implementation, while the highest R 2 score (0.9949) was obtained with two models, (i) the histogram gradient boosting regression algorithm and (ii) in two of the stacking implementations, (a) the combination of random forest regression, gradient boosting regression and histogram-based gradient boosting regression with final estimator being the gradient boosting regression and (b) the combination on linear regression, decision tree regression, random forest regression, gradient boosting regression, and histogram-based gradient boosting regression with final estimator being the gradient boosting regression. Among the foundational models, the model with the highest R 2 score (0.9569) was the decision tree. The model with the lowest R 2 score (0.7341) is SVR (i.e., this was also the lowest overall R 2 score).
In ensemble methods, the training/validation with histogram gradient boosting algorithm and stacking resulted in the highest R 2 value (0.9949). Combinations of which the highest R 2 value was obtained in the stacking model, were the combinations in which multiple ensemble methods were used together as base learners. As the low error rate is an indication of better success in regression algorithms, among all models, the success of the histogram gradient boosting algorithm and stacking models can also be confirmed In ensemble methods, the stacking models have the lowest error rates. The combination of which the lowest MSE is obtained (NMSE = −8.93) was the Stacking model with the combination of random forest regression, gradient boosting regression, and histogrambased gradient boosting regression with the final estimator being the gradient boosting regression. The combination in which the lowest MAE is obtained (NMAE = −1.77) was the combination of linear regression, decision tree regression, random forest regression, gradient boosting regression, and histogram-based gradient boosting regression with the final estimator being the gradient boosting regression.
The analyses were performed with a PC equipped with IntelI CoreI I3-5005U CPU @ 2.00 GHz (4Cores). When the time performance of the models was evaluated, the SVR model was the model with the worst time performance (54 s for 10-fold cv process) among the foundational models and is the model with the lowest overall R 2 score. Among the models with the best performance, the time performance of the Histogram Gradient Boosting model (24 s for 10-fold cv process) was much better than the best-performing Stacking model (178 s for 10-fold cv process). A comparison of the models is shown in Table 6.

Discussion and Conclusions
Due to the rapid population growth throughout the world, the energy demand is increasing day by day. The built environment is one of the key consumers of energy. Especially for buildings in tropical climates, the energy load required for cooling is very high. Therefore, it is necessary to know the energy load required for the cooling of the building to develop building designs with a focus on energy efficiency.
In the study, various machine learning algorithms are implemented to predict the cooling load of a tropical building based on its architectural attributes such as floor area, aspect ratio, ceiling height, window material, external wall material, roof material, window wall ratio north faced, window wall ratio south faced, horizontal shading, and orientation. The main findings of this study are as follows: (1) Ensemble learning algorithms/models are superior to foundational algorithms models in the prediction of the cooling load of the building through regression. Among the ensemble models, stacking-based models were found to be most successful when compared to others. Ensemble models have been more successful (high R 2 , low error) than base models as they combine decisions from multiple models to improve their overall performance. (2) It is observed that Support Vector Regression was the least efficient model among all foundational and ensemble models, not only in terms of performance/accuracy but also in terms of time performance in the training/validation stages. (3) When only the foundational algorithms were compared, Decision Tree Regression was the model with the best performance. This indicates that Tree Based approaches can be efficient in the prediction of the cooling load of buildings based on their architectural properties. (4) In a similar study, Guo et al. [75] predicted heating and cooling loads based on light gradient boosting machine algorithms. Common models in our study and [75] are Random Forest and SVR. The same R 2 values were obtained for Random Forest in both studies, but SVR had a higher R 2 value in Guo et al. [75]. This indicates (a) that based on the nature of the dataset, SVR can also provide accurate results, so tests with SVR should not be neglected in studies for developing cooling load prediction models, (b) that Tree Based approaches and Ensemble models are very promising in cooling load prediction. (5) When the time performance of the models is taken into account, the Histogram Gradient Boosting algorithm appears as the optimal model, as it also provides a good prediction performance.
In summary, the results of the study have demonstrated that Ensemble Learning algorithms can be successfully used to establish the relationship between the architectural properties of tropical buildings and their cooling load because ensemble methods come to a conclusion by using more than one predictor in the same prediction task. In this method, the results of predictors with different metric scores are combined with different methods (voting, stacking, etc.). Thus, more successful performance is achieved. Furthermore, the cooling load of tropical buildings can be accurately predicted through the use of ensemble learning algorithms. Future research will focus on how hyperparameter optimization would enhance the performance of the provided models. The accuracy of the prediction model provided in this paper can be further enhanced through the addition of other predictor variables such as the occupancy status of the rooms, occupancy schedule, space usage conditions, and characteristics.