Next Article in Journal
Combined Application of Nitrogen and Phosphorus Promotes the Growth and Nutrient Accumulations of Cinnamomum camphora Container Seedlings
Next Article in Special Issue
Prediction Model of Flavonoids Content in Ancient Tree Sun−Dried Green Tea under Abiotic Stress Based on LASSO−Cox
Previous Article in Journal
Advances in Molecular Breeding of Forage Crops: Technologies, Applications and Prospects
Previous Article in Special Issue
Intelligent Cotton Pest and Disease Detection: Edge Computing Solutions with Transformer Technology and Knowledge Graphs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automated Model Selection Using Bayesian Optimization and the Asynchronous Successive Halving Algorithm for Predicting Daily Minimum and Maximum Temperatures

by
Dilip Kumar Roy
1,*,
Mohamed Anower Hossain
1,
Mohamed Panjarul Haque
1,
Abed Alataway
2,
Ahmed Z. Dewidar
2 and
Mohamed A. Mattar
2,*
1
Irrigation and Water Management Division, Bangladesh Agricultural Research Institute, Gazipur 1701, Bangladesh
2
Prince Sultan Bin Abdulaziz International Prize for Water Chair, Prince Sultan Institute for Environmental, Water and Desert Research, King Saud University, Riyadh 11451, Saudi Arabia
*
Authors to whom correspondence should be addressed.
Agriculture 2024, 14(2), 278; https://doi.org/10.3390/agriculture14020278
Submission received: 3 November 2023 / Revised: 15 January 2024 / Accepted: 31 January 2024 / Published: 8 February 2024
(This article belongs to the Special Issue Application of Machine Learning and Data Analysis in Agriculture)

Abstract

:
This study addresses the crucial role of temperature forecasting, particularly in agricultural contexts, where daily maximum ( T m a x ) and minimum ( T m i n ) temperatures significantly impact crop growth and irrigation planning. While machine learning (ML) models offer a promising avenue for temperature forecasts, the challenge lies in efficiently training multiple models and optimizing their parameters. This research addresses a research gap by proposing advanced ML algorithms for multi-step-ahead T m a x and T m i n forecasting across various weather stations in Bangladesh. The study employs Bayesian optimization and the asynchronous successive halving algorithm (ASHA) to automatically select top-performing ML models by tuning hyperparameters. While both the Bayesian and ASHA optimizations yield satisfactory results, ASHA requires less computational time for convergence. Notably, different top-performing models emerge for T m a x and T m i n across various forecast horizons. The evaluation metrics on the test dataset confirm higher accuracy, efficiency coefficients, and agreement indices, along with lower error values for both T m a x and T m i n forecasts at different weather stations. Notably, the forecasting accuracy decreases with longer horizons, emphasizing the superiority of one-step-ahead predictions. The automated model selection approach using Bayesian and ASHA optimization algorithms proves promising for enhancing the precision of multi-step-ahead temperature forecasting, with potential applications in diverse geographical locations.

1. Introduction

The temperature variable stands out as one of the atmospheric parameters with the highest accuracy, contributing to enhanced reliability in weather forecasts. Precisely predicting the air temperature at a specific location and time is a crucial research challenge with diverse applications, spanning from energy generation to agriculture. Climate scientists anticipate that the escalating air temperatures in the forthcoming decades may lead to adverse environmental impacts [1]. The demand for temperature forecasts in the agricultural sector is growing for both maximum ( T m a x ) and minimum ( T m i n ) temperatures, as these significantly influence crop growth and potential yield. It is a key research topic in atmospheric sciences, with potential applications to the agricultural sector [2]. Daily T m a x and T m i n serve as valuable indicators of crop growth and yield, as they can be used to predict the irrigation water requirements for crops. Consequently, predicting the daily T m a x is a significant issue, with practical applications in crop science, as it has been demonstrated that irrigation water requirements depend greatly on weather conditions. According to Allen et al. [3] and Ali [4], irrigation water requirements are closely linked to weather parameters such as temperature, relative humidity, sunshine hours, wind speed, and rainfall. Among these, temperature stands out as a key factor, influencing not only the determination of irrigation water requirements but also facilitating plant growth through the photosynthesis process. Notably, both maximum and minimum temperatures play a crucial role in shaping irrigation water requirements and crop scheduling, as expounded by Wang et al. [5] and Haque et al. [6]. The connection between temperature forecasting and agriculture is profound, as temperature significantly influences various aspects of crop growth and development [7,8]. Temperature forecasts play a crucial role in agriculture by providing valuable information for planning and decision making across various agricultural activities. Temperature forecasts offer essential insights into various aspects of agriculture, including crop growth, pest management, irrigation, harvest timing, climate change adaptation, and resource management [9]. By incorporating accurate temperature forecasts into agricultural practices, farmers can make informed decisions, optimize resource allocation, enhance productivity, and mitigate risks, ultimately contributing to the development of sustainable and efficient agricultural systems. Therefore, the present study aims to provide multi-step-ahead forecasts for both T m a x and T m i n in three distinct climatic zones of Bangladesh.
The precise prediction of air temperatures has attracted the attention of researchers in recent years. This is because precise temperature forecasting has a wide range of applications in fields such as climate science, agriculture, energy management, and urban planning [10]. Temperature forecasting has been continuously evolving for nearly a century, since the inception of weather forecasting. According to the literature, there are two fundamental methods for predicting weather forecasting, including air temperature and precipitation forecasting: general circulation or physically based simulations and statistical modeling [11,12,13,14]. Physically based models are classic approaches that utilize computer simulations based on mathematical equations and are often referred to as numerical weather prediction models [15,16]. However, physically based models are constrained by the need for significant computing power and a clear understanding of the system being modeled. On the other hand, statistical models aim to reduce the reliance on physically based models. They are easier to understand and less computationally complex than their physically based counterparts. Typically, statistical models are applied to the outputs of numerical weather prediction models. Most studies have demonstrated that results from statistical models and physical models are generally consistent. There are two types of statistical analysis [17]: correlation techniques and regression approaches.
Regression approaches primarily rely on machine learning (ML)-based data-driven methodologies, which have gained popularity in the prediction and forecasting of air temperatures [1,18,19,20]. However, it is observed that in the absence of transient weather systems, the daily cycle of temperature is more or less well defined. Therefore, a classic regression model could often be utilized to forecast the air temperature when there is no cloud cover during the data acquisition process. Nevertheless, accurately predicting the air temperature using classic regression methods poses a challenge, given the chaotic nature and nonlinear trends of weather parameters. In such scenarios, ML-based methods have proven to be viable alternatives to classic regression models. In recent years, various soft computing approaches have been applied to address temperature prediction challenges in diverse areas. Many of these approaches have harnessed the power of neural computing techniques, known for their speed and accuracy [10]. Specifically, ML-based approaches to air temperature prediction involve the application of various methods, including Artificial Neural Network (ANN) [21], genetic algorithm-tuned ANN [22], Honey Badger Algorithm-tuned ANN [23], Gene Expression Programming [23], Support Vector Regression [14,17,21,24,25], Multi-Layer Perceptron [1,14], Multi-Variate Adaptive Regression Spline [26], Extreme Learning Machine [26,27], M5 Prime [28], Random Forest [17,26,29,30], Lasso Regression [29], Regression Tree [17], Long Short-Term Memory Network (LSTM) [1,31], GRU-LSTM [32], Convolutional Neural Network (CNN) [29], CNN-LSTM [1,33,34], Simple Recurrent Neural Network with Convolutional Filters [35], and Stochastic Adversarial Video Prediction [35]. Cifuentes et al. [18] provided a detailed review of air temperature forecasting approaches using ML techniques. These forecasting methods have consistently demonstrated improved prediction results.
Implementing new techniques in temperature forecasting is crucial for reducing modeling errors and mitigating model parameter uncertainties [36]. However, achieving higher forecast accuracy with a desired model remains a complex scientific challenge. In this article, we are not referring to the relationship between NWP and ML models; rather, we propose a new approach based on ML-based modeling to address the inherent limitations of NWPs, such as the need for well-defined prior knowledge and extensive computational capacity. In contrast, ML techniques excel at identifying hidden patterns in the dataset without requiring prior knowledge. Thus, these approaches may serve as suitable alternatives to NWPs in weather forecasting. Previous studies on temperature forecasting employed various ML approaches and optimization algorithm-tuned ML models, including the deep learning approaches. These studies compared a few modeling approaches and proposed the best predictive model based on the comparison results. However, this approach is limited by the need to select appropriate candidate models for comparison and to identify the top-performing model. In other words, traditional ML-based approaches to temperature forecasting involve manual model selection, which can be time-consuming and subjective. To address these limitations, it is often beneficial to compare multiple approaches while optimizing their tunable hyperparameters using optimization algorithms to identify the most suitable prediction or forecast model for a given dataset. This process of automatic model selection involves automatically choosing the most appropriate regression model for a given dataset. The objective is to find the model that best fits the data and provides the most accurate predictions. To this end, the present study proposes an automated model selection technique to enhance forecasting accuracy and streamline the modeling process. Bayesian optimization [37] and the asynchronous successive halving algorithm (ASHA) [38] were employed to search for the top-performing model by tuning hyperparameters.
The selection of significant input variables is a critical step in ML-based modeling applications. It involves the identification and selection of the most relevant and informative features from the available dataset [39,40]. This process is fundamental in ML-based forecasting models, as it plays a vital role in improving predictive performance, reducing overfitting, enhancing computational efficiency, and improving the interpretability of ML-based models. It enables models to leverage the most relevant and informative features, leading to more accurate, efficient, and interpretable predictions in various applications and domains. Previous studies have utilized both linear methods [41] and nonlinear techniques [40], including Minimum Redundancy Maximum Relevance (MRMR) [42] approaches, to identify the most significant input variables for forecast models. One promising approach to identifying the most influential input variables is the utilization of F-tests. F-tests are a common approach to assessing the importance of individual features in predicting a continuous target variable. In this approach, each feature is evaluated independently based on its relationship with the target variable, using the F-statistic and associated p-value. It is important to note that while univariate feature ranking provides insights into an individual feature’s importance, it does not capture potential interactions or dependencies between features. Therefore, it is essential to complement this analysis with other feature selection or dimensionality reduction techniques to consider the combined effects of multiple features and capture complex relationships in the regression model. Another approach to significant input variable selection is the use of MRMR, a popular technique for selecting a subset of features that are both informative and minimally redundant. MRMR aims to maximize the relevance of features to the target variable while minimizing the redundancy between the selected features. Neighborhood Component Analysis (NCA) is another feature selection technique that aims to find an optimal subset of features for a given classification or regression task. NCA is a distance-based feature selection method that learns a linear transformation of the original feature space to maximize the discriminability of the data points. In the present study, a combination of all variables that were selected by the individual variable selection approaches was used to include all possible contributing variables affecting the outputs. In this proposed approach, the common variables that were determined by all approaches were used only once.
Traditional machine learning (ML) approaches for temperature prediction typically entail manual model selection, a process that is known to be time-consuming and subjective. Addressing these challenges, it proves advantageous to compare multiple approaches and optimize their tunable hyperparameters through optimization algorithms. This helps identify the most suitable prediction or forecast model for a given dataset. This research introduces an innovative approach to automated model selection by employing Bayesian optimization and the asynchronous successive halving algorithm, providing a systematic and efficient method for choosing the most suitable models in predicting daily minimum and maximum temperatures. The incorporation of the ASHA algorithm contributes to enhancing the efficiency of the model selection process, particularly in scenarios where computational resources are limited or asynchronous evaluations are necessary. The research addresses the challenge of model selection in a robust and adaptive manner through optimizing model hyperparameters, providing a valuable contribution to the broader field of machine learning and climate science. Finally, by emphasizing the application of advanced optimization techniques in the field of temperature prediction, the study contributes to a broader understanding of automated model selection strategies in environmental forecasting, potentially paving the way for similar methodologies in related domains. To the best of the authors’ knowledge, there have been no prior attempts to predict and forecast temperatures using optimization algorithms such as Bayesian optimization and ASHA for automatic model selection. This underscores the novel contribution of the current study.
The agricultural sector stands to benefit significantly from the evolution of machine learning (ML) models, especially with a focus on enhancing computational efficiency. The advancement of ML models, particularly in the context of agriculture, has led to a growing interest in developing less computationally expensive models to their enhance scalability and accessibility [37]. Bayesian optimization emerges as a promising approach in this regard, offering a systematic method for optimizing the hyperparameters of ML models with a reduced computational burden. Utilizing Bayesian optimization empowers researchers to refine temperature models with increased effectiveness, thereby facilitating advancements in precision agriculture and the accuracy of climate-related predictions. On the other hand, the ASHA optimization algorithm further complements this effort by parallelizing the model training process, thereby expediting the optimization procedure [38]. This strategy offers an effective means of optimizing hyperparameters, streamlining the training process, and making models more accessible in the context of agriculture. In agricultural applications, where the need for real-time decision making is crucial, the implementation of less computationally intensive ML models can significantly improve efficiency [43]. The incorporation of Bayesian and ASHA optimization techniques contributes to the sustainable evolution of precision agriculture and climate-related predictions.
This study aims to train several regression models using Bayesian and ASHA optimizations on a given training dataset and identify the best-performing model on a test dataset. The objective is to investigate the effectiveness of the Bayesian and ASHA optimization algorithms in forecasting daily T m a x and T m i n values and to provide a comparison of these two optimization algorithm-tuned models. By automating the model selection process, we aim to overcome the limitations of manual selection, such as subjectivity and suboptimal choices. Our proposed approach automates and eliminates manual steps that are required to go from a dataset to a predictive model. We present a novel approach that automates the model selection process for temperature forecasting, offering a more objective and efficient alternative to manual selection. Therefore, the contributions of this study encompass (a) building multiple regression models for a given training dataset of T m a x and T m i n by optimizing their hyperparameters using Bayesian and ASHA optimization algorithms, (b) performing a comparative analysis of models tuned by Bayesian and ASHA optimization algorithms, and (c) identifying the top-performing models for multiple forecast horizons at three weather stations.

2. Materials and Methods

2.1. Study Area and the Data

Daily T m a x and T m i n data were collected from three meteorological stations, namely, Barishal, Gazipur, and Ishurdi stations. The selection of these stations was based on their representation of three distinct climatic regions in Bangladesh: (1) Gazipur station represents the central region of Bangladesh, situated at approximately 24.00° N latitude and 90.43° E longitude, with an elevation of 14 m above mean sea level; (2) Ishurdi station represents the northern climatic regions of Bangladesh, located at around 24.04° N latitude and 90.07° E longitude, with an elevation of 18 m above mean sea level; and (3) Barishal station represents the southern part of Bangladesh, positioned at approximately 22.60° N latitude and 90.36° E longitude, with an elevation of 1.0 m above mean sea level. The study area, including the locations of these three weather stations, is depicted in Figure 1.
This study utilizes medium-term daily temperature data obtained from the Bangladesh Meteorological Department (BMD) across three weather stations to provide multi-step-ahead temperature forecasts. The maximum and minimum temperatures of the day were measured using Zeal P1000 maximum and minimum thermometers from G. H. Zeal Ltd., London, UK. The thermometers have an accuracy of ±0.2 °C and a range and resolution of −50 to +70  °C with 0.1  °C increments, and they were positioned at a measurement height of 2 m. The geographical distribution of meteorological stations is illustrated in Figure 1, depicting a reasonable coverage of three distinct regions across the country. Notably, Bangladesh’s topography is predominantly flat, with some elevated regions in the northeast and southeast. Consequently, it is reasonable to infer that the selected meteorological stations comprehensively represent the climatic conditions of the three regions of the country [44].
The temperature distribution in Bangladesh is significantly influenced by various local conditions. The country’s proximity to the equator contributes to a predominantly tropical climate, characterized by high temperatures throughout the year. The Bay of Bengal, bordering the southern coastline, acts as a key influencer, moderating temperatures along the coastal regions compared to the inland areas. The flat topography, interspersed with some upland regions, further contributes to temperature variations. Seasonal monsoons, driven by distinct wind patterns, play a crucial role in shaping temperature dynamics and precipitation patterns, thus exerting a notable impact on the overall temperature distribution across different geographical regions of Bangladesh [5,6].
Gazipur station exhibits a tropical wet and dry or savanna climate (classification: Aw). The district records an annual average temperature of 28.95 °C, marking a 1.21% increase compared to the national averages in Bangladesh. Gazipur usually receives around 71.24 mm of precipitation annually. Ishurdi station also features a tropical wet and dry or savanna climate (classification: Aw). The district maintains an annual temperature of 29.52 °C, which is 1.78% higher than the national averages in Bangladesh. Ishurdi typically receives around 98.38 mm of precipitation annually. Barisal station experiences a tropical climate, characterized by significantly less rainfall in winter compared to the summer months. According to Köppen and Geiger [45], this location falls under the Aw classification. Statistical analysis reveals an average temperature of approximately 25.6 °C, with an annual rainfall of around 2005 mm. The temperate characteristics of Barisal poses challenges in clearly categorizing distinct seasons in the region.
The acquired temperature forecast results remain unaffected by local atmospheric conditions. While the mentioned local factors are implicitly embedded in the data collected from weather stations, it is crucial to note that the modeling outcomes presented in this research solely derive from historical temperature data from BMD.
A diurnal cycle, also known as a diel cycle, manifests as a recurring pattern every 24 h due to the complete rotation of the Earth around its axis. The Earth’s rotation gives rise to temperature variations on the surface during both day and night, contributing to seasonal weather changes. The primary determinant of the diurnal cycle is the influx of solar radiation [46]. The atmospheric seasonal cycle is influenced by the Earth’s axial tilt. The Earth’s seasonal cycle arises from its 23° axial tilt, causing varying solar radiation at different latitudes throughout the year. Equinoxes align the sun with the equator, the June solstice with the Tropic of Cancer, and the December solstice with the Tropic of Capricorn, creating hemispheric temperature disparities in summer and winter. The Annual Temperature Cycle (ATC) encompasses seasonal temperature changes that are influenced by fluctuations in solar radiation reaching the Earth’s surface throughout the year [47]. Typically, evaluating the ATC relies on sparse and unevenly distributed air temperature observations or numerical model simulations. This study utilized temperature data collected by the Bangladesh Meteorological Department (BMD) at specific intervals, encompassing daily maximum and minimum values. The data were sourced from three designated weather stations for analysis.
In this study, the daily T m a x and T m i n data were collected from three weather stations, located in distinct climatic regions of Bangladesh. Ensuring the quality of the temperature datasets is essential to enhance the reliability of temperature forecasts using ML algorithms [48]. While a comprehensive quality assurance process was not conducted for this specific dataset, the accuracy and completeness of the recorded temperature data were systematically assessed using range/limit tests. Range testing is a fundamental quality control method that involves verifying that every observation falls within a specified range [48]. Only values within the predefined limits are considered valid [49,50], while readings outside the specified range are appropriately marked as invalid. The valid readings within the allowable range were used to simulate future temperature fluctuations in the selected weather stations, with a particular focus on generating multi-step forward temperature forecasts for both T m a x and T m i n .
A small portion of the collected data, amounting to less than 2% of the total data from all weather stations, contained missing values. To address this issue, the ‘moving median’ imputation technique was employed. This technique utilizes a moving median with a predetermined window length to fill in the missing values. After applying the imputation method, the weather stations Barishal, Gazipur, and Ishurdi had 2677 readings (from 1 January 2015 to 30 April 2022), 6695 readings (from 1 January 2004 to 30 April 2022), and 2041 readings (from 1 June 2015 to 31 December 2020) of daily T m a x and T m i n values.
Table 1 presents the descriptive statistics for daily T m a x and T m i n at the three weather stations. It can be seen from Table 1 that the data exhibited left (negative) skewness, suggesting that the distribution had an extended left tail compared to the right tail. Additionally, kurtosis values included both positive and negative values, indicating that the datasets had both ‘heavy-tailed’ (positive kurtosis) and ‘light-tailed’ (negative kurtosis) distributions.
The Gaussian distribution, commonly known as a normal distribution, is characterized by a bell-shaped curve. It is often assumed that measurements will adhere to a normal distribution, featuring an equal number of measurements above and below the mean. However, in real-world scenarios, data may not precisely conform to the Gaussian distribution and could exhibit slight deviations. In practice, achieving exact conformity to a Gaussian distribution is rare. Ideally, if a distribution is truly normal, the mean, median, and mode values would be identical—an occurrence seldom observed in real-world situations. In instances where the values of the mean, median, and mode differ, the distribution is considered skewed and deviates from the Gaussian norm [51]. The presented data in Table 1 reveal that the mean, median, and mode values of the temperature measurements exhibit slight differences. These numerical variations suggest that the data do not perfectly adhere to a Gaussian distribution and show indications of being slightly skewed. The specific numeric values can be found in Table 1 for reference. Skewness quantifies the asymmetry that is present in data relative to their sample mean. Negative skewness indicates that the data are more dispersed to the left of the mean, while positive skewness suggests greater dispersion to the right. A perfectly symmetric distribution, such as the normal distribution, has a skewness of zero. Kurtosis serves as a metric for gauging the susceptibility of a distribution to outliers. The normal distribution exhibits a kurtosis of 3. Distributions with kurtosis values exceeding 3 are more prone to outliers than the normal distribution, while those with values below 3 are less susceptible. Some definitions of kurtosis involve subtracting 3 from the calculated value, resulting in a kurtosis of 0 for the normal distribution. This study adopts this later definition (subtracting 3 from the calculated value) to compute the kurtosis value.

2.2. Data Preprocessing for the Lagged Input and Output Variables

Time-lagged information was gathered from the collected time series of the daily T m a x and T m i n values. The outputs from the models were the day-ahead T m a x and T m i n values. Therefore, the inputs to the models (five models for five-days-ahead forecasts) were
T d ,   d 1 , d 2 , d 3 , , d n  
The outputs were
T d + 1 ,   d + 2 , d + 3 , , d + n  
Due to time-lagging of input variables and the target, the observed daily T m a x and T m i n records were reduced at each weather station. At Barishal station, a total of 2642 historical records remained (from 1 January 2015 to 26 March 2022) after removing 35 records due to time-lagging (5-time lag forward and 30-time lag backward) from the entire time series of 2677 readings (from 1 January 2015 to 30 April 2022). At Gazipur station, a total of 6660 historical records remained (from 1 January 2004 to 26 March 2022) after removing 35 records due to time-lagging (5-time lag forward and 30-time lag backward) from the entire time series of 6695 readings (from 1 January 2004 to 30 April 2022). At Ishurdi station, a total of 2006 historical records remained (from 1 June 2015 to 26 November 2020) after removing 35 records due to time-lagging (5-time lag forward and 30-time lag backward) from the entire time series of 2041 readings (from 1 June 2015 to 31 December 2020). Each station’s remaining dataset was divided into two sets: 80% for model training and 20% for model testing. While there is no established rule for data splitting during model learning and testing [52], it is recommended that testing data comprise between 10% and 40% of the total dataset size [53].

2.3. Input Variable Selection

The first step in developing forecast models using ML-based approaches is the selection of the most significant input variables [39,40]. For the purpose of choosing input variables in hydrology and water resource modeling [39], both linear methods [41] and nonlinear techniques [40] have been employed. However, because hydrology and water resource modeling issues are frequently nonlinear in nature [54], linear methods based on Partial Autocorrelation Function (PACF) and Autocorrelation Function (ACF) are often less suitable techniques. For modeling of hydrological and water resources as well as other fields of science and engineering applications, nonlinear approaches that utilize mutual information (MI) [55] typically outperform linear techniques [35,40,56].
Since the only data used in this effort are the daily T m a x and T m i n values at three weather stations (Barishal, Gazipur, and Ishurdi), the time-lagged versions of the acquired temperature data (from each station) were used as potential inputs. To extract the time-lagged information from the T m a x and T m i n time series and choose which lags to include as prospective inputs, the PACF was employed. The PACF is a technique that is commonly used in time series analysis to identify the most influential lagged features for predicting a target variable [57]. The PACF measures the correlation between a time series variable and its lagged values, while accounting for the influence of intermediate lags. By applying PACF, one can identify the lagged features that have the most significant impact on the target variable. These lagged features capture the historical patterns and dependencies that can be exploited for accurate time series forecasting. Figure 2 displays the PACF plots for the data from three weather stations, revealing that the current and past 30-time lags are essential for forecasting temperatures for the next five days ( T d + 1 ,   T d + 2 ,   T d + 3 ,   T d + 4 ,   a n d   T d + 5 ). The PACF provided an initial guess of the candidates for input variables. However, this initial selection of input variables may include unnecessary or redundant features, which could hinder the training of ML-based forecasting models. Therefore, it is important to determine the most significant input variables to ensure proper training and computational efficiency.
Next, this study used the F-statistic [58,59] to identify the most influential input variables. The F-test is a statistical test that assesses how much the variability in the target variable is explained by a specific feature compared to the variability that is not explained by that feature. It examines the importance of each predictor individually using an F-test, testing the hypothesis that the response values that are grouped by predictor variable values are drawn from populations with the same mean against the alternative hypothesis that the population means are not all the same. A small p-value of the F-test indicates the importance of the corresponding predictor. The F-statistic is calculated by dividing the mean square regression (MSR), representing explained variability, by the mean square error (MSE), representing unexplained variability. Univariate feature ranking with F-tests helps identify the most influential features for a regression task, emphasizing the most relevant variables and potentially enhancing interpretability and model performance [58,59].
The variable importance is also determined using the MRMR technique [60,61]. The MRMR algorithm aims to find a balance between selecting informative features while avoiding redundant ones. MRMR considers both relevance and redundancy when identifying a subset of features that collectively offer maximum information while minimizing duplication. There are variations and extensions of MRMR, such as weighted MRMR or incremental MRMR, which provide additional flexibility and adaptability to different scenarios. Overall, MRMR is a valuable feature selection technique that enhances the interpretability, generalization, and efficiency of machine learning models by focusing on the most relevant and nonredundant features. The MRMR algorithm [61] identifies a best-possible set of characteristics that are maximally and mutually different and are useful for representing the response variable. The approach maximizes a feature set’s relevance to the response variable while minimizing its redundancy. Using pairwise mutual information of attributes and mutual information of an attribute and the response, the technique measures the degree of duplication as well as the relevance of variables.
The MRMR technique seeks to identify an ideal set S of features that maximizes V s , the relevance of S pertaining to a response variable y , and minimizes W s , the redundancy of S , where V s and W s are defined with mutual information (MI) I :
V s = 1 S x S I x , y ,
W s = 1 S 2 x , z S I x , z .
where S denotes the quantity of attributes in S . The amount of uncertainty in one variable that can be lowered by understanding the other variable is measured by the MI between two variables.
The MI ( I ) of the discrete random variables X and Z can be represented by [60]
I X , Z = i , j P X = x i , Z = z i log P X = x i , Z = z i P X = x i P Z = z i .
If X and Z are independent, then I equals 0. If X and Z are the same random variable, then I equals the entropy of X .
It is necessary to take into account all 2 Ω combinations in order to find the optimum set S , where Ω is the set of all features. The MRMR approach utilizes an alternative approach. In this approach, the MRMR approach scores attributes by employing the forward addition technique, which necessitates O ( | Ω | · | S | ) computations, by utilizing the MI quotient (MIQ) value.
M I Q x = V x W x
where V x and W x are the relevance and redundancy of a feature, respectively:
V x = I x , y
W x = 1 S z S I x , z .
The MRMR function uses a heuristic technique to quantify the significance of a characteristic and then generates a score. A high score value denotes the significance of the associated predictor. A degree of trust in feature selection is also indicated by a decline in the feature significance score. The score value of the second most essential attribute, for instance, is significantly lower than the score value of x if the algorithm is confident in choosing it. The results can be used to identify an ideal set S for a particular collection of features.
The features were ranked using the MRMR algorithm according to the following steps:
Step 1: Select the feature with the largest relevance, max x Ω V x . Add the selected feature to an empty set S .
Step 2: Find the features with nonzero relevance and zero redundancy in the complement of S , S c .
  • If S c does not include a feature with nonzero relevance and zero redundancy, go to step 4.
  • Otherwise, select the feature with the largest relevance, max x S c , W x = 0 V x . Add the selected feature to the set S .
Step 3: Repeat Step 2 until the redundancy is not zero for all features in S c .
Step 4: Select the feature that has the largest MIQ value with nonzero relevance and nonzero redundancy in S c , and add the selected feature to the set S .
max x S c M I Q x = max x S c I x , y 1 S z S I x , z .
Step 5: Repeat Step 4 until the relevance is zero for all features in S c .
Step 6: Add the features with zero relevance to S in random order.
Feature selection using Neighborhood Component Analysis (NCA) [62] focuses on learning a transformation matrix that preserves the discriminative information in the data while reducing the dimensionality. By considering the local neighborhood relationships between data points, NCA can identify the most informative features for the task at hand. It is important to note that NCA assumes linearity in the data and may not capture complex nonlinear relationships. Therefore, it is advisable to combine NCA with nonlinear dimensionality reduction techniques or explore other feature selection methods for capturing nonlinear feature interactions if needed. The NCA feature selection for regression can be mathematically represented as follows [62]:
Given n observations S = x i , y i ,   i = 1,2 , , n , where the response values y i R are continuous, the aim is to predict the response y given the training set S .
Consider a randomized regression model that:
  • Randomly picks a point ( R e f ( x ) ) from S as the ‘reference point’ for x ;
  • Sets the response value at x equal to the response value of the reference point R e f ( x ) .
Again, the probability P R e f x = x j | S that point x j is picked from S as the reference point for x is
P R e f x = x j | S = k d w x , x j j = 1 n k d w x , x j
Now consider the leave-one-out application of this randomized regression model, that is, predicting the response for x i using the data in S 1 and the training set S excluding the point x i , y i . The probability that point x j is picked as the reference point for x i is
P i j = P R e f x i = x j | S 1 = k d w x i , x j j = 1 , j i n k d w x i , x j
Let y i ^ be the response value that the randomized regression model predicts and y i be the actual response for x i . And let l = R 2 R be a loss function that measures the disagreement between y i ^ and y i . Then, the average value of l y i , y i ^ is
l i = E l y i , y i ^ | S i = j = 1 , j i n P i j l y i , y j .
After adding the regularization term, the objective function for minimization is
f w = 1 n i = 1 n l i + λ r = 1 p w r 2 .
The default loss function l y i , y j for NCA for regression is mean absolute deviation.
In this study, a combination of variable selection methods (F-tests, MRMR, and NCA) was used to select the most significant input variables from an initial pool of 30 candidate inputs determined by PACF. To include all possible contributing variables affecting the outputs, all variables selected by the individual variable selection approaches were considered, while the common variables determined by all approaches were used only once. Doing so eliminates the possibility of excluding any variables which might have been deemed important to forecast the output. Using this combinatory variable selection scheme, the possible number of candidate variables of the inputs for one-, two-, three-, four-, and five-days-ahead forecast of T m a x at Barishal station were 26, 27, 26, 22, and 25, respectively. The number of candidate variables of the inputs for one-, two-, three-, four-, and five-days-ahead forecast of T m i n at Barishal station were 21, 22, 24, 24, and 23, respectively. Similarly, the possible number of candidate input variables for one-, two-, three-, four-, and five-days-ahead forecast of T m a x at Gazipur station were 25, 26, 25, 24, and 25, respectively. The number of candidate input variables for one-, two-, three-, four-, and five-days-ahead forecast of T m i n at Gazipur station were 26, 25, 26, 25, and 25, respectively. Likewise, the possible number of candidate variables of the inputs for one-, two-, three-, four-, and five-days-ahead forecast of T m a x at Ishurdi station were 23, 23, 26, 26, and 26, respectively. The number of candidate variables of the inputs for one-, two-, three-, four-, and five-days-ahead forecast of T m i n at Ishurdi station were 24, 28, 21, 23, and 21, respectively.

2.4. Model Development

The accuracy and robustness of any forecasting model largely depend on the right selection of models and their forecasting accuracy. Forecasting models are developed using training and testing of the state-of-the-art machine learning algorithms, the optimal parameters of which were decided here based on parameters tuning using the Bayesian and ASHA optimization algorithms. Training several models and finding their optimal parameter sets through hyperparameter tuning can often be a challenging and time-consuming task. This task was made easier and faster by developing and comparing multiple models automatically through tuning their hyperparameters using optimization algorithms. In this approach, instead of training each model with different sets of hyperparameters, we selected a few different models and tuned their default hyperparameters using Bayesian and ASHA optimizations. These optimization algorithms search for an optimal set of hyperparameters for a particular model by minimizing the objective function of the model (minimization of the mean squared error (MSE)). The optimization algorithms deliberately selected new hyperparameters in each iteration, produced an optimal set of hyperparameters for a given training dataset, and identified the model that performed best on a test dataset. With the Bayesian and ASHA optimization algorithms, the function randomly selected several models with various hyperparameter values and trained them on a small subset of the training data. If the log (1 + valLoss) value for a particular model was found promising, where valLoss is the cross-validation MSE, the model was promoted and trained on a larger amount of the training data. This process was repeated, and successful models were trained on progressively larger amounts of data. To reiterate, our proposed approach executed the following three steps simultaneously in the model development process:
  • Data exploration and preprocessing: Identify variables with low predictive power that should be eliminated.
  • Feature extraction and selection: Extract features automatically and—among a large feature set—identify those with high predictive power.
  • Model selection and tuning: Automatically tune model hyperparameters and identify the best-performing model.
The flow diagram of the entire model building process can be illustrated as Figure 3.

2.5. Hyperparameter Optimization

The majority of ML algorithms require a careful selection of hyperparameters [63]. The choice of hyperparameter settings significantly impacts the performance of ML models [64], and unconscious hyperparameter selection can result in low-performing models. Many studies employed trial-and-error selection of hyperparameters [64,65], grid search, and/or random search [66], while some employed heuristic optimization algorithms like particle swarm optimization and genetic algorithms [66]. Nevertheless, a precise and effective automated hyperparameter optimization method is highly desirable [64] and is essential for ensuring a fair comparison across ML alternatives. Furthermore, when evaluating various ML models, fair assessments can only be made if they are equally optimized (or receive the same level of attention) for the specific task at hand.
Bayesian optimization and asynchronous successive halving algorithm (ASHA) are advanced optimization techniques that are employed to automate and enhance the efficiency of model selection processes. In Bayesian optimization, a probabilistic model is iteratively updated to capture the relationship between model hyperparameters and performance, guiding the search toward promising regions of the parameter space. This process allows for intelligent and resource-efficient exploration, especially in scenarios with limited computational resources. On the other hand, ASHA introduces an asynchronous approach to hyperparameter optimization, enabling parallel evaluations and efficient resource utilization. By continually pruning underperforming models, ASHA converges to optimal hyperparameter configurations. ASHA is particularly beneficial in scenarios with limited computational resources or asynchronous evaluations. It employs a successive halving strategy to efficiently allocate resources to promising configurations, eliminating less favorable ones. Both Bayesian optimization and ASHA contribute significantly to automating the selection of suitable models, particularly in the context of temperature forecasting, where manual selection can be time-consuming and subjective. The combination of these algorithms in this study represents a novel and valuable contribution to the field, showcasing their effectiveness in optimizing hyperparameters and improving the overall accuracy of temperature prediction models.
A state-of-the-art approach for both global and local hyperparameter optimization is known as Bayesian optimization (BO) [37]. BO has been shown to outperform alternative methods like grid search and random search on various challenging optimization benchmarks [66]. In the quest for discovering the best hyperparameters, BO often surpasses the abilities of domain experts [67]. BO is versatile and applicable to a wide range of problem scenarios, accommodating both integer and real-valued hyperparameters. BO relies on the selection of a prior function and an acquisition function. The acquisition function, typically employing Expected Improvement (EI), works in conjunction with a Gaussian process prior [37]. The choice of covariance function, particularly the use of the Matern52 kernel, is crucial In determining the effectiveness of Gaussian processes [37]. With respect to x in a bounded domain, the BO algorithm process seeks to minimize a scalar objective function, f ( x ) . Depending on whether the function is stochastic or deterministic, it may yield different results when evaluated at the same point x . The variable x can have continuous real values, integers, or categorical components, referring to a discrete set of names. The key components in the minimization process include [68]:
  • A Gaussian process model of f ( x ).
  • A Bayesian update procedure for modifying the Gaussian process model at each new evaluation of f ( x ) .
  • An acquisition function a ( x ) (based on the Gaussian process model of f ) that is maximized to determine the next point x for evaluation.
The ASHA [38] optimization algorithm is an upgraded variant of the successive halving algorithm (SHA) [69,70]. ASHA is a user-friendly hyperparameter optimization technique that leverages aggressive early stopping and is particularly suited for tackling large-scale hyperparameter optimization problems [38]. It has demonstrated superior performance on a workload employing 500 workers, exhibits linear scalability with the number of workers in distributed environments, and is well-suited for tasks involving substantial parallelism [38]. An advantage of ASHA is that the user does not need to specify in advance how many configurations they want to evaluate, because it operates asynchronously. However, it still requires the same inputs as SHA. A comprehensive explanation of ASHA can be found in the original work by Li et al. [38] and is not repeated in this effort.
To the best of the authors’ knowledge, this is the first instance of the Bayesian and ASHA optimization algorithms being employed to tune the hyperparameters of multiple ML algorithms to automatically select the best model for forecasting multi-step-ahead daily T m a x and T m i n . In this study, forecasting models were developed by fine-tuning hyperparameters of seven widely utilized ML algorithms, aiming to identify the optimal model for forecasting daily T m a x and T m i n . Table 2 outlines the candidate ML algorithms and their adjustable hyperparameters. The hyperparameters were tuned using both the Bayesian and ASHA optimization algorithms, and a comparison was performed with respect to training time and accuracy. The best model was selected based on its ability to yield the lowest training and test errors, employing either the Bayesian or ASHA optimization algorithms.
Forecasting models were developed for one-, two-, three-, four-, and five-days-ahead forecasting of daily T m a x and T m i n using data from three weather stations. Consequently, a total of 30 forecasting models were chosen by applying Bayesian and ASHA optimization techniques. Each training dataset, consisting of input–output pairs, was divided into a training set, containing 80% of the data, and a test set, containing the remaining 20%. To enhance computational efficiency, both algorithm runs were executed in the parallel computing environment of MATLAB (MATLAB, 2021a).

2.6. Statistical Indices for Performance Evaluation

The following statistical indices were used to evaluate the performances of the developed temperature forecast models. The accuracy index is an evaluation metric that compares the proportion of accurate forecasts made by a model to all forecasts made. The higher the accuracy score is, the better the model performance will be. The ideal value of accuracy is 1.0. The correlation coefficient, R, denotes the strength of linear regression between the observed and forecasted values; however, for this linear relationship, the highest possible value (ideal) of R = 1.0 can be obtained despite the fact that the slope and ordinate intercept are different from 1.0 and 0, respectively [53]. Therefore, other indices need to be used to justify the model performance. A normalized/dimensionless measure of residual variance, Nash–Sutcliffe Efficiency Coefficient (NS) metric, is calculated by dividing residual variance by variance of observed dataset. NS ≤ 0.4, 0.40 < NS ≤ 0.50, 0.50 < NS ≤ 0.65, 0.65 < NS ≤ 0.75, and 0.75 < NS ≤ 1.00 are categories that are labeled as unsatisfactory, acceptable, satisfactory, good, and exceptionally good, respectively [71,72]. Willmott’s Index of Agreement (IOA) [73] is able to detect additive and proportional differences in the observed and model-forecasted means and variances. The IOA usually ranges from −1 to +1, with higher values indicating greater model performance. Nevertheless, the IOA is often overly sensitive to extreme values due to the squared differences [74]. The Kling Gupta Efficiency (KGE) [75,76], which combines the three components of model errors (i.e., correlation, bias, and ratio of variances or coefficients of variation) in a more balanced way, has been widely used for evaluating the prediction ability of models in recent years.
Generally, the RMSE criterion measures the error of the model. A lower value of RMSE indicates a higher forecasting power of the model. However, the value of RMSE largely depends on the magnitude of the data, and therefore, a lower value of RMSE does not necessarily mean a better forecasting performance. To overcome this issue, the NRMSE criterion was used to eliminate the dimensionality effect of the data. Model performance is said to be excellent when NRMSE is less than 0.1, good when NRMSE is between 0.1 and 0.2, fair when NRMSE is between 0.2 and 0.3, and poor when NRMSE is greater than 0.3 [77,78]. The Mean Absolute Percentage Relative Error (MAPRE) is the most common measure used to evaluate a model’s prediction performance, probably because the variable’s units are scaled to percentage units, which makes it easier to understand. It works best if there are no extremes to the data (and no zeros). It is often used as a loss function in regression analysis and model evaluation. Median Absolute Deviation (MAD) is a resistant measure of variability, as it relies on the median as the estimate of the center of the distribution and on the absolute difference rather than the squared difference. Because the MAD is the median deviation of scores from the overall median, not all observations are equally weighted in this measure of dispersion. The clear advantage of MAD is the avoidance of influence by outliers. However, it has its own problems: if the distribution is actually normal, there is a loss of efficiency, in that it does not make as much use of the information as what is available in the data [79]. The MBE criterion provides an estimation of whether the developed model systematically under- or over-predicts the actual values. The MBE is usually not used as a measure of the model error, as high individual errors in prediction can also produce a low MBE. MBE is primarily used to estimate the average bias in the model and to decide if any steps need to be taken to correct the model bias [80]. The Percentage Bias (PBIAS) measures the average tendency of the simulated values to be larger or smaller than their observed ones. The optimal value of PBIAS is 0.0, with low-magnitude values indicating accurate model simulation. Positive values indicate overestimation bias, whereas negative values indicate model underestimation bias.
Accuracy [81]:
A c c = 1 a b s m e a n T i , P T i , S T i , S
Correlation coefficient (R) [82]:
R = i = 1 n T i A T A ¯ T i A T P ¯ i = 1 n T i A T A ¯ 2 i = 1 n T i P T P ¯ 2
Nash–Sutcliffe Efficiency Coefficient (NS) [83]:
N S = 1 i = 1 n T i A T i P 2 i = 1 n T i A T A ¯ 2
Willmott’s Index of Agreement (IOA) [73]:
d = 1 i = 1 n T i A T i P 2 i = 1 n T i P T A ¯ T i A T A ¯ 2
Kling–Gupta Efficiency (KGE) [75,76]:
K G E = 1 E D = 1 R 1 2 + 1 2 + β 1 2
= 1 n i = 1 n T i , P T P ¯ 2 1 n i = 1 n T i , S T S ¯ 2
β = 1 n i = 1 n T i , P 1 n i = 1 n T i , S
Root mean squared error (RMSE) [74]:
R M S E = 1 n i = 1 n T i A T i P 2 T A ¯
Normalized RMSE [84]:
N R M S E = R M S E T A ¯
Mean Absolute Percentage Relative Error (MAPRE) (MAPE, n.d.):
M A P R E = 1 n i = 1 n T i , S T i , P T i , S × 100
Median Absolute Deviation (MAD) [85]:
M A D T i A , T i P = m e d i a n T 1 A T 1 P , T 2 A T 2 P , , T n A T n P f o r   i = 1 ,   2 , , n
Mean Bias Error (MBE) [86]:
M B E = 1 n i = 1 n T i , P T i , S
Percentage Bias (PBIAS) [87,88]:
P B I A S = i 1 n T i , P T i , S i = 1 n T i , S × 100
where T i A and T i P are the observed and forecasted T ( T m a x   &   T m i n ) for the i t h data point in the daily maximum and minimum temperature dataset, respectively; T A ¯ and T P ¯ are the means of the observed and forecasted T ( T m a x   &   T m i n ), respectively; and n represents the total number of entries in the dataset.

3. Results and Discussion

A range of statistical performance evaluation indices, as discussed in the previous section, were employed to assess the performance of the various forecasting models. In the subsequent paragraphs, the performances of the identified top-performing models on the test dataset (selected based on their performance on both the training and test datasets) for the five forecasting horizons at the three stations are presented.
Table 3 provides a comparison of the best models that were selected by the Bayesian and ASHA optimization algorithms, considering log(1 + valLoss) values and training time for multi-step-ahead T m a x forecasting. As can be observed from the log(1 + valLoss) value presented in Table 3, the Bayesian optimization algorithm generally outperformed the ASHA algorithm in selecting the best model. However, in terms of training time, the ASHA algorithm demonstrated faster convergence in finding the optimal model parameters compared to the Bayesian algorithm. However, while the ASHA algorithm showed competitive performance in terms of training time, the Bayesian algorithm also achieved convergence within acceptable time limits. Therefore, in cases where training time is not a critical factor, the Bayesian optimization can be used to find optimal model parameters for selecting the best models for a specific task. On the other hand, when training time is a more important consideration than model accuracy, the ASHA optimization algorithm is advisable. Notably, there were instances where the ASHA algorithm outperformed the Bayesian algorithm in both log(1 + valLoss) and training time, such as for the one-step-ahead T m a x forecast at Barishal station and four-step-ahead T m a x forecast at Ishurdi station. In cases where the Bayesian algorithm showed superior performance based on log(1 + valLoss) values, the differences were not substantial, while there was a significant difference in training times between the two algorithms.
Table 4 presents the training and test performance results of the Bayesian and ASHA algorithm-tuned forecast models for maximum temperatures ( T m a x ) at the weather stations. At this stage, the final models were selected based on the RMSE criterion: the best models were those producing the lowest difference between the training and test RMSE values. This ensures that the selected best models were neither over-trained nor under-trained. The data in Table 4 reveal that the ASHA algorithm-tuned best models produced the lowest difference between the training and test RMSE values in most instances (forecasting horizons and weather stations) for forecasting T m a x values. Although the Bayesian algorithm required a longer time to converge to optimal solutions for selecting the best models, it excelled in three specific instances: the one-day- and five-days-ahead T m a x forecast at Barishal station, and the three-days-ahead T m a x forecast at Gazipur station (as indicated in Table 4). A complete list of the selected top-performing models can be found in Table 7.
Table 5 provides a comparative performance evaluation of the Bayesian and ASHA algorithm-tuned best models for forecasting T m i n at the weather stations, based on log (1 + valLoss) values and training time requirements.
The results in Table 5 reveal that, in terms of log (1 + valLoss) values, the Bayesian algorithm-tuned best models outperformed the ASHA algorithm-tuned models in all instances except for the two-days-ahead forecast at the Gazipur station, where the ASHA algorithm-tuned GPR model was found to be the top-performing best model. Moreover, the ASHA algorithm-tuned models exhibited faster convergence to optimal solutions for parameter values in comparison to the Bayesian algorithm-tuned models in all instances. Additionally, the differences between the training errors (log (1 + valLoss)) produced by the Bayesian and ASHA algorithms for selecting the best models were relatively small for all instances. In summary, when computational time is not a limiting factor, the Bayesian algorithm is a suitable choice for searching for the best models (Table 5). However, it is essential to carefully evaluate the differences between the training and test errors of the best models that are produced by the Bayesian and ASHA algorithms before making a definitive decision on model selection.
The comparison of the training and testing performance between the Bayesian and ASHA optimization-tuned best models for forecasting T m i n at the weather stations, as assessed by the RMSE criterion, is presented in Table 6. It is evident from Table 6 that both the Bayesian and ASHA algorithms exhibited similar performance in selecting the top-performing best models across all forecasting horizons and at all weather stations. The Bayesian algorithm outperformed in seven instances, while the ASHA algorithm provided top-performing forecast models in eight instances based on the lowest differences between the training and test RMSE values (Table 6). Overall, the selected top-performing best models demonstrated acceptable results according to the RMSE criterion. However, further validation of the forecasting performance of the selected models is required by computing other performance evaluation indices on the test dataset. A comprehensive performance evaluation of the selected top-performing models based on several statistical indices can be found in Table 8 and Table 9 and Figure 3 and Figure 4.
The complete list of the selected best models for different weather stations under five forecasting horizons, based on log (1 + valLoss), RMSE, training time, and differences between the training and test RMSE (as presented in Table 3, Table 4, Table 5 and Table 6), is provided in Table 7.
Following the selection of the best-performing models, they were utilized to forecast multi-step-ahead T m a x and T m i n values on the test dataset at the respective weather stations. To evaluate the forecasting performance, various statistical performance indices were computed and are presented in Table 8 and Table 9. Additionally, the results are visualized in Figure 3 and Figure 4.
Table 8 presents a comprehensive overview of the performance of the best models in forecasting T m a x under five different forecast horizons at the weather stations. It can be observed from Table 8 that the selected models consistently produced lower values of various performance metrics such as RMSE, NRMSE, MAPRE, MAD, MBE, and PBIAS. This indicates that the models have demonstrated improved accuracy in forecasting T m a x . It is also inferred from the results presented in Table 8 that the forecasting performance slightly decreased with an increase in the forecast horizon. This observation aligns with prior findings reported in studies by Rahman et al. [42] and Barzegar et al. [89], which indicated that ML-based forecast models tend to exhibit reduced accuracy as the forecast horizon extends further into the future. The RMSE values were pretty small for all selected models across the weather stations. Small RMSE values generally indicate that the models’ predictions are closely aligned with the actual observations, suggesting a high level of accuracy in the forecasts. The NRMSE values were consistently less than 0.1 for all instances. An NRMSE below 0.1 is considered excellent performance in forecasting, as it indicates that the forecasted values are very close to the actual values [77,78]. The MAPRE values are also within acceptable ranges. For Barishal, Gzipur, and Ishurdi stations, the MAPRE values were below 8% (ranging from 5.091% to 7.954% for different forecast horizons), 7% (ranging from 4.737% to 6.468% for different forecast horizons), and 6% (ranging from 4.728% to 5.426% for different forecast horizons), respectively, depending on the forecast horizons. Since a MAPRE value below 10% is deemed acceptable for ML-based forecast models [90], these results suggest that the selected models are producing forecasts that meet or exceed acceptable standards. In summary, these performance indices demonstrate that the selected best forecast models are capable of producing accurate and reliable forecasts for T m a x values at the weather stations, as evidenced by their small RMSE and NRMSE values, as well as MAPRE values that are well within the acceptable range.
The MAD values are also reported as being acceptable for all models. These values ranged from 0.634 °C for a one-day-ahead forecast at Gazipur station to 1.070 °C for a five-days-ahead forecast at Barishal station. The lower the MAD is, the closer the forecasts are to the actual values, indicating accurate model forecasts. Additionally, the models produced smaller values of MBE and PBIAS, which were pretty close to the optimal value of 0.0. MBE quantifies the average bias (overestimation or underestimation) in the forecasts, while PBIAS provides a measure of the Percentage Bias. The models produced low-magnitude values of both MBE and PBIAS, indicating that they are making reasonably accurate forecasts. Some models show negative MBE and PBIAS values, indicating a slight underestimation bias, particularly for Gazipur and Ishurdi stations and for three-days-ahead forecasts at Barishal station. On the other hand, positive MBE and PBIAS values are observed for one-day-, two-days-, four-days-, and five-days-ahead T m a x forecasts at the Barishal station, suggesting a slight overestimation bias (Table 8). These biases are of smaller magnitude, indicating that the models tend to slightly under-predict or over-predict temperature values, but the deviations from the actual values are not substantial. In summary, the forecast models are delivering forecasts with acceptable levels of accuracy, as indicated by the MAD, MBE, and PBIAS values. While some models exhibit slight underestimation or overestimation biases, these biases are relatively small, and the forecasts are still considered accurate. These results provide confidence in the performance of the selected forecast models.
Table 9 presents a comprehensive assessment of the performance of the top-performing models for forecasting T m i n values under the five forecast horizons at the weather stations. It can be seen from Table 9 that the overall forecasting performances of the models are acceptable, although the models showed slightly poor performance with respect to the computed RMSE, NRMSE, and MAPRE values when compared to the forecasting of T m a x (Table 8). Nevertheless, the models produced lower RMSE values (Table 9). The computed NRMSE criterion suggests that the model performances were excellent (NRMSE < 0.1) to good (NRMSE values slightly higher than 0.1, i.e., NRMSE > 0.1 and NRMSE < 0.2) according to the ranges reported in Heinemann et al. [77] and Li et al. [78]. On the other hand, the models produced lower MAD, MBE, and PBIS values compared to those for the models developed for forecasting T m a x (Table 8 and Table 9). It is noted that the computed MBE and PBIAS values were pretty close to the ideal value of 0.0. According to the MBE criterion, the models produced very small amounts of underestimated biases, as indicated by the negative MBE values for all forecasting horizons and at all weather stations (Table 9). Similarly, the PBIAS criterion also suggests that mostly underestimated biases existed, except for the model developed to forecast T m i n for two-step-ahead at the Barishal weather station.
The findings are reported in the form of bar diagrams, especially to demonstrate the model performance based on some other statistical performance indices. Figure 4 and Figure 5 illustrate the models’ performance with respect to accuracy, R, NS, IOA, and KGE criteria. These performance indices are referred to as benefit indices, because higher values of these indices indicate improved model performance. Figure 4 shows the performance of the best models on the test dataset when forecasting T m a x across different forecast horizons at the respective weather stations. This visual representation allows for a quick and easy assessment of how well the models are performing and how their performance varies with different lead times (forecast horizons).
Figure 4 reveals important insights into the performance of the best models for forecasting T m a x under different forecast horizons at the weather stations. It is perceived from Figure 4 that the best models consistently demonstrate excellent performance with regard to the accuracy and IOA criteria, where accuracy values are close to 1, and IOA values exceed 0.8 for all forecast horizons. This indicates that the models produce forecasts that closely align with observed data and show a strong agreement with the reference measurements. Notably, the accuracy criterion does not exhibit a decreasing trend with an increase in the forecast horizon. In other words, the models maintain high accuracy regardless of the lead time. This is a positive finding, suggesting that the models are reliable for both short-term and longer-term forecasts. However, the R, NS, IOA, and KGE values indicate that the model performance is indeed influenced by the forecast horizon. These indices suggest that the forecasting performance tends to decrease as the forecast horizon increases, which is consistent with prior research [42,89]. R values are higher than 0.8 for the first and second forecast horizons (one- and two-step-ahead forecasts) at all weather stations. However, a decrease in R values is observed as the forecast horizon extends, with the lowest R value (0.771) occurring at the fifth forecast horizon for the Ishurdi station. In general, model performances were relatively poor with respect to the NS and KGE criteria (Figure 4). The findings of this research is in good agreement with the findings presented in Müller and Piché [91], who stated that ML-based models often showed contrasting performance with respect to different performance evaluation indices. These findings collectively suggest that the models are especially well suited for shorter-term forecasts, but they still provide valuable forecasts for longer lead times, albeit with slightly reduced performance.
In order to forecast T m i n at the weather stations over a range of forecast horizons, the best models were tested against the test dataset. The findings are presented in Figure 5, which shows the performance of the best models for forecasting T m i n under various forecast horizons at the weather stations. Figure 5 suggests that the selected top models performed exceptionally well for all forecast horizons and at all the weather stations. This is particularly evident from the accuracy and IOA values, which consistently exceed 0.95. Such high accuracy and IOA values indicate excellent model performance, suggesting that the models generate forecasts that closely match the observed data. Similar to the T m a x forecast, the accuracy of the models remains high across different forecast horizons, with accuracy values close to 1. This indicates that the models maintain their high forecasting accuracy regardless of the lead time. Contrary to accuracy, other performance evaluation indices, including NS, IOA, and KGE, display a diminishing trend as the prediction horizon is extended. This is in line with the common observation that the forecasting performance tends to decrease as the forecast horizon increases, which aligns with the results observed for T m a x forecasting. It is important to note that the R values appear to be consistent across all weather stations. These values are high, exceeding 0.90, indicating a strong correlation between the model forecasts and the observed data. In general, the best models appeared to perform better across all forecast horizons based on the R (>0.90), NS (>0.81), and KGE (>0.87) criterion values. Overall, the findings from Figure 5 suggest that the selected models perform remarkably well in forecasting T m i n over various forecast horizons and at different weather stations. The models exhibit high accuracy, strong correlations, and consistent performance across forecast horizons, even though other indices show a slight decrease with an increasing forecast lead time.
In a comparative context, the results for T m i n forecasting seem to outperform those for T m a x forecasting. This improvement in T m i n forecasting performance might be attributed to the quality and volume of the collected data. High-quality and abundant data often lead to more accurate forecasts.
While direct comparisons among the results presented in this research are hindered by the diverse study conditions and modeling approaches employed, an indirect evaluation was conducted by contrasting the computed performance indices in this study with those in previous research. For one-day-ahead minimum temperature forecasts, R2 values of 0.939, 0.911, and 0.901 were achieved across the Barishal, Gazipur, and Ishurdi stations, respectively. Conversely, for one-day-ahead maximum temperature forecasts, R2 values of 0.819, 0.764, and 0.633 were obtained for the same stations. These results surpass the outcomes of previous studies utilizing CNN (~0.5), LSTM (~0.6), and CNN-LSTM (~0.7) for one-day-ahead temperature forecasts [33].
Moreover, our findings compare favorably or even outperform those of Ebtehaz et al. [27], who utilized IORELM for 10 h ahead temperature forecasts (R = 0.95, NSE = 0.89, RMSE = 3.74, MAE = 1.92). Our best models yielded RMSE values of approximately 2.0 °C and 1.5 °C for one-day-ahead maximum and minimum temperature forecasts, respectively. The detailed statistical performance indices for the proposed best models are provided in Figure 4 and Figure 5, as well as in Table 8 and Table 9.
Furthermore, our research outcomes stand up well against those of Fister et al. [29], focusing on the temperature dataset of the Paris region. Our proposed best model exhibited superior performance compared to Lasso Regression, Decision Tree, Adaboost, RF, and CNN in terms of MSE values. Alomar et al. [17] identified the SVR model as the top performer for daily temperature forecasting, achieving an RMSE value of 3.592 °C, which is higher than the RMSE values produced by our proposed best models for both minimum and maximum temperature forecasts across the three weather stations (refer to Table 8 and Table 9).
In terms of the RMSE criterion, our proposed best models demonstrated comparable or superior performance (RMSE ~ 2 °C for both minimum and maximum temperatures) compared to the ANN (RMSE ~ 3 °C), GEP (RMSE ~ 3 °C), and HBA-ANN (RMSE ~ 2 °C) models that were developed for the coldest and warmest regions globally [23]. Based on this comprehensive comparison, it can be argued that our proposed best models exhibit acceptable and sometimes superior performance compared to recently proposed machine and deep learning models for temperature forecasting. However, it is important to note that direct comparisons are challenging due to variations in data and study locations.
The research on automated model selection using Bayesian optimization and the asynchronous successive halving algorithm for predicting daily minimum and maximum temperatures holds crucial implications for the agricultural domain. Accurate temperature predictions are fundamental to agricultural planning, impacting crop growth, yield estimation, and resource allocation. The application of Bayesian optimization ensures a thorough exploration of model parameters, enhancing the precision of temperature forecasts, which is crucial for optimal crop management. The incorporation of the asynchronous successive halving algorithm contributes to computational efficiency in finding the optimal hyperparameters for the selected best models. As a result, this research has the potential to significantly improve agricultural productivity, resource utilization, and resilience to climate variability, ultimately benefiting farmers and stakeholders across the agricultural supply chain.

4. Conclusions

Accurate and reliable forecasting of daily maximum ( T m a x ) and minimum ( T m i n ) temperatures can be effectively utilized in the development of a sustainable and efficient agricultural water management strategy. However, due to nonlinear interactions between temperatures and other explanatory variables, as well as their multi-scale behavior that changes over time, producing reliable temperature ( T m a x   a n d   T m i n ) forecasts is often challenging. The prerequisites for creating accurate ML-based forecast models include selecting only the most influential input variables from a list of prospective input variables and optimizing model parameters. To address these challenges, this study proposes an innovative approach for selecting the most influential input variables and determining the best predictive models for forecasting daily T m a x and T m i n values. These methods were combined with Bayesian and ASHA hyperparameter tuning to perform automated model parameter estimation. Notably, this study is the first to utilize the Bayesian and ASHA algorithms for automating the model selection process to provide accurate T m a x and T m i n forecasts at different weather stations in Bangladesh. Furthermore, the study provides a comparison of the best models that are tuned with Bayesian and ASHA algorithms. The selected best models were explored for one-, two-, three-, four-, and five-days-ahead T m a x and T m i n forecasting. The top-performing models for different forecasting horizons (1-day-, 2-days-, 3-days-, 4-days-, and 5-days-ahead) at the three weather stations were identified. The results demonstrate the suitability of these models in forecasting multi-step-ahead (5-days-ahead) daily T m a x and T m i n values, as indicated by the computed performance evaluation indices. The findings of this research demonstrated the ability and practical applicability of the proposed models in forecasting days-ahead T m a x and T m i n values at the weather stations.
The primary objective of this research is to propose an ML-based methodology that is capable of accurately approximating daily temperature fluctuations and providing multi-step-ahead temperature forecasts. Importantly, the proposed methodology can be applied to other regions with diverse data ranges. Given the varying time intervals in the data from the three weather stations, the ML-based modeling approaches were developed separately for each station. The duration of data collection was determined based on the availability of data from the selected weather stations within the study area. Despite the absence of data for specific time intervals at some weather stations, the available dataset, spanning approximately a reasonable duration, remains sufficient and valuable for addressing the research objectives. Therefore, we believe that our findings are relevant and contribute significantly to the advancement of the field. Indeed, utilizing similar interval data for all stations and developing a model for one station, then validating its generalization capability at other stations, would be an interesting topic for future research. This approach could provide insights into the transferability and robustness of the proposed ML-based methodology across different weather stations and regions.
The research paper presented a novel approach for automated model selection using Bayesian and ASHA algorithms. The findings contribute to the field of climate science and weather forecasting, providing valuable insights into improving temperature prediction models through automation and optimization techniques. The proposed methodology can be further extended and applied to other domains requiring accurate and efficient model selection.
In this research, a limited set of ML algorithms was employed for selecting the best model, with the assistance of Bayesian and ASHA optimization algorithms. To broaden the scope of future studies, a more comprehensive array of ML algorithms could be explored for hyperparameter tuning using optimization algorithms. Additionally, the inclusion of a few deep learning algorithms in the pool of prospective models could be considered, enabling a more thorough exploration of the best-performing model through parameter tuning. The use of the Bayesian and ASHA optimization algorithms to fine-tune hyperparameters across multiple ML algorithms facilitates the automatic selection of the most effective forecasting model. It is noteworthy that alternative optimization algorithms, such as the genetic algorithm (GA) or particle swarm optimization (PSO), could also be investigated in future studies.
However, it is essential to mention that while incorporating a diverse set of ML and optimization algorithms could enhance the depth of the study, it comes with the trade-off of increased complexity and time consumption in hyperparameter tuning across multiple ML models. This potential intricacy may pose challenges in achieving optimum results within the set parameters.

Author Contributions

Conceptualization, methodology, investigation, resources, data curation, project administration, formal analysis, and writing—original draft preparation: D.K.R., M.A.H. and M.P.H.; data curation, project administration, investigation, and writing—review and editing: A.A., A.Z.D. and M.A.M. All authors have read and agreed to the published version of the manuscript.

Funding

The authors extend their appreciation to the Deputyship for Research & Innovation, Ministry of Education in Saudi Arabia for funding this research (IFKSURC-1-4106).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Acknowledgments

The authors extend their appreciation to the Deputyship for Research & Innovation, Ministry of Education in Saudi Arabia for funding this research (IFKSURC-1-4106).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Roy, D.S. Forecasting the air temperature at a weather station using deep neural networks. Procedia Comput. Sci. 2020, 178, 38–46. [Google Scholar] [CrossRef]
  2. Parker, D.J.; Blyth, A.M.; Woolnough, S.J.; Dougill, A.J.; Bain, C.L.; de Coning, E.; Diop-Kane, M.; Kamga Foamouhoue, A.; Lamptey, B.; Ndiaye, O.; et al. The African SWIFT Project: Growing Science Capability to Bring about a Revolution in Weather Prediction. Bull. Am. Meteorol. Soc. 2022, 103, E349–E369. [Google Scholar] [CrossRef]
  3. Allen, R.G.; Pereira, L.S.; Raes, D.; Smith, M. Crop Evapotranspiration—Guidelines for Computing Crop Water Requirements—FAO Irrigation and Drainage Paper 56; FAO: Rome, Italy, 1998. [Google Scholar]
  4. Ali, M.H. Crop Water Requirement and Irrigation Scheduling. In Fundamentals of Irrigation and On-Farm Water Management: Volume 1; Ali, M.H., Ed.; Springer New York: New York, NY, USA, 2010; pp. 399–452. ISBN 978-1-4419-6335-2. [Google Scholar]
  5. Wang, J.; Xin, L.; Wang, X.; Jiang, M. The Impact of Climate Change and Grain Planting Structure Change on Irrigation Water Requirement for Main Grain Crops in Mainland China. Land 2022, 11, 2174. [Google Scholar] [CrossRef]
  6. Haque, M.P.; Hossain, M.; Muhammad, A. Climate Change Effect on Irrigation Water Requirement of Wheat and Maize in Northern Part of Bangladesh. Int. J. Clim. Res. 2021, 5, 25–34. [Google Scholar] [CrossRef]
  7. Lobell, D.; Field, C. Global Scale Climate–Crop Yield Relationships and the Impacts of Recent Warming. Environ. Res. Lett. 2007, 2, 14002. [Google Scholar] [CrossRef]
  8. Tao, F.; Zhang, S.; Zhang, Z.; Rötter, R.P. Maize Growing Duration Was Prolonged across China in the Past Three Decades under the Combined Effects of Temperature, Agronomic Management, and Cultivar Shift. Glob. Chang. Biol. 2014, 20, 3686–3699. [Google Scholar] [CrossRef] [PubMed]
  9. Gregory, P.J.; Johnson, S.N.; Newton, A.C.; Ingram, J.S.I. Integrating Pests and Pathogens into the Climate Change/Food Security Debate. J. Exp. Bot. 2009, 60, 2827–2838. [Google Scholar] [CrossRef] [PubMed]
  10. Paniagua-Tineo, A.; Salcedo-Sanz, S.; Casanova-Mateo, C.; Ortiz-García, E.G.; Cony, M.A.; Hernández-Martín, E. Prediction of daily maximum temperature using a support vector regression algorithm. Renew. Energy 2011, 36, 3054–3060. [Google Scholar] [CrossRef]
  11. Attar, N.F.; Khalili, K.; Behmanesh, J.; Khanmohammadi, N. On the reliability of soft computing methods in the esti-mation of dew point temperature: The case of arid regions of Iran. Comput. Electron. Agric. 2018, 153, 334–346. [Google Scholar] [CrossRef]
  12. Curceac, S.; Ternynck, C.; Ouarda, T.B.M.J.; Chebana, F.; Niang, S.D. Short-term air temperature forecasting using Nonparametric Functional Data Analysis and SARMA models. Environ. Model. Softw. 2019, 111, 394–408. [Google Scholar] [CrossRef]
  13. Naganna, S.R.; Deka, P.C.; Ghorbani, M.A.; Biazar, S.M.; Al-Ansari, N.; Yaseen, Z.M. Dew point temperature estima-tion: Application of artificial intelligence model integrated with nature-inspired optimization algorithms. Water 2019, 11, 742. [Google Scholar] [CrossRef]
  14. Salcedo-Sanz, S.; Deo, R.C.; Carro-Calvo, L.; Saavedra-Moreno, B. Monthly prediction of air temperature in Australia and New Zealand with machine learning algorithms. Theor. Appl. Climatol. 2016, 125, 13–25. [Google Scholar] [CrossRef]
  15. Marchuk, G.I. 5—A weather prediction scheme based on conservation laws. In Numerical Methods in Weather Prediction; Marchuk, G.I., Ed.; Academic Press: Cambridge, MA, USA, 1974; pp. 161–199. [Google Scholar] [CrossRef]
  16. Richardson, L.F. The fundamental equations. In Weather Prediction by Numerical Process, Cambridge Mathematical Library; Richardson, L.F., Ed.; Cambridge Mathematical Library, Cambridge University Press: Cambridge, UK, 2007; pp. 21–114. [Google Scholar] [CrossRef]
  17. Alomar, M.; Khaleel, F.; Aljumaily, M.; Masood, A.; Mohd Razali, S.F.; Alsaadi, M.; Al-Ansari, N.; Hameed, M. Data-driven models for atmospheric air temperature forecasting at a continental climate region. PLoS ONE 2022, 17, e0277079. [Google Scholar] [CrossRef] [PubMed]
  18. Cifuentes, J.; Marulanda, G.; Bello, A.; Reneses, J. Air temperature forecasting using machine learning techniques: A review. Energies 2020, 13, 4215. [Google Scholar] [CrossRef]
  19. Deng, T.; Duan, H.-F.; Keramat, A. Spatiotemporal characterization and forecasting of coastal water quality in the semi-enclosed Tolo Harbour based on machine learning and EKC analysis. Eng. Appl. Comput. Fluid Mech. 2022, 16, 694–712. [Google Scholar] [CrossRef]
  20. Zhu, S.; Hadzima-Nyarko, M.; Gao, A.; Wang, F.; Wu, J.; Wu, S. Two hybrid data-driven models for modeling wa-ter-air temperature relationship in rivers. Environ. Sci. Pollut. Res. 2019, 26, 12622–12630. [Google Scholar] [CrossRef]
  21. Papacharalampous, G.; Tyralis, H.; Koutsoyiannis, D. Univariate time series forecasting of temperature and precipita-tion with a focus on machine learning algorithms: A multiple-case study from Greece. Water Resour. Manag. 2018, 32, 5207–5239. [Google Scholar] [CrossRef]
  22. Gómez-Orellana, A.M.; Guijo-Rubio, D.; Pérez-Aracil, J.; Gutiérrez, P.A.; Salcedo-Sanz, S.; Hervás-Martínez, C. One month in advance prediction of air temperature from Reanalysis data with explainable Artificial Intelligence techniques. Atmos. Res. 2023, 284, 106608. [Google Scholar] [CrossRef]
  23. Zhou, J.; Wang, D.; Band, S.S.; Mirzania, E.; Roshni, T. Atmosphere air temperature forecasting using the honey badger optimization algorithm: On the warmest and coldest areas of the world. Eng. Appl. Comput. Fluid Mech. 2023, 17, 2174189. [Google Scholar] [CrossRef]
  24. Mirarabi, A.; Nassery, H.R.; Nakhaei, M.; Adamowski, J.; Akbarzadeh, A.H.; Alijani, F. Evaluation of data-driven models (SVR and ANN) for groundwater-level prediction in confined and unconfined systems. Environ. Earth Sci. 2019, 78, 489. [Google Scholar] [CrossRef]
  25. Radhika, Y.; Shashi, M. Atmospheric temperature prediction using support vector machines. Int. J. Comput. Theory Eng. 2009, 1, 55–58. [Google Scholar] [CrossRef]
  26. Sanikhani, H.; Deo, R.C.; Samui, P.; Kisi, O.; Mert, C.; Mirabbasi, R.; Gavili, S.; Yaseen, Z.M. Survey of different da-ta-intelligent modeling strategies for forecasting air temperature using geographic information as model predictors. Comput. Electron. Agric. 2018, 152, 242–260. [Google Scholar] [CrossRef]
  27. Ebtehaj, I.; Bonakdari, H.; Gharabaghi, B.; Khelifi, M. Time-series-based air temperature forecasting based on the outli-er robust extreme learning machine. Environ. Sci. Proc. 2023, 25, 51. [Google Scholar] [CrossRef]
  28. Shahdad, M.; Saber, B. Multistep-ahead forecasting for maximum and minimum air temperatures using a new hybrid intelligence tree-based filter classifier. Model. Earth Syst. Environ. 2022, 8, 5449–5465. [Google Scholar] [CrossRef]
  29. Fister, D.; Pérez-Aracil, J.; Peláez-Rodríguez, C.; Del Ser, J.; Salcedo-Sanz, S. Accurate long-term air temperature pre-diction with Machine Learning models and data reduction techniques. Appl. Soft Comput. 2023, 136, 110118. [Google Scholar] [CrossRef]
  30. Handler, S.L.; Reeves, H.D.; McGovern, A. Development of a probabilistic subfreezing road temperature nowcast and forecast using machine learning. Weather Forecast. 2020, 35, 1845–1863. [Google Scholar] [CrossRef]
  31. Yang, Q.; Lee, C.-Y.; Tippett, M.K. A long short-term memory model for global rapid intensification prediction. Weather Forecast. 2020, 35, 1203–1220. [Google Scholar] [CrossRef]
  32. Sari, Y.; Arifin, Y.F.; Novitasari, N.; Faisal, M.R. Deep learning approach using the GRU-LSTM hybrid model for air temperature prediction on daily basis. Int. J. Intell. Syst. Appl. Eng. 2022, 10, 430–436. [Google Scholar]
  33. Hou, J.; Wang, Y.; Zhou, J.; Tian, Q. Prediction of hourly air temperature based on CNN–LSTM. Geomat. Nat. Hazards Risk 2022, 13, 1962–1986. [Google Scholar] [CrossRef]
  34. Khan, M.I.; Maity, R. Hybrid deep learning approach for multi-step-ahead prediction for daily maximum temperature and heatwaves. Theor. Appl. Climatol. 2022, 149, 945–963. [Google Scholar] [CrossRef]
  35. Gong, Y.; Wang, Z.; Xu, G.; Zhang, Z. A Comparative study of groundwater level forecasting using data-driven models based on ensemble empirical mode decomposition. Water 2018, 10, 730. [Google Scholar] [CrossRef]
  36. Moravej, M.; Amani, P.; Hosseini-Moghari, S.-M. Groundwater level simulation and forecasting using interior search algorithmleast square support vector regression (ISA-LSSVR). Groundw. Sustain. Dev. 2020, 11, 100447. [Google Scholar] [CrossRef]
  37. Snoek, J.; Larochelle, H.; Adams, R. Practical Bayesian optimization of machine learning algorithms. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Nice, France, 2012; Volume 4. [Google Scholar]
  38. Li, L.; Jamieson, K.; Rostamizadeh, A.; Gonina, E.; Ben-Tzur, J.; Hardt, M.; Recht, B.; Talwalkar, A. A system for massively parallel hyperparameter tuning. In Proceedings of the 3 Rd MLSys Conference, Austin, TX, USA, 2–4 March 2020. [Google Scholar]
  39. Galelli, S.; Humphrey, G.B.; Maier, H.R.; Castelletti, A.; Dandy, G.C.; Gibbs, M.S. An evaluation framework for input variable selection algorithms for environmental data-driven models. Environ. Model. Softw. 2014, 62, 33–51. [Google Scholar] [CrossRef]
  40. Quilty, J.; Adamowski, J.; Khalil, B.; Rathinasamy, M. Bootstrap rank-ordered conditional mutual information (broCMI): A nonlinear input variable selection method for water resources modeling. Water Resour. Res. 2016, 52, 2299–2326. [Google Scholar] [CrossRef]
  41. Yaseen, Z.M.; Jaafar, O.; Deo, R.C.; Kisi, O.; Adamowski, J.; Quilty, J.; El-Shafie, A. Stream-flow forecasting using extreme learning machines: A case study in a semiarid region in Iraq. J. Hydrol. 2016, 542, 603–614. [Google Scholar] [CrossRef]
  42. Rahman, A.T.M.S.; Hosono, T.; Quilty, J.M.; Das, J.; Basak, A. Multiscale groundwater level forecasting: Coupling new machine learning approaches with wavelet transforms. Adv. Water Resour. 2020, 141, 103595. [Google Scholar] [CrossRef]
  43. Lam, R.; Willcox, K.; Wolpert, D.H. Bayesian optimization with a finite budget: An approximate dynamic program-ming approach. In Advances in Neural Information Processing Systems, Proceedings of the 29th Conference on Neural Information Processing Systems (NIPS 2016), Barcelona, Spain, 5–10 December 2016; Neural Information Processing Systems Foundation, Inc.: San Diego, CA, USA, 2017. [Google Scholar]
  44. Shahid, S.; Harun, S.; Katimon, A. Changes in Diurnal Temperature Range in Bangladesh during the Time Period 1961–2008. Atmos. Res. 2012, 118, 260–270. [Google Scholar] [CrossRef]
  45. Arnfield, A.J. Köppen Climate Classification. Encyclopedia Britannica. 2023. Available online: https://www.britannica.com/science/Koppen-climate-classification (accessed on 2 November 2023).
  46. Betts, A.K.; Ball, J.H. The FIFE Surface Diurnal Cycle Climate. J. Geophys. Res. Atmos. 1995, 100, 25679–25693. [Google Scholar] [CrossRef]
  47. Thomson, D.J. The Seasons, Global Temperature, and Precession. Science 1995, 268, 59–68. [Google Scholar] [CrossRef] [PubMed]
  48. Estévez, J.; García-Marín, A.P.; Morábito, J.A.; Cavagnaro, M. Quality assurance procedures for validating meteoro-logical input variables of reference evapotranspiration in mendoza province (Argentina). Agric. Water Manag. 2016, 172, 96–109. [Google Scholar] [CrossRef]
  49. Feng, S.; Hu, Q.; Qian, W. Quality control of daily meteorological data in China, 1951–2000: A new dataset. Int. J. Climatol. 2004, 24, 853–870. [Google Scholar] [CrossRef]
  50. Shafer, M.A.; Fiebrich, C.A.; Arndt, D.S.; Fredrickson, S.E.; Hughes, T.W. Quality Assurance Procedures in the Okla-homa Mesonetwork. J. Atmos. Ocean. Technol. 2000, 17, 474–494. [Google Scholar] [CrossRef]
  51. Dasgupta, A.; Wahed, A. Chapter 4—Laboratory Statistics and Quality Control. In Clinical Chemistry, Immunology and Laboratory Quality Control; Dasgupta, A., Wahed, A., Eds.; Elsevier: Berlin/Heidelberg, Germany, 2014; pp. 47–66. ISBN 978-0-12-407821-5. [Google Scholar]
  52. Deo, R.C.; Downs, N.; Parisi, A.V.; Adamowski, J.F.; Quilty, J.M. Very short-term reactive forecasting of the solar ultra-violet index using an extreme learning machine integrated with the solar zenith angle. Environ. Res. 2017, 155, 141–166. [Google Scholar] [CrossRef] [PubMed]
  53. Barzegar, R.; Ghasri, M.; Qi, Z.; Quilty, J.; Adamowski, J. Using bootstrap ELM and LSSVM models to estimate river ice thickness in the Mackenzie River Basin in the Northwest Territories, Canada. J. Hydrol. 2019, 577, 123903. [Google Scholar] [CrossRef]
  54. Hadi, S.J.; Abba, S.I.; Sammen, S.S.; Salih, S.Q.; Al-Ansari, N.; Yaseen, Z.M. Non-linear input variable selection ap-proach integrated with non-tuned data intelligence model for streamflow pattern simulation. IEEE Access 2019, 7, 141533–141548. [Google Scholar] [CrossRef]
  55. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  56. Taormina, R.; Galelli, S.; Karakaya, G.; Ahipasaoglu, S.D. An information theoretic approach to select alternate subsets of predictors for data-driven hydrological models. J. Hydrol. 2016, 542, 18–34. [Google Scholar] [CrossRef]
  57. Box, G.E.P.; Jenkins, G.M.; Reinsel, G.C. Time Series Analysis: Forecasting and Control, 3rd ed.; Prentice Hall: Englewood Cliffs, NJ, USA, 1994. [Google Scholar]
  58. Nash, W.J.; Sellers, T.L.; Talbot, S.R.; Cawthorn, A.J.; Ford, W.B. The Population Biology of Abalone (Haliotis species) in Tasmania. I. Blacklip Abalone (H. rubra) from the north coast and islands of Bass Strait; Technical Report no. 48; Department of Primary Industry and Fisheries: Tasmania, Australia, 1994.
  59. Waugh, S. Extending and Benchmarking Cascade-Correlation: Extensions to the Cascade-Correlation Architecture and Benchmarking of Feed-Forward Supervised Artificial Neural Networks; University of Tasmania: Tasmania, Australia, 1995. [Google Scholar]
  60. Darbellay, G.A.; Vajda, I. Estimation of the information by an adaptive partitioning of the observation space. IEEE Trans. Inf. Theory 1999, 45, 1315–1321. [Google Scholar] [CrossRef]
  61. Ding, C.; Peng, H. Minimum redundancy feature selection from microarray gene expression data. J. Bioinform. Comput. Biol. 2005, 3, 185–205. [Google Scholar] [CrossRef]
  62. Yang, W.; Wang, K.; Zuo, W. Neighborhood component feature selection for high-dimensional data. J. Comput. 2012, 7, 161–168. [Google Scholar] [CrossRef]
  63. Probst, P.; Wright, M.N.; Boulesteix, A.-L. Hyperparameters and tuning strategies for random forest. WIREs Data Min. Knowl. Discov. 2019, 9, e1301. [Google Scholar] [CrossRef]
  64. Zhang, F.; Deb, C.; Lee, S.E.; Yang, J.; Shah, K.W. Time series forecasting for building energy consumption using weighted Support Vector Regression with differential evolution optimization technique. Energy Build. 2016, 126, 94–103. [Google Scholar] [CrossRef]
  65. Jeong, J.; Park, E. Comparative applications of data-driven models representing water table fluctuations. J. Hydrol. 2019, 572, 261–273. [Google Scholar] [CrossRef]
  66. Bergstra, J.; Bengio, Y. Random search for hyper-parameter optimization. J. Mach. Learn. Res. 2012, 13, 281–305. [Google Scholar]
  67. Falkner, S.; Klein, A.; Hutter, F. BOHB: Robust and efficient hyperparameter optimization at scale. In Proceedings of the 35th International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018. [Google Scholar]
  68. Gelbart, M.; Snoek, J.; Adams, R. Bayesian optimization with unknown constraints. arXiv 2014, arXiv:1403.5607. [Google Scholar]
  69. Jamieson, K.; Talwalkar, A. Non-stochastic best arm identification and hyperparameter optimization. In Proceedings of the 19th International Conference on Artificial Intelligence and Statistics, Cadiz, Spain, 9–11 May 2016. [Google Scholar]
  70. Karnin, Z.; Koren, T.; Somekh, O. Almost optimal exploration in multi-armed bandits. In Proceedings of the 30th International Conference on Machine Learning, ICML 2013, Atlanta, GA, USA, 16–21 June 2013; pp. 2275–2283. [Google Scholar]
  71. Gupta, H.V.; Sorooshian, S.; Yapo, P.O. Status of automatic calibration for hydrologic models: Comparison with multi-level expert calibration. J. Hydrol. Eng. 1999, 4, 135–143. [Google Scholar] [CrossRef]
  72. Moriasi, D.N.; Arnold, J.G.; Van Liew, M.W.; Bingner, R.L.; Harmel, R.D.; Veith, T.L. Model evaluation guidelines for systematic quantification of accuracy in watershed simulations. Trans. ASABE 2007, 50, 885–900. [Google Scholar] [CrossRef]
  73. Willmott, C.J. On the validation of models. Phys. Geogr. 1981, 2, 184–194. [Google Scholar] [CrossRef]
  74. Legates, D.R.; McCabe, G.J., Jr. Evaluating the use of “goodness-of-fit” measures in hydrologic and hydroclimatic model validation. Water Resour. Res. 1999, 35, 233–241. [Google Scholar] [CrossRef]
  75. Gupta, H.V.; Kling, H.; Yilmaz, K.K.; Martinez, G.F. Decomposition of the mean squared error and NSE performance criteria: Implications for improving hydrological modelling. J. Hydrol. 2009, 377, 80–91. [Google Scholar] [CrossRef]
  76. Kling, H.; Fuchs, M.; Paulin, M. Runoff conditions in the upper Danube basin under an ensemble of climate change scenarios. J. Hydrol. 2012, 424–425, 264–277. [Google Scholar] [CrossRef]
  77. Heinemann, A.; Oort, P.; Fernandes, D.; Maia, A. Sensitivity of APSIM/ORYZA model due to estimation errors in solar radiation. Bragantia 2011, 71, 572–582. [Google Scholar] [CrossRef]
  78. Li, M.-F.; Tang, X.-P.; Wu, W.; Liu, H.-B. General models for estimating daily global solar radiation for different solar radiation zones in mainland China. Energy Convers. Manag. 2013, 70, 139–148. [Google Scholar] [CrossRef]
  79. Rindskopf, D.; Shiyko, M. Measures of dispersion, skewness and kurtosis. In International Encyclopedia of Education, 3rd ed.; Peterson, P., Baker, E., McGaw, B., Eds.; Elsevier: Oxford, UK, 2010; pp. 267–273. [Google Scholar] [CrossRef]
  80. Pal, R. Validation Methodologies. In Predictive Modeling of Drug Sensitivity; Elsevier: Amsterdam, The Netherlands, 2017; pp. 83–107. [Google Scholar] [CrossRef]
  81. Elbeltagi, A.; Deng, J.; Wang, K.; Malik, A.; Maroufpoor, S. Modeling long-term dynamics of crop evapotranspiration using deep learning in a semi-arid environment. Agric. Water Manag. 2020, 241, 106334. [Google Scholar] [CrossRef]
  82. Kirch, W. Pearson’s correlation coefficient. In Encyclopedia of Public Health; Kirch, W., Ed.; Springer: Dordrecht, The Netherlands, 2008; pp. 1090–1091. [Google Scholar] [CrossRef]
  83. Nash, J.E.; Sutcliffe, J.V. River flow forecasting through conceptual models part I—A discussion of principles. J. Hydrol. 1970, 10, 282–290. [Google Scholar] [CrossRef]
  84. Hyndman, R.J.; Koehler, A.B. Another look at measures of forecast accuracy. Int. J. Forecast. 2006, 22, 679–688. [Google Scholar] [CrossRef]
  85. Pham-Gia, T.; Hung, T.L. The mean and median absolute deviations. Math. Comput. Model. 2001, 34, 921–936. [Google Scholar] [CrossRef]
  86. Pledger, S. Unified maximum likelihood estimates for closed capture–recapture models using mixtures. Biometrics 2000, 56, 434–442. [Google Scholar] [CrossRef] [PubMed]
  87. Sorooshian, S.; Duan, Q.; Gupta, V.K. Calibration of rainfall-runoff models: Application of global optimization to the Sacramento Soil Moisture Accounting Model. Water Resour. Res. 1993, 29, 1185–1194. [Google Scholar] [CrossRef]
  88. Yapo, P.O.; Gupta, H.V.; Sorooshian, S. Automatic calibration of conceptual rainfall-runoff models: Sensitivity to cali-bration data. J. Hydrol. 1996, 181, 23–48. [Google Scholar] [CrossRef]
  89. Barzegar, R.; Fijani, E.; Asghari Moghaddam, A.; Tziritis, E. Forecasting of groundwater level fluctuations using en-semble hybrid multi-wavelet neural network-based models. Sci. Total Environ. 2017, 599–600, 20–31. [Google Scholar] [CrossRef] [PubMed]
  90. Roy, D.K.; Datta, B. Multivariate adaptive regression spline ensembles for management of multilayered coastal aqui-fers. J. Hydrol. Eng. 2017, 22, 4017031. [Google Scholar] [CrossRef]
  91. Müller, J.; Piché, R. Mixture surrogate models based on Dempster-Shafer theory for global optimization problems. J. Glob. Optim. 2011, 51, 79–104. [Google Scholar] [CrossRef]
Figure 1. Study area indicating the positioning of the weather stations.
Figure 1. Study area indicating the positioning of the weather stations.
Agriculture 14 00278 g001
Figure 2. PACF plots of the T m a x and T m i n data.
Figure 2. PACF plots of the T m a x and T m i n data.
Agriculture 14 00278 g002
Figure 3. Flow diagram of the proposed automatic model selection scheme.
Figure 3. Flow diagram of the proposed automatic model selection scheme.
Agriculture 14 00278 g003
Figure 4. Performance of the best models on test dataset in terms of forecasting maximum temperatures ( T m a x ) under various forecast horizons at the weather stations.
Figure 4. Performance of the best models on test dataset in terms of forecasting maximum temperatures ( T m a x ) under various forecast horizons at the weather stations.
Agriculture 14 00278 g004
Figure 5. Performance of the best models on test dataset to forecast minimum temperatures ( T m i n ) under various forecast horizons at the weather stations.
Figure 5. Performance of the best models on test dataset to forecast minimum temperatures ( T m i n ) under various forecast horizons at the weather stations.
Agriculture 14 00278 g005
Table 1. Descriptive statistics of the daily maximum and minimum temperatures at the three stations (Barishal, Gazipur, and Ishurdi stations), Bangladesh.
Table 1. Descriptive statistics of the daily maximum and minimum temperatures at the three stations (Barishal, Gazipur, and Ishurdi stations), Bangladesh.
VariablesMeanMedianModeStandard DeviationSkewnessKurtosis
Barishal station
Minimum temperature, °C21.56723.6026.005.324−0.730−0.779
Maximum temperature, °C30.42031.2032.003.907−0.584−0.030
Gazipur station
Minimum temperature, °C21.20123.0026.005.628−0.624−0.859
Maximum temperature, °C30.97532.0034.003.947−1.0621.849
Ishurdi station
Minimum temperature, °C21.37423.5027.005.984−0.735−0.761
Maximum temperature, °C31.46332.6034.004.163−0.8270.284
Table 2. Machine learning algorithms and their tunable hyperparameters.
Table 2. Machine learning algorithms and their tunable hyperparameters.
Machine Learning AlgorithmHyperparameters
Ensemble Regression (ER) ModelMethod (least-squares boosting, bootstrap aggregation), number of ensemble learning cycles, learning rate for shrinkage, minimum number of leaf node observations, maximal number of decision splits, number of predictors to select at random for each split
Gaussian Process Regression (GPR) ModelSigma, basis function, kernel function, kernel scale, kernel parameters, standardization
Kernel Regression (KR) ModelEpsilon, kernel scale, Lambda, learner, number of dimensions of expanded space
Linear Regression (LR) Model for High-Dimensional DataLambda, learner, regularization
Artificial Neural Network (ANN) ModelActivations, Lambda, layer sizes, standardization, layer bias initializer, layer weight initializer
Support Vector Machine (SVM) Regression ModelBox constraint, Epsilon, kernel scale, kernel function, polynomial order, standardization
Binary Decision Regression Tree (BDRT)Minimum number of leaf node observations, maximal number of decision splits
Table 3. Comparison of the Bayesian and ASHA optimization algorithm-tuned best models for forecasting maximum temperatures ( T m a x ) at the weather stations. Boldface indicates models with lowest log (1 + valLoss) values.
Table 3. Comparison of the Bayesian and ASHA optimization algorithm-tuned best models for forecasting maximum temperatures ( T m a x ) at the weather stations. Boldface indicates models with lowest log (1 + valLoss) values.
Forecast HorizonWeather StationBayesian OptimizationASHA Optimization
ModelError (log (1 + valLoss))
(°C)
Training Time
(s)
ModelError (log (1 + valLoss))
(°C)
Training Time
(s)
OneBarishalGPR1.5337609GPR1.529525
GazipurEnsemble1.56513,480Ensemble1.5681249
IshurdiEnsemble1.4154015GPR1.55256
TwoBarishalGPR1.7336547GPR1.803408
GazipurGPR1.77712,787Ensemble1.7821865
IshurdiGPR1.6054774GPR1.64899
ThreeBarishalGPR1.8546803GPR1.904225
GazipurEnsemble1.86415,197Ensemble1.875805
IshurdiGPR1.6814458Ensemble1.73047
FourBarishalGPR1.9131277GPR1.919139
GazipurGPR1.89919,626GPR1.9201242
IshurdiGPR1.7294023GPR1.709136
FiveBarishalGPR1.9467056GPR1.949474
GazipurGPR1.9366553GPR1.9571512
IshurdiGPR1.7274022GPR1.845169
valLoss is the cross-validation MSE, GPR is the Gaussian Process Regression model, and Ensemble is the Ensemble Regression model (Ensemble of Regression Trees).
Table 4. Comparison of the training and testing performance of the Bayesian and ASHA optimization-tuned best models to forecast maximum temperatures ( T m a x ) at the weather stations. Boldface indicates lowest difference between the training and test RMSE values.
Table 4. Comparison of the training and testing performance of the Bayesian and ASHA optimization-tuned best models to forecast maximum temperatures ( T m a x ) at the weather stations. Boldface indicates lowest difference between the training and test RMSE values.
Weather StationForecasting HorizonBayesian OptimizationASHA Optimization
ModelTrain RMSE (°C)Test RMSE (°C)ModelTrain RMSE (°C)Test RMSE (°C)
BarishalOneGPR1.61211.9047GPR0.00071.8889
TwoGPR0.00062.3676GPR0.01502.3149
ThreeGPR0.05152.5895GPR0.40222.6621
FourGPR0.00062.7271GPR0.00062.6808
FiveGPR0.00212.8052GPR0.00062.7171
GazipurOneEnsemble1.39221.9439Ensemble1.85401.9448
TwoGPR0.00072.2239Ensemble1.39292.2227
ThreeEnsemble1.70472.3472Ensemble1.61462.3694
FourGPR0.00062.4826GPR2.16992.4381
FiveGPR0.00072.4981GPR0.00122.4543
IshurdiOneEnsemble0.99255.3209GPR0.00112.1540
TwoGPR0.00081.9553GPR0.00061.9039
ThreeGPR0.15432.1110Ensemble1.67682.1259
FourGPR0.00062.1952GPR0.00062.1799
FiveGPR0.00112.2693GPR0.02732.2214
GPR is the Gaussian Process Regression model, and Ensemble is the Ensemble Regression model (Ensemble of Regression Trees).
Table 5. Comparison of the Bayesian and ASHA optimization algorithm-tuned best models for forecasting minimum temperatures ( T m i n ) at the weather stations. Boldface indicates models with lowest log (1 + valLoss) values.
Table 5. Comparison of the Bayesian and ASHA optimization algorithm-tuned best models for forecasting minimum temperatures ( T m i n ) at the weather stations. Boldface indicates models with lowest log (1 + valLoss) values.
Forecast HorizonWeather StationBayesian OptimizationASHA Optimization
ModelError (log (1 + valLoss))
(°C)
Training Time
(s)
ModelError (log (1 + valLoss))
(°C)
Training Time
(s)
OneBarishalGPR1.0246234GPR1.040368
GazipurEnsemble1.36711,467GPR1.385978
IshurdiEnsemble1.4404476LR1.450591
TwoBarishalGPR1.2627501GPR1.236395
GazipurGPR1.57612,662GPR1.5821728
IshurdiGPR1.6504297Ensemble1.73295
ThreeBarishalGPR1.3097376GPR1.409418
GazipurGPR1.67516,816GPR1.7121121
IshurdiGPR1.7243824Ensemble1.89341
FourBarishalGPR1.3428320GPR1.511388
GazipurGPR1.72917,931GPR1.7451176
IshurdiGPR1.8154494Ensemble1.95748
FiveBarishalGPR1.3616991GPR1.428315
GazipurGPR1.78027,554GPR1.8241021
IshurdiGPR1.8764056GPR2.03122
valLoss is the cross-validation MSE, GPR is the Gaussian Process Regression model, Ensemble is the Ensemble Regression model (Ensemble of Regression Trees), and LR is the linear regression model for high-dimensional data.
Table 6. Comparison of the training and testing performance of the Bayesian and ASHA optimization-tuned best models to forecast minimum temperatures ( T m i n ) at the weather stations. Boldface indicates lowest difference between the training and test RMSE values.
Table 6. Comparison of the training and testing performance of the Bayesian and ASHA optimization-tuned best models to forecast minimum temperatures ( T m i n ) at the weather stations. Boldface indicates lowest difference between the training and test RMSE values.
Weather StationForecasting HorizonBayesian OptimizationASHA Optimization
ModelTrain RMSE (°C)Test RMSE (°C)ModelTrain RMSE (°C)Test RMSE (°C)
BarishalOneGPR1.30761.3847GPR0.02221.4450
TwoGPR0.00191.9051GPR0.00271.8921
ThreeGPR0.11172.0781GPR0.00152.0474
FourGPR0.00202.1936GPR1.82082.1215
FiveGPR0.07582.3109GPR0.00142.2469
GazipurOneEnsemble1.70951.6419GPR0.00341.7077
TwoGPR1.34171.9519GPR0.00161.9545
ThreeGPR0.97982.1314GPR0.08062.1278
FourGPR0.00182.2215GPR0.00212.2187
FiveGPR0.00212.2384GPR0.00252.2151
IshurdiOneEnsemble1.48511.6217Linear R1.76321.6510
TwoGPR0.00181.8667Ensemble1.53281.8879
ThreeGPR0.05102.0699Ensemble1.24742.0755
FourGPR0.00152.2289Ensemble2.03232.2240
FiveGPR0.00152.2337GPR1.13692.2228
GPR is the Gaussian Process Regression model, and Ensemble is the Ensemble Regression model (Ensemble of Regression Trees).
Table 7. Selected best models at different weather stations under the five forecasting horizons.
Table 7. Selected best models at different weather stations under the five forecasting horizons.
Forecast HorizonWeather Stations
BarishalGazipurIshurdi
Tmax
OneBayesian-GPRASHA-EnsembleASHA-GPR
TwoASHA-GPRASHA-EnsembleASHA-GPR
ThreeASHA-GPRBayesian-EnsembleASHA-Ensemble
FourASHA-GPRASHA-GPRASHA-GPR
FiveBayesian-GPRASHA-GPRASHA-GPR
Tmin
OneBayesian-GPRBayesian-EnsembleBayesian-Ensemble
TwoASHA-GPRBayesian-GPRASHA-Ensemble
ThreeBayesian-GPRBayesian-GPRASHA -Ensemble
FourASHA-GPRASHA-GPRASHA -Ensemble
FiveBayesian-GPRASHA-GPRBayesian-Ensemble
Bayesian-GPR is the Bayesian optimization algorithm-tuned Gaussian Process Regression model, ASHA-GPR is the asynchronous successive halving algorithm-tuned Gaussian Process Regression model, Bayesian-Ensemble is the Bayesian optimization algorithm-tuned Ensemble Regression model (Ensemble of Regression Trees), ASHA-Ensemble is the asynchronous successive halving algorithm-tuned Ensemble Regression model (Ensemble of Regression Trees).
Table 8. Performance of the best models on test dataset to forecast maximum temperatures ( T m a x ) under various forecast horizons at the weather stations.
Table 8. Performance of the best models on test dataset to forecast maximum temperatures ( T m a x ) under various forecast horizons at the weather stations.
Forecast
Horizon
Performance Indicators
RMSE (°C)NRMSEMAPRE (%)MAD (°C)MBE (°C)PBIAS (%)
Barishal
One1.9050.0645.0910.7020.0060.021
Two2.3150.0786.3610.8990.0900.303
Three2.6620.0897.3181.029−0.030−0.102
Four2.6810.0907.5121.0300.2310.774
Five2.8050.0947.9541.0700.2080.696
Gazipur
One1.9450.0624.7370.634−0.072−0.230
Two2.2230.0715.7520.811−0.152−0.490
Three2.3470.0756.1160.865−0.179−0.575
Four2.4380.0786.3970.859−0.101−0.326
Five2.4540.0796.4680.897−0.104−0.336
Ishurdi
One2.1540.0695.3790.817−0.173−0.726
Two1.9040.0614.7280.693−0.079−0.333
Three2.1260.0685.2460.783−0.180−0.756
Four2.1800.0705.3980.833−0.105−0.441
Five2.2210.0715.4260.820−0.141−0.591
RMSE is the root mean squared error, NRMSE is the normalized RMSE, MAPRE is the Mean Absolute Percentage Relative Error, MAD is the Median Absolute Deviation, MBE is the Mean Bias Error, and PBIAS is the Percentage Bias.
Table 9. Performance of the best models on test dataset to forecast minimum temperatures ( T m i n ) under various forecast horizons at the weather stations.
Table 9. Performance of the best models on test dataset to forecast minimum temperatures ( T m i n ) under various forecast horizons at the weather stations.
Forecast
Horizon
Performance Indicators
RMSE (°C)NRMSEMAPRE (%)MAD (°C)MBE (°C)PBIAS (%)
Barishal
One1.3850.0685.1680.383−0.046−0.223
Two1.8920.0927.9530.6880.0160.076
Three2.0780.1018.8570.834−0.052−0.255
Four2.1220.1039.2060.915−0.102−0.498
Five2.3110.11210.0060.971−0.049−0.240
Gazipur
One1.6420.0787.0810.505−0.018−0.084
Two1.9520.0938.6880.643−0.080−0.379
Three2.1310.1019.7250.727−0.044−0.209
Four2.2190.10610.1280.831−0.035−0.166
Five2.2150.10510.2210.876−0.018−0.086
Ishurdi
One1.6220.0745.9050.520−0.014−0.082
Two1.8880.0867.1690.661−0.036−0.218
Three2.0760.0957.8350.699−0.068−0.411
Four2.2240.1028.2340.667−0.048−0.288
Five2.2340.1028.3980.736−0.095−0.572
RMSE is the root mean squared error, NRMSE is the normalized RMSE, MAPRE is the Mean Absolute Percentage Relative Error, MAD is the Median Absolute Deviation, MBE is the Mean Bias Error, and PBIAS is the Percentage Bias.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Roy, D.K.; Hossain, M.A.; Haque, M.P.; Alataway, A.; Dewidar, A.Z.; Mattar, M.A. Automated Model Selection Using Bayesian Optimization and the Asynchronous Successive Halving Algorithm for Predicting Daily Minimum and Maximum Temperatures. Agriculture 2024, 14, 278. https://doi.org/10.3390/agriculture14020278

AMA Style

Roy DK, Hossain MA, Haque MP, Alataway A, Dewidar AZ, Mattar MA. Automated Model Selection Using Bayesian Optimization and the Asynchronous Successive Halving Algorithm for Predicting Daily Minimum and Maximum Temperatures. Agriculture. 2024; 14(2):278. https://doi.org/10.3390/agriculture14020278

Chicago/Turabian Style

Roy, Dilip Kumar, Mohamed Anower Hossain, Mohamed Panjarul Haque, Abed Alataway, Ahmed Z. Dewidar, and Mohamed A. Mattar. 2024. "Automated Model Selection Using Bayesian Optimization and the Asynchronous Successive Halving Algorithm for Predicting Daily Minimum and Maximum Temperatures" Agriculture 14, no. 2: 278. https://doi.org/10.3390/agriculture14020278

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop