You are currently viewing a new version of our website. To view the old version click .
Computation
  • Article
  • Open Access

14 March 2025

Short-Term Load Forecasting in Distribution Substation Using Autoencoder and Radial Basis Function Neural Networks: A Case Study in India

,
,
and
1
Center for AI and Deep Learning, School of Computer Science and AI, SR University, Warangal 506371, India
2
Department of Electrical and Electronics Engineering, SR University, Warangal 506371, India
3
Department of Electrical and Electronics Engineering, Kakatiya Institute of Technology and Science (KITS), Warangal 506371, India
4
Department of Railroad and Electrical Engineering, Woosong University, Daejeon 34606, Republic of Korea

Abstract

Electric load forecasting is an essential task for Distribution System Operators in order to achieve proper planning, high integration of small-scale production from renewable energy sources, and to define effective marketing strategies. In this framework, machine learning and data dimensionality reduction techniques can be useful for building more efficient tools for electrical energy load prediction. In this paper, a machine learning model based on a combination of a radial basis function neural network and an autoencoder is used to forecast the electric load on a 33/11 kV substation located in Godishala, Warangal, India. One year of historical data on an electrical substation and weather are considered to assess the effectiveness of the proposed model. The impact of weather, day, and season status on load forecasting is also considered. The input dataset dimensionality is reduced using autoencoder to build a light-weight machine learning model to be deployed on edge devices. The proposed methodology is supported by a comparison with the state of the art based on extensive numerical simulations.

1. Introduction

Utility companies must estimate electrical power consumption in order to reduce operating and maintenance costs, manage demand and supply, increase dependability, properly plan for future investments, and engage in efficient energy trading [1,2,3]. Based on the forecasting time, electrical load forecasting may be broadly classified into four primary categories: extremely short-term and short-term [4,5], medium-term, and long-term load forecasting [6]. Depending on the time horizon, load forecasting can be used for a variety of purposes, including optimal operations, grid stability, demand-side management (DSM), or long-term strategic planning [7].
One of the more difficult problems in a deregulated power system is anticipating short-term active power load. Load consumption patterns are more commonly adjusted as a result of weather, cultural events, and people’s social routines, according to [8]. The active power load in low-voltage distribution systems is very unpredictable, particularly in India, because the majority of customers are residential and commercial customers who are heavily impacted by the considerations stated above. As a result, accurate forecasting of active power load is a crucial responsibility for distribution system operators for proper distribution network operational planning and energy trading. As shown in Table 1, much work has been carried out on the demand-side forecasting problem for accurate prediction using artificial intelligence (AI) techniques.
Table 1. Literature survey on short-term load forecasting.
All the instances of the literature mentioned in Table 1 have made a significant contribution to dealing with short-term electric power load forecasting issues. A new strategy is proposed in this study to increase forecasting accuracy and to design a light-weight model for active power load forecasting applied to a 33/11 kV substation by employing autoencoder (AE) for dimensionality reduction and a Radial Basis Function Neural Network (RBFNN) for load forecasting.
The major contributions and constraints considered in this paper can be summarized as follows:
  • Electric power is forecasted one hour ahead by considering the status of the day, i.e., weekend/weekday. As the 33/11 kV substation is located in Godishala, Warangal, where all offices and colleges are closed on Sunday, only Sunday is considered the weekend.
  • Electric power is forecasted one hour ahead by considering the season. In India, there are three seasons, i.e., Winter (November–February), Summer (March–June), and Rainy (July–October).
  • Electric power is forecasted one hour ahead by considering the weather condition in terms of temperature and humidity.
  • Electric power is forecasted one hour ahead by considering the last four hours of load data, i.e., P(T-1), P(T-2), P(T-3), and P(T-4), and by considering the load of the last four days, i.e., P(T-24), P(T-48), P(T-72), and P(T-96).
  • A new optimal autoencoder architecture is developed to reduce the dimensions of the dataset from 8664 × 12 to 8664 × 9.
  • A new optimal RBFNN architecture is developed to forecast the load using the compressed data.
The remainder of the paper is organized as follows: Section 2 demonstrates methodology developed to forecast the load and to reduce the dimensions of the dataset; Section 3 presents simulation results; and in Section 4, some conclusions are drawn.

2. Materials and Methods

This section presents the detailed procedure we used for dimensionality reduction of the input dataset using AE for forecasting temperature, load, and humidity using RBFNN. The complete architecture for forecasting the load one hour ahead for effective trading an hour ahead in the energy market is shown in Figure 1. RBFNN and Autoencoder models are implemented in Google Colab with Python programming language having version 3.10.
Figure 1. Short-term load forecasting architecture.

2.1. Active Power Load Dataset

The active power load dataset based on load data was collected from a 33/11 kV substation located in Godishala, Warangal, India, during the period from January 2021 to December 2021. In order to train and test the AE+RBFNN model, normalized data [25] were used and generated based on Equation (1). In this paper, forecasting the load depends upon previous load samples. Hence, the last four hours and last four days prior to the time of forecasting are considered. Also, load consumption varies based on season and day status (weekend/weekday) as well as temperature and humidity. In the region of Warangal, India, the summer temperature reaches 47 °C; in winter, it reaches 22 °C. These values indicate high impact load consumption. The data were collected during the period of January 2021 to December 2021, covering all weather conditions during the year and all temperature and humidity changes throughout the year. As our models were trained with these data, they are generalized across all types of load variations and seasonal variations
X N o r m = x x m i n x m a x x m i n

2.2. Dimensionality Reduction Using Autoencoder (AE)

Autoencoder (AE) is a deep neural network that is used to reduce the dimensionality of input data and reconstruct the original data from the compressed representation. It has three parts: an encoder, latent space, and decoder. The encoder compresses the input data into a lower dimension in the latent space, whereas the decoder reconstructs the original data from lower-dimension compressed data available in the latent space [26]. In this paper, AE is used to convert higher-dimension original data of volume 8664 × 12 into lower-dimension compressed data of volume 8664 × 9. The complete architecture of AE with a ReLu activation function [27] in the latent space is shown in Figure 2.
Figure 2. AE model architecture [28].
The architecture of the encoder part of the autoencoder, which provides lower-dimension data from higher-dimension original data, is shown in Figure 3. The data are compressed to 10 features from 12, while the fourth neuron in the latent space gives a zero value for all samples in the original dataset. Hence, the feature corresponding to the fourth neuron is removed, and the compressed data only have 9 features. The optimal AE model is identified by tuning hyperparameters like the number of hidden layers and neurons minimizing the following cost value:
C = 0.5 · NVL + 0.5 · NVD
Figure 3. Encoder architecture.
The normalized validation loss (NVL) is defined by the following equation:
NVL = VL VL m i n VL m a x VL m i n
where VL m i n and VL m a x are the minimum and maximum validation loss among all observed architectures, and NVD is the normalized variance deviation of the compressed dataset from the original dataset.
The validation loss (VL) is computed by
VL = 1 n s a m p l e t e s t s a m p l e = 1 n s a m p l e t e s t ( d o ) 2
where d and o represent the desired and actual outputs, respectively.

2.3. Radial Basis Function Neural Network (RBFNN)

RBFNN [29] is a machine learning model that can be used for both regression and classification problems. It has three layers: input, hidden, and output layers [30,31]. Unlike simple artificial neural networks, it uses both unsupervised and supervised learning strategies. Unsupervised learning, i.e., competitive learning [32] based on Euclidean distance (also called the k-means clustering algorithm) is used between the input and hidden layer, whereas supervised learning based on the back-propagation algorithm [33] is used between hidden and output layers. The Gaussian function [34,35] shown in Equation (5) is used as an activation function in the hidden layer and the linear activation function is used in the output layer. The optimal RBFNN architecture is identified based on validation loss shown in Equation (4) by tuning the hyperparameters of the RBFNN model as mentioned in [28]:
f ( x ) = e ( x μ ) 2 2 σ 2
The advantages of RBFNN [36,37] for short-term laod forecasting are as indicated below:
  • Electricity load patterns are highly nonlinear and depend on multiple factors (e.g., time, temperature, holidays). RBFNN uses Gaussian radial basis functions that allow it to approximate complex functions more effectively than simple linear models.
  • Unlike deep learning models (e.g., LSTMs, CNNs), RBFNNs require fewer training epochs because they rely on localized activation functions.
  • Traditional RBFNNs rely on raw input features, which may contain noise or irrelevant information.
In this paper, the optimal architecture for RBFNN for temperature forecasting is shown in Figure 4; for humidity forecasting, it is shown in Figure 5. In this paper, RBFNN is used to forecast the load for a 33/11 kV substation based on features extracted by AE from original features. Temperature and humidity are the input features for AE and the result that comes from the output neuron of the RBFNN model is used for temperature and load forecasting, respectively, as shown in Figure 1. The optimal architecture of RBFNN for active power load forecasting is shown in Figure 6. The complete training algorithm for RBFNN is taken from [38].
Figure 4. RBFNN is used to forecast temperature.
Figure 5. RBFNN is used to forecast humidity.
Figure 6. RBFNN is used to forecast active power load for 33/11 kV substation.
The combination of AE and RBFNN is useful to extract strengths from each model and to develop a better model [39,40]. The features of each individual model, AE and RBFNN, and the combined hybrid model are presented in Table 2.
Table 2. Features of RBFNN, AE, and AE + RBFNN models.

3. Results

To train and validate the RBFNN model for load forecasting and the AE for dimensionality reduction, historical load data were collected over the period 1 January to 31 December 2021 from a 33/11 kV substation in Godishala, Telangana State, India. This load dataset comprises 8760 total active power load samples (24 h × 365 days = 8760) rearranged into an 8664 × 12 matrix. This weather dataset was prepared based on the temperature and humidity data available at Weather Data-Online, available at https://www.wunderground.com/history/daily/in/kazipet/VOWA/date/2021-1-2 (accessed on 29 January 2022).The datasets were split 70%:30% for training and testing, respectively. Based on this ratio, the shape of the training and testing data for each model is shown in Table 3. The following procedure is used for the active power load forecasting.
Table 3. Training and testing data size.
  • Dimensionality reduction using AE;
  • Temperature forecasting using RBFNN;
  • Humidity forecasting using RBFNN;
  • Load forecasting using RBFNN;
  • Web application development.

3.1. Data Insights

The stochastic features of the initial dataset that is used to forecast the load are shown in Table 4.
Table 4. Statistical information of original dataset.

3.2. Optimal AE for Dimensionality Reduction

In order to identify the optimal AE model, the AE model was trained and validated with various numbers of neurons in the hidden layer and various batch sizes. The mean square errors on the training and testing data are reported in Table 5.
Table 5. Error metric during AE training.
Among all the trained architectures with the minimum objective function, Equation (2) is considered the optimal architecture. From Table 6, it can be identified that the AE’s having ten neurons in the hidden layer results in a lower cost function value, i.e., 0.0061; hence, this is considered the optimal AE architecture for developing the compressed dataset from the original dataset to properly forecast the load.
Table 6. Reconstructed vs. original data.
The statistical parameters of compressed data for load forecasting are presented in Table 7. Among all features, feature X4 has only zero values; hence, this feature was removed from the compressed data, and the size of the data became 8664 × 9.
Table 7. Statistical information of compressed dataset.
The proposed autoencoder approach is compared with principal component analysis (PCA) by considering a total of 10 principal components. A comparision between PCA [41] and the autoencoder approach in terms of the variance gap from the original data is shown below in Table 8. From Table 8, it can be observed that autoencoder provides data closer to the original data than the data produced by PCA, which resulted in 33% data reduction.
Table 8. Comparison between AE and PCA.
Optimal RBFNN architectures with minimum validation errors are designed by tuning hyper-parameters like the number of centroids and the width parameter to forecast the temperature and humidity as presented in Table 9 for temperature and in Table 10 for humidity. From Table 9, the RBFNN model with 12 neurons in the hidden layer and a width parameter of 11 had a lower validation error (0.0022)and training error (0.0037) and thus was considered the optimal model to estimate temperature. Similarly, in Table 10, the RBFNN model with 12 neurons in the hidden layer and a width parameter of 10 had a lower validation error, 0.0078, and training error, 0.011, and thus was considered the optimal model to forecast the humidity parameter.
Table 9. Validation error of RBFNN for temperature estimation.
Table 10. Validation error of RBFNN for humidity estimation.

3.3. Optimal RBFNN Model for Forecasting Active Power Load

The optimal RBFNN architecture with the minimum validation error is designed by tuning hyper-parameters like the number of centroids and the width parameter to forecast the load on the 33/11 kV substation, as presented in Table 11. From Table 11, the RBFNN model with 19 neurons in the hidden layer and a width parameter of 12 had a lower validation error (0.0065) and corresponding training error (0.0119) as shown in Table 12; the model was thus considered optimal for forecasting load.
Table 11. RBFNN validation loss for active power load forecasting.
Table 12. RBFNN training loss for active power load forecasting.
The RBFNN model with 19 centroids and 12 parameters was trained with different batch sizes, i.e., 8, 16, and 32. The converging characteristics of the RBFNN architecture for various batch sizes—8, 16, and 32—are shown in Figure 7, Figure 8 and Figure 9, respectively. According to observations, the RBFNN model that was trained with batch size 32 had a minimal validation loss of 0.0065. Additionally, there was reduced variance between the errors in the training and validation phases, indicating that the model was well trained and free of under- and over-fit issues.
Figure 7. Training and validation plots of RBFNN for load forecasting with batch size 8.
Figure 8. Training and validation plots of RBFNN for load forecasting with batch size 16.
Figure 9. Training and validation plots of RBFNN for load forecasting with batch size 32.
Thus, a web application for forecasting the load associated with the given 33/11 kV substation located in Godishala, Warangal, India, was developed and integrated by running the trained AE-RBFNN combination on the backend; this web application is shown in Figure 10.
Figure 10. Web application to forecast load.

3.4. Comparative Analysis

A comparison between the actual load in the training and testing data with the forecast load using the combined AE and RBFNN model is shown in Figure 11 and Figure 12, respectively. It can be observed that the predicted load using our combined AE and RBFNN model is close to the actual load available in the training and testing datasets for most of the samples.
Figure 11. Comparison with actual load available on training dataset.
Figure 12. Comparison with actual load available on testing dataset.
The proposed combined AE and RBFNN model for forecasting the load at the given 33/11 kV substation is compared with the RBFNN model without AE in terms of training and validation loss as shown in Figure 13. From Figure 13, it can be observed that the combined AE and RBFNN model performs better than the no-AE RBFNN model, showing less training and a lower validation loss.
Figure 13. AE + RBFNN vs. RBFNSS.
A comparison between the actual load and the load forecasted using the combined AE and RBFNN model is shown in Figure 14, where it can be observed that the load predicted using the combined AE and RBFNN model was closer to the actual load than the load predicted using the only-RBFNN model. A sudden and abnormal deviation of load patterns occurs between the 7th hour and the 16th hour, certainly due to the planning of New Year (2024) celebrations in Warangal, India.
Figure 14. Comparison with actual load on 31 December 2021.
The proposed combined AE and RBFNN model for forecasting the load at the given 33/11 kV substation was validated by comparing it with a regression tree model [42] as shown in Figure 15. Here, it is observed that the regression tree model results in a higher testing loss than the proposed model, while also suffering from over-fitting.
Figure 15. AE + RBFNN vs. regression tree.
The proposed model is also compared in terms of testing loss with the models developed in [21,24] as mentioned in Table 1; the comparison is presented in Table 13. From Table 13, it can be observed that the proposed model predicts the load with the lowest testing loss in comparison with the other models considered.
Table 13. Validation of AE + RBFNN model.

4. Conclusions

In this study, a combined AE and RBFNN model was designed to predict the active power load for energy markets one hour ahead of time. AE was utilized as a dimensionality reduction approach to building our light-weight RBFNN model. The AE derives the new features from the original input data. The complexity of the RBFNN model decreases as the number of features retrieved through the AE decreases when compared to the original input features. With respect to the RBFNN model alone, the combined AE and RBFNN model predicts load with higher accuracy. The developed combined AE-RBFNN model was validated by comparing it to the regression tree model in terms of mean square error value, demonstrating that the suggested model is capable of predicting the load one hour in advance with a reduced mean square error value. Temperature and humidity are the two main input factors affecting the load forecasting model; both input variables were forecasted individually using RBFNN. Finally, a web-based application was created to forecast load by executing the suggested combined AE and RBFNN model in the backend. Future studies can extend the physical model, considering more external parameters and exploring different neural structures with respect to different application domains.

Author Contributions

Conceptualization, V.V. and P.K.K.; methodology, V.V. and P.K.K.; software, P.K.K.; validation, V.V. and P.K.K.; formal analysis, V.V.; investigation, V.V.; resources, V.V., P.K.K. and R.C.D.; data curation, V.V. and P.K.K.; writing—original draft preparation, V.V.; writing—review and editing, S.R.S.; visualization, V.V.; supervision, V.V.; project administration, V.V. All authors have read and agreed to the published version of the manuscript.

Funding

Woosong University’s Academic Research Funding—2025.

Data Availability Statement

The active power load data used to train and test the relevant neural network models are available at https://data.mendeley.com/datasets/7vdt5rz47x/1 (accessed on 2 February 2023).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Almeshaiei, E.; Soltan, H. A methodology for electric power load forecasting. Alex. Eng. J. 2011, 50, 137–144. [Google Scholar] [CrossRef]
  2. Vanting, N.B.; Ma, Z.; Jørgensen, B.N. A scoping review of deep neural networks for electric load forecasting. Energy Inform. 2021, 4, 49. [Google Scholar] [CrossRef]
  3. Giglio, E.; Luzzani, G.; Terranova, V.; Trivigno, G.; Niccolai, A.; Grimaccia, F. An efficient artificial intelligence energy management system for urban building integrating photovoltaic and storage. IEEE Access 2023, 11, 18673–18688. [Google Scholar] [CrossRef]
  4. Mansoor, M.; Grimaccia, F.; Leva, S.; Mussetta, M. Comparison of echo state network and feed-forward neural networks in electrical load forecasting for demand response programs. Math. Comput. Simul. 2021, 184, 282–293. [Google Scholar] [CrossRef]
  5. Dagdougui, H.; Bagheri, F.; Le, H.; Dessaint, L. Neural network model for short-term and very-short-term load forecasting in district buildings. Energy Build. 2019, 203, 109408. [Google Scholar] [CrossRef]
  6. Su, P.; Tian, X.; Wang, Y.; Deng, S.; Zhao, J.; An, Q.; Wang, Y. Recent trends in load forecasting technology for the operation optimization of distributed energy system. Energies 2017, 10, 1303. [Google Scholar] [CrossRef]
  7. Neusser, L.; Canha, L.N. Real-time load forecasting for demand side management with only a few days of history available. In Proceedings of the IEEE 4th International Conference on Power Engineering, Energy and Electrical Drives, Istanbul, Turkey, 13–17 May 2013; pp. 911–914. [Google Scholar]
  8. Jahromi, A.J.; Mohammadi, M.; Afrasiabi, S.; Afrasiabi, M.; Aghaei, J. Probability density function forecasting of residential electric vehicles charging profile. Appl. Energy 2022, 323, 119616. [Google Scholar] [CrossRef]
  9. Massaoudi, M.; S. Refaat, S.; Abu-Rub, H.; Chihi, I.; Oueslati, F.S. PLS-CNN-BiLSTM: An end-to-end algorithm-based Savitzky–Golay smoothing and evolution strategy for load forecasting. Energies 2020, 13, 5464. [Google Scholar] [CrossRef]
  10. Yin, L.; Xie, J. Multi-temporal-spatial-scale temporal convolution network for short-term load forecasting of power systems. Appl. Energy 2021, 283, 116328. [Google Scholar] [CrossRef]
  11. Syed, D.; Abu-Rub, H.; Ghrayeb, A.; Refaat, S.S.; Houchati, M.; Bouhali, O.; Bañales, S. Deep learning-based short-term load forecasting approach in smart grid with clustering and consumption pattern recognition. IEEE Access 2021, 9, 54992–55008. [Google Scholar] [CrossRef]
  12. Munkhammar, J.; van der Meer, D.; Widén, J. Very short term load forecasting of residential electricity consumption using the Markov-chain mixture distribution (MCM) model. Appl. Energy 2021, 282, 116180. [Google Scholar] [CrossRef]
  13. Guo, W.; Che, L.; Shahidehpour, M.; Wan, X. Machine-Learning based methods in short-term load forecasting. Electr. J. 2021, 34, 106884. [Google Scholar] [CrossRef]
  14. Timur, O.; Üstünel, H.Y. Short-Term Electric Load Forecasting for an Industrial Plant Using Machine Learning-Based Algorithms. Energies 2025, 18, 1144. [Google Scholar] [CrossRef]
  15. Eskandari, H.; Imani, M.; Moghaddam, M.P. Convolutional and recurrent neural network based model for short-term load forecasting. Electr. Power Syst. Res. 2021, 195, 107173. [Google Scholar] [CrossRef]
  16. Guo, W.; Liu, S.; Weng, L.; Liang, X. Power Grid Load Forecasting Using a CNN-LSTM Network Based on a Multi-Modal Attention Mechanism. Appl. Sci. 2025, 15, 2435. [Google Scholar] [CrossRef]
  17. Perçuku, A.; Minkovska, D.; Hinov, N. Enhancing Electricity Load Forecasting with Machine Learning and Deep Learning. Technologies 2025, 13, 59. [Google Scholar] [CrossRef]
  18. Cao, W.; Liu, H.; Zhang, X.; Zeng, Y.; Ling, X. Short-Term Residential Load Forecasting Based on the Fusion of Customer Load Uncertainty Feature Extraction and Meteorological Factors. Sustainability 2025, 17, 1033. [Google Scholar] [CrossRef]
  19. Sheng, Z.; Wang, H.; Chen, G.; Zhou, B.; Sun, J. Convolutional residual network to short-term load forecasting. Appl. Intell. 2021, 51, 2485–2499. [Google Scholar] [CrossRef]
  20. Hong, T.; Pinson, P.; Fan, S.; Zareipour, H.; Troccoli, A.; Hyndman, R.J. Probabilistic energy forecasting: Global energy forecasting competition 2014 and beyond. Int. J. Forecast. 2016, 32, 896–913. [Google Scholar] [CrossRef]
  21. Veeramsetty, V.; Mohnot, A.; Singal, G.; Salkuti, S.R. Short term active power load prediction on a 33/11 kv substation using regression models. Energies 2021, 14, 2981. [Google Scholar] [CrossRef]
  22. Li, B.; Liao, Y.; Liu, S.; Liu, C.; Wu, Z. Research on Short-Term Load Forecasting of LSTM Regional Power Grid Based on Multi-Source Parameter Coupling. Energies 2025, 18, 516. [Google Scholar] [CrossRef]
  23. Zhao, Y.; Li, J.; Chen, C.; Guan, Q. A Diffusion–Attention-Enhanced Temporal (DATE-TM) Model: A Multi-Feature-Driven Model for Very-Short-Term Household Load Forecasting. Energies 2025, 18, 486. [Google Scholar] [CrossRef]
  24. Veeramsetty, V.; Sai Pavan Kumar, M.; Salkuti, S.R. Platform-Independent Web Application for Short-Term Electric Power Load Forecasting on 33/11 kV Substation Using Regression Tree. Computers 2022, 11, 119. [Google Scholar] [CrossRef]
  25. Abumohsen, M.; Owda, A.Y.; Owda, M. Electrical load forecasting using LSTM, GRU, and RNN algorithms. Energies 2023, 16, 2283. [Google Scholar] [CrossRef]
  26. Subray, S.; Tschimben, S.; Gifford, K. Towards Enhancing Spectrum Sensing: Signal Classification Using Autoencoders. IEEE Access 2021, 9, 82288–82299. [Google Scholar] [CrossRef]
  27. Atienza, R. Advanced Deep Learning with Keras: Apply Deep Learning Techniques, Autoencoders, GANs, Variational Autoencoders, Deep Reinforcement Learning, Policy Gradients, and More; Packt Publishing Ltd.: Birmingham, UK, 2018. [Google Scholar]
  28. Veeramsetty, V.; Kiran, P.; Sushma, M.; Babu, A.M.; Rakesh, R.; Raju, K.; Salkuti, S.R. Active power load data dimensionality reduction using autoencoder. In Power Quality in Microgrids: Issues, Challenges and Mitigation Techniques; Springer: Berlin/Heidelberg, Germany, 2023; pp. 471–494. [Google Scholar]
  29. Wurzberger, F.; Schwenker, F. Learning in deep radial basis function networks. Entropy 2024, 26, 368. [Google Scholar] [CrossRef]
  30. Yu, Q.; Hou, Z.; Bu, X.; Yu, Q. RBFNN-based data-driven predictive iterative learning control for nonaffine nonlinear systems. IEEE Trans. Neural Networks Learn. Syst. 2019, 31, 1170–1182. [Google Scholar] [CrossRef]
  31. Liu, J. Radial Basis Function (RBF) Neural Network Control for Mechanical Systems: Design, Analysis and Matlab Simulation; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2013. [Google Scholar]
  32. Lu, Y.; Cheung, Y.M.; Tang, Y.Y. Self-adaptive multiprototype-based competitive learning approach: A k-means-type algorithm for imbalanced data clustering. IEEE Trans. Cybern. 2019, 51, 1598–1612. [Google Scholar] [CrossRef]
  33. Lillicrap, T.P.; Santoro, A.; Marris, L.; Akerman, C.J.; Hinton, G. Backpropagation and the brain. Nat. Rev. Neurosci. 2020, 21, 335–346. [Google Scholar] [CrossRef]
  34. Kandilogiannakis, G.; Mastorocostas, P.; Voulodimos, A.; Hilas, C. Short-Term Load Forecasting of the Greek Power System Using a Dynamic Block-Diagonal Fuzzy Neural Network. Energies 2023, 16, 4227. [Google Scholar] [CrossRef]
  35. Zaini, F.A.; Sulaima, M.F.; Razak, I.A.W.A.; Othman, M.L.; Mokhlis, H. Improved Bacterial Foraging Optimization Algorithm with Machine Learning-Driven Short-Term Electricity Load Forecasting: A Case Study in Peninsular Malaysia. Algorithms 2024, 17, 510. [Google Scholar] [CrossRef]
  36. Ng, W.W.; Xu, S.; Wang, T.; Zhang, S.; Nugent, C. Radial basis function neural network with localized stochastic-sensitive autoencoder for home-based activity recognition. Sensors 2020, 20, 1479. [Google Scholar] [CrossRef]
  37. Yang, Y.; Wang, P.; Gao, X. A novel radial basis function neural network with high generalization performance for nonlinear process modelling. Processes 2022, 10, 140. [Google Scholar] [CrossRef]
  38. Veeramsetty, V.; Kiran, P.; Sushma, M.; Salkuti, S.R. Weather Forecasting Using Radial Basis Function Neural Network in Warangal, India. Urban Sci. 2023, 7, 68. [Google Scholar] [CrossRef]
  39. Lee, C.M.; Ko, C.N. Short-term load forecasting using adaptive annealing learning algorithm based reinforcement neural network. Energies 2016, 9, 987. [Google Scholar] [CrossRef]
  40. Liu, P.; Zheng, P.; Chen, Z. Deep learning with stacked denoising auto-encoder for short-term electric load forecasting. Energies 2019, 12, 2445. [Google Scholar] [CrossRef]
  41. Vaccaro Benet, P.; Ugalde-Ramírez, A.; Gómez-Carmona, C.D.; Pino-Ortega, J.; Becerra-Patiño, B.A. Identification of Game Periods and Playing Position Activity Profiles in Elite-Level Beach Soccer Players Through Principal Component Analysis. Sensors 2024, 24, 7708. [Google Scholar] [CrossRef]
  42. Tsionas, M.; Assaf, A.G. Regression trees for hospitality data analysis. Int. J. Contemp. Hosp. Manag. 2023, 35, 2374–2387. [Google Scholar] [CrossRef]
  43. Song, C.Y. A Study on Learning Parameters in Application of Radial Basis Function Neural Network Model to Rotor Blade Design Approximation. Appl. Sci. 2021, 11, 6133. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.