You are currently viewing a new version of our website. To view the old version click .
Processes
  • Article
  • Open Access

19 October 2023

Research of Carbon Emission Prediction: An Oscillatory Particle Swarm Optimization for Long Short-Term Memory

,
,
,
,
,
,
,
,
and
1
Hainan Academy of Forestry, Haikou 571100, China
2
Key Laboratory of Tropical Forestry Resources Monitoring and Application of Hainan Province, Haikou 571100, China
3
Geodetic Data Processing Centre of Mininstry of Natural Resources of the People’s Republic of China, Xi’an 710054, China
4
MNR Key Laboratory for Geo-Environmental Monitoring of Great Bay Area & Guangdong Key Laboratory of Urban Informatics & Shenzhen Key Laboratory of Spatial Smart Sensing and Services & Guangdong-Hong Kong-Macau Joint Laboratory for Smart Cities, Shenzhen University, Shenzhen 518060, China

Abstract

Carbon emissions play a significant role in shaping social policy-making, industrial planning, and other critical areas. Recurrent neural networks (RNNs) serve as the major choice for carbon emission prediction. However, year-frequency carbon emission data always results in overfitting during RNN training. To address this issue, we propose a novel model that combines oscillatory particle swarm optimization (OPSO) with long short-term memory (LSTM). OPSO is employed to fine-tune the hyperparameters of LSTM, utilizing an oscillatory strategy to effectively mitigate overfitting and consequently improve the accuracy of the LSTM model. In validation tests, real data from Hainan Province, encompassing diverse dimensions such as gross domestic product, forest area, and ten other relevant factors, are used. Standard LSTM and PSO-LSTM are selected in the control group. The mean absolute error (MAE), root mean square error (RMSE), and mean absolute percentage error (MAPE) are used to evaluate the performance of these methods. In the test dataset, the MAE of OPSO-LSTM is 117.708, 65.72% better than LSTM and 29.48% better than PSO-LSTM. The RMSE of OPSO-LSTM is 149.939, 68.52% better than LSTM and 41.90% better than PSO-LSTM. The MAPE of OPSO-LSTM is 0.017, 65.31% better than LSTM, 29.17% better than PSO-LSTM. The experimental results prove that OPSO-LSTM can provide reliable predictions for carbon emissions.

1. Introduction

Carbon emission projections (CEPs) are pivotal in achieving dual carbon goals, and they will play a critical role in driving future low-carbon and environmentally friendly development within the local economy and society. CEPs also serve as a valuable guide for businesses and investors operating in energy, transportation, manufacturing, and various other sectors. Foreseeing future carbon emission levels enables companies and investors to evaluate the environmental risks and opportunities associated with different investment projects and determine the strategic direction for sustainable development. Additionally, some financial institutions and investors utilize carbon emissions projections to assess the carbon footprint and climate-related risks within their portfolios.
Lastly, carbon emission projections empower companies and organizations to more effectively manage their resource and energy consumption. By gaining insight into future carbon emissions, companies can take proactive measures to reduce energy consumption, lower carbon emissions, enhance resource efficiency, and develop sustainable strategies for their business operations and supply chain management.
Recurrent neural networks (RNNs) are a popular choice for time-series data prediction tasks. They are well suited for this purpose due to their inherent ability to process sequential data while maintaining a memory of past information. RNNs excel in capturing temporal dependencies and patterns within time series, making them effective for tasks like stock price forecasting, weather prediction, and natural language processing.
However, traditional RNN algorithms applied to carbon emissions prediction tasks can face several challenges. Firstly, they struggle with capturing long-term dependencies in the data, hindering their ability to model extended trends and seasonal patterns. Secondly, these models are prone to overfitting, especially when dealing with complex and high-dimensional datasets. Additionally, selecting appropriate hyperparameters, such as the size of hidden layers and learning rates, can be a challenging task. Lastly, RNN models may be susceptible to getting stuck in local minima during training, making it difficult to achieve global optimization and accurate predictions.
To address these challenges and improve the accuracy of carbon emission prediction, we propose a novel oscillatory particle swarm optimization (OPSO)-based long short-term Memory (LSTM) model. Specifically, this paper makes several key contributions:
Firstly, we introduce the concept of OPSO. OPSO employs an oscillatory strategy to enhance the search capability of traditional particle swarm optimization (PSO) in high-dimensional spaces.
Secondly, we integrate OPSO with the conventional LSTM model. OPSO is employed to fine-tune the hyperparameters of LSTM. Our method effectively addresses issues associated with PSO, such as limited search range, susceptibility to early local optima, and reduced search accuracy in later stages. By combining OPSO with LSTM, we optimize LSTM’s parameters, thereby improving its efficiency when confronted with a larger number of network layers.
The rest of this paper is organized as follows: in Section 2, recent works related to carbon emission are presented; in Section 3, the structure and process of the proposed method are introduced; in Section 4, experimental details and results are presented; in Section 5, the conclusion of this work is presented.

3. Materials and Methods

Traditional RNN models demand substantial training data to guarantee model accuracy. Nevertheless, carbon emission datasets typically consist of only 30 annual data points, rendering traditional RNNs susceptible to overfitting during training. In this part, we introduce a novel solution to address this challenge by employing the oscillating particle swarm algorithm. Additionally, the model presented in this article offers the capability to monitor statistical outcomes from relevant agencies and even provide feedback on unusual data patterns, enhancing its versatility and practicality.

3.1. Long Short-Term Memory

The long short-term memory (LSTM) network is a special version of the traditional recurrent neural network (RNN). LSTM effectively solves the problem of gradient vanishing and gradient explosion in long series prediction. Compared with the traditional neural network, LSTM relies on the gate mechanism. The LSTM through the gate mechanism controls the flow of information input and output. The cell unit form of LSTM and the network structure of LSTM are shown in Figure 1. In Figure 1, the yellow part represents the fully connected layer, the purple part represents the LSTM layer, and the yellow part represents features.
Figure 1. The cell unit form of LSTM and the network structure of LSTM. Source: own creation.
The gate mechanisms are proposed for LSTM, including the forget gate, the input gate, and the output gate. At a specific time, the cell input of LSTM consists of the sequence input x t , the state of the previous hidden layer h t 1 , and the state of the previous cell C t 1 . In the first step, the forget gate forgets useless input information by Equations (1) and (2).
v t = [ h t 1 , x t ]
f t = σ ( W f · v t + b f )
where f t , W f , and b f are the output, weight, and bias of the forget gate. The σ ( ) represents a sigmoid function. v t denotes a mix of h t 1 and x t information.
In the second step, the information stream is the input gate. The input gate includes two parts, one part called the update part, used to determine which information needs to be updated, and the other part called the candidate part, used to create a new vector of candidate values, generating candidate memory. Details are described by Equations (3) and (4).
i t = σ ( W i · v t + b i )
C t = t a n h ( W c · v t + b c )
where i t , w i , and b i are the output, weight, and bias of the update part in the input gate. σ ( ) represents a sigmoid function. v t denotes a mix of h t 1 and x t information. C t , w c , and b c are the output, weight, and bias of the candidate part in the input gate. t a n h ( ) represents an activation function.
The last step involves information through the output gate. The output gate controls how much information is output to the next unit. Details are described by Equations (5) and (6).
O t = σ ( W o · v t + b o )
h t = O t · t a n h ( C t )
where O t , w o , and b o are the output, weight, and bias of the output gate. The σ ( ) represents a sigmoid function. v t denotes a mix of h t 1 and x t information. t a n h ( ) represents an activation function.

3.2. Oscillatory Particle Swarm Optimization

Traditional RNNs typically demand extensive training data to enhance model accuracy. Yet, in the context of carbon emission prediction, data scarcity is common, increasing the risk of overfitting. To address this challenge, researchers have explored utilizing the particle swarm algorithm (PSO) for hyperparameter tuning in the model.
PSO simulates the foraging behavior of birds or fish and is usually used in single-objective optimization problems. Traditional PSO is vulnerable to local optima and difficult to jump out of local optima in the early stage of evolution and has low accuracy in the later iterations. We propose an oscillating particle swarm optimization (OPSO) to solve these problems. In OPSO, the velocity update formula follows that of PSO. However, to enhance the ability of PSO to jump out of the local optimum at an early stage and to find accurate results at a later stage, we insert an adaptive parameter related to the number of population evolutions that does not need to be set in the position update formula. The schematic diagram for the different search ranges in the early and late stages of OPSO is shown in Figure 2. The OPSO velocity and position update equations are shown in Equations (7) and (8).
v t + 1 ( i ) = w · v t ( i ) + c 1 · r a n d · ( p b e s t t ( i ) x t ( i ) ) + c 2 · r a n d · ( g b e s t t x t ( i ) )
x t + 1 ( i ) = x t ( i ) + ( 1 + 1 / ( e x p ( t ) ) ) · v t ( i ) ;
where v t + 1 ( i ) represents the velocity of the ( i ) th particle at the t + 1 iteration. p b e s t t ( i ) and x t ( i ) are, respectively, the optimal position of the ( i ) th particle in personal history and its own position after the t evolution. g b e s t t denotes the global best position in the t evolution. e x p ( t ) represents the oscillation coefficient of the t evolution. c 1 and c 2 are factors of learning. w is the weight of the velocity. r a n d is a random number from 0 to 1.
Figure 2. The schematic diagram for the search ranges in the early and late stages of OPSO. Source: own creation.

3.3. Framework of OPSO-LSTM Model

Enhancing the precision of carbon emission prediction models necessitates meticulous parameter optimization, specifically refining elements such as the number of neurons and the learning rate within the LSTM architecture. Nonetheless, the conventional approach of manual debugging proves arduous and time-intensive, prompting us to seek an innovative solution. We seamlessly integrate the OPSO algorithm into the LSTM framework, resulting in the OPSO-LSTM model. Within this framework, OPSO assumes the pivotal role of fine-tuning LSTM’s neuron count and learning rate, yielding significant enhancements in prediction accuracy.
The OPSO-LSTM model’s construction entails the orchestration of four fundamental components, each playing a crucial role in its overall efficacy. These components encompass data pre-processing, where raw data are refined and prepared for analysis; OPSO’s relentless pursuit of the optimal solution, model training, which involves iterative adjustments for optimal performance; and the eventual generation of predictions. To provide a visual aid in understanding this intricate process, we have included a detailed flowchart in Figure 3, shedding light on the sequence of operations. A comprehensive exploration of each of these components is shown as follows:
Figure 3. The flowchart of OPSO-LSTM. Source: own creation.
  • Step 1: Normalize all the data within the range −1 to 1 according to the formula. The normalization formula is shown in Equation (9). The normalization can eliminate the influence of the scale between indicators to solve the comparability between data indicators.
    X n o r m = 2 ( X X m i n ) / ( X m a x X m i n ) 1
    where X n o r m is the output after normalization. X m i n and X m a x are the minimum and maximum values of the input class of data. X represents the input data.
  • Step 2: In the OPSO optimization process, the number of neurons and initial learning rate of LSTM are taken as optimization objects, and the mean square error (MSE) is taken as a fitness function of OPSO. The fitness function of OPSO is as follows:
    F i t n e s s = M S E = 1 n i = 1 n ( Y i Y i ¯ ) 2
    where n is the number of samples. Y i and Y i ¯ are the true value and the predicted value, respectively.
  • Step 3: The optimal number of neurons and the initial learning rate obtained by searching by OPSO are used as parameters during LSTM training.

4. Experiment and Results

To assess the performance of the OPSO-LSTM model, we conducted a series of simulation experiments utilizing authentic data. In these experiments, standard LSTM and PSO-LSTM are selected in the control group. Real data of carbon emission of Hainan Provenience are used in training and testing these models. Experimental details are presented in subsections.

4.1. Data Preparation

Understanding the intricate relationship between carbon dioxide (CO2) emissions from energy consumption and a multitude of factors such as local gross domestic product (GDP), resident population, passenger and freight traffic, civilian vehicle ownership, and residential consumption levels is essential. To comprehensively explore this relationship, we have meticulously curated a comprehensive dataset spanning the years 1991 to 2021, specific to Hainan Province. This dataset encompasses a diverse array of ten crucial data categories, including GDP, retail sales of consumer goods (RSCG), production value of agriculture, forestry, animal husbandry, and fishery (PVAFAF), sown area of crops (SAC), value of building enterprises (VBE), resident population (RP), passenger traffic (PT), cargo volume (CV), civilian vehicle ownership (CVO), and resident consumption level (RCL). The detailed data tables, labeled as Table 1 and Table 2, contain the specific information. Our model training process utilized data from 1991 to 2016, while the subsequent evaluation of the OPSO-LSTM model’s reliability was conducted using data from the years 2017 to 2021. These data can be downloaded from the CEInet Statistics Database (https://db.cei.cn/jsps/Home) and Hainan Statistical Yearbook(HSY).
Table 1. The specific data of GDP, RSCG, PVAFAF, SAC, VBE, and RP (Source: CEInet Statistics Database and HSY).
Table 2. The specific data of PT, CV, CVO, RSL, and CO 2 emissions from energy consumption (Source: CEInet Statistics Database and HSY).

4.2. OPSO-LSTM Forecasting Results

PSO and OPSO are used to optimize parameters of LSTM, indicate that OPSO is better than PSO in optimal LSTM, and verify the performance of the OPSO-LSTM in predicting the CO 2 emissions from energy consumption of Hainan Province. We chose LSTM and POS-LSTM as the control group. To ensure fairness in the comparison, we used the same population size of 50 and number of iterations of 100 as PSO in our experiments, as well as the same range of optimized objects. The detailed parameters of the PSO and OPSO are shown in Table 3.
Table 3. The detailed parameters of the PSO and OPSO. Source: own creation.
We put the sample data into the control group and the proposed OPSO-LSTM for training and prediction. The curves of the control group and the proposed OPSO-LSTM training and prediction results are shown in Figure 4, Figure 5 and Figure 6. Data from 1991 to 2015 are used in training, and data from 2016 to 2021 are used in tests. With the accumulation of years of input data, the training results of OPSO-LSTM are basically consistent with the actual results and significantly better than PSO-LSTM and LSTM. In the prediction of the test, the OPSO-LSTM curves significantly fit the true values better. The specific results for the test data are shown in Table 4. Compared with the control group, the OPSO-LSTM model has the best prediction in 2017, 2019, and 2021. In the prediction of 2021, the error of OPSO-LSTM is only 1.29% lower than the 12.9% and 7.57% of LSTM and PSO-LSTM.
Figure 4. The training and prediction result of LSTM. Source: own creation.
Figure 5. The training and prediction result of PSO-LSTM. Source: own creation.
Figure 6. The training and prediction result of OPSO-LSTM. Source: own creation.
Table 4. The specific results for the test data. Source: own creation.
To further demonstrate the advantages of OPSO-LSTM in predicting CO 2 emissions from energy consumption in Hainan Province, we chose the three most common forecast evaluation metrics, including mean absolute error (MAE), root mean square error (RMSE), and mean absolute percentage error (MAPE), to calculate specific data for LSTM, PSO-LSTM, and OPSO-LSTM on the test set. The MEA is used to calculate the mean error between the true and predicted values. The RMSE is a metric that presents the fitting effect metric. Compared to the MAE, the MAPE normalizes the error between the true and predicted values for each pair, decreasing the contribution of absolute errors for individual outliers. The detailed calculations of these indicators are shown in Table 5, and the equations of the MAE, RMSE, and MAPE are as follows:
M A E = 1 n i = 1 n | P i R i |
R M S E = 1 n i = 1 n ( P i R i ) 2
M A P E = 1 n i = 1 n | P i R i R i |
where n is the number of samples. P i and R i are the prediction value and reality value, respectively.
Table 5. The LSTM, PSO-LSTM and OPSO-LSTM with specific MEA, RMSE and MAPE on the test set. Source: own creation.
Compared with the LSTM and PSO-LSTM models, the MAE of OPSO-LSTM is 117.708, 65.72% better than LSTM and 29.48% better than PSO-LSTM. The RMSE of OPSO-LSTM is 149.939, 68.52% better than LSTM and 41.90% better than PSO-LSTM. The MAPE of OPSO-LSTM is 0.017, 65.31% better than LSTM and 29.17% better than PSO-LSTM. All the indicators indicate that the OPSO-LSTM is superior to the LSTM and PSO-LSTM. However, it is worth noticing that in the years 2020 and 2021, there is a significant change in the prediction accuracy of all methods. This is caused by the dramatic changes in the economy caused by COVID-19. Therefore, improving model stability against drastic economic changes is a valuable future work.

5. Conclusions

In this paper, a novel OPSO-LSTM model for carbon emissions prediction is proposed. OPSO-LSTM combines two key components: OPSO (oscillatory particle swarm optimization) and LSTM (long short-term memory). OPSO is employed to fine-tune the hyperparameters of the LSTM, utilizing an oscillatory strategy to effectively mitigate overfitting and consequently improve the accuracy of the LSTM model. To verify the predictive ability of the proposed model, simulation experiments are implemented. The standard LSTM and PSO-LSTM are selected in the control group. Real data from Hainan Province, encompassing diverse dimensions such as gross domestic product, forest area, and ten other relevant factors, are used in model training and testing. The mean absolute error (MAE), root mean square error (RMSE), and mean absolute percentage error (MAPE) are used to evaluate the performance of these methods. In the test dataset, the MAE of OPSO-LSTM is 117.708, 65.72% better than LSTM and 29.48% better than PSO-LSTM. The RMSE of OPSO-LSTM is 149.939, 68.52% better than LSTM and 41.90% better than PSO-LSTM. The MAPE of OPSO-LSTM is 0.017, 65.31% better than LSTM and 29.17% better than PSO-LSTM. Compared with the LSTM model and the PSO-LSTM model, OPSO-LSTM is able to overcome the model instability problem caused by the difficulty of hyperparameter selection. The experimental results show that OPSO-LSTM is able to provide high-precision results for carbon emission prediction tasks.
The OPSO-LSTM model empowers industries, policymakers, and environmentalists with the capability to make data-driven decisions, optimize energy consumption, allocate resources efficiently, and advance sustainable practices, thereby playing a pivotal role in addressing climate change and environmental impact assessment and shaping a more sustainable future.
Considering the ongoing technological advancements in neural networks, we believe that exploring the impact of high-dimensional input data on OPSO’s prediction accuracy is a crucial future endeavor. Additionally, choosing a more suitable adaptive function for optimizing LSTM parameters can significantly affect model accuracy. Furthermore, improving the training speed of OPSO-LSTM for real-time predictions is an achievable future goal.

Author Contributions

Conceptualization, Y.C.; methodology, Z.C.; software, K.L.; validation, T.S.; formal analysis, X.C.; investigation, J.L.; resources, T.W.; data curation, Y.L.; writing—original draft preparation, Y.C.; writing—review and editing, J.G.; visualization, Q.L.; supervision, B.S.; project administration, J.G.; funding acquisition, J.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Hainan Provincial Finance Science and Technology Program (SQKY2022-0003), the Education Department Scientific Research Program Project of Hubei Province of China (Q20222208), the Open Fund Hubei Internet Finance Information Engineering Technology Research Center (IFZX2209), the Natural Science Foundation of China (52201363), Guangdong-Hong Kong-Macau Joint Laboratory for Smart Cities, and Shenzhen Science and Technology Program (Shenzhen Key Laboratory of Digital Twin Technologies for Cities) (Grant Number ZDSYS20210623101800001).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Data available on request.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Abeydeera, L.H.U.W.; Mesthrige, J.W.; Samarasinghalage, T.I. Global research on carbon emissions: A scientometric review. Sustainability 2019, 11, 3972. [Google Scholar] [CrossRef]
  2. Wang, X.; Chen, H.; Gan, C.; Lin, H.; Dou, Q.; Tsougenis, E.; Huang, Q.; Cai, M.; Heng, P.A. Weakly Supervised Deep Learning for Whole Slide Lung Cancer Image Analysis. IEEE Trans. Cybern. 2020, 50, 3950–3962. [Google Scholar] [CrossRef] [PubMed]
  3. Jin, H. Prediction of direct carbon emissions of Chinese provinces using artificial neural networks. PLoS ONE 2021, 16, e0236685. [Google Scholar] [CrossRef]
  4. Gao, P.; Yue, S.; Chen, H. Carbon emission efficiency of China’s industry sectors: From the perspective of embodied carbon emissions. J. Clean. Prod. 2021, 283, 124655. [Google Scholar] [CrossRef]
  5. Cai, H.; Wang, Z.; Zhu, Y. Understanding the structure and determinants of intercity carbon emissions association network in China. J. Clean. Prod. 2022, 352, 131535. [Google Scholar] [CrossRef]
  6. Kılıç, M.; Kuzey, C. The effect of corporate governance on carbon emission disclosures: Evidence from Turkey. Int. J. Clim. Chang. Strateg. Manag. 2019, 11, 35–53. [Google Scholar] [CrossRef]
  7. Waheed, R.; Sarwar, S.; Wei, C. The survey of economic growth, energy consumption and carbon emission. Energy Rep. 2019, 5, 1103–1115. [Google Scholar] [CrossRef]
  8. Lee, J.H.; Cho, J.H. Firm-value effects of carbon emissions and carbon disclosures—Evidence from korea. Int. J. Environ. Res. Public Health 2021, 18, 12166. [Google Scholar] [CrossRef]
  9. Alam, M.S.; Safiullah, M.; Islam, M.S. Cash-rich firms and carbon emissions. Int. Rev. Financ. Anal. 2022, 81, 102106. [Google Scholar] [CrossRef]
  10. Bai, J.; Chen, Z.; Yan, X.; Zhang, Y. Research on the impact of green finance on carbon emissions: Evidence from China. Econ. Res.-Ekon. Istraz. 2022, 35, 6965–6984. [Google Scholar] [CrossRef]
  11. Ott, C.; Schiemann, F. The market value of decomposed carbon emissions. J. Bus. Financ. Account. 2023, 50, 3–30. [Google Scholar] [CrossRef]
  12. Chen, X.; Cheng, F.; Liu, C.; Cheng, L.; Mao, Y. An improved Wolf pack algorithm for optimization problems: Design and evaluation. PLoS ONE 2021, 16, e0254239. [Google Scholar] [CrossRef] [PubMed]
  13. Mou, X.; Zhang, Y.; Jiang, J.; Sun, H. Achieving Low Carbon Emission for Dynamically Charging Electric Vehicles through Renewable Energy Integration. IEEE Access 2019, 7, 118876–118888. [Google Scholar] [CrossRef]
  14. Ahvar, E.; Ahvar, S.; Mann, Z.A.; Crespi, N.; Glitho, R.; Garcia-Alfaro, J. DECA: A Dynamic Energy Cost and Carbon Emission-Efficient Application Placement Method for Edge Clouds. IEEE Access 2021, 9, 70192–70213. [Google Scholar] [CrossRef]
  15. Zhou, F.; Huang, Z.; Zhang, C. Carbon price forecasting based on CEEMDAN and LSTM. Appl. Energy 2022, 311, 118601. [Google Scholar] [CrossRef]
  16. Kong, F.; Song, J.; Yang, Z. A novel short-term carbon emission prediction model based on secondary decomposition method and long short-term memory network. Environ. Sci. Pollut. Res. 2022, 29, 64983–64998. [Google Scholar] [CrossRef]
  17. Hu, Y.; Lv, K. Hybrid prediction model for the interindustry carbon emissions transfer network based on the grey model and general vector machine. IEEE Access 2020, 8, 20616–20627. [Google Scholar] [CrossRef]
  18. Wang, J.; Xie, Y.; Xie, S.; Chen, X. Cooperative particle swarm optimizer with depth first search strategy for global optimization of multimodal functions. Appli. Intell. 2022, 52, 10161–10180. [Google Scholar] [CrossRef]
  19. Gao, M.; Yang, H.; Xiao, Q.; Goh, M. A novel fractional grey Riccati model for carbon emission prediction. J. Clean. Prod. 2021, 282, 124471. [Google Scholar] [CrossRef]
  20. Wang, T.; Li, Z.; Geng, X.; Jin, B.; Xu, L. Time Series Prediction of Sea Surface Temperature Based on an Adaptive Graph Learning Neural Model. Future Internet 2022, 14, 171. [Google Scholar] [CrossRef]
  21. Cai, J.; Ma, X. Carbon emission prediction model of agroforestry ecosystem based on support vector regression machine. Appl. Ecol. Environ. Res. 2019, 17, 6397–6413. [Google Scholar] [CrossRef]
  22. Kong, F.; Song, J.; Yang, Z. A daily carbon emission prediction model combining two-stage feature selection and optimized extreme learning machine. Environ. Sci. Pollut. Res. 2022, 29, 87983–87997. [Google Scholar] [CrossRef] [PubMed]
  23. Huang, W.; Dong, S. Improved short-term prediction of significant wave height by decomposing deterministic and stochastic components. Renew. Energy 2021, 177, 743–758. [Google Scholar] [CrossRef]
  24. Zhao, S.; Patuano, A. International chinese students in the uk: Association between use of green spaces and lower stress levels. Sustainability 2022, 14, 89. [Google Scholar] [CrossRef]
  25. Lu, X.; Ota, K.; Dong, M.; Yu, C.; Jin, H. Predicting Transportation Carbon Emission with Urban Big Data. IEEE Trans. Sustain. Comput. 2017, 2, 333–344. [Google Scholar] [CrossRef]
  26. Greff, K.; Srivastava, R.K.; Koutnik, J.; Steunebrink, B.R.; Schmidhuber, J. LSTM: A Search Space Odyssey. IEEE Trans. Neural Netw. Learn. Syst. 2017, 28, 2222–2232. [Google Scholar] [CrossRef] [PubMed]
  27. Abbasimehr, H.; Shabani, M.; Yousefi, M. An optimized model using LSTM network for demand forecasting. Comput. Ind. Eng. 2020, 143, 106435. [Google Scholar] [CrossRef]
  28. Zhao, J.; Kou, L.; Wang, H.; He, X.; Xiong, Z.; Liu, C.; Cui, H. Carbon Emission Prediction Model and Analysis in the Yellow River Basin Based on a Machine Learning Method. Sustainability 2022, 14, 6153. [Google Scholar] [CrossRef]
  29. Guo, J.; Sato, Y. A Pair-wise Bare Bones Particle Swarm Optimization Algorithm for Nonlinear Functions. Int. J. Netw. Distrib. Comput. 2017, 5, 143. [Google Scholar] [CrossRef]
  30. Tian, H.; Guo, J.; Xiao, H.; Yan, K.; Sato, Y. An electronic transition-based bare bones particle swarm optimization algorithm for high dimensional optimization problems. PLoS ONE 2022, 17, e0271925. [Google Scholar] [CrossRef]
  31. Guo, X.; He, J.; Wang, B.; Wu, J. Prediction of Sea Surface Temperature by Combining Interdimensional and Self-Attention with Neural Networks. Remote Sens. 2022, 14, 4737. [Google Scholar] [CrossRef]
  32. Guo, J.; Zhou, G.; Yan, K.; Shi, B.; Di, Y.; Sato, Y. A novel hermit crab optimization algorithm. Sci. Rep. 2023, 13, 9934. [Google Scholar] [CrossRef] [PubMed]
  33. Xiao, H.; Guo, J.; Shi, B.; Di, Y.; Pan, C.; Yan, K.; Sato, Y. A Twinning Memory Bare-bones Particle Swarm Optimization Algorithm for No-linear Functions. IEEE Access 2022, 11, 25768–25785. [Google Scholar] [CrossRef]
  34. Li, J.; Tian, Z.; Zhang, G.; Li, W. Multi-AUV Formation Predictive Control Based on CNN-LSTM under Communication Constraints. J. Mar. Sci. Eng. 2023, 11, 873. [Google Scholar] [CrossRef]
  35. Guo, J.; Sato, Y. A fission-fusion hybrid bare bones particle swarm optimization algorithm for single-objective optimization problems. Appl. Intell. 2019, 49, 3641–3651. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.