Next Article in Journal
Improvement of Luminescence and Photocatalytic Performance of ZnO:Eu3+ Nanocrystals Activated by Na+ Ions
Previous Article in Journal
Research on a BP Neural Network Slope Safety Coefficient Prediction Model Based on Improved Sparrow Algorithm Optimization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Estimation of Methane Gas Production in Turkey Using Machine Learning Methods

by
Güler Ferhan Ünal Uyar
1,
Mustafa Terzioğlu
2,
Mehmet Kayakuş
3,*,
Burçin Tutcu
2,
Ahmet Çoşgun
4,
Güray Tonguç
5 and
Rüya Kaplan Yildirim
6
1
Department of Business Administration, Faculty of Economics and Administrative Sciences, Akdeniz University, Antalya 07058, Turkey
2
Accounting and Tax Department, Korkuteli Vocational School, Akdeniz University, Antalya 07800, Turkey
3
Department of Management Information Systems, Faculty of Manavgat Social Sciences and Humanities, Akdeniz University, Antalya 07600, Turkey
4
Department of Mechanical Engineering, Faculty of Engineering, Akdeniz University, Antalya 07058, Turkey
5
Department of Management Information Systems, Faculty of Applied Sciences, Akdeniz University, Antalya 07058, Turkey
6
Management and Organization Department, Aydin Vocational School, Adnan Menderes University, Aydın 09010, Turkey
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(14), 8442; https://doi.org/10.3390/app13148442
Submission received: 16 May 2023 / Revised: 18 June 2023 / Accepted: 20 July 2023 / Published: 21 July 2023
(This article belongs to the Topic Clean Energy Technologies and Assessment)

Abstract

:
Methane gas emission into the atmosphere is rising due to the use of fossil-based resources in post-industrial energy use, as well as the increase in food demand and organic wastes that comes with an increasing human population. For this reason, methane gas, which is among the greenhouse gases, is seen as an important cause of climate change along with carbon dioxide. The aim of this study was to predict, using machine learning, the emission of methane gas, which has a greater effect on the warming of the atmosphere than other greenhouse gases. Methane gas estimation in Turkey was carried out using machine learning methods. The R2 metric was calculated as logistic regression (LR) 94.9%, artificial neural networks (ANNs) 93.6%, and support vector regression (SVR) 92.3%. All three machine learning methods used in the study were close to ideal statistical criteria. LR had the least error and highest prediction success, followed by ANNs and then SVR. The models provided successful results, which will be useful in the formulation of policies in terms of animal production (especially cattle production) and the disposal of organic human wastes, which are thought to be the main causes of methane gas emission.

1. Introduction

In order to ensure the continuity of living things in the world, there must be suitable climatic conditions and sufficient water resources. Although the world’s climate has always been in a state of change, this change gained rapid momentum in today’s world due to the activities of human beings, especially after the Industrial Revolution. Although its effects and consequences vary from region to region, the threat of climate change concerns all humanity. The first international step in this context is known as the United Nations Framework Convention on Climate Change, formed in 1992 [1,2,3,4].
One of the most important factors affecting climate change is the increase in emissions of natural greenhouse gases such as methane (CH4), carbon dioxide (CO2), and diazo monoxide (N2O), which cannot be prevented due to continuing human activities. In order to reduce global warming by 1.5°, an emission reduction of 45 per cent is required. From 2022 to 2023, this rate was found to be 1%. It was determined that the increase in methane gas emissions is very high, and 52 of 198 countries have not reported this under the United Nations Framework Convention on Climate Change [5].
When the literature was examined, many studies were found to focus on greenhouse gas and carbon dioxide emissions. However, methane gas, which plays a major role in global warming, is less well studied. In order to reach the target reduction of 1.5° in 2050, methane gas emissions, which are 27 times more effective than carbon dioxide in causing global warming, must be urgently taken under control [6,7].
The motivation of this study was to provide a guiding scientific approach in the formulation of policies necessary to control methane gas production by estimating its production. For this purpose, Turkey, which has both agricultural and industrial activities, and has not been able to reduce the use of coal in energy production to the desired level, was selected as a sample country for the study. The data set covering the years 1990–2020 was used in the study. When designing the model, the variables affecting the emission of methane gas, and the contributions of these variables to methane emission, were formed around three main categories: agricultural activities, fossil energy sources, and wastes. In the architecture of the model, the general variables of population and gross domestic product, which are considered highly influential to the above three categories, were added to the model. With these models, created using machine learning methods, Turkey’s methane gas production was predicted with an average success rate of 93.6%. This study will create a vision for countries to reduce emissions by predicting methane gas. It is also expected to be a guide in the road map that Turkey will follow for emission reduction at COP 28, to be held in Dubai on 30 November 2023. The study is expected to contribute to the literature due to the lack of studies on methane gas.
In the first part of the study, the importance and purpose of the research are emphasized. In the second part, similar studies in the literature are reviewed. Next, the characteristics of the data set and methodology are given and the design and results of the models are discussed. The results of the models are analyzed according to three different statistical methods and compared graphically. Finally, the contribution of the study’s results to the literature are described.
When the literature was examined, it was seen that most studies were situated within the framework of greenhouse gas and carbon dioxide emissions. By considering the effect of methane gas emissions on global warming, this study identified a gap in the literature. Table 1 shows the systematic literature review.

2. Materials and Methods

In this study, three different machine learning methods, logistic regression, artificial neural networks, and support vector regression, were used to predict methane gas production in Turkey. Models with eleven independent variables were designed to predict the dependent variable methane gas. The data were reduced to 0–1 using the min-max normalization technique. Cross-validation was used to increase the success of the study. R2, MAE, and MSE statistical metrics were used to determine the success and error of the models.

2.1. Data Set

Turkey, which has both agricultural and industrial activities and cannot reduce coal use in energy production to the desired level, was selected as a sample country. Data covering the annual period between 1990 and 2020 were used. The data set was created from the data obtained from the Turkish Statistical Institute (TUIK) database. The data set consisted of 372 data in total. While designing the model in the study, the variables affecting the emission of methane gas and the contribution of these variables to methane emission were formed around three main categories. The first and most important factor influencing methane gas emission is the agricultural activity of a country. The largest methane gas emission resulting from agricultural activities is caused by enteric fermentation. Enteric fermentation is the digestion of food by microorganisms in the digestive system of cattle. As a result of this process, methane gas is formed as a by-product. For this reason, the number of cattle in Turkey was determined as the input variable for red meat production in the model. Another agricultural activity that causes methane gas emissions is paddy cultivation. During the cultivation process of paddy, a large amount of water must be kept on the field surface for a long time. As a result, the organic materials in the soil remain in an oxygen-free environment and cause methane gas emission. Considering this situation in paddy cultivation, the surface area of cultivated paddy fields was added to the model. The last methane gas emission factor, which was accepted as an input variable in the category of agricultural activities, is the burning of biomass in agricultural areas. As a result of this combustion, methane gas is released to the atmosphere from the organic residues in these areas. The size of agricultural surface area and agricultural greenhouse gas emissions were, therefore, included in the model. The second main category of causative factors for methane gas emissions is the production and utilization processes of coal and natural gas, which are fossil energy sources. During the production (including enrichment), storage, and distribution phases of natural gas, methane gas, which is the main component of this gas, is released and causes emissions. At the same time, methane emissions occur during the extraction of coal, another fossil-based fuel (especially in closed underground mines). The share of these two fossil fuels in total energy production was added to the model as a variable. Since these two fossil fuels are used especially in manufacturing and heavy industry, the industrial production index was also added to the model. The third main factor category in the formation of methane gas emission is waste. Especially in landfill areas, where domestic wastes accumulate, there is a high rate of organic waste, which increases methane gas emission. For this reason, greenhouse gas emissions of wastes were also added to the model. The general variables of population and gross domestic product of the country, which are considered to affect all three main categories of emission causes, were also added to the model. The model is shown in Table 2.
In designing machine learning models, the properties of the independent variables are important. The average, maximum, minimum, and standard deviation (SD) of the data used in the model are given in Table 3.

2.2. Machine Learning

Machine learning is basically an algorithm-based approach to obtain information used for data classification and prediction. Machine learning is a scientific field of study used to develop various algorithms, modelling, and techniques to enable computers to learn like humans. It deals with learning methods and the performance of these methods by applying mathematical and statistical operations on data and making inferences from predictions [21,22]. Different mathematical and statistical methods are used to reach the solution. They are divided into two groups, supervised learning and unsupervised learning. Each method should be chosen according to the data set used in machine learning. In supervised learning, data are labelled, classified, and dependent and independent variables are used together. In unsupervised learning, the data set has independent variables but no dependent variables. The data set is unlabeled, and the problem and its results are not known beforehand [23].

2.2.1. Logistic Regression

Logistic regression is used in classification problems to predict the outcomes of categorical dependent variables depending on one or more pre-indicator variables. It is a regression method that helps to make classifications and predictions. Logistic regression analysis is a method that calculates the estimated values of the dependent variable as probabilities and allows classification in accordance with probability rules [24].
There are three basic methods in logistic regression: binary, ordinal, and nominal. Binary logistic regression is a logistic regression analysis with dependent variables containing two possible answers. Ordinal logistic regression is where the dependent variables have an ordinal scale and at least three categories. In nominal logistic regression, the dependent variable has a nominal scale and at least three categories [25].
Odds ratio is used in logistic regression. Odds ratio (OR) is the ratio of the probability of success or occurrence “P” to the probability of failure or non-occurrence “1-P”. Odds values take values within the range (0, +∞). If two separate odds ratios are compared, the odds ratio is obtained. The odds ratio (OR) cannot be negative according to the formula and can be a value between 0 and infinity. When OR = 1, it can be said that the factor of interest (according to the reference) has no effect on increasing or decreasing the probability of the situation under investigation. When OR < 1, the factor of interest (according to the reference) has a decreasing effect on the probability of the situation under investigation. When OR > 1, the factor of interest (according to the reference) has an increasing effect on the probability of the investigated situation [26].

2.2.2. Artificial Neural Networks

Artificial neural networks (ANNs) are a machine learning technique that imitates the learning method of the human brain and performs functions such as learning, remembering, and generating new data from the data obtained using generalization methods. ANNs are synthetic systems that mimic biological neural networks. As in the human biological structure, the aim is to train machines to learn through artificial neural networks and to make decisions with what they have learned. Artificial neural networks can be trained and adapted to be self-organized, self-learning, and self-evaluating to model the learning structure of the human brain [27].
ANNs are mathematical models consisting of many neurons connected to each other by weights. The network consists of input, intermediate (hidden), and output layers. Each input is multiplied by a connection weight. After the neurons weigh the input information, they sum it linearly and the threshold converts this information into output information by processing it in a linear or non-linear function. Other neurons connected to the cell receive this output as input information [28].
ANNs can also be classified structurally. ANNs are categorized into two main groups, feed-forward networks and feedback networks, in terms of the structure of the connections between neurons, depending on the direction of information flow. In feed-forward networks, the flow of information from input to output layer through the intermediate layer proceeds in only one direction. They consist of an input, hidden, and output layer. Information flow on the network proceeds from the input layer to the output layer. In other words, neurons are fed one after the other. In feedback artificial neural networks, the information flows from the output of any neuron to its input. In this type of network structure, feedback connections are in question [29].

2.2.3. Support Vector Regression

Support vector machines (SVMs) are a supervised learning algorithm used for both classification and regression analyses. The algorithm was developed by Vladimir Vapnik [30]. It can be used for continuous dependent variables and categorical variables. It is a supervised learning method that analyzes data, recognizes models and patterns, and is used in classification and regression analysis. Using the training data, it produces a mapping function between input and output. SVMs are high dimensional and distributed. Unlike classical models, the parameters are not predefined, and their number varies according to the training data. By keeping the error value in the training data constant, the confidence interval is minimized [31].
The line dividing the data set into two is called the hyperplane. It creates one or more hyperplanes in multidimensional space. Although it is possible to draw infinite hyperplanes, it is important to determine the optimal hyperplane, in other words, the most appropriate hyperplane. The distance between the hyperplane and the nearest data point in the data set is called the margin. The aim is to select a hyperplane with the highest margin between the hyperplane and a point in the data set. This is to increase the chances of correctly classifying new data. There should be no data points within the separated region [32].
A data set may often contain a set of data that cannot be linearly separated. In this case, in order to classify the data set, it is necessary to pass the data from a two-dimensional view to a three-dimensional view. This process is called the two-dimensional process. It takes a low dimensional input field and transforms it into a higher dimensional field. Maximizing the distances between the nearest data point and the hyperplane helps to determine the correct hyperplane. This is called the margin distance. Another important reason for choosing a high margin hyperplane is robustness. If we choose a low margin hyperplane, the probability of misclassification will be high [33].

3. Results and Discussion

In this study, three different methods, namely, logistic regression, artificial neural network, and support vector regression, were used to predict methane gas in Turkey. The study system is shown in Figure 1.
Eleven independent variables were used to predict methane gas, which was the dependent variable of the study. The independent variables were Turkey’s population, GDP, coal share in electricity, number of cattle, natural gas share in electricity, industrial production index, agriculture greenhouse gas emission, agricultural area, waste greenhouse gas emission, cattle meat production, and paddy production area. For each independent variable, there were thirty-one annual data between 1990 and 2020.
The min-max method was used for normalization in the study. A linear transformation was performed on the original data. In this normalization, the relationships between the original data values were preserved, and the largest and smallest values in a group of data were considered [34]. All other data were normalized according to these values. This method is a scaling technique where the data are rescaled to be between 0 and 1. The result of having a limited range between 0 and 1 in this process is that it suppresses the effect of outliers and allows for smaller standard deviations [35].
Cross-validation measures the success of the model by creating validation clusters from the data set. Testing with a single data set may not be sufficient. For this reason, it is very important to perform k-fold cross-validation, that is, to create more than one machine learning model from a single data set, to check these models with different test sets, and to average the accuracy of all of them. The data set was clustered according to the k-layer cross-validation applied as training and test data. When ten layers of cross-validation were applied to the data, nine parts were used to develop the model and the remaining one part was used to test the model. This process was repeated 10 times. Each time, it took a different validation set as the test set and used the remaining 9 sets to improve the model [36,37].
The coefficient of determination (R2) is a statistical measure that examines how differences in one variable can be explained by differences in a second variable when predicting the outcome of a particular event. In other words, the coefficient of determination is a statistical metric that measures how well a statistical model predicts an outcome. This coefficient assesses how strong the linear relationship between two variables is. The outcome is represented by the dependent variable of the model. The lowest value of R2 is 0 and the highest value is 1. It is the proportion of variance in the dependent variable that is explained by the model [38,39]. The R2 is shown in Equation (1):
R 2 = 1 U n e x p l a i n e d V a r i a t i o n T o t a l V a r i a t i o n
The MAE (mean absolute error) is a statistical metric that measures the average magnitude of errors in a set of predictions without considering their direction. It is calculated by taking the absolute difference between the predicted values and the actual values and averaging it across the data set. MAE only measures the magnitude of errors and is not concerned with their direction. The lower the MAE, the higher the accuracy of a model [40,41]. The MAE formula is shown in Equation (2):
M A E = 1 n i = 1 n e i
The MSE (mean squared error) is a model evaluation metric often used in regression models. MSE is equal to the mean squared difference between predicted values and actual values. Since MSE squares the error, it causes large errors to be clearly highlighted. MSE is a metric that ranges from 0 to infinity and values closer to zero are considered better. For MSE values calculated for the same data set, the lower the MSE value, the more accurate the model [42,43]. The MSE is shown in Equation (3):
M S E = 1 n i = 1 n e i 2
where n is the number of datapoints and e is the error value.
Three different machine learning techniques were used in this study: logistic regression, artificial neural networks, and support vector regression.
In the artificial neural network method, a feedback model consisting of eleven input neurons and one output neuron was developed. As a result of trial-and-error methods, two hidden layers were used, as this gave the most successful result. There were three neurons in each hidden layer. The model is shown in Figure 2.
After testing various functions for the activation function, the sigmoid function was preferred because it gave the most successful result. One of the parameters that has an effect on the success of the model in artificial neural networks is learning. The feedback learning algorithm was used in this study, and 100 iterations were performed. Figure 3 shows that the error curve decreased non-linearly. In the figure, the x-axisshows the number of iterations, and the y-axisshows the error. The scale of the x-axiswas between 1 and 100 and the y-axiswas between 0 and 5. The error curve stabilized between the 50th and 62th iteration and reached the ideal value.
The maximum likelihood method was used to estimate the parameters for the logistic regression model. In this study, the log likelihood value was −51.022 as a result of 100 iterations. Nonlinear SVR was preferred in the method where support vector regression was used. As a result of the tests for the kernel function, it was decided to use the radial basis function (RBF). The overlapping penalty value of the model was chosen as 10 and the RBF sigma value was chosen as 0.5.
The evaluation of regression models is given in Table 4.
Figure 4 shows the coefficient of determination (R2) graphs. R2 shows the rate of explanation of independent variables. In other words, it refers to the variance ratio of the dependent variable explained by the independent variables. Here, zero indicates that the model has 0% explanatory power, and one indicates that the model has 100% explanatory power and that the independent variables are strong in explaining the dependent variable and provide a linear curve [38,39]. R2 was 94.9% for logistic regression, 93.6% for artificial neural networks, and 92.3% for support vector regression. These results show that the ideal values were met.
The mean absolute error (MAE) graphs are presented in Figure 5. MAE is calculated by averaging the absolute values of the prediction errors. It is often preferred for determining the error values of models because it can be easily interpreted. The MAE value can vary from 0 to ∞. The lower the MAE value, the lower the error value of the model [40,41]. The MAE metric gives the magnitude of the error as a quantity. In the study, it was seen that the MAE was 0.052 for logistic regression, 0.053 for artificial neural networks, and 0.065 for support vector regression. The MAE was considered to be successful in all three models.
Figure 6 shows the mean squared error (MSE) graphs. MSE is a statistical metric that evaluates the error of machine learning methods. MSE is equal to the mean squared difference between predicted values and actual values. MSE is often used in regression models. Because MSE squares the error, it causes major errors to be clearly highlighted. MSE is a metric that ranges from 0 to infinity, and values close to zero are considered better [42,43]. It was seen that the MSE was 0.005 for logistic regression, 0.006 for artificial neural networks, and 0.009 for support vector regression. The error rate of the models was found to be low and acceptable.
According to the error and success evaluation, the order of the models in terms of their success is logistic regression, then artificial neural network and, finally, support vector regression. Figure 7 shows the scatter plots of the models.
Figure 5 shows the scatter plot of the machine learning models used for methane gas prediction. The x-axisin the figure shows the actual values and the y-axisshows the predicted values. The scales of the x and y axes were between 0 and 1. In all three models, there was a positive correlation between the methane gas value and the predicted results. Moreover, this link was quite strong. As the value of one of the variables increased, the other increased and the points clustered near the line. Figure 8 shows the line plots of the models.
Figure 8 shows the relationship between actual and predicted values. The x-axis in the figure shows the years and the y-axisshows the actual and predicted values. The x-axisscale covered the period between 1990 and 2020; the y-axiscovered the period between 0 and 1. The relationship was strong for all three models. When the figures were analyzed in detail, it was seen that the model with the strongest relationship was logistic regression, followed by artificial neural network and support vector regression.
The boundaries of the study were contained to countries that have coal mines and continue to produce energy from coal, and support cattle breeding and paddy production. In this respect, the model may need to be modified in order to be applied in continental European countries with serious regulations on coal production and very limited paddy production.

4. Conclusions

The aim of this study was to use machine learning to predict the emission of methane gas, which has a greater effect on the warming of the atmosphere than other greenhouse gases. The differentiating aspect of this study from other studies in the literature is that greenhouse gas emission studies have so far either predicted greenhouse gas emissions in general or focused on carbon dioxide emissions. However, the atmospheric heating potential of methane gas is higher than other greenhouse gases. This study was also unique in its use of not only technical data but also economic and environmental factors, which were accepted as variables in the design of the model.
In the study, three different machine learning techniques were applied to find the best machine learning technique. The aim was to provide preliminary information for researchers that the machine learning technique works successfully, so they can use this model in the future. In terms of the usability of the model and the accessibility of the data set, it was thought to be a guiding study, especially in making decisions to prevent methane gas and in the creation of sustainable policies. In particular, the model provided statistically very useful and successful results that can be used in the formulation of policies in terms of animal production (especially cattle production) and the disposal of organic human waste, which is considered to be the main cause of methane gas emission.
The results revealed that the methane gas emission variables used in the study can be considered as variables for future greenhouse gas emission estimation studies.
In this study, methane gas production in Turkey was successfully estimated using machine learning methods. Three different supervised machine learning methods—logistic regression, artificial neural networks, and support vector regression—were used in the study. To analyze and evaluate these methods, R2, MAE, and MSE metrics were calculated. R2 was found to be 94.9% for logistic regression, 93.6% for artificial neural networks, and 92.3% for support vector regression. These results show that the ideal values were met. MSE was 0.052 for logistic regression, 0.053 for artificial neural networks, and 0.065 for support vector regression. MAE was considered successful in all three models. MSE was 0.005 for logistic regression, 0.006 for artificial neural networks, and 0.009 for support vector regression. The error rate of the models was found to be low and acceptable. According to the results of the analyses, it can be seen that the success of all three models was high, and the error was within the acceptable range. The model was also more successful than that of similar studies. According to the statistical metrics, logistic regression, artificial neural networks, and support vector regression were all successful models, with the least error obtained with logistic regression, for methane gas production forecasting in Turkey.

Author Contributions

Methodology, M.K., B.T., G.T. and G.F.Ü.U.; formal analysis, M.K. and A.Ç.; data curation, M.T., B.T. and G.T.; writing—original draft, M.K., A.Ç., B.T., M.T., R.K.Y. and G.F.Ü.U.; writing—review and editing, B.T., G.T., M.T., R.K.Y., G.F.Ü.U., A.Ç. and R.K.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sands, P. The United Nations framework convention on climate change. Rev. Eur. Comp. Int’l Envtl. L. 1992, 1, 270. [Google Scholar] [CrossRef] [Green Version]
  2. Bodansky, D. The United Nations framework convention on climate change: A commentary. Yale J. Int’l l. 1993, 18, 451. [Google Scholar]
  3. Lindzen, R.S. Climate dynamics and global change. Annu. Rev. Fluid Mech. 1994, 26, 353–378. [Google Scholar] [CrossRef]
  4. Thuiller, W. Climate change and the ecologist. Nature 2007, 448, 550–552. [Google Scholar] [CrossRef] [PubMed]
  5. C. (COP27). Assessment Reports. Şarm El-Şeyh Egypt. 2022. Available online: https://cop27.eg/#/ (accessed on 10 March 2023).
  6. Anika, O.C.; Nnabuife, S.G.; Bello, A.; Okoroafor, R.E.; Kuang, B.; Villa, R. Prospects of Low and Zero-Carbon Renewable fuels in 1.5-Degree Net Zero Emission Actualisation by 2050: A Critical Review. Carbon Capture Sci. Technol. 2022, 5, 100072. [Google Scholar] [CrossRef]
  7. Pierrehumbert, R. There is no Plan B for dealing with the climate crisis. Bull. At. Sci. 2019, 75, 215–221. [Google Scholar] [CrossRef] [Green Version]
  8. Oertel, C.; Matschullat, J.; Zurba, K.; Zimmermann, F.; Erasmi, S. Greenhouse gas emissions from soils—A review. Geochemistry 2016, 76, 327–352. [Google Scholar] [CrossRef] [Green Version]
  9. Garip, E.; Oktay, A.B. Forecasting CO2 Emission with Machine Learning Methods. In Proceedings of the 2018 International Conference on Artificial Intelligence and Data Processing (IDAP), Malatya, Turkey, 28–30 September 2018. [Google Scholar]
  10. Baareh, A.K. Solving the carbon dioxide emission estimation problem: An artificial neural network model. J. Softw. Eng. Appl. 2013, 6, 338–342. [Google Scholar] [CrossRef] [Green Version]
  11. Kalra, S.; Lamba, R.; Sharma, M. Machine learning based analysis for relation between global temperature and concentrations of greenhouse gases. J. Inf. Optim. Sci. 2020, 41, 73–84. [Google Scholar] [CrossRef]
  12. Hamrani, A.; Akbarzadeh, A.; Madramootoo, C.A. Machine learning for predicting greenhouse gas emissions from agricultural soils. Sci. Total Environ. 2020, 741, 140338. [Google Scholar] [CrossRef] [PubMed]
  13. Saha, D.; Basso, B.; Robertson, G.P. Machine learning improves predictions of agricultural nitrous oxide (N2O) emissions from intensively managed cropping systems. Environ. Res. Lett. 2020, 16, 024004. [Google Scholar] [CrossRef]
  14. Gholami, H.; Mohamadifar, A.; Sorooshian, A.; Jansen, J.D. Machine-learning algorithms for predicting land susceptibility to dust emissions: The case of the Jazmurian Basin, Iran. Atmos. Pollut. Res. 2020, 11, 1303–1315. [Google Scholar] [CrossRef]
  15. Şişeci Çeşmeli, M.; Pençe, I. Forecasting of Greenhouse Gas Emissions in Turkey using Machine Learning Methods. Acad. Platf. J. Eng. Sci. 2020, 8, 332–348. [Google Scholar]
  16. Aydin, S.G.; Aydoğdu, G. CO2 Emissions in Turkey and EU Countries Using Machine Learning Algorithms. Eur. J. Sci. Technol. 2022, 37, 42–46. [Google Scholar]
  17. Abbasi, N.A.; Hamrani, A.; Madramootoo, C.A.; Zhang, T.; Tan, C.S.; Goyal, M.K. Modelling carbon dioxide emissions under a maize-soy rotation using machine learning. Biosyst. Eng. 2021, 212, 1–18. [Google Scholar] [CrossRef]
  18. Kerimov, B.; Chernyshev, R. Review of machine learning methods in the estimation of greenhouse gas emissions. In Proceedings of the International Conference of Young Scientists Modern Problems of Earth Sciences, Tbilisi, Georgia, 21–22 November 2022. [Google Scholar]
  19. Saleh, C.; Dzakiyullah, N.R.; Nugroho, J.B. Carbon dioxide emission prediction using support vector machine. IOP Conf. Ser. Mater. Sci. Eng. 2016, 114, 012148. [Google Scholar] [CrossRef]
  20. Jiang, Z.; Yang, S.; Smith, P.; Pang, Q. Ensemble machine learning for modeling greenhouse gas emissions at different time scales from irrigated paddy fields. Field Crop. Res. 2023, 292, 108821. [Google Scholar] [CrossRef]
  21. Ghaderzadeh, M.; Asadi, F.; Hosseini, A.; Bashash, D.; Abolghasemi, H.; Roshanpour, A. Machine Learning in Detection and Classification of Leukemia Using Smear Blood Images: A Systematic Review. Sci. Program. 2021, 2021, 9933481. [Google Scholar] [CrossRef]
  22. El Naqa, I.; Murphy, M.J. What is Machine Learning? Springer: Berlin/Heidelberg, Germany, 2015. [Google Scholar]
  23. Haldorai, A.; Ramu, A.; Suriya, M. Organization internet of things (IoTs): Supervised, unsupervised, and reinforcement learning. In Business Intelligence for Enterprise Internet of Things; Springer: Berlin/Heidelberg, Germany, 2020; pp. 27–53. [Google Scholar]
  24. Sperandei, S. Understanding logistic regression analysis. Biochem. Med. 2014, 24, 12–18. [Google Scholar] [CrossRef]
  25. Bayaga, A. Multinomial Logistic Regression: Usage and Application in Risk Analysis. J. Appl. Quant. Methods 2010, 5, 288–298. [Google Scholar]
  26. Hailpern, S.M.; Visintainer, P.F. Odds Ratios and Logistic Regression: Further Examples of their use and Interpretation. Stata J. Promot. Commun. Stat. Stata 2003, 3, 213–225. [Google Scholar] [CrossRef] [Green Version]
  27. Travassos, X.L.; Avila, S.L.; Ida, N. Artificial Neural Networks and Machine Learning techniques applied to Ground Penetrating Radar: A review. Appl. Comput. Inform. 2018, 17, 296–308. [Google Scholar] [CrossRef]
  28. Dongare, A.D.; Kharde, R.R.; Kachare, A.D. Introduction to artificial neural network. Int. J. Eng. Innov. Technol. 2012, 2, 189–194. [Google Scholar]
  29. Mansoor, M.; Grimaccia, F.; Leva, S.; Mussetta, M. Comparison of echo state network and feed-forward neural networks in electrical load forecasting for demand response programs. Math. Comput. Simul. 2020, 184, 282–293. [Google Scholar] [CrossRef]
  30. Vapnik, V. The Nature of Statistical Learning Theory; Springer Science & Business Media: Berlin/Heidelberg, Germany, 1999. [Google Scholar]
  31. Vapnik, V.N. An overview of statistical learning theory. IEEE Trans. Neural Netw. 1999, 10, 988–999. [Google Scholar] [CrossRef] [Green Version]
  32. Hao, P.-Y.; Kung, C.-F.; Chang, C.-Y.; Ou, J.-B. Predicting stock price trends based on financial news articles and using a novel twin support vector machine with fuzzy hyperplane. Appl. Soft Comput. 2020, 98, 106806. [Google Scholar] [CrossRef]
  33. Gu, B.; Sheng, V.S.; Tay, K.Y.; Romano, W.; Li, S. Cross Validation Through Two-Dimensional Solution Surface for Cost-Sensitive SVM. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1103–1121. [Google Scholar] [CrossRef]
  34. Migilinskas, D.; Ustinovichius, L. Normalization in the selection of construction alternatives. Int. J. Manag. Decis. Mak. 2007, 8, 623–639. [Google Scholar]
  35. Saranya, C.; Manikandan, G. A study on normalization techniques for privacy preserving data mining. Int. J. Eng. Technol. 2013, 5, 2701–2704. [Google Scholar]
  36. Stone, M. Cross-validation: A review. Stat. A J. Theor. Appl. Stat. 1978, 9, 127–139. [Google Scholar]
  37. Picard, R.P.; Cook, R.D. Cross-validation of regression models. J. Am. Stat. Assoc. 1984, 79, 575–583. [Google Scholar] [CrossRef]
  38. Ozer, D.J. Correlation and the coefficient of determination. Psychol. Bull. 1985, 97, 307–315. [Google Scholar] [CrossRef]
  39. Di Bucchianico, A. Coefficient of determination (R2). In Encyclopedia of Statistics in Quality and Reliability; Wiley Online Library: Hoboken, NJ, USA, 2008. [Google Scholar]
  40. Willmott, C.J.; Matsuura, K. Advantages of the mean absolute error (MAE) over the root mean square error (RMSE) in assessing average model performance. Clim. Res. 2005, 30, 79–82. [Google Scholar] [CrossRef]
  41. Chai, T.; Draxler, R.R. Root mean square error (RMSE) or mean absolute error (MAE)?—Arguments against avoiding RMSE in the literature. Geosci. Model Dev. 2014, 7, 1247–1250. [Google Scholar] [CrossRef] [Green Version]
  42. Wallach, D.; Goffinet, B. Mean squared error of prediction as a criterion for evaluating and comparing system models. Ecol. Model. 1989, 44, 299–306. [Google Scholar] [CrossRef]
  43. Tuchler, M.; Singer, A.; Koetter, R. Minimum mean squared error equalization using a priori information. IEEE Trans. Signal Process. 2002, 50, 673–683. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Working system.
Figure 1. Working system.
Applsci 13 08442 g001
Figure 2. ANN model.
Figure 2. ANN model.
Applsci 13 08442 g002
Figure 3. Error plot for ANN.
Figure 3. Error plot for ANN.
Applsci 13 08442 g003
Figure 4. R2 evaluation.
Figure 4. R2 evaluation.
Applsci 13 08442 g004
Figure 5. MAE evaluation.
Figure 5. MAE evaluation.
Applsci 13 08442 g005
Figure 6. MSE evaluation.
Figure 6. MSE evaluation.
Applsci 13 08442 g006
Figure 7. Scatter plot of methods.
Figure 7. Scatter plot of methods.
Applsci 13 08442 g007
Figure 8. Line plot of machine learning models.
Figure 8. Line plot of machine learning models.
Applsci 13 08442 g008
Table 1. Systematic literature review.
Table 1. Systematic literature review.
Author DetailsTopicInput/Output VariablesMachine Learning TechniqueConclusion
(Oertel et al., 2016) [8]Greenhouse gas emission and parameters affecting the emission processInput: Literature review was conducted.

Output: According to the result of the literature review, it was revealed that 300 mg CO2 leads to a global annual net soil emission of ≥350 Pg CO2e.
The study is based on a literature review.The author discussed the parameters related to soil emission (by reviewing the literature).
(Garip and Oktay, 2018) [9]Estimation of carbon dioxide emissionsInput: Oil, natural gas, coal, hydropower, renewable energy.

Output: In the CO2 forecasts between 2004 and 2014, it was observed that the forecasts were successful in the first years and the success in the forecasts decreased in the last two years.
  • Random Forest Regression
  • Support Vector Regression
It was observed that support vector machines gave better results than the random forest regression method.
(Baareh, 2013) [10]Estimation of carbon dioxide emissionsInputs: Oil, natural gas, coal, primary energy consumption.

Output: Data from 1982 to 2000 were used and found to have high predictive ability.
  • Artificial neural network
The artificial neural network model was found to have high forecasting ability.
(Kalra et al., 2020) [11]Measurement of the relationship between global temperature and greenhouse gas concentrationInput: Carbon dioxide, nitrous oxide, methane.

Output: In the 65 years of global temperature data between 1850 and 2016, they found that CO2 increased to the maximum in the global temperature increase.
  • Linear regression
  • Decision tree regression
  • Random forest regression
  • Artificial neural networks
It was observed that the best performance was obtained from the artificial neural network method.
(Hamrani et al., 2020) [12]Estimation of greenhouse gas emissionsInput: Carbon dioxide and nitrous oxide.

Output: In the 5-year period from 2012 to 2017, the LSTM model was found to give the most accurate performance.
  • Classical regression
  • Shallow learning
  • Deep learning
It was observed that the best performance was obtained from the LSTM model.
(Saha et al., 2021) [13]Estimation of agricultural nitrous oxide emissionsInput: Nitrous oxide.

Output: 1576 daily observations over a 6-year period from 2002 to 2014 were used. The random forest regression model was able to predict nitrous oxide data in two different fields.
  • Random forest regression
It was found that the random forest regression model can improve nitrous oxide flux predictions with limited data.
(Gholami et al., 2020) [14]Using machine learning for dust emission sensitization in IranInputs: Land use, lithology, digital elevation, vegetation cover, wind speed, rainfall, bulk density, organic matter, electrical conductivity, texture, soil texture, exchangeable sodium content and calcium carbonate content, soil property.

Output: The analysis of these 14 inputs revealed that there was no multicollinearity between the inputs. CForest was found to make the best prediction.
  • XGBoost
  • Cubist
  • BMARS
  • ANFIS
  • Cforest
  • Elasticnet
It was observed that CForest algorithm made the best prediction.
(Şişeci Çeşmeli and Pençe, 2020) [15]Greenhouse gas estimation for Turkey using machine learning algorithmsInput: Greenhouse gases data set for 1967–2017 (as time series).

Output: Tested with time series and 10-fold cross-validation, the LTSM model was the most successful and estimated 15.67 billion tons of greenhouse gas emissions prospectively.
  • Poisson regression
  • Linear regression
  • Artificial neural networks
  • Adaptive network based fuzzy inference system (ANFIS)
  • LSTM
It was observed that LSTM made the best estimation.
(Gümüştekin Aydın and Aydoğdu, 2022) [16]Carbon dioxide emission estimation of Turkey and EU countries using machine learning algorithmsInput: Consumption of “total population, solid fossil fuel, natural gas, oil, solar energy, biogas, primary solid biofuel, renewable municipal waste, geothermal energy and hydropower” from 2000 to 2019.

Output: It was observed that the amount of CO2 decreased in EU countries and the amount of CO2 increased exponentially in Turkey.
  • Support vector machines
  • Decision tree modelling
  • Artificial neural networks
It was seen that the support vector machines method made the best prediction.
(Abbasi et al., 2021) [17]Estimation of carbon dioxide emissions through machine learning algorithmsInputs: Soil moisture, temperature, organic matter, total carbon, nitrogen, air temperature, solar radiation, rainfall, pan evaporation.

Output: Soil temperature, organic matter, carbon, nitrogen, air temperature, radiation and pan evaporation were found to be highly correlated parameters.
  • Support vector machines
  • Random forest regression
  • Least absolute shrinkage selection
  • Feed forward neural network
  • Radial basis functional neural network
  • Extreme learning machine
It was observed that the random forest regression method made the best prediction.
(Kerimov and Chernyshev, 2022) [18]Estimation of greenhouse gas emissions through machine learning algorithmsA research study was conducted to define the model structure.
  • Decision tree
  • Random forest regression
  • Deep neural networks
The study compared the advantages and disadvantages of the techniques.
(Saleh et al., 2016) [19]Estimation of carbon dioxide emissions using support vector machinesInput: Electric power and coal

Output: The support vector machines method was used by trial and error. The results showed that the RMSE value was 0.004, but the aim was to obtain a value lower than this.
  • Support vector machines
In the study, it was concluded that the RMSE value was higher than expected. It was predicted that a lower value would lead to a higher prediction and a higher prediction would provide information about CO2 emission.
(Jiang et al., 2023) [20]Estimation of greenhouse gas emissions in paddy fields in China through machine learning algorithmsInput: 3-year data set from 2009 to 2011.

Output: The stacking model improved R2 and reduced RMSE.
  • Linear regression
  • Random forest regression
  • Nearest neighbor regression
  • Gradient boosting regression
  • Stacking group
It was found that the stacking model was the best prediction method, and the linear regression model was the worst prediction method.
Table 2. Input/Output variables.
Table 2. Input/Output variables.
Output VariableInput Variables
Methane Gas Emissions (tons)Agricultural ActivitiesNumber of Cattle (pcs)
Red Meat Production (tons)
Agricultural Area (hectare)
Agricultural Greenhouse Gas Emissions (tons)
Paddy Production (hectare)
EnergyShare of Coal in Total Energy Production (%)
Share of Natural Gas in Total Energy Production (%)
Industrial Production Index (2003 = 100)
WasteWaste Greenhouse Gas Emission (tons)
GeneralPopulation (Number of People)
GDP ($)
Table 3. Characteristics of independent variables.
Table 3. Characteristics of independent variables.
AverageMaximumMinimumSD
Methane gas48,868,49263,988,98040,945,9437,131,161
Population68,251,38783,614,36253,921,7589,039,080
GDP506,221,633,744957,783,020,853130,690,172,297301,529,901,430
Coal share in electricity30.7237.2022.804.05
Natural gas share in electricity33.8849.7014.6012.25
Industrial production index70.59144.6933.1334.52
Number of cattle12,580,16818,157,9719,901,4582,275,392
Agricultural area39,389,93542,033,00037,716,0001,160,451
Agriculture greenhouse gas emission48,621,59973,155,37237,607,7949,063,809
Waste greenhouse gas emission15,132,17717,786,98911,080,8262,168,268
Cattle meat production614,8821,341,445303,120313,293
Paddy production area82,125126,41940,40029,149
Table 4. Evaluation of models.
Table 4. Evaluation of models.
Logistic RegressionArtificial Neural NetworkSupport Vector Regression
R20.9490.9360.923
MAE0.0520.0530.065
MSE0.0050.0060.009
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ünal Uyar, G.F.; Terzioğlu, M.; Kayakuş, M.; Tutcu, B.; Çoşgun, A.; Tonguç, G.; Kaplan Yildirim, R. Estimation of Methane Gas Production in Turkey Using Machine Learning Methods. Appl. Sci. 2023, 13, 8442. https://doi.org/10.3390/app13148442

AMA Style

Ünal Uyar GF, Terzioğlu M, Kayakuş M, Tutcu B, Çoşgun A, Tonguç G, Kaplan Yildirim R. Estimation of Methane Gas Production in Turkey Using Machine Learning Methods. Applied Sciences. 2023; 13(14):8442. https://doi.org/10.3390/app13148442

Chicago/Turabian Style

Ünal Uyar, Güler Ferhan, Mustafa Terzioğlu, Mehmet Kayakuş, Burçin Tutcu, Ahmet Çoşgun, Güray Tonguç, and Rüya Kaplan Yildirim. 2023. "Estimation of Methane Gas Production in Turkey Using Machine Learning Methods" Applied Sciences 13, no. 14: 8442. https://doi.org/10.3390/app13148442

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop