Next Article in Journal
Diagnostics of Inter-Turn Short Circuit Fault in Dry-Type Air-Core Reactor Based on Lissajous Graph and Lightweight Network Model
Previous Article in Journal
Investigating Coastal Effects on Offshore Wind Conditions in Japan Using Unmanned Aerial Vehicles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Comparative Analysis of Machine Learning Algorithms in Energy Poverty Prediction

by
Elpida Kalfountzou
1,
Lefkothea Papada
1,
Christos Tourkolias
2,
Sevastianos Mirasgedis
3,
Dimitris Kaliampakos
1 and
Dimitris Damigos
1,*
1
School of Mining and Metallurgical Engineering, National Technical University of Athens, 15780 Athens, Greece
2
Centre for Renewable Energy Sources and Saving, 19009 Pikermi, Greece
3
Institute for Environmental Research & Sustainable Development, National Observatory of Athens, 15236 Palea Penteli, Greece
*
Author to whom correspondence should be addressed.
Energies 2025, 18(5), 1133; https://doi.org/10.3390/en18051133
Submission received: 24 January 2025 / Revised: 14 February 2025 / Accepted: 23 February 2025 / Published: 25 February 2025
(This article belongs to the Section C: Energy Economics and Policy)

Abstract

:
Given the limited potential of conventional statistical models, machine learning (ML) techniques in the field of energy poverty have attracted growing interest, especially during the last five years. The present paper adds new insights to the existing literature by exploring the capacity of ML algorithms to successfully predict energy poverty, as defined by different indicators, for the case of the “Urban Region of Athens” in Greece. More specifically, five energy poverty indicators were predicted on the basis of socio-economic/technical variables through training different machine learning classifiers. The analysis showed that almost all classifiers managed to successfully predict three out of five energy poverty indicators with a remarkably good level of accuracy, i.e., 81–94% correct predictions of energy-poor households for the best models and an overall accuracy rate of over 94%. The most successful classifier in terms of energy poverty prediction proved to be the “Random Forest” classifier, closely followed by “Trees J48” and “Multilayer Perceptron” classifiers (decision tree and neural network approaches). The impressively high accuracy scores of the models confirmed that ML is a promising tool towards understanding energy poverty drivers and shaping appropriate energy policies.

1. Introduction

1.1. The Problem of Energy Poverty

Research on energy poverty is gaining more and more interest over the last years, with numerous new enquiries on the definition, measurement, and analysis of the phenomenon. In general, energy poverty is a serious socio-economic problem, with different conceptual approaches throughout the world. In the developed world, it is connected to the lack of access to clean energy services and electricity for 2 billion people and 750 million people, respectively [1], whereas, in the developed world, it is mainly associated with the inability of households to adequately meet their energy needs at an affordable cost.
A policy and academic acknowledgement of the issue has been developed recently, arguing that energy poverty is rather a social than an inherently economic or environmental issue, following global issues [2]. Indeed, vulnerable consumers were inevitably affected by the negative impact of the COVID-19 pandemic outbreak on the energy market [3,4]. The increase in energy prices along with the pandemic led crowds of people to difficulty or inability to pay their energy bills [5]. Hence, final energy use was drastically reduced by the household sector, primarily among vulnerable parts of the population, such as low-income households [6]. This issue has intensified in Europe throughout the last years, as a consequence of financial crises and global geopolitical crises (e.g., war in Ukraine, conflicts in the Middle East), which have brought about great instability and great increases in energy prices.
Greece is among the European countries with severe energy poverty issues [7], also coming 5th in ranking with people “at risk of poverty or social exclusion” (26.1% of the country’s population) [8]. Over the past decade, a significant rise in fuel costs has taken place, combined with a remarkable decrease in the real value of households’ disposable incomes. At the same time, a dramatic rise in the level of food prices has been recorded, and environmental and health issues have multiplied due to the use of inappropriate and low-cost or free materials as fuels for space heating.
Research on energy poverty in Greece has gained increasing attention over the last decade, mainly due to the deterioration of the problem recently. Indicatively, Atsalis et al. [9] analyzed energy poverty as being expressed by the 10% indicator (ratio of energy expenses to the household’s income), based on data from the Household Budget Survey of the Hellenic Statistical Authority, showing an energy poverty rate of about 20–25% in Greece in 2013. Papada and Kaliampakos [10] carried out a primary survey in the whole country, revealing that 58% of households in Greece faced energy poverty in 2015, according to the 10% indicator, while a similar study in the mountainous areas of the country pointed out the intensity of the problem in these areas, affecting 73.5% of mountainous households [11]. Further works have studied the problem in specific geographical areas, such as in Athens [12,13], in the Attica region [14], in Western and Central Macedonia [15], in the Thessaloniki Urban Complex [16], etc. Moreover, new indicators have been suggested over time, such as, for instance, the indicator of “Degree of Coverage of Energy Needs” [17], demonstrating that 46% of Greek households compress their energy needs.
Numerous works have used traditional statistical tools to measure energy poverty in Greece. Indicatively, logistic regression models were used by Lyra et al. [18], showing that 40% of households in Greece face energy poverty, while also highlighting certain variables, i.e., household income, dwelling type, location of residence, and educational level, as decisive predictors of the problem. Similarly, binary logistic regression models were used by Kalfountzou et al. [19] to predict various indicators (10%, 2M, M/2) based on certain socio-economic variables. Halkos and Kostakis [20] employed a probit model [15], according to which 9–10% of Greek households are steadily vulnerable to energy poverty, and factors such as income level, educational level, housing characteristics, migration background, and employment proved to be the ones that mostly affect the problem.
However, given the limited potential of conventional statistical models [21] and the identification that current challenges entail large and complex datasets, there is a growing need for models with improved predictive power [22], a finding that has been especially confirmed, among other research areas, for the field of energy poverty mitigation [23]. In this context, machine learning algorithms predicting energy poverty have started attracting increasing attention recently.

1.2. Machine Learning and Energy Poverty

Although conventional regression models can be used to explore relationships between socio-economic variables and energy poverty risk, the employment of machine learning (ML) offers three (3) distinct advantages for such types of analyses:
  • Modern-day issues often include large amounts of data with multiple features, which are challenging for traditional regression models to handle. ML algorithms though are fundamentally designed to process and analyze large and complex datasets effectively.
  • Conventional regression models typically require prior assumptions regarding the correlations between the factors, potentially inserting bias in the analysis or limiting its scope. ML algorithms though learn from real data and identify correlations through the training process, without requiring any prior assumptions.
  • ML algorithms are inherently able to address non-linear dependencies, which is a main requirement for studying and analyzing complex phenomena such as energy poverty, whereas traditional regression models are better suited to address linear relationships.
The literature shows that a large number of works have used machine learning techniques to study partial aspects of energy poverty, including high energy prices, low household income, and low energy efficiency of residences [24], but a limited quantity of such works focus on energy poverty as a multidimensional phenomenon [24]. Actually, machine learning techniques addressing energy poverty alleviation have begun to gain popularity, mainly during the last five years, with the most famous among them being decision tree-based and neural network-based approaches [24].
As regards decision tree-based approaches, Longa et al. [25] used the XGBoost classifier to predict energy poverty, defined by the ratio of household energy costs to household income (four risk categories) in the Netherlands, using as input variables income and socio-economic factors (household size, economic value of the property, property status, house age, and average population density). Al Kez et al. [26] used a random forest classifier to predict energy poverty, defined by the “Low Income Low Energy Efficiency” indicator (four risk categories) in the UK, based on socio-economic data and satellite remote sensing, using secondary statistical data (UK English Housing Survey). A random forest classifier was also used by Spandagos et al. [27] to predict energy poverty for all the EU-28 countries, based on data from Eurostat’s EU-SILC survey. van Hove et al. [28] trained a gradient boosting classifier (CatBoost model) in order to identify energy poverty risk (four risk categories), based on socio-economic factors and using data from a primary survey within 11 European countries. The same algorithm (CatBoost) was also used by Mukalabai et al. [29] in order to explore energy poverty in the global south (South America, Africa, and Asia).
As regards neural network-based approaches, Pino-Mejías et al. [30] used artificial neural networks (ANNs) to predict the “Fuel Poverty Potential Risk Index” in Chile and, particularly, in the Bio-Bio Region. Building upon this work, Bienvenido-Huertas et al. [31] applied ANNs to predict the same indicator (FPPRI) in Chile’s three cities with different climate zones (Valparaiso, Concepción, and Santiago). Papada and Kaliampakos [32] used a multilayer perceptron classifier to predict energy poverty defined by various “objective” indicators, using as inputs “subjective” indicators. In the developing world, Abbas et al. [33] employed a multilayer perceptron to predict the “Multidimensional Energy Poverty Index” in Africa and Asia, using as input variables socio-economic variables of survey data.
Apart from studies using a distinct ML algorithm, there are numerous references employing two or more ML algorithms in terms of energy poverty prediction. Indicatively, Bienvenido-Huertas et al. [31] employed three ML algorithms, i.e., random forest, multilayer perceptron (ANN), and M5P, to predict the “2M” indicator of energy poverty in Spain. Hong and Park [22] applied different ML algorithms, i.e., ANN, decision tree, random forest, bagging, support vector machine, and extreme gradient boosting, to predict energy poverty based on secondary statistical data, with the random forest algorithm presenting the best performance among the predictive models. Three ML algorithms were employed by Gawusu et al. [34], including XGBoost—multiple linear regression (MLR), XGBoost—random forest (RF), and XGBoost—artificial neural network (ANN), to predict the “Multidimensional energy poverty index” in Wa, located in Ghana, concluding that the XGBoost—random forest (RF) algorithm achieved the best accuracy. Different ML algorithms were also applied by Balkissoon et al. [35], namely, decision tree, random forest, support vector machine, and extreme gradient boosting, to forecast energy poverty in Missouri, with the extreme gradient boosting showing the best performance among all the models. Grzybowska et al. [36] applied logistic regression, along with three ML algorithms, i.e., CatBoost, extreme gradient boosting, and balanced random forest, to forecast energy poverty—as being defined by the “inability to keep the home warm” indicator—in the Visegrad Group countries (Hungary, Slovakia, and Poland), based on data from Eurostat’s EU-SILC survey.
As a result, there is a growing number of studies adopting ML techniques to forecast energy poverty in the last five years. In an attempt to enhance the existing knowledge of ML performance in the field, in the present paper, different ML algorithms were employed to predict energy poverty depending on various socio-economic/technical variables within the “Urban Region of Athens” (URA) in Greece. More precisely, the Urban Region of Athens (URA) is the heart of Greater Athens, the capital of Greece, and the most densely populated area of the country. It covers a space of 412 km2 and has a population of more than 3 million people (30% of the country’s population), making it the biggest urban area in Greece. Furthermore, apart from large parts of the population, the study area concentrates most services and facilities; hence, it is a research area of particular importance.
The rest of the paper is structured as follows: Section 2 introduces the methodology employed to conduct the survey. It presents the energy poverty indicators, as well as the predictors used. Additionally, this section describes the various ML algorithms and provides a description of how the accuracy of the models was evaluated. Section 3 provides the results of the research and discusses how they can be interpreted with respect to previous studies. Section 4 presents the main conclusions of the research.

2. Materials and Methods

Data on energy poverty were collected using the Household Budget Survey (HBS) datasets that were obtained by the Hellenic Statistical Authority. The HBS survey microdata—provided at the household level—were obtained for the period 2017–2021. From the whole dataset and, in order to limit the sample to the Urban Region of Athens (URA), only observations related to the region “EL30–Attiki” (variable HA08) and to densely populated areas (variable HA09) were taken into account, amounting to 6645 observations.
The analysis aimed to predict energy poverty, as defined by different indicators (output indicators), on the basis of particular socio-economic/technical factors (input variables), using machine learning. In more detail, the authors identified five (5) energy poverty indicators as the most representative of the energy poverty problem, as they capture its crucial aspects (low income and high energy cost) from many different sides. The five (5) energy poverty indicators (output indicators) were calculated as described below:
  • The 2M indicator classifies households as energy poor if their ratio of equivalised energy expenses to equivalised disposable income exceeds twice the national median ratio. Both the variables of energy expenses and income were equivalised in order to take into account differences in size and composition of the household, i.e., energy cost was divided by the equivalisation factors and disposable income by the equivalent household size based on the scale of equivalence of the modified OECD, accordingly.
  • The M/2 indicator classifies households as energy poor if their absolute equivalised energy expenses are less than half the national median value.
  • The official national energy poverty indicator (NEPI) classifies households as energy poor if (i) the annual expenditure on the total final energy consumed in the dwelling is less than 80% of the theoretically required amount of energy and (ii) the equivalised total household income is less than 60% of the median equivalised national income, based on the respective poverty definition in Greece.
  • The modified NEPI indicator differs from the NEPI index only in point (i), by setting a different limit for classifying a household as energy poor, i.e., the limit of 60% vs. 80% used in the NEPI indicator [13]. In more detail, the share of 60% of the theoretical energy consumption, needed to guarantee an adequate level of energy services at home, represents the actual energy consumption of Greek households according to Greek circumstances.
  • The modified LIHC (low income high cost) classifies households as energy poor if they have an equivalised residual income below 60% of the equivalised median national income. In order to calculate the equivalised residual income, 60% of equivalised required energy costs was deducted from the equivalised total income of the household. Then, this value was compared to 60% of the equivalised median national income, and if it was lower, the household was regarded as energy poor. Energy costs were equivalised according to the equivalisation factors employed in the official LIHC index.
The eight (8) socio-economic/technical factors used as main predictors of energy poverty are described below:
  • Household income (HH099): continuous variable in HBS that represents net income, i.e., total income of all sources including non-monetary elements, before income taxes.
  • Household size (HB05): integer variable in HBS that represents the number of household members.
  • House area (DS017): continuous variable in HBS representing the square meters (m2) of the area of the residence.
  • Heating system efficiency: continuous variable expressed in percentage. Estimates were made by the authors based on typical efficiency ratio values of heating systems in buildings.
  • Elderly people: binary variable that indicates the presence of elderly people in households (1 if people over 65 years old live in the residence and 0 otherwise).
  • Unemployed people: binary variable that indicates the presence of unemployed people in households (1 if unemployed people live in the residence and 0 otherwise).
  • Fuel cost for heating: continuous variable that is expressed in EUR/kWh. Estimates were made by the authors based on statistics from the “Liquid Fuel Prices Observatory” of the Ministry of Development [37], as well as from other sources [38].
  • U-factor: building continuous technical variable that is expressed in W/m2 K. Estimates were made by the authors based on the proposed coefficients from the “Regulation of Energy Efficiency in Buildings” in Greece [39].
The final selection of the predictors (input variables) was based on numerous tests that examined several different variables to identify the most decisive factors affecting energy poverty. The predictors that gave the best results in the tested models were finally selected.
After the calculation of indicators and variables, the data selected were modified, i.e., missing values were detected, the file was converted to an “arff” file, etc., in order to be transferred in the appropriate format to “WEKA”, i.e., the application used to train machine learning algorithms. More precisely, written in Java at the University of New Zealand, WEKA (Waikato Environment for Knowledge Analysis) provides several machine learning algorithms and a set of data preprocessing tools [40]. The classifiers included in WEKA are generally divided into six (6) main categories, i.e., functions, decision trees, Bayesian classifiers, rules, meta-learning algorithms, and lazy classifiers.
  • From the group of functions’ classifiers, multi-layer perceptron (MLP) and logistic regression were selected for the present analysis. More specifically:
    • The multilayer perceptron is a kind of artificial neural network, which uses a nonlinear activation function within its hidden layer, providing a nonlinear mapping between the input and the output layer. Its basic structure is divided into three layers, of which one layer defines the input values, another one or more the mathematical function, and, finally, the output layer that captures the final result. A large number of neurons in each layer are linked with weighted connections, in which entire layers of neurons are interlinked through weighted connections with every neuron [41].
    • Logistic regression is a type of classification algorithm that is used under supervised learning and determines the likelihood of an event, with a binary output. In contrast with linear regression, which returns a continuous output, logistic regression returns a probability score of the outcome of the binary event. A logistic regression model identifies all the underlying “weights” and “biases” based on the provided dataset during training and understands how to correctly classify the dataset using the dependent categorical variables.
  • From the group of decision trees’ classifiers, the J48 classification algorithm and Random Forest were selected for the present analysis. In general, a decision tree builds a tree that utilizes the branching technique to show every possible outcome of a decision. Each internal node in its decision tree representation symbolizes a feature, every branch symbolizes the result of the parent node, and, finally, each leaf symbolizes the class label. In order to classify a case, a top-down “divide and conquer” strategy begins at the root of the tree [42]. More specifically:
    • The J48 classification algorithm is based on the C4.5 algorithm, an extension of Ross Quinlan’s ID3, i.e., a statistical decision tree classifier. J48 is extremely useful for examining categorical and continuous data. In WEKA, J48 includes many additional features that can be configured by the user, such as decision tree pruning, missing value estimation, and attribute value range. Pruning, for instance, is the act of selecting the largest tree and removing all the branches below this level. It is an essential process that handles the phenomenon of overfitting [43].
    • Random forest classification algorithm is based on forming multiple individual decision trees that are derived from a different sample taken from the dataset. At every tree node, a random subgroup is utilized in order to split the dataset into increasingly uniform subgroups. The subgroup of variables achieving the best performance in terms of data purity is chosen at this tree node. After creating the forest of decision trees, the ultimate outcome of classification is based on voting by every tree. The most votes give the final classification [44].
  • From the group of Bayesian classifiers, naïve Bayes was selected for the present analysis. Bayes’ theorem with independent hypotheses among the predictors is the basic concept of the naïve Bayes classifier. The simplest approach to the Bayes network in WEKA is the naïve Bayes, in which all attributes of a dataset, given the target class, are independent of each other. Thus, in naïve Bayes’ network, the class has no “parent”, i.e., it does not receive its properties from another dataset, and each attribute has the class as a unique “parent” [45]. The classification model is created without complex iterative parameter estimation. Due to this feature, Bayesian classifiers are useful even for very large datasets and provide reliable results for complex, real-world problems [41].
  • From the group of rules’ classifiers, Decision Table was selected for the present analysis. This algorithm constructs a classifier represented in the form of a decision table and represents a particularly simple classifier regarding its classification methodology. It evaluates subsets of features using a sequential search for the optimal initial choice and is able to use cross-validation for the evaluation procedures. Basically, through the decision table, a comprehensive system is created that includes all possible scenarios of conditions under which classification can be performed, from which the algorithm determines the scenario that presents the highest accuracy and probability of occurrence [46].
  • From the group of meta-learning algorithms, AdaBoost.M1 was selected for the present analysis. Meta-learning algorithms transform classification algorithms into more powerful classification tools, significantly improving their learning capabilities. This improvement is achieved by different methods, such as by combining different output data from separate classifiers. WEKA includes numerous meta-learning algorithms, which follow different forms of algorithm improvement. The main ones are boosting, bagging, and randomization, combining classifiers and cost-sensitive learning. From the available improvement algorithms, a boosting-type algorithm was selected. Boosting is a general improvement method that creates a strong classifier from a number of weak classifiers. This is carried out by creating a model from the training data and then creating a second model that tries to correct the errors of the first model. Models are added until the training set is optimally predicted or until a maximum number of models are added. A classic case of a boosting algorithm is AdaBoost.M1 (or AdaBoost). AdaBoost.M1 can be used to boost the performance of any machine learning algorithm, but it works most efficiently on algorithms with low training capacity. Thus, AdaBoost delivers maximum performance when used to boost decision trees [47].
  • From the group of lazy classifiers, IBk (instance-based learning with parameter k) was selected for the present analysis. The principle of an IBk algorithm is essentially equivalent to the kNN (k-nearest neighbor) algorithm, using the same distance metric. When classifying with kNN, the value is classified according to the majority presence of its near neighbors. The value is then assumed to belong to the most popular class among its nearest neighbors. A different type of search algorithm can be implemented to enhance the efficiency of nearest neighbor detection. While linear search is the standard approach, alternatives such as ball trees, K-D trees, and so-called “cover trees” are also available. The parameter of this classifier is the distance function. Predictions from multiple neighbors are weighted from the case tested, according to their distance [48].
In order to optimize the models, numerous different types of the above classifiers were tested in terms of their characteristics in order to detect the best features of each one. Indicatively, the “SMOTE” supervised filter (synthetic minority oversampling technique) was applied to all cases in order to handle the unbalanced classes of the dataset and minimize bias in the results. In addition, some adjustments were made to the SMOTE application parameters, e.g., to the k-nearest neighbor approach used to create synthetic cases in order to normalize data (selection of k = 10, so that the filter is instructed to create synthetic cases based on the ten nearest neighboring cases). Furthermore, the “Randomize” unsupervised filter was applied to all cases in order to avoid bias and overfitting. It should be noted though that no normalization or standardization was performed on the data. In order to train and test the models, the “Percentage Split” method was selected, and, more specifically, it was determined that 70% of the data will be used for training and 30% for testing the model.
To assess the accuracy of the models, a set of indicators was taken into account:
  • Accuracy: It is the main statistical result and the first value taken into account for the evaluation. WEKA presents the number of correct predictions and their percentage over the total predictions made. Τhis percentage is referred to simply as the “Accuracy” of the model. It is defined by the following equation:
Accuracy = True Instances/(True Instances + False Instances)
  • Detailed accuracy by class: The main statistical metrics that represent the accuracy level by class are the following:
    • Precision or positive predictive: It refers to the ratio of true positive instances to the total positive (true and false) instances of each class. It is defined by the following equation:
      Precision = True Positive Instances/(True Positive Instances + False Positive Instances)
    • Recall or sensitivity: It refers to the ratio of true positive instances to the sum of true positive and false negative instances of each class. It is defined by the following equation:
      Recall = True Positive Instances/(True Positive Instances + False Negative Instances)
    • F-measure: It expresses the harmonic mean of “Recall” and “Precision” values. It allows a model to be evaluated taking into account symmetrically the above factors in one metric, which is useful when describing the performance of the model. It is defined by the following equation:
      F–Measure = (2 × Precision × Recall)/(Precision + Recall)
    • ROC (receiver operating characteristic) area: It is an accuracy measure, which indicates the ability of the model to accurately predict random data. It aims to be as high as possible for a high-performing model.
  • Confusion matrix: It is a 2 × 2 matrix in which the percentages of true and false instances of each class are presented, with the diagonal elements representing the true instances (true positive and true negative) for each class of the output variable (Figure 1).
Figure 1. The confusion matrix.
Figure 1. The confusion matrix.
Energies 18 01133 g001
A detailed flowchart displaying the entire methodology performed is illustrated in Figure 2.
Figure 2. Flowchart of the methodology performed.
Figure 2. Flowchart of the methodology performed.
Energies 18 01133 g002

3. Results and Discussion

Table 1, Table 2, Table 3, Table 4 and Table 5 present the performance of the machine learning classifiers tested per indicator. Among the hundreds of tests performed aiming at improving performance levels, the best models are the only ones presented, i.e., those with the best performance metrics, along with the highest accuracy scores.

3.1. Prediction of the Indicator “NEPI”

Regarding the prediction of the indicator “NEPI”, almost all classifiers present highly valuable models, with the accuracy score ranging from 87.77% to 94.00% (Table 1). Small differences appear between the models presented, as the performance metrics (precision, recall, F-measure, and ROC area) exceed 0.75 for the weighted average in all models, or even 0.87 if not taking into account the “Lazy IBk” classifier. The correctly predicted instances range between 95% and 97% for the class of the non-energy-poor households and between 57% and 82% for the class of the energy-poor households. Overall, the best performance of the indicator “NEPI” is achieved by the “Random Forest” (94.02%) and the “Multilayer Perceptron” (93.34%) classifiers, with negligible differences between them, followed by “Trees J48” (92.79%), “Rules Decision table” (92.15%), and “Logistic” (91.84%) classifiers (also high predictive models) and, at a longer distance by “Meta AdaBoost M1”, “Naïve Bayes” and “Lazy IBk” classifiers. A graphical interface of classifiers is given indicatively in Figure 3, displaying the ANN (multilayer perceptron) predicting the indicator “NEPI”.
The model resulting from the “Random Forest” classifier presents the best results of the seven classifiers, with an accuracy rate of 94.02%. The performance metrics (precision, recall, F-measure, and ROC area) of this model are remarkably good, exceeding 0.93 for the weighted average. ROC area is equal to 0.97, almost approaching 1.0, implying that the examined model is approaching an ideal model. Similarly, 97% of non-energy-poor households and 81% of energy-poor households are correctly predicted, according to the diagonal elements of the confusion matrix. In the best model, the supervised filter “SMOTE” was used to adapt the relative frequency between majority and minority classes, i.e., increase twofold the instances of the minority class based on the k-nearest neighbor method (10 nearest neighbors selected), and then the unsupervised filter “Randomize” was used in order to shuffle the order of instances.
When removing the variable of “household income” as an input variable, the predictive power of the model with respect to energy-poor households falls drastically. In more detail, in this case, the “Random Forest” classifier manages to correctly predict only 32% of energy-poor households while maintaining significantly high percentages in terms of correctly predicting the class of non-energy-poor households (95%), according to the confusion matrix. This model presents an overall accuracy score of 83.67% and good performance metrics (precision, recall, F-measure, and ROC area), exceeding 0.79 for the weighted average, yet it is regarded as an unreliable model due to the extremely low rates of the second class. The above finding highlights the principal role of income in energy poverty prediction, as without it, the model loses its predictive power of the one class by almost 50 percentage points.
Table 1. Prediction of indicator “NEPI” and confusion matrices (test sets presented).
Table 1. Prediction of indicator “NEPI” and confusion matrices (test sets presented).
Prediction of Indicator NEPI–2192 Instances
ClassifierPrecisionRecallF-MeasureROC AreaClassAccuracyConfusion Matrix
Multilayer perceptron0.9590.9600.9590.973Non-energy poor93.34%96%4%
0.8200.8160.8180.973Energy poor18%82%
0.9330.9330.9330.973(weighted avg)
Naïve Bayes0.9200.9370.9290.925Non-energy poor88.23%94%6%
0.6950.6390.6660.925Energy poor36%64%
0.8790.8820.8800.925(weighted avg)
Rules decision table0.9370.9690.9530.968Non-energy poor92.15%97%3%
0.8380.7090.7680.968Energy poor29%71%
0.9190.9220.9190.968(weighted avg)
Meta AdaBoost M10.9200.9610.9400.938Non-energy poor90.06%96%4%
0.7860.6290.6990.938Energy poor37%63%
0.8960.9010.8960.938(weighted avg)
Logistic0.9390.9630.9510.968Non-energy poor91.84%96%4%
0.8120.7210.7640.968Energy poor28%72%
0.9160.9180.9160.968(weighted avg)
Trees J480.9470.9650.9560.923Non-energy poor92.79%97%3%
0.8320.7610.7950.923Energy poor24%76%
0.9260.9280.9270.923(weighted avg)
Lazy IBk0.9070.9470.9270.758Non-energy poor87.77%95%5%
0.7070.5700.6310.758Energy poor43%57%
0.8710.8780.8720.758(weighted avg)
Random forest0.9580.9690.9640.971Non-energy poor94.02%97%3%
0.8560.8110.8330.971Energy poor19%81%
0.9390.9400.9400.971(weighted avg)
In other words, the “Random Forest” classifier, using the eight (8) input socio-economic/technical variables, i.e., household income, household size, house area, heating fuel cost, heating system efficiency, presence of elderly people, presence of unemployed people, and building U-factor, can correctly predict energy poverty, as being defined by the indicator “NEPI”, at a remarkably good level (correct prediction of energy-poor households of 81%), with the overall accuracy rate of the model reaching 93.34%.
Compared with previous relevant research, Papada and Kaliampakos [21] used a neural network (multilayer perceptron classifier) to predict the same output variable (NEPI) in Greece with the same machine learning tool (WEKA), using as independent variables socio-economic/geographical factors, i.e., household size, house area, house age, ownership status, and elevation. The best model of these tests showed an overall accuracy score lower compared to the present outcomes of the “Random Forest” classifier (82.72% vs. 93.34%) and lower performance metrics (over 0.82 vs. over 0.93 of the present model for the weighted average of all metrics). As per the confusion matrix, the best model of [32] showed fewer correct predictions, i.e., 73% vs. 82% for the class of energy-poor households and 88% vs. 96% for the class of non-energy-poor households. The two models are comparable as they use the same methodology for predicting the same indicator but different samples and different input variables. Overall, the performance of the present model is significantly higher, probably because of the use of more (in number) and more detailed input variables.

3.2. Prediction of the Indicator “Modified NEPI”

As regards the prediction of the indicator “modified NEPI”, highly valuable models arise, as also was the case with the “NEPI” indicator. More precisely, all classifiers present high accuracy scores, ranging from 86.62% to 94.45% (Table 2). Small differences appear between most of the models presented, as the performance metrics (precision, recall, F-measure, and ROC area) exceed 0.77 for the weighted average in all models, or even 0.87 if not taking into account the “Lazy IBk” classifier. The correctly predicted instances range between 93% and 97% for the class of non-energy-poor households and between 57% and 83% for the class of energy-poor households. Overall, the best performance of the indicator “modified NEPI” is achieved by the “Random Forest” (94.45%) and “Trees J48” (93.36%) classifiers with negligible differences between them, followed by “Multilayer Perceptron” (92.54%), “Logistic” (92.30%), and “Rules Decision table” (91.53%) classifiers (also high predictive models) and, at a longer distance, by “Meta AdaBoost M1”, “Naïve Bayes”, and “Lazy IBk” classifiers.
The model resulting from the use of the “Random Forest” classifier presents the best results, with an accuracy rate of 94.45%. The performance metrics (precision, recall, F-measure, and ROC area) are remarkably good, exceeding 0.94 for the weighted average. ROC area equals 0.98, almost approaching the ideal unit of 1.0. Similarly, 97% of non-energy-poor households and 83% of energy-poor households are correctly predicted, according to the diagonal elements of the confusion matrix. In the best model, the supervised filter “SMOTE” was used to adapt the relative frequency between majority and minority classes, based on the k-nearest neighbor method (10 nearest neighbors selected), and then the unsupervised filter “Randomize” was used in order to shuffle the order of instances.
If removing the variable of “household income” as an input variable, the predictive power of the model for energy-poor households falls drastically, as also was the case with the “NEPI” indicator. Similarly, in the absence of income, the “Random Forest” classifier manages to correctly predict only 37% of energy-poor households, and, despite its impressive predictive power of non-energy-poor households (97%) with an accuracy rate of 85.75%, the model is regarded as unreliable in total.
In other words, the “Random Forest” classifier, using the eight (8) input socio-economic/technical variables, can correctly predict energy poverty, as being defined by the indicator “modified NEPI”, at a remarkably good level (correct prediction of energy-poor households of 83%), with the overall accuracy rate of the model reaching 94.45%.
Table 2. Prediction of indicator “modified NEPI” and confusion matrices (test sets presented).
Table 2. Prediction of indicator “modified NEPI” and confusion matrices (test sets presented).
Prediction of Indicator Modified NEPI–2197 Instances
ClassifierPrecisionRecallF-MeasureROC
Area
ClassAccuracyConfusion Matrix
Multilayer
perceptron
0.9470.9620.9540.972Non-energy poor92.54%96%4%
0.8260.7710.7980.972Energy poor23%77%
0.9240.9250.9240.972(weighted avg)
Naïve Bayes0.9050.9320.9190.914Non-energy poor86.62%93%7%
0.6700.5870.6260.914Energy poor41%59%
0.8610.8660.8630.914(weighted avg)
Rules decision table0.9310.9670.9490.952Non-energy poor91.53%97%3%
0.8320.6970.7580.952Energy poor30%70%
0.9120.9150.9120.952(weighted avg)
Meta AdaBoost M10.9040.9590.9310.902Non-energy poor88.44%96%4%
0.7650.5680.6520.902Energy poor43%57%
0.8780.8840.8780.902(weighted avg)
Logistic0.9400.9660.9530.970Non-energy poor92.30%97%3%
0.8380.7400.7860.970Energy poor26%74%
0.9210.9230.9210.970(weighted avg)
Trees J480.9560.9620.9590.945Non-energy poor93.36%96%4%
0.8340.8140.8240.945Energy poor19%81%
0.9330.9340.9330.945(weighted avg)
Lazy IBk0.9090.9430.9260.774Non-energy poor87.80%94%6%
0.7140.6010.6530.774Energy poor40%60%
0.8720.8780.8740.774(weighted avg)
Random forest0.9600.9720.9660.975Non-energy poor94.45%97%3%
0.8740.8280.8500.975Energy poor17%83%
0.9440.9440.9440.975(weighted avg)

3.3. Prediction of the Indicator “2M”

Regarding the prediction of the indicator “2M”, classifiers present reasonable models, with high accuracy scores but average correct predictions of the class of energy-poor households. More precisely, the accuracy score ranges from 93.34% to 97.24% (Table 3), while the performance metrics (precision, recall, F-measure, and ROC area) exceed 0.71 for the weighted average in all models, or even 0.86 if not taking into account the “Lazy IBk” classifier. The confusion matrix provides remarkably high values for the class of non-energy-poor households in all models (of the order of 98–100%), but average rates for the class of energy-poor households (43% to 65%). As a result, the models derived from “Naïve Bayes”, “Lazy IBk”, and “Logistic” classifiers are considered non-meaningful at all, due to the low rates (below and around 50%) of the confusion matrix of the second class.
Overall, the best performance of the indicator “2M” is achieved by the “Rules Decision table” (97.24%) and the “Random Forest” (96.70%) classifiers with minor differences between them, followed by “Trees J48” (96.00%), “Meta AdaBoost M1” (95.34%), and “Multilayer Perceptron” (93.34%) classifiers.
Table 3. Prediction of indicator “2M” and confusion matrices (test sets presented).
Table 3. Prediction of indicator “2M” and confusion matrices (test sets presented).
Prediction of Indicator 2M–2062 Instances
ClassifierPrecisionRecallF-MeasureROC
Area
ClassAccuracyConfusion Matrix
Multilayer perceptron0.9680.9860.9770.934Non-energy poor93.34%99%1%
0.7720.5870.6670.934Energy poor41%59%
0.9540.9570.9550.934(weighted avg)
Naïve Bayes0.9570.9830.9700.914Non-energy poor94.33%98%2%
0.6700.4330.5260.914Energy poor57%43%
0.9360.9430.9380.914(weighted avg)
Rules decision table0.9730.9970.9850.965Non-energy poor97.24%100%0%
0.9510.6530.7750.965Energy poor35%65%
0.9720.9720.9700.965(weighted avg)
Meta AdaBoost M10.9680.9830.9750.928Non-energy poor95.34%98%2%
0.7250.5800.6440.928Energy poor42%58%
0.9500.9530.9510.928(weighted avg)
Logistic0.9630.9890.9760.944Non-energy poor95.49%99%1%
0.7880.5200.6270.944Energy poor48%52%
0.9510.9550.9510.944(weighted avg)
Trees J480.9670.9850.9760.865Non-energy poor95.54%99%1%
0.7540.5730.6520.865Energy poor43%57%
0.9520.9550.9530.865(weighted avg)
Lazy IBk0.9600.9790.9690.716Non-energy poor94.23%98%2%
0.6400.4730.5440.716Energy poor53%47%
0.9360.9420.9380.716(weighted avg)
Random forest0.9710.9940.9820.942Non-energy poor96.70%99%1%
0.8940.6200.7320.942Energy poor38%62%
0.9650.9670.9640.942(weighted avg)
The model resulting from the use of the “Rules Decision table” classifier displays the best results, with an accuracy rate of 97.24%. The performance metrics (precision, recall, F-measure, and ROC area) are remarkably good, exceeding 0.96 for the weighted average. ROC area is equal to 0.97, implying that the examined model is almost an ideal model. Similarly, as per the diagonal parts of the confusion matrix, 100% of non-energy-poor households and 65% of energy-poor households are correctly predicted. In the best model, the supervised filter “SMOTE” was used to adapt the relative frequency between majority and minority classes, based on the k-nearest neighbor method (10 nearest neighbors selected), and then the unsupervised filter “Randomize” was used in order to shuffle the order of instances.
The best model was also tested for the case of removing the variable of “household income” as an input variable. The findings of the previous indicators—in terms of the crucial role of income in energy poverty prediction—were once again validated, as the new model without income manages to correctly predict only 31% of energy-poor households; hence, it is regarded as unreliable in total.
In other words, the “Rules Decision table” classifier, using the eight (8) input socio-economic/technical variables selected, can correctly predict energy poverty, as being defined by the indicator “2M”, at a moderately satisfactory level (correct prediction of energy-poor households of 65%), with the overall accuracy rate of the model reaching 97.24%.

3.4. Prediction of the Indicator “M/2”

As regards the prediction of the indicator “M/2”, almost all classifiers present unreasonable models, with satisfactory and high accuracy scores but particularly low correct predictions of the second class (energy-poor households). More precisely, the accuracy score ranges from 75.46% to 86.81% (Table 4), and the performance metrics (precision, recall, F-measure, and ROC area) exceed 0.69 for the weighted average in all models.
Table 4. Prediction of indicator “M/2” and confusion matrices (test sets presented).
Table 4. Prediction of indicator “M/2” and confusion matrices (test sets presented).
Prediction of Indicator M/2–2274 Instances
ClassifierPrecisionRecallF-MeasureROC AreaClassAccuracyConfusion Matrix
Multilayer perceptron0.8310.9820.9000.809Non-energy poor83.47%98%2%
0.8690.3730.5220.809Energy poor63%37%
0.8400.8350.8090.809(weighted avg)
Naïve Bayes0.8190.8690.8430.709Non-energy poor75.46%87%13%
0.4910.3960.4390.709Energy poor60%40%
0.7390.7550.7450.709(weighted avg)
Rules decision table0.8570.9830.9160.846Non-energy poor86.28%98%2%
0.9020.4850.6310.846Energy poor51%49%
0.8680.8630.8470.846(weighted avg)
Meta AdaBoost M10.8140.9660.8840.818Non-energy poor80.74%97%3%
0.7460.3090.4370.818Energy poor69%31%
0.7980.8070.7760.818(weighted avg)
Logistic0.7900.9500.8620.725Non-energy poor77.00%95%5%
0.5670.2070.3040.725Energy poor79%21%
0.7360.7700.7270.725(weighted avg)
Trees J480.8650.9340.8980.818Non-energy poor83.95%93%7%
0.7240.5440.6210.818Energy poor46%54%
0.8310.8390.8310.818(weighted avg)
Lazy IBk0.8520.8610.8570.697Non-energy poor78.14%86%14%
0.5500.5330.5410.697Energy poor47%53%
0.7790.7810.7800.697(weighted avg)
Random forest0.8760.9630.9170.875Non-energy poor86.81%96%4%
0.8310.5710.6770.875Energy poor43%57%
0.8650.8680.8590.875(weighted avg)
The confusion matrix provides high rates for the class of non-energy-poor households in all models (86% to 98%), but considerably low rates for the class of energy-poor households (21% to 57%). As a result, the models derived from almost all classifiers, i.e., “Logistic”, “Meta AdaBoost M1”, “Multilayer Perceptron”, “Naïve Bayes”, “Rules Decision table”, and “Lazy IBk” classifiers, are considered non-meaningful at all, due to the low rates (below and around 50%) of the confusion matrix of the second class.
The models resulting from the “Random Forest” and “Trees J48” classifiers are the only ones acceptable, both presenting high accuracy scores (86.81% and 83.95%, respectively) but low correct predictions of the second class (57% and 54%, respectively). Especially the “Random Forest” classifier displays the best results between the two, with an accuracy rate of 86.81% and good performance metrics (precision, recall, F-measure, and ROC area), exceeding 0.85 for the weighted average. In this model, 96% of non-energy-poor households and 57% of energy-poor households are correctly predicted, according to the diagonal parts of the confusion matrix. In this model, the supervised filter “SMOTE” was used to adapt the relative frequency between majority and minority classes, i.e., increase twofold the instances of the minority class based on the k-nearest neighbor method (10 nearest neighbors selected), and then the unsupervised filter “Randomize” was used to shuffle the order of instances. When removing the input variable of “household income”, the new model—without income—manages to correctly predict even fewer energy-poor households (49%), which is not meaningful at all.
In other words, the “Random Forest” classifier, using the eight (8) input socio-economic/technical variables selected, can correctly predict energy poverty, as being defined by the indicator “M/2”, at a marginally accepted level (correct prediction of energy-poor households of 57%), with the overall accuracy rate of the model reaching 86.81%. Practically, this model fails to successfully predict energy poverty as being defined by the indicator “M/2”.

3.5. Prediction of the Indicator “Modified LIHC”

Regarding the prediction of the indicator “modified LIHC”, all classifiers present highly valuable models, with the accuracy score ranging from 84.84% to 96.37% (Table 5). Small differences appear between the models presented, as the performance metrics (precision, recall, F-measure, and ROC area) are considerably good, exceeding 0.84 for the weighted average in all models. The correctly predicted instances range between 86% and 97% for the class of non-energy-poor households and between 80% and 97% for the class of energy-poor households. Although differences are slight, overall, the best performance of the indicator “modified LIHC” is achieved by the “Random Forest” classifier (96.37%), followed by “Trees J48” (95.47%), “Multilayer Perceptron” (94.69%), and “Logistic” (94.69%) classifiers and subsequently by “Rules Decision table” (93.33%), Meta AdaBoost M1” (89.45%), “Lazy IBk” (89.37%), and “Naïve Bayes” classifiers (84.84%).
The model resulting from the “Random Forest” classifier presents the best results of the seven classifiers, with an accuracy rate of 96.37%. The performance metrics (precision, recall, F-measure, and ROC area) are remarkably good, exceeding 0.96 for the weighted average. ROC area is equal to 0.99, i.e., an ideal model in terms of random classification. Similarly, 97% of non-energy-poor households and 94% of energy-poor households are correctly predicted, according to the diagonal elements of the confusion matrix. The supervised filter “SMOTE” was used to adapt the relative frequency between majority and minority classes based on the k-nearest neighbor method (10 nearest neighbors selected), and then the unsupervised filter “Randomize” was used to shuffle the order of instances.
When removing the input variable of “household income”, the predictive power of the model with respect to energy-poor households falls considerably, as happened with all previous indicators. Hence, in the absence of income, the model manages to correctly predict 64% of energy-poor households and 87% of non-energy-poor households, with an accuracy rate of 78.90%, which is accepted as a model according to metrics, but it is far less valuable compared to the case of incorporating “household income” as an input variable.
In other words, the “Random Forest” classifier, using the eight (8) input socio-economic/technical variables selected, can correctly predict energy poverty, as being defined by the indicator “modified LIHC” at a remarkably good level (correct prediction of energy-poor households of 94%), with the overall accuracy rate of the model reaching 96.37%.
Comparing with previous relevant research, a similar decision tree classifier (Random Forest) was employed by Al Kez et al. [26] in order to predict—not the same but—a similar energy poverty indicator in the UK (the “LILEE indicator”, or low income low energy efficiency). This model, which used as input variables income, efficiency, and eight (8) additional variables, presented an impressive accuracy score close to 100%, but when income and efficiency were excluded from the model, the accuracy score decreased to 67%. The rates are very close to those of the present paper, confirming the great performance of a decision tree (especially of “Random Forest”) classifier in energy poverty prediction, as well as the decisive role of income in energy poverty prediction via machine learning.
Table 5. Prediction of indicator “modified LIHC” and confusion matrices (test sets presented).
Table 5. Prediction of indicator “modified LIHC” and confusion matrices (test sets presented).
Prediction of Indicator Modified LIHC–2427 Instances
ClassifierPrecisionRecallF-MeasureROC
Area
ClassAccuracyConfusion Matrix
Multilayer perceptron0.9830.9360.9590.991Non-energy poor94.69%94%6%
0.8850.9680.9240.991Energy poor3%97%
0.9500.9470.9470.991(weighted avg)
Naïve Bayes0.9030.8640.8830.913Non-energy poor84.84%86%14%
0.7530.8170.7840.913Energy poor18%82%
0.8530.8480.8500.913(weighted avg)
Rules decision table0.9400.9600.9500.982Non-energy poor93.33%96%4%
0.9180.8800.8980.982Energy poor12%88%
0.9330.9330.9330.982(weighted avg)
Meta AdaBoost M10.9620.8760.9170.973Non-energy poor89.45%88%12%
0.7910.9310.8560.973Energy poor7%93%
0.9050.8950.8960.973(weighted avg)
Logistic0.9590.9610.9600.987Non-energy poor94.69%96%4%
0.9220.9190.9210.987Energy poor8%92%
0.9470.9470.9470.987(weighted avg)
Trees J480.9690.9620.9660.968Non-energy poor95.47%96%4%
0.9260.9400.9330.968Energy poor6%94%
0.9550.9550.9550.968(weighted avg)
Lazy IBk0.9040.9390.9210.871Non-energy poor89.37%94%6%
0.8700.8040.8350.871Energy poor20%80%
0.8930.8940.8930.871(weighted avg)
Random forest0.9720.9740.9730.994Non-energy poor96.37%97%3%
0.9480.9440.9460.994Energy poor6%94%
0.9640.9640.9640.994(weighted avg)

4. Conclusions

The present research enlightens an innovative practice of analyzing energy poverty, i.e., machine learning. A distinctive aspect of the research is that a model is authorized to learn from real data on how to weight the selected drivers/variables and determine their interrelationships, contrary to traditional statistical techniques in which the weighting of variables is a prerequisite for running a model. Hence, the present work addresses a gap in the literature regarding the absence of an “established protocol” for weighting drivers of energy poverty, as highlighted by Walker et al. [49] and Longa et al. [25].
The analysis showed that almost all classifiers managed to successfully predict three (3) out of five (5) energy poverty indicators with a high level of accuracy, namely, “NEPI”, “modified NEPI”, and “modified LIHC” (81%, 83%, and 94% correct predictions of energy-poor households of the best models, respectively, with an overall accuracy rate of the models over 94% in all cases). However, considerably lower rates were achieved in the case of “2M” and “M/2” indicators, i.e., 65% and 57% correct predictions of energy-poor households, with 97% and 86% total accuracy scores of the models, respectively. Overall, the most successful classifier in terms of energy poverty prediction proved to be the “Random Forest” classifier (decision tree approach), closely followed by “Trees J48” and “Multilayer Perceptron” classifiers (decision tree and neural network approach, respectively) and, at a longer distance, by “Logistic”, “Rules Decision table”, and “Meta AdaBoost M1” classifiers. The weakest classifiers, or else, classifiers with the lowest performance, proved to be “Lazy IBk” and “Naïve Bayes” classifiers.
Another interesting finding is the validation of the principal role of income in energy poverty prediction when applying machine learning techniques. More specifically, when removing the variable of “household income” from the input variables, the predictive power of the best models concerning energy-poor households drops drastically in almost all cases (by 30–50 percentage points, accompanied also by a total lower accuracy score of the models). With the inclusion of income, though, “Random Forest”, “Trees J48”, and “Multilayer Perceptron” classifiers presented an impressive performance, achieving particularly high success rates.
Moreover, it is noteworthy that, although numerous different types of classifiers were tested in terms of their characteristics, the best results were reached by their simplest types, e.g., by selecting one (1) instead of two (2) or more hidden layers in the “Multilayer Perceptron” classifier, etc. Furthermore, the filter “SMOTE” was applied to all cases, again through numerous tests of the filter’s characteristics, in order to adapt the relative frequency between majority and minority classes and thus minimize bias in the results.
The particularly high prediction rates and accuracy scores that came up in the study area confirmed that ML algorithms could be a useful tool in terms of detecting energy poverty roots, without knowing their exact relationships with the problem. This information can in turn benefit decision-making. Practically, the present research showed that eight (8) specific socio-economic/technical variables, i.e., household income, household size, house area, heating fuel cost, heating system efficiency, presence of elderly people, presence of unemployed people, and building U-factor, are of particular importance when combined all together, as they can accurately predict energy poverty by over 80%. By this means, machine learning can serve as a valuable policy tool for decision-makers, in order to spot vulnerable households and, hence, take more focused policy measures, e.g., lower energy prices, tax reliefs, allowances, motives towards upgrading the building’s energy performance, etc.
As regards future research, one possible direction would include the exploration of whether the high rates of the predictive models or the predictors themselves would significantly change depending on the sample, i.e., by testing different study areas/countries with various geographical characteristics that could affect energy poverty in different ways, such as mountainous areas, lowlands, island areas, etc. In any case, due to the exact complex nature of energy poverty, machine learning seems to be a promising tool towards experimenting quickly with large amounts of data (variables/characteristics/causes of the problem) and shaping appropriate energy policies.

Author Contributions

Conceptualization, D.D., E.K. and L.P.; methodology, L.P., D.K. and E.K.; formal analysis, L.P. and E.K.; resources, E.K., C.T. and S.M.; data curation, C.T. and S.M.; writing—original draft preparation, E.K. and L.P.; visualization, E.K. and L.P.; writing—review and editing, L.P.; supervision, D.D.; funding acquisition, D.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was co-funded by the European Union (LIFE programme) under Grant Agreement no. 101076277. Views and opinions expressed are however those of the author(s) only and do not necessarily reflect those of the European Union or CINEA. Neither the European Union nor the granting authority can be held responsible for them.

Data Availability Statement

Restrictions apply to the availability of these data. Data were obtained from the Hellenic Statistical Authority and are available with the permission of the Hellenic Statistical Authority.

Acknowledgments

The authors would like to acknowledge the provision of data by the Hellenic Statistical Authority.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. IEA. SDG7: Data and Projections. 2023. Available online: https://www.iea.org/reports/sdg7-data-and-projections (accessed on 19 December 2024).
  2. Samarakoon, S. A justice and wellbeing centered framework for analysing energy poverty in the Global South. Ecol. Econ. 2019, 165, 106385. [Google Scholar] [CrossRef]
  3. Hesselman, M.; Varo, A.; Guyet, R.; Thomson, H. Energy poverty in the COVID-19 era: Mapping global responses in light of momentum for the right to energy. Energy Res. Soc. Sci. 2021, 81, 102246. [Google Scholar] [CrossRef] [PubMed]
  4. Carfora, A.; Scandurra, G.; Thomas, A. Forecasting the COVID-19 effects on energy poverty across EU member states. Energy Policy 2022, 161, 112597. [Google Scholar] [CrossRef] [PubMed]
  5. Baker, S.H.; Carley, S.; Konisky, D.M. Energy insecurity and the urgent need for utility disconnection protections. Energy Policy 2021, 159, 112663. [Google Scholar] [CrossRef]
  6. Clark, I.; Chun, S.; O’Sullivan, K.; Pierse, N. Energy Poverty among Tertiary Students in Aotearoa New Zealand. Energies 2021, 15, 76. [Google Scholar] [CrossRef]
  7. Papada, L.; Kaliampakos, D. A Stochastic Model for energy poverty analysis. Energy Policy 2018, 116, 153–164. [Google Scholar] [CrossRef]
  8. Eurostat. People at Risk of Poverty or Social Exclusion. 2023. Available online: https://ec.europa.eu/eurostat/databrowser/view/sdg_01_10/default/table?lang=en (accessed on 19 December 2024).
  9. Atsalis, A.; Mirasgedis, S.; Tourkolias, C.; Diakoulaki, D. Fuel poverty in Greece: Quantitative analysis and implications for policy. Energy Build. 2016, 131, 87–98. [Google Scholar] [CrossRef]
  10. Papada, L.; Kaliampakos, D. Measuring energy poverty in Greece. Energy Policy 2016, 94, 157–165. [Google Scholar] [CrossRef]
  11. Papada, L.; Kaliampakos, D. Energy poverty in Greek mountainous areas: A comparative study. J. Mt. Sci. 2017, 14, 1229–1240. [Google Scholar] [CrossRef]
  12. Spiliotis, E.; Arsenopoulos, A.; Kanellou, E.; Psarras, J.; Kontogiorgos, P. A multi-sourced data based framework for assisting utilities identify energy poor households: A case-study in Greece. Energy Sources Part B Econ. Plan. Policy 2020, 15, 49–71. [Google Scholar] [CrossRef]
  13. Kalfountzou, E.; Tourkolias, C.; Mirasgedis, S.; Damigos, D. Identifying Energy-Poor Households with Publicly Available Information: Promising Practices and Lessons Learned from the Athens Urban Area, Greece. Energies 2024, 17, 919. [Google Scholar] [CrossRef]
  14. Ntaintasis, E.; Mirasgedis, S.; Tourkolias, C. Comparing different methodological approaches for measuring energy poverty: Evidence from a survey in the region of Attika, Greece. Energy Policy 2019, 125, 160–169. [Google Scholar] [CrossRef]
  15. Boemi, S.-N.; Avdimiotis, S.; Papadopoulos, A.M. Domestic energy deprivation in Greece: A field study. Energy Build. 2017, 144, 167–174. [Google Scholar] [CrossRef]
  16. Palmos Analysis. Thessaloniki: 190,000 Households Vulnerable or in a State of Energy Poverty. Available online: https://parallaximag.gr/life/energiaki-ftochia-stopoleodomiko-sigkrotima-thessalonikis (accessed on 20 December 2024).
  17. Papada, L.; Kaliampakos, D. Being forced to skimp on energy needs: A new look at energy poverty in Greece. Energy Res. Soc. Sci. 2020, 64, 101450. [Google Scholar] [CrossRef]
  18. Lyra, K.; Mirasgedis, S.; Tourkolias, C. From measuring fuel poverty to identification of fuel poor households: A case study in Greece. Energy Effic. 2022, 15, 6. [Google Scholar] [CrossRef]
  19. Kalfountzou, E.; Papada, L.; Damigos, D.; Degiannakis, S. Predicting energy poverty in Greece through statistical data analysis. Int. J. Sustain. Energy 2022, 41, 1605–1622. [Google Scholar] [CrossRef]
  20. Halkos, G.; Kostakis, I. Exploring the persistence and transience of energy poverty: Evidence from a Greek household survey. Energy Effic. 2023, 16, 50. [Google Scholar] [CrossRef]
  21. Papada, L.; Kaliampakos, D. Artificial Neural Networks as a Tool to Understand Complex Energy Poverty Relationships: The Case of Greece. Energies 2024, 17, 3163. [Google Scholar] [CrossRef]
  22. Hong, Z.; Park, I.K. Comparative Analysis of Energy Poverty Prediction Models Using Machine Learning Algorithms. J. Korea Plan. Assoc. 2021, 56, 239–255. [Google Scholar] [CrossRef]
  23. Hassani, H.; Yeganegi, M.R.; Beneki, C.; Unger, S.; Moradghaffari, M. Big Data and Energy Poverty Alleviation. Big Data Cognit. Comput. 2019, 3, 50. [Google Scholar] [CrossRef]
  24. López-Vargas, A.; Ledezma-Espino, A.; Sanchis-de-Miguel, A. Methods, data sources and applications of the Artificial Intelligence in the Energy Poverty context: A review. Energy Build. 2022, 268, 112233. [Google Scholar] [CrossRef]
  25. Dalla Longa, F.; Sweerts, B.; Van Der Zwaan, B. Exploring the complex origins of energy poverty in The Netherlands with machine learning. Energy Policy 2021, 156, 112373. [Google Scholar] [CrossRef]
  26. Al Kez, D.; Foley, A.; Abdul, Z.K.; Del Rio, D.F. Energy poverty prediction in the United Kingdom: A machine learning approach. Energy Policy 2024, 184, 113909. [Google Scholar] [CrossRef]
  27. Spandagos, C.; Tovar Reaños, M.A.; Lynch, M.Á. Energy poverty prediction and effective targeting for just transitions with machine learning. Energy Econ. 2023, 128, 107131. [Google Scholar] [CrossRef]
  28. Van Hove, W.; Dalla Longa, F.; Van Der Zwaan, B. Identifying predictors for energy poverty in Europe using machine learning. Energy Build. 2022, 264, 112064. [Google Scholar] [CrossRef]
  29. Mukelabai, M.D.; Wijayantha, K.G.U.; Blanchard, R.E. Using machine learning to expound energy poverty in the global south: Understanding and predicting access to cooking with clean energy. Energy AI 2023, 14, 100290. [Google Scholar] [CrossRef]
  30. Pino-Mejías, R.; Pérez-Fargallo, A.; Rubio-Bellido, C.; Pulido-Arcas, J.A. Artificial neural networks and linear regression prediction models for social housing allocation: Fuel Poverty Potential Risk Index. Energy 2018, 164, 627–641. [Google Scholar] [CrossRef]
  31. Bienvenido-Huertas, D.; Pérez-Fargallo, A.; Alvarado-Amador, R.; Rubio-Bellido, C. Influence of climate on the creation of multilayer perceptrons to analyse the risk of fuel poverty. Energy Build. 2019, 198, 38–60. [Google Scholar] [CrossRef]
  32. Papada, L.; Kaliampakos, D. Exploring Energy Poverty Indicators Through Artificial Neural Networks. In Artificial Intelligence and Sustainable Computing; Pandit, M., Gaur, M.K., Rana, P.S., Tiwari, A., Eds.; Algorithms for Intelligent Systems; Springer Nature: Singapore, 2022; pp. 231–242. [Google Scholar] [CrossRef]
  33. Abbas, K.; Butt, K.M.; Xu, D.; Ali, M.; Baz, K.; Kharl, S.H.; Ahmed, M. Measurements and determinants of extreme multidimensional energy poverty using machine learning. Energy 2022, 251, 123977. [Google Scholar] [CrossRef]
  34. Gawusu, S.; Jamatutu, S.A.; Ahmed, A. Predictive Modeling of Energy Poverty with Machine Learning Ensembles: Strategic Insights from Socioeconomic Determinants for Effective Policy Implementation. Int. J. Energy Res. 2024, 2024, 9411326. [Google Scholar] [CrossRef]
  35. Balkissoon, S.; Fox, N.; Lupo, A.; Haupt, S.E.; Penny, S.G.; Miller, S.J.; Beetstra, M.; Sykuta, M.; Ohler, A. Forecasting energy poverty using different machine learning techniques for Missouri. Energy 2024, 313, 133904. [Google Scholar] [CrossRef]
  36. Grzybowska, U.; Wojewódzka-Wiewiórska, A.; Vaznonienė, G.; Dudek, H. Households Vulnerable to Energy Poverty in the Visegrad Group Countries: An Analysis of Socio-Economic Factors Using a Machine Learning Approach. Energies 2024, 17, 6310. [Google Scholar] [CrossRef]
  37. Ministry of Development. Liquid Fuel Prices Observatory. Available online: http://www.fuelprices.gr/ (accessed on 7 January 2025). (In Greek).
  38. General Secretariat of Commerce & Consumer Protection. Refinery Prices. Available online: http://oil.gge.gov.gr/ (accessed on 20 December 2024). (In Greek)
  39. Ministry of the Environment and Energy. Greek Regulation of Energy Efficiency in Buildings-Τ.Ο.Τ.Ε.Ε. KENAK 20701-1/2017’. 2017. Available online: https://www.kenak.gr/files/TOTEE_20701-1_2017.pdf (accessed on 10 January 2025). (In Greek).
  40. Frank, E.; Hall, M.A.; Witten, I.H. The WEKA Workbench. Practical Machine Learning Tools and Techniques. In Data Mining, 4th ed.; Morgan Kaufmann: Waikato, New Zealand, 2016. [Google Scholar]
  41. Kumar, Y. Analysis of Bayes, Neural Network and Tree Classifier of Classification Technique in Data Mining using WEKA. In Proceedings of the Computer Science & Information Technology (CS & IT); Academy & Industry Research Collaboration Center (AIRCC): Chennai, India, 2012; pp. 359–369. [Google Scholar] [CrossRef]
  42. Hangloo, S.; Kour, S.; Kumar, S. A Survey on Machine Learning: Concept, Algorithms, and Applications. Int. J. Sci. Res. Comput. Sci. Eng. Inf. Technol. (IJSRCSEIT) 2017, 2, 293–301. [Google Scholar]
  43. Kaur, G.; Chhabra, A. Improved J48 Classification Algorithm for the Prediction of Diabetes. IJCA 2014, 98, 13–17. [Google Scholar] [CrossRef]
  44. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  45. Cerquides, J.; de M’antaras, R.L. Maximum a Posteriori Tree Augmented Naive Bayes Classifiers. October 2003. Available online: https://www.iiia.csic.es/~mantaras/ReportIIIA-TR-2003-10.pdf (accessed on 8 January 2025).
  46. Kalmegh, S.R. Comparative Analysis of the WEKA Classifiers Rules Conjunctiverule & Decisiontable on Indian News Dataset by Using Different Test Mode. Int. J. Eng. Sci. Invent. 2018, 7, 01–09. [Google Scholar]
  47. Devi, T.; Sundaram, K.M. A comparative analysis of meta and tree classification algorithms using weka. Comput. Sci. 2016, 3, 77–83. [Google Scholar]
  48. Vijayarani, S.; Muthulakshmi, M. Comparative Analysis of Bayes and Lazy Classification Algorithms. Int. J. Adv. Res. Comput. Commun. Eng. 2013, 2, 3118–3124. [Google Scholar]
  49. Walker, R.; McKenzie, P.; Liddell, C.; Morris, C. Area-based targeting of fuel poverty in Northern Ireland: An evidenced-based approach. Appl. Geogr. 2012, 34, 639–649. [Google Scholar] [CrossRef]
Figure 3. Graphical interface of ANN (multilayer perceptron) predicting the indicator “NEPI”.
Figure 3. Graphical interface of ANN (multilayer perceptron) predicting the indicator “NEPI”.
Energies 18 01133 g003
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kalfountzou, E.; Papada, L.; Tourkolias, C.; Mirasgedis, S.; Kaliampakos, D.; Damigos, D. A Comparative Analysis of Machine Learning Algorithms in Energy Poverty Prediction. Energies 2025, 18, 1133. https://doi.org/10.3390/en18051133

AMA Style

Kalfountzou E, Papada L, Tourkolias C, Mirasgedis S, Kaliampakos D, Damigos D. A Comparative Analysis of Machine Learning Algorithms in Energy Poverty Prediction. Energies. 2025; 18(5):1133. https://doi.org/10.3390/en18051133

Chicago/Turabian Style

Kalfountzou, Elpida, Lefkothea Papada, Christos Tourkolias, Sevastianos Mirasgedis, Dimitris Kaliampakos, and Dimitris Damigos. 2025. "A Comparative Analysis of Machine Learning Algorithms in Energy Poverty Prediction" Energies 18, no. 5: 1133. https://doi.org/10.3390/en18051133

APA Style

Kalfountzou, E., Papada, L., Tourkolias, C., Mirasgedis, S., Kaliampakos, D., & Damigos, D. (2025). A Comparative Analysis of Machine Learning Algorithms in Energy Poverty Prediction. Energies, 18(5), 1133. https://doi.org/10.3390/en18051133

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop