Next Article in Journal
Preparation of Cu-Containing Substances via an Ultrasonic-Assisted Solvothermal Approach and Their Catalytic Effects on the Thermal Decomposition of Ammonium Perchlorate
Previous Article in Journal
Investigation of the Properties of Low Water-to-Solid Ratio Vibro-Press-Formed Alkali-Activated Municipal Solid Waste Incineration Bottom-Ash Concrete
Previous Article in Special Issue
The Design and Fabrication of Shear-Mode Piezoelectric Accelerometers with High Bandwidth Using High Piezoelectric g-Coefficient NKN-Based Ceramics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Prediction of ABX3 Perovskite Formation Energy Using Machine Learning

by
Ziliang Deng
1,†,
Kailing Fang
1,†,
Chong Guo
1,
Zhichao Gong
1,
Haojie Yue
1,
Huacheng Zhang
1,
Kang Li
1,
Kun Guo
1,*,
Zhiyong Liu
1,
Bing Xie
1,
Jinshan Lu
1,
Kui Yao
2 and
Francis Eng Hock Tay
3
1
School of Power and Energy, Nanchang Hangkong University, Nanchang 330063, China
2
Institute of Materials Research and Engineering, A*STAR (Agency for Science, Technology and Research), 2 Fusionopolis Way, Singapore 138634, Singapore
3
Department of Mechanical Engineering, National University of Singapore, 9 Engineering Drive 1, Singapore 117575, Singapore
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Materials 2025, 18(13), 2927; https://doi.org/10.3390/ma18132927
Submission received: 24 April 2025 / Revised: 11 June 2025 / Accepted: 17 June 2025 / Published: 20 June 2025
(This article belongs to the Special Issue Advances in Ferroelectric and Piezoelectric Materials)

Abstract

:
Materials with perovskite phases are widely used in solar cells and ferroelectric, piezoelectric, dielectric and superconducting devices due to their various notable functions. However, structural instability limits some compositions in forming robust perovskite phases for device applications. The analytical approach using the tolerance factor (t) can only guarantee prediction accuracy within a limited range, ascribed to its nature of overlooking the atomic interaction. Hence, here we establish a prediction model using formation energy as the target parameter for its reflection of the reaction of atoms and apply machine learning as the analysis method since it has been successfully employed in plenty of material property prediction studies. Machine learning employs statistical methodologies to identify correlative patterns within large-scale datasets, enabling accurate predictions with robust generalization. In this work, we built a model to predict the formation energy of ABX3 perovskite using machine learning and achieved a model with an R-squared value of 0.928 and a root mean square error of 0.301 eV/atom, validated by first-principles computations. In total, 75% of the values were correctly predicted within an error lower than 0.06. This work could contribute to accelerating the study of solving perovskites’ instability.

1. Introduction

Inorganic perovskites are widely used in solar cells and ferroelectric, piezoelectric, dielectric and superconducting devices due to their various notable functions [1,2]. However, researchers have found a significant concern that has prevented perovskites from being widely used in certain applications, which is their instability towards phase, light, and moisture. For instance, Liu et al. [3] found that all inorganic lead halide perovskites (ILHPs) exhibit excellent thermal stability compared to hybrid organic–inorganic perovskites. Yet the low structural stability against ambient conditions still limits their practical applications. Xiang et al. [4] pointed out that although the power conversion efficiency of inorganic perovskite-based solar cells has reached over 20%, inorganic perovskite materials still suffer from instability towards phase, light, moisture, etc. These instability issues seriously affect the device’s performance and pose as one of the major concerns [2,5]. Ouedraogo et al. studied the stability of all-inorganic perovskite solar cells and concluded that under the operating temperature on earth, the stability of inorganic lead halide perovskite materials remains the most crucial issue to overcome [6]. The biggest challenge in Wang et al.’s work in studying inorganic perovskites, especially CsPbI3, is phase stability at room temperature (RT) [7]. The structural stability issue also exists in the oxide perovskite phase for ferroelectric, piezoelectric and high dielectric applications, such as in Pb(Zn,Nb)O3 and Pb(Zn,Nb)O3 systems [8,9]. Therefore, future work should focus on solving the stability issue for the broader development of inorganic perovskites.
Some studies tried to solve this through chemical modification methods such as doping and indeed made some progress. For example, Li et al. [10] used large-radius cations in the CsPbI3 perovskite to replace Cs+, significantly improving the film morphology as well as the phase stability. Other methods include improving the material processing method [11].
Nevertheless, these approaches require tedious experimental exploration and resources in the lab but have little appeal. Hence, another train of thought is to use some parameter as a reflection of the stability of the target material, offering an insight to researchers before they invest effort. The tolerance factor (t) has become the most extensive research object. In the ABX3-type perovskite structure (covering ABO3), the tolerance factor (t) proposed by Goldschmidt can be expressed as follows:
t = r A + r X 2 ( r B + r X )
where rA, rB and rX are the ionic radii on A-site, B-site and X-site, respectively. In the case of mixed occupancies, it is the average ionic radius at the designated position. Yan et al. [12] used the tolerance factor t in their work to conjecture the distortion level in the lattice of BT-based perovskites to obtain a reflection of their stability performance. In the research of Coondoo et al. [13], Yan et al. [12], and Yu et al. [8], t also played a similar role.
The aforementioned tolerance factor t can be used to estimate whether a stable perovskite structure is expected to be formed dependent on the composition [4]; there exists some situations that are beyond the prediction ability of t. Fundamentally, t is a geometric approach based on the assumption of a hard sphere model, which becomes less valid when it comes to iodide anions with their lower electronegativity of the heavier halides and greater chemical softness [14]. This is confirmed in the work of Travis et al., where they applied a range of stabilities for hybrid iodide perovskites, roughly 0.8 ≤ t ≤ 1, as suggested by Cheetham and co-workers [15,16], which is very similar to that found for oxides and fluorides, but it turns out that this criterion is not that satisfying due to the low accuracy. Hence, it can be concluded that t may be a necessary condition for perovskite formation, but not a sufficient condition, and the traditional approach that works reasonably effectively for fluoride and oxide compounds cannot be used to explain the known structures of the inorganic ABI3 compounds [14]. Besides the tolerance factor t, there is another parameter, the octahedral factor μ, which is the ratio of the radii of the B-site cation and the X-site anion, often being employed alongside t to evaluate the stability of perovskites. Zhao and co-workers [17] used both μ and t in their work, finding out that in a complex perovskite system where there exist mixed ionic and covalent bonding and multinary features, it is challenging to assign realistic t and μ for evaluating crystallographic stability. These results mean that, although t or μ can provide a qualitative range for the formability of perovskite, it is not a good stability descriptor, i.e., quantitative correlation with stability [18]. In addition to this, some researchers tried another path by carrying out a mathematical process on t and μ and creating new parameters. For example, Sun and co-workers [18] created (μ + t)η and improved the prediction with an accuracy of around 90% by bringing the atomic packing fraction (APF) into consideration for correcting the error, which is much better than using t or μ alone.
The studies mentioned above have demonstrated research efforts in improving the generalization ability of various relevant parameters to represent the stability of perovskite structures, yet all these approaches have a commonality of overlooking the correlation between the ions since the theory of t and μ assumes the maximum anion contact with cations at the very first moment, which means they can only improve the prediction of stability in a limited way. Therefore, we herein address a more effective parameter that can not only recapitulate the interaction of the atoms but also is capable of generalization. We choose the formation energy (Eform) because it can reflect the chemical interactions between elements [19] and represent the bonding strength of atoms within a material system. It correlates with many thermodynamic and kinetic properties, and thus the stability [20] and synthesizability [21] of a compound [22]. Furthermore, the states and distribution of defects are determined by the formation energy of the configuration and the entropy of the system. Thus, this makes detailed calculations regarding the formation energies of different defect configurations and entropies of various systems essential for defect configuration design [23].
It is also noteworthy that some researchers already utilized Eform as a basis for determining whether a compound is stable and for comparing the stability of different compounds in their work [24]. For example, Xiang et al. [4] proved that the increasing intrinsic formation energies of vacancies is a way to decrease their concentration and make perovskites more stable against external factors, as suggested by Saidaminov et al. [25]. Zhou et al. [26] pointed out that the substitution of A-, B-, and X-site ions with other corresponding ions can tune the formation energies, resulting in more stable compounds. Wang and co-workers [7] addressed this in the work of Hu et al. [27], where CsPb0.6Sn0.4I3 exhibited higher oxidation stability compared with its hybrid counterpart, as the smaller Cs+ leads to stronger antibonding of Sn 5s with I 5p and thus the smaller formation energy of Sn defects in inorganic perovskites. Liu et al. calculated Eform before deciding which site in their targeted perovskite system to dope to ensure their outcome compound was stable [3]. From another perspective, Eform could also be applied to confirm that compounds are not stable. For example, Sutton et al. [28] calculated the formation energies concerning the precursors of CsI and PbI2, finding out that the compounds are unstable. In summary, Eform plays an essential role in evaluating perovskite stabilities [29,30].
However, despite the fact that Eform is capable of evaluating perovskite stabilities, it is not easy to obtain; usually, researchers use first-principles computation (FPC) to calculate it, with a huge workload. This often makes the result deviate from the essence of prediction for saving unnecessary work input. We will use a machine learning (ML) tool, which is a big data-driven approach covering predictive analytics, clustering, relationship mining, and anomaly detection [31,32]. Usually, building an ML model should include the following steps: data collection, feature generation and screening, and model training and evaluation, with the workflow shown in Figure 1. In general, the underlying logic of ML is essentially a statistical method that finds correlations between massive features of certain systematic materials and their properties to obtain a mutual correlation. Attributed to this, ML can reach accurate predictions and possesses natural generalization capabilities. More details about ML will be discussed in the following Section 2.
Motivated by these considerations, we believe that ABX3 perovskites and the instability issues in certain applications are very worthy of studying [33,34,35,36,37,38]. In addition, Eform as an indicator of stability also has lots of application potential [30,39]. Starting from the instability problem of perovskite in certain applications, this work demonstrates the limitations of the existing tolerance factor and octahedral factor, as well as the rationality of using the formation energy as a stability indicator. We selected machine learning as the main research method, and through the optimization training of multiple model algorithms, we found a balance between computing power consumption, accuracy and error, and combined this with the SHAP method to analyze the descriptor to improve the interpretability of the model. Finally, FPC calculation was used to produce an intuitive display of the application scenarios of the model.

2. Methodology

As the fourth paradigm of materials science [31], ML is essentially a method of materials informatics [32,40]. Materials informatics is defined as the implementation of data science in the problems inherent in materials science to accelerate the design and discovery of materials [41]. After years of development, the field of materials science has accumulated massive databases, which contain huge value as there are inherent correlations between them [42,43]. From the perspective of statistics and computational science, these correlations can be mined to accelerate the discovery and performance optimization of new materials [44,45,46]. This concept echoes the Materials Genome Initiative (MGI), and together they promote the progress of materials science [32,47,48,49].
Although ML has become more and more mature as a result of the research of many scholars in recent years, we still need to point out that due to the working principle of ML methods, they still have some limitations [50,51,52]. We summarize these into three aspects: first, data dependence and generalization ability issues, which are caused by the size of the dataset [53] and experimental data noise or annotation bias [51]; second, the black box effect and interpretability limitations, primarily due to the inherent lack of transparency in their decision-making processes [53]; third, challenges in practical applications, including, but not limited to, computational resource constraints, deployment scalability issues, and robustness deficiencies in real-world environments [54,55].
In our work, we aim to address the limitations of the ML methods mentioned above by introducing SHAP feature analysis to enhance the interpretability of the model (Section 3.3) and combining it with FPC calculation to intuitively reflect the model’s application scenarios in practice (Section 3.4).

2.1. Dataset Establishment

The dataset is vital for the model performance, as the data quality could directly affect the model training. Our data includes information on the material’s crystal structure, elemental composition, physical properties, chemical properties, and synthesis methods; this information can be found in published research papers and material databases in the public domain, such as the Inorganic Crystal Structure Database (ICSD) [56], Materials Project [57], Open Quantum Materials Database (OQMD) [58], and AiiD [59]. They can be extracted from open source software libraries, for example, SciKit-learn [60], XGBoost [61], Keras [62], Magpie [63], and Pymatgen [64]. These datasets collect data from various research experiments and FPC. Thus, they have different ways of expressing data and different scales of capacity, which require us to preprocess the material data, including removing null values and outliers, to make the data distribution more reasonable and uniform so as to improve the predictive ability of the model. All the software and packages for this work are listed in Table S3, Supplementary Materials.

2.2. Feature Generation

After a decent dataset is established, feature engineering is required, i.e., generating and screening features that are key to the target attributes. Usually, corresponding feature descriptors, such as ionic radius, atomic number, and element electronegativity, can be generated based on the material’s structural information and chemical composition to expand the dataset. Based on the feature selection method, the corresponding important features are screened to train the ML model. The whole process may involve operations such as feature selection, conversion, and dimensionality reduction. After this process, the dataset is capable of being processed by ML models.

2.3. Model Selection

We start the model selection to obtain the most suitable algorithm model or model combination to train the processed dataset. There are plenty of algorithms for selection like linear regression (LR), support vector regression (SVR), neural network multilayer perceptron (MLP), decision tree regression (DTR), random forest (RF), and extreme gradient boosting (XGBoost). The algorithm model selection depends on the characteristics of the data, the complexity of the problem, and the research objectives. Hence, we need to implement algorithm models on our dataset, which is randomly divided into two portions by a certain proportion based on capacity, named the training set and the test set. The test set outcome is used for comparison with the training set outcome for the evaluation and cross-validation of the generalization ability and prediction accuracy of the model. To determine the accuracy, there are four common indicators, mean absolute error (MAE), mean square error (MSE), root mean square error (RMSE), and goodness of fit (R-squared, R2), where the smaller the values of the indicators MAE, MSE, and RMSE, and the closer R2 is to 1, the more reliable. The expressions are
M A E = 1 N i = 1 N y i t r u e y i p r e d
M S E = 1 N i = 1 N y i t r u e y i p r e d 2
R M S E = 1 N i = 1 N y i t r u e y i p r e d 2
R 2 = 1 i = 1 N y i t r u e y i p r e d 2 i = 1 N y i t r u e y i t r u e ¯ 2
where y i t r u e I and y i p r e d represent the experimentally measured and model-predicted formation energies, respectively, for each composition in our dataset.

2.4. Model Verification Means

By finishing the work above, we have acquired a well-functioning prediction model. To make it more persuasive, the models need to be verified with other methods according to the research subject. For example, He and co-workers [65] built a model with ML to predict the morphotropic phase boundary (MPB) to search for high-performance piezoelectric devices; they verified the results of the ML model by synthesizing compounds of the MPB region they predicted. In our work, synthesizing compounds would not be appropriate as Eform is not one of the properties that can be directly reflected by their detection. Therefore, we chose FPC to verify our model since Eform is usually calculated based on it. FPC is mainly based on density functional theory (DFT) [66]. The core idea of DFT is to express the ground-state energy of the system as a function of the electron density. The ground-state energy and electron distribution of the system can be obtained by solving the energy minimum value at a given electron density. This enables DFT to accurately describe the system without explicitly considering the wave function. Therefore, using FPC can naturally help us gain a reliable Eform, and the work of Bartel et al. [67], using FPC to calculate the formation enthalpies, can help prove that.

3. Results and Discussion

3.1. Data Processing and Feature Screening

Our database is mainly based on the Materials Project database since it has a broad capability to realize a model with generalization ability through its open API port. With the screening conditions of the ABX3 structure, we obtained a total of 4358 compounds. After preprocessing, i.e., removing less helpful data, this was reduced to 2703 compounds. Except for the data for which we have do not directly differentiate the A, B, and X compositions of the compounds, for facilitating subsequent feature generation, we employed Python regular expressions to batch-select the most abundant X-site in the chemical formulas and, among the two elements left, assigned the element with a larger radius to the A-site and the other one to the B-site, leveraging the characteristics of perovskite compounds. Thereafter, we generated features of the 2703 compounds and created 168 features of each compound for future training.
With 168 features, there are bound to be redundant ones indicating similar properties, and therefore we used the Pearson correlation coefficient as a reflection of each feature against the others to remove those with a correlation greater than 0.8. In addition, the outlier data of t and μ and corresponding features were also removed, leaving 63 features for each compound. So far, our features have been simplified in the dimension of redundant features. We will further simplify the existing features based on the Pearson correlation coefficient from the perspective of the actual importance of each feature in describing the target feature. As shown in Figure 2, we chose the top 30 features for further model construction. Usually, from the aspects of computing power savings and efficiency considerations, the number of features should be reduced to the lowest possible number before model training; however, this method likely overlooks some features that have benefits to the model training. We looked into the probability distribution of the 30 remaining features, as shown in Figure S1 (Supplementary Materials). The result demonstrates a uniform distribution of the features within their respective ranges without notable anomalies.

3.2. Model Training and Performance Evaluation

We divide the dataset consisting of 30 screened features from previous work as input features plus one target feature (Eform) into the training set and test set in a 4:1 ratio for model training and model verification, respectively. For model training, certain ML algorithms are selected, namely linear regression (LR), support vector regression (SVR), neural network multilayer perceptron (MLP), decision tree regression (DTR), random forest (RF), and extreme gradient boosting (XGBoost), to examine their matching ability with the 30 features for predicting Eform. The specific results are shown in Figure 3.
As shown in Table 1, comparing the evaluation indicators of each model, the training effects of the models are ranked as follows: XGBoost, RF, MLP, SVR, DTR, and LR. Among them, the training effects of the RF and XGBoost models are relatively close. The goodness of fit R2 reaches 0.922 and 0.928, and the root mean square error is 0.313 and 0.301 eV/atom, respectively. In addition, to quantify the relative prediction error, we employed a normalized metric defined as the ratio of RMSE/average ( y i p r e d ). The XGBoost model still exhibited the best performance of 0.175. The model training achieved the ideal effect of higher accuracy and smaller deviation.
Figure 3a–f show the discreteness of the training set and test set of our model; the more the data distribution converges to the red diagonal line, the more accurate the model’s prediction results are. Figure 3e,f show a significantly better fit than the other models, which can be more intuitively illustrated by its R2. In addition, it is worth mentioning that the fit of the yellow points to the red diagonal line indicates that the model has excellent generalization ability to unknown data; in other words, it performs better in terms of predictive capacity.

3.3. Model Optimization and Feature Analysis

Our model is now functioning well. To further develop our model to a level of even better accuracy and to diminish the size of the model to save computing power, we use Shapley Additive Explanations (SHAP) to evaluate from another standard how many of the features we choose are actually contributing to the algorithms and if it is possible that there are still redundant features. By using the SHAP tool to sort the features of the two best-performing algorithm models, RF and XGBoost, according to their importance, as discussed above, we can obtain the relationship between the number of features and the accuracy and deviation of the two models, allowing us to remove some features that we kept for model selection. This helps us finally select the model, as shown in Figure 4.
As we assumed, the trends of R and RMSE against the number of features tend to stabilize or even decrease after reaching extreme values, indicating that there are indeed redundant features to be removed. We marked the lowest feature to realize the best R and the smallest RMSE for both models with red dots. Clearly, the RF model has the highest accuracy (R) and the smallest deviation (RMSE) in the test set at 13 features, while the XGBoost model only requires 9 features and even reaches a slightly higher R2. Therefore, we picked XGBoost as the ultimate model and further optimized it by discovering the nine best features required.
We used SHAP to determine the nine key features with greater importance. As depicted in Figure 5c, the impact on Eform of each feature is shown in Figure 5d. Features with data points on the positive axis indicate a positive correlation with the target property; otherwise, they indicate a negative correlation. The higher the clustering of red points, the higher the correlation. With the nine features filtered, we retrained the XGBoost model results, as shown in Figure 5a,b. The model effectiveness of fit R2 has increased from the original 0.928 to 0.939, and the root mean square error (RMSE) has decreased from 0.301 eV/atom to 0.278 eV/atom, indicating that the prediction performance of the model has been improved with the selection of key features and the elimination of redundant features. And compared with the relevant ABX3 perovskite stability work of Zhu et al. [5], our R2 of 0.94 is higher than theirs, 0.91, using the same algorithm. From this, we can see that the nine most important features are HF_X, PA_EN_AX, PA_EN_BX, GP_A, N_VE_B, IR_X, HF_B, SumDens, SumGP and SumN_VS (standing for fusion-heat of X-site element, Pauling electronegativity of A-site and B-site, the A-site group, the number of valence electrons at the B-site, the X-site ion radius, the heat of fusion at the B-site, the combined density, the combined group, and the combined number of p valence electrons; all the features and corresponding physical meanings used in this work are illustrated in Table S2, Supplementary Materials), among which HF_X is positively correlated, and PA_EN_AX and PA_EN_BX are negatively correlated. This can also be confirmed by the results of the Pearson correlation heat map in Figure 5b.

3.4. Model Validation

In order to present the prediction results of our model for Eform more intuitively and evaluate it based on the traditional model evaluation criteria in the above work, we further validate the model. Figure 6 exhibits the comparison between the Eform distribution of compounds acquired randomly from the Materials Project database and the corresponding prediction results generated with our model. It can be seen that there is no significant difference between the two sets of distributions, indicating that our model has a high prediction accuracy. We used the acquired data to check the distribution of the nine features to verify whether the correlation between the features and Eform is consistent with our conjecture, as shown in Figure S2 (Supplementary Materials). Obviously, all nine features showed a high correlation with Eform, and the trends are consistent with our conjecture, especially PA_EN_AX and PA_EN_BX, which showed high aggregation and clear negative relevance. We also randomly generated some compounds, performed FPC and model predictions simultaneously, and compared the results obtained (see Figure 7; detailed results in Table S1, Supplementary Materials).

4. Conclusions

In summary, to assess the instability of ABX3-type perovskites, we used ML methods to build a model to predict the formation energy of ABX3-type perovskites to guide the design of the perovskite structure. To save computing resources, we simplified the model while retaining the desired reliability and accuracy. Through multiple rounds of screening, our ML model features were gradually reduced from 168 per compound to only 9, achieving a goodness of fit R2 of 0.939 and a root mean square error of 0.278 eV/atom. We further verified the actual working of the model by comparing the FPC results with the model prediction results of random ABX3-type perovskite compounds. Our results showed that 75% of the values were correctly predicted within an error lower than 0.06. The model proposed here provides a powerful tool for the materials science community to accelerate the discovery and design of perovskite materials with minimized burden on computation.

Supplementary Materials

The following supporting information can be downloaded at https://www.mdpi.com/article/10.3390/ma18132927/s1. Figure S1. The probability distribution of the top 30 input features in the order of Pearson correlation importance from large to small. The 30 selected descriptors, whether they are continuous or scattered, are evenly distributed within their respective intervals, with no obvious anomalies, which means our descriptors are good; Figure S2. The scatter density distribution of the formation energy and 9 input features of the XGBoost model: (a–c) HF_X, PA_EN_AX, PA_EN_BX; (d–f) GP_A, N_VE_B, IR_X; (g–i) HF_B, SumDens, SumGP; Table S1. Eform calculated by FPC and predicted by ML model and differences in the 20 generated compounds; Table S2. Initial features and corresponding physical meanings; Table S3. Necessary software and packages and their version.

Author Contributions

Conceptualization, Z.D. and K.F.; methodology, Z.D. and K.F.; software, Z.D. and K.F.; validation, Z.D., K.F. and C.G.; formal analysis, Z.D.; investigation, K.F.; resources, Z.D. and Z.G.; data curation, K.F., H.Y. and H.Z.; writing—original draft preparation, Z.D. and K.F.; writing—review and editing, Z.D., Z.G. and K.G.; visualization, K.F., Z.G. and K.L.; supervision, K.G., F.E.H.T., B.X., Z.L., J.L. and K.Y.; project administration, K.G., F.E.H.T., B.X., Z.L., J.L. and K.Y.; funding acquisition, K.G., F.E.H.T., B.X., Z.L., J.L. and K.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This project is partially supported by the National Natural Science Foundation of China (Grant No. 52002164, No. 52060020, No. 52162018, No. 52162019) and the Natural Science Foundation of Jiangxi Province of China (Grant No. 20242BAB25224). The author from IMRE acknowledges support from A*STAR, under the RIE2020 AME Programmatic Fund (Grant No. A20G9b0135) and RIE2025, IAF-ICP Grant I2301E0027.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article/Supplementary Material. Further inquiries can be directed to the corresponding author.

Acknowledgments

This study was partially supported by the High Performance Computing Service of the Modern Educational Technology and Information Center of Nanchang Hangkong University.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Tao, Q.; Xu, P.; Li, M.; Lu, W. Machine learning for perovskite materials design and discovery. Npj Comput. Mater. 2021, 7, 23. [Google Scholar] [CrossRef]
  2. Liu, Y.; Tan, X.; Liang, J.; Han, H.; Xiang, P.; Yan, W. Machine Learning for Perovskite Solar Cells and Component Materials: Key Technologies and Prospects. Adv. Funct. Mater. 2023, 33, 2214271. [Google Scholar] [CrossRef]
  3. Liu, D.; Shao, Z.; Li, C.; Pang, S.; Yan, Y.; Cui, G. Structural Properties and Stability of Inorganic CsPbI3 Perovskites. Small Struct. 2021, 2, 2000089. [Google Scholar] [CrossRef]
  4. Xiang, W.; Liu, S.F.; Tress, W. A review on the stability of inorganic metal halide perovskites: Challenges and opportunities for stable solar cells. Energy Environ. Sci. 2021, 14, 2090–2113. [Google Scholar] [CrossRef]
  5. Zhu, Y.; Zhang, J.; Qu, Z.; Jiang, S.; Liu, Y.; Wu, Z.; Yang, F.; Hu, W.; Xu, Z.; Dai, Y. Accelerating stability of ABX3 perovskites analysis with machine learning. Ceram. Int. 2024, 50, 6250–6258. [Google Scholar] [CrossRef]
  6. Ouedraogo, N.A.N.; Chen, Y.; Xiao, Y.Y.; Meng, Q.; Han, C.B.; Yan, H.; Zhang, Y. Stability of all-inorganic perovskite solar cells. Nano Energy 2020, 67, 104249. [Google Scholar] [CrossRef]
  7. Wang, Z.; Zhang, J.; Guo, W.; Xiang, W.; Hagfeldt, A. Formation and Stabilization of Inorganic Halide Perovskites for Photovoltaics. Matter 2021, 4, 528–551. [Google Scholar] [CrossRef]
  8. Yu, S.; Yao, K.; Tay, F.E.H. Observations and Analyses on the Thermal Stability of (1-x)Pb(Zn1/3Nb2/3)O3−xPbTiO3 Thin Films. Chem. Mater. 2007, 19, 4373–4377. [Google Scholar] [CrossRef]
  9. Yu, S.; Yao, K.; Tay, F.E.H. Structure and Properties of (1−x)PZN−xPT Thin Films with Perovskite Phase Promoted by Polyethylene Glycol. Chem. Mater. 2006, 18, 5343–5350. [Google Scholar] [CrossRef]
  10. Li, F.; Pei, Y.; Xiao, F.; Zeng, T.; Yang, Z.; Xu, J.; Sun, J.; Peng, B.; Liu, M. Tailored dimensionality to regulate the phase stability of inorganic cesium lead iodide perovskites. Nanoscale 2018, 10, 6318–6322. [Google Scholar] [CrossRef]
  11. Yao, K.; Yu, S.; Tay, F.E.H. Preparation of perovskite Pb(Zn1∕3Nb2∕3)O3-based thin films from polymer-modified solution precursors. Appl. Phys. Lett. 2006, 88, 052904. [Google Scholar] [CrossRef]
  12. Yan, S.; Cao, Z.; Liu, Q.; Gao, Y.; Zhang, H.; Li, G. Enhanced piezoelectric activity around orthorhombic-tetragonal phase boundary in multielement codoping BaTiO3. J. Alloys Compd. 2022, 923, 166398. [Google Scholar] [CrossRef]
  13. Coondoo, I.; Alikin, D.; Abramov, A.; Figueiras, F.G.; Shur, V.Y.; Miranda, G. Exploring the effect of low concentration of stannum in lead-free BCT-BZT piezoelectric compositions for energy related applications. J. Alloys Compd. 2023, 960, 170562. [Google Scholar] [CrossRef]
  14. Travis, W.; Glover, E.N.K.; Bronstein, H.; Scanlon, D.O.; Palgrave, R.G. On the application of the tolerance factor to inorganic and hybrid halide perovskites: A revised system. Chem. Sci. 2016, 7, 4548–4556. [Google Scholar] [CrossRef]
  15. Kieslich, G.; Sun, S.; Cheetham, A.K. Solid-state principles applied to organic–inorganic perovskites: New tricks for an old dog. Chem. Sci. 2014, 5, 4712–4715. [Google Scholar] [CrossRef]
  16. Kieslich, G.; Sun, S.; Cheetham, A.K. An extended Tolerance Factor approach for organic–inorganic perovskites. Chem. Sci. 2015, 6, 3430–3433. [Google Scholar] [CrossRef]
  17. Zhao, X.; Yang, J.; Fu, Y.; Yang, D.; Xu, Q.; Yu, L.; Wei, S.-H.; Zhang, L. Design of Lead-Free Inorganic Halide Perovskites for Solar Cells via Cation-Transmutation. J. Am. Chem. Soc. 2017, 139, 2630–2638. [Google Scholar] [CrossRef]
  18. Sun, Q.; Yin, W. Thermodynamic Stability Trend of Cubic Perovskites. J. Am. Chem. Soc. 2017, 139, 14905–14908. [Google Scholar] [CrossRef]
  19. King, D.J.M.; Middleburgh, S.C.; McGregor, A.G.; Cortie, M.B. Predicting the formation and stability of single phase high-entropy alloys. Acta Mater. 2016, 104, 172–179. [Google Scholar] [CrossRef]
  20. Ong, S.P.; Wang, L.; Kang, B.; Ceder, G. Li−Fe−P−O2 Phase Diagram from First Principles Calculations. Chem. Mater. 2008, 20, 1798–1807. [Google Scholar] [CrossRef]
  21. Miura, A.; Bartel, C.J.; Goto, Y.; Mizuguchi, Y.; Moriyoshi, C.; Kuroiwa, Y.; Wang, Y.; Yaguchi, T.; Shirai, M.; Nagao, M.; et al. Observing and Modeling the Sequential Pairwise Reactions that Drive Solid-State Ceramic Synthesis. Adv. Mater. 2021, 33, 2100312. [Google Scholar] [CrossRef]
  22. Liang, Y.; Chen, M.; Wang, Y.; Jia, H.; Lu, T.; Xie, F.; Cai, G.; Wang, Z.; Meng, S.; Liu, M. A universal model for accurately predicting the formation energy of inorganic compounds. Sci. China Mater. 2022, 66, 343–351. [Google Scholar] [CrossRef]
  23. Feng, Y.; Wu, J.; Chi, Q.; Li, W.; Yu, Y.; Fei, W. Defects and Aliovalent Doping Engineering in Electroceramics. Chem. Rev. 2020, 120, 1710–1787. [Google Scholar] [CrossRef]
  24. Monika; Pachori, S.; Agrawal, R.; Choudhary, B.L.; Verma, A.S. An efficient and stable lead-free organic–inorganic tin iodide perovskite for photovoltaic device: Progress and challenges. Energy Rep. 2022, 8, 5753–5763. [Google Scholar] [CrossRef]
  25. Saidaminov, M.I.; Kim, J.; Jain, A.; Quintero-Bermudez, R.; Tan, H.; Long, G.; Tan, F.; Johnston, A.; Zhao, Y.; Voznyy, O.; et al. Suppression of atomic vacancies via incorporation of isovalent small ions to increase the stability of halide perovskite solar cells in ambient air. Nat. Energy 2018, 3, 648–654. [Google Scholar] [CrossRef]
  26. Zhou, Y.; Zhao, Y. Chemical stability and instability of inorganic halide perovskites. Energy Environ. Sci. 2019, 12, 1495–1511. [Google Scholar] [CrossRef]
  27. Hu, M.; Chen, M.; Guo, P.; Zhou, H.; Deng, J.; Yao, Y.; Jiang, Y.; Gong, J.; Dai, Z.; Qian, F.; et al. Sub-1.4eV bandgap inorganic perovskite solar cells with long-term stability. Nat. Commun. 2020, 11, 151. [Google Scholar] [CrossRef]
  28. Sutton, R.J.; Filip, M.R.; Haghighirad, A.A.; Sakai, N.; Wenger, B.; Giustino, F.; Snaith, H.J. Cubic or Orthorhombic? Revealing the Crystal Structure of Metastable Black-Phase CsPbI3 by Theory and Experiment. ACS Energy Lett. 2018, 3, 1787–1794. [Google Scholar] [CrossRef]
  29. Nagabhushana, G.P.; Shivaramaiah, R.; Navrotsky, A. Direct calorimetric verification of thermodynamic instability of lead halide hybrid perovskites. Proc. Natl. Acad. Sci. USA 2016, 113, 7717–7721. [Google Scholar] [CrossRef]
  30. Alhashmi, A.; Kanoun, M.B.; Goumri-Said, S. Machine Learning for Halide Perovskite Materials ABX3 (B = Pb, X = I, Br, Cl) Assessment of Structural Properties and Band Gap Engineering for Solar Energy. Materials 2023, 16, 2657. [Google Scholar] [CrossRef]
  31. Raccuglia, P.; Elbert, K.C.; Adler, P.; Falk, C.; Wenny, M.B.; Mollo, A.; Zeller, M.; Friedler, S.A.; Schrier, J.; Norquist, A.J. Machine-learning-assisted materials discovery using failed experiments. Nature 2016, 533, 73–76. [Google Scholar] [CrossRef]
  32. Agrawal, A.; Choudhary, A. Perspective: Materials informatics and big data: Realization of the “fourth paradigm” of science in materials science. APL Mater. 2016, 4, 053208. [Google Scholar] [CrossRef]
  33. Singh, M.; Tiwari, J.P. Tailoring of the Band Gap of MA3 Bi2 I9 through Doping at A as well as X Sites (of ABX3 Structure): Futuristic Material for Multijunction Solar Cells. ACS Appl. Energy Mater. 2025, 8, 6264–6269. [Google Scholar] [CrossRef]
  34. Shimul, A.I.; Sarker, S.R.; Ghosh, A.; Zaman, M.T.U.; Alrafai, H.A.; Hassan, A.A. Examining the optoelectronic and photovoltaic characteristics of Mg3SbM3 (M = F, Cl, Br) perovskites with diverse charge transport layers through numerical optimization and machine learning techniques. Inorg. Chem. Commun. 2025, 179, 114737. [Google Scholar] [CrossRef]
  35. Zhai, X.; Chen, M. Accelerated Design for Perovskite-Oxide-Based Photocatalysts Using Machine Learning Techniques. Materials 2024, 17, 3026. [Google Scholar] [CrossRef]
  36. Shafiq, M.; Amin, B.; Jehangir, M.A.; Chaudhry, A.R.; Murataza, G. First-principle calculations to investigate mechanical and acoustical properties of predicted stable halide Perovskite ABX3. J. Mol. Graph. Model. 2024, 133, 108861. [Google Scholar] [CrossRef]
  37. Pyun, D.; Lee, S.; Lee, S.; Jeong, S.-H.; Hwang, J.-K.; Kim, K.; Kim, Y.; Nam, J.; Cho, S.; Hwang, J.-S.; et al. Machine Learning-Assisted Prediction of Ambient-Processed Perovskite Solar Cells’ Performances. Energies 2024, 17, 5998. [Google Scholar] [CrossRef]
  38. Hongyu, C.; Liang, C.; Wensheng, Y. Stability Challenges in Industrialization of Perovskite Photovoltaics: From Atomic-Scale View to Module Encapsulation. Adv. Funct. Mater. 2024, 35, 2412389. [Google Scholar] [CrossRef]
  39. Ahamed, A.I.; Siam, J.; Rabah, B.; Riyad, K.; Rifat, R.; Moamen, S.R.; Hasan, M.S.S.; Amnah, M.A.; Azizur, M.R.; Alamgir, M.H.; et al. Exploring ACdX3 Perovskites: DFT Analysis of Stability, Electronic, Optical, and Mechanical Properties for Solar Applications. J. Inorg. Organomet. Polym. Mater. 2025, 1–26. [Google Scholar] [CrossRef]
  40. Himanen, L.; Geurts, A.; Foster, A.S.; Rinke, P. Data-Driven Materials Science: Status, Challenges, and Perspectives. Adv. Sci. 2019, 6, 1900808. [Google Scholar] [CrossRef]
  41. Vivanco-Benavides, L.E.; Martínez-González, C.L.; Mercado-Zúñiga, C.; Torres-Torres, C. Machine learning and materials informatics approaches in the analysis of physical properties of carbon nanotubes: A review. Comput. Mater. Sci. 2022, 201, 110939. [Google Scholar] [CrossRef]
  42. Bauer, S.; Benner, P.; Bereau, T.; Blum, V.; Boley, M.; Carbogno, C.; Catlow, C.R.A.; Dehm, G.; Eibl, S.; Ernstorfer, R.; et al. Roadmap on Data-Centric Materials Science. arXiv 2024, arXiv:2402.10932. [Google Scholar] [CrossRef]
  43. Chong, S.S.; Ng, Y.S.; Wang, H.-Q.; Zheng, J.-C. Advances of machine learning in materials science: Ideas and techniques. Front. Phys. 2023, 19, 13501. [Google Scholar] [CrossRef]
  44. Obasi, C.; Oranu, O. Exploring Machine Learning Algorithms and Their Applications in Materials Science. J. Comput. Intell. Mater. Sci. 2024, 2, 23–35. [Google Scholar] [CrossRef]
  45. Liu, Y.; Zhao, T.; Ju, W.; Shi, S. Materials discovery and design using machine learning. J. Mater. 2017, 3, 159–177. [Google Scholar] [CrossRef]
  46. Batra, R. Accurate machine learning in materials science facilitated by using diverse data sources. Nature 2021, 589, 524–525. [Google Scholar] [CrossRef]
  47. Liu, Y.; Niu, C.; Wang, Z.; Gan, Y.; Zhu, Y.; Sun, S.; Shen, T. Machine learning in materials genome initiative: A review. J. Mater. Sci. Technol. 2020, 57, 113–122. [Google Scholar] [CrossRef]
  48. Suh, C.; Fare, C.; Warren, J.A.; Pyzer-Knapp, E.O. Evolving the Materials Genome: How Machine Learning Is Fueling the Next Generation of Materials Discovery. Annu. Rev. Mater. Res. 2020, 50, 1–25. [Google Scholar] [CrossRef]
  49. Ramprasad, R.; Batra, R.; Pilania, G.; Mannodi-Kanakkithodi, A.; Kim, C. Machine learning in materials informatics: Recent applications and prospects. npj Comput. Mater. 2017, 3, 54. [Google Scholar] [CrossRef]
  50. Jianxin, X. Prospects of materials genome engineering frontiers. Mater. Genome Eng. Adv. 2023, 1, e17. [Google Scholar] [CrossRef]
  51. Jiang, X.; Fu, H.; Bai, Y.; Jiang, L.; Zhang, H.; Wang, W.; Yun, P.; He, J.; Xue, D.; Lookman, T.; et al. Interpretable Machine Learning Applications: A Promising Prospect of AI for Materials. Adv. Funct. Mater. 2025, 2507734. [Google Scholar] [CrossRef]
  52. Xie, J.; Su, Y.; Zhang, D.; Feng, Q. A Vision of Materials Genome Engineering in China. Engineering 2022, 10, 10–12. [Google Scholar] [CrossRef]
  53. Xu, P.; Ji, X.; Li, M.; Lu, W. Small data machine learning in materials science. npj Comput. Mater. 2023, 9, 42. [Google Scholar] [CrossRef]
  54. Pouchard, L.U.; Lin, Y.U.; Van Dam, H.U. Replicating Machine Learning Experiments in Materials Science. Adv. Parallel Comput. 2020, 36, 743–755. [Google Scholar] [CrossRef]
  55. Andrew, S.; Vinayak, B.; Qianxiang, A.; Chad, R. Challenges in Information-Mining the Materials Literature: A Case Study and Perspective. Chem. Mater. 2022, 34, 4821–4827. [Google Scholar] [CrossRef]
  56. Belsky, A.; Hellenbrandt, M.; Karen, V.L.; Luksch, P. New developments in the Inorganic Crystal Structure Database (ICSD): Accessibility in support of materials research and design. Acta Crystallogr. Sect. B Struct. Sci. 2002, 58, 364–369. [Google Scholar] [CrossRef]
  57. Jain, A.; Ong, S.; Hautier, G.; Chen, W.; Richards, W.; Dacek, S.; Cholia, S.; Gunter, D.; Skinner, D.; Ceder, G.; et al. Commentary: The Materials Project: A materials genome approach to accelerating materials innovation. APL Mater. 2013, 1, 011002. [Google Scholar] [CrossRef]
  58. Kirklin, S. The Open Quantum Materials Database (OQMD): Assessing the accuracy of DFT formation energies. npj Comput. Mater. 2015, 15, 15010. [Google Scholar] [CrossRef]
  59. Pizzi, G.; Cepellotti, A.; Sabatini, R.; Marzari, N.; Kozinsky, B. AiiDA: Automated interactive infrastructure and database for computational science. Comput. Mater. Sci. 2016, 111, 218–230. [Google Scholar] [CrossRef]
  60. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  61. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD ’16), San Francisco, CA, USA, 13–17 August 2016; Association for Computing Machinery: New York, NY, USA, 2016; pp. 785–794. [Google Scholar] [CrossRef]
  62. Ketkar, N.; Moolayil, J. Deep Learning with Python: Learn Best Practices of Deep Learning Models with PyTorch; Apress: Berkeley, CA, USA, 2021. [Google Scholar] [CrossRef]
  63. Ward, L.; Agrawal, A.; Choudhary, A.; Wolverton, C. A general-purpose machine learning framework for predicting properties of inorganic materials. npj Comput. Mater. 2016, 2, 16028. [Google Scholar] [CrossRef]
  64. Ong, S.P.; Richards, W.D.; Jain, A.; Hautier, G.; Kocher, M.; Cholia, S.; Gunter, D.; Chevrier, V.L.; Persson, K.A.; Ceder, G. Python Materials Genomics (pymatgen): A robust, open-source python library for materials analysis. Comput. Mater. Sci. 2013, 68, 314–319. [Google Scholar] [CrossRef]
  65. He, J.; Yu, C.; Hou, Y.; Su, X.; Li, J.; Liu, C.; Xue, D.; Cao, J.; Su, Y.; Qiao, L.; et al. Accelerated discovery of high-performance piezocatalyst in BaTiO3-based ceramics via machine learning. Nano Energy 2022, 97, 107218. [Google Scholar] [CrossRef]
  66. Orio, M.; Pantazis, D.A.; Neese, F. Density functional theory. Photosynth. Res. 2009, 102, 443–453. [Google Scholar] [CrossRef]
  67. Bartel, C.J.; Weimer, A.W.; Lany, S.; Musgrave, C.B.; Holder, A.M. The role of decomposition reactions in assessing first-principles predictions of solid stability. npj Comput. Mater. 2019, 5, 4. [Google Scholar] [CrossRef]
Figure 1. Workflow of the ML strategy to predict the formation energy of ABX3 perovskites. (a) Dataset collection. (b) Feature generation. (c) Feature screening by Pearson correlation coefficient. (d) Model training. (e) Make predictions with trained model. (f) Validate prediction by comparing with first-principles computations.
Figure 1. Workflow of the ML strategy to predict the formation energy of ABX3 perovskites. (a) Dataset collection. (b) Feature generation. (c) Feature screening by Pearson correlation coefficient. (d) Model training. (e) Make predictions with trained model. (f) Validate prediction by comparing with first-principles computations.
Materials 18 02927 g001
Figure 2. Importance ranking of features in describing the target feature.
Figure 2. Importance ranking of features in describing the target feature.
Materials 18 02927 g002
Figure 3. Thirty input features selected to compare the predicted value and the true value of various ML models: (a) LR; (b) SVR; (c) MLP; (d) DTR; (e) RF; (f) XGBoost.
Figure 3. Thirty input features selected to compare the predicted value and the true value of various ML models: (a) LR; (b) SVR; (c) MLP; (d) DTR; (e) RF; (f) XGBoost.
Materials 18 02927 g003
Figure 4. The relationship between the number of features and the accuracy and deviation of the model: (a) R of RF algorithm; (b) RMSE of RF algorithm; (c) R of XGBoost algorithm; (d) RMSE of XGBoost algorithm.
Figure 4. The relationship between the number of features and the accuracy and deviation of the model: (a) R of RF algorithm; (b) RMSE of RF algorithm; (c) R of XGBoost algorithm; (d) RMSE of XGBoost algorithm.
Materials 18 02927 g004
Figure 5. Results of retraining of the top 9 features on XGBoost model: (a) comparison of real formation energy and predicted value; (b) Pearson correlation heat map of the top 9 features and formation energy; (c) ranking of feature importance based on SHAP interpretation method; (d) feature analysis based on SHAP interpretation method.
Figure 5. Results of retraining of the top 9 features on XGBoost model: (a) comparison of real formation energy and predicted value; (b) Pearson correlation heat map of the top 9 features and formation energy; (c) ranking of feature importance based on SHAP interpretation method; (d) feature analysis based on SHAP interpretation method.
Materials 18 02927 g005
Figure 6. Distribution diagram of true value and ML-predicted value of formation energy for different features in 2D coordinates: (ac) formation energy from Materials Project; (df) predicted formation energy obtained using XGBoost algorithm.
Figure 6. Distribution diagram of true value and ML-predicted value of formation energy for different features in 2D coordinates: (ac) formation energy from Materials Project; (df) predicted formation energy obtained using XGBoost algorithm.
Materials 18 02927 g006
Figure 7. Comparison between ML-predicted results and FPC results.
Figure 7. Comparison between ML-predicted results and FPC results.
Materials 18 02927 g007
Table 1. Thirty input features selected and evaluated using the training results of various machine learning models.
Table 1. Thirty input features selected and evaluated using the training results of various machine learning models.
ModelMAE
(eV/atom)
RMSE
(eV/atom)
MSER2Pearson’s
R Value
RMSE/Average ( y i p r e d )
LR0.4730.6070.3690.7060.8430.353
SVR0.2290.3750.1400.8880.9420.221
MLP0.2210.3520.1240.9010.9500.214
DTR0.2290.4070.1650.8680.9340.235
RF0.1940.3130.0980.9220.9610.188
XGBoost0.1860.3010.0900.9280.9630.175
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Deng, Z.; Fang, K.; Guo, C.; Gong, Z.; Yue, H.; Zhang, H.; Li, K.; Guo, K.; Liu, Z.; Xie, B.; et al. Prediction of ABX3 Perovskite Formation Energy Using Machine Learning. Materials 2025, 18, 2927. https://doi.org/10.3390/ma18132927

AMA Style

Deng Z, Fang K, Guo C, Gong Z, Yue H, Zhang H, Li K, Guo K, Liu Z, Xie B, et al. Prediction of ABX3 Perovskite Formation Energy Using Machine Learning. Materials. 2025; 18(13):2927. https://doi.org/10.3390/ma18132927

Chicago/Turabian Style

Deng, Ziliang, Kailing Fang, Chong Guo, Zhichao Gong, Haojie Yue, Huacheng Zhang, Kang Li, Kun Guo, Zhiyong Liu, Bing Xie, and et al. 2025. "Prediction of ABX3 Perovskite Formation Energy Using Machine Learning" Materials 18, no. 13: 2927. https://doi.org/10.3390/ma18132927

APA Style

Deng, Z., Fang, K., Guo, C., Gong, Z., Yue, H., Zhang, H., Li, K., Guo, K., Liu, Z., Xie, B., Lu, J., Yao, K., & Tay, F. E. H. (2025). Prediction of ABX3 Perovskite Formation Energy Using Machine Learning. Materials, 18(13), 2927. https://doi.org/10.3390/ma18132927

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop