Next Article in Journal
Pressure-Induced Structural Stabilities and Superconductivity in Rhodium Borides
Previous Article in Journal
Use of BOIvy Optimization Algorithm-Based Machine Learning Models in Predicting the Compressive Strength of Bentonite Plastic Concrete
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Application of Soft Computing Represented by Regression Machine Learning Model and Artificial Lemming Algorithm in Predictions for Hydrogen Storage in Metal-Organic Frameworks

1
SINOPEC Research Institute of Petroleum Engineering, Beijing 100101, China
2
School of Intelligent Software and Engineering, Nanjing University, Suzhou 215163, China
3
School of Resources and Safety Engineering, Central South University, Changsha 410083, China
4
State Key Laboratory of Geomechanics and Geotechnical Engineering, Institute of Rock and Soil Mechanics, Chinese Academy of Sciences, Wuhan 430071, China
*
Authors to whom correspondence should be addressed.
Materials 2025, 18(13), 3122; https://doi.org/10.3390/ma18133122
Submission received: 12 May 2025 / Revised: 8 June 2025 / Accepted: 16 June 2025 / Published: 1 July 2025
(This article belongs to the Special Issue Hydrides for Energy Storage: Materials, Technologies and Applications)

Abstract

Metal-organic frameworks (MOFs) have been extensively studied for hydrogen storage due to their unique properties. This paper aims to develop several regression-based machine learning models to predict the hydrogen storage capacity of MOFs, including artificial neuron network (ANN), support vector regression (SVR), random forest (RF), extreme learning machine (ELM), kernel extreme learning machine (KELM), and generalized regression neural network (GRNN). An improved population-based metaheuristic optimization algorithm, the artificial lemming algorithm (ALA), is employed to select the hyperparameters of these machine learning models, enhancing their performance. All developed models are trained and tested using experimental data from multiple studies. The performance of the models is evaluated using various statistical metrics, complemented by regression plots, error analysis, and Taylor graphs to further identify the most effective predictive model. The results show that the ALA-RF model obtains the best performance in predicting hydrogen storage, with optimal values of coefficient of determination (R2), root mean square error (RMSE), Willmott’s index (WI), and weighted average percentage error (WAPE) in both training and testing phases (0.9845 and 0.9840, 0.2719 and 0.2828, 0.9961 and 0.9959, and 0.0667 and 0.0714, respectively). Additionally, pressure is identified as the most significant feature for predicting hydrogen storage in MOFs. These findings provide an intelligent solution for the selection of MOFs and optimization of operational conditions in hydrogen storage processes.

1. Introduction

The accelerating global energy crisis, driven by the depletion of fossil fuels and growing pressure to achieve carbon neutrality, has led to a search for sustainable, high energy density solutions [1]. Hydrogen energy stands out as a promising alternative due to its zero-emission profile and high gravimetric energy density, though its practical application is hampered by low volumetric densities and demanding storage requirements such as high-pressure or extremely low-temperature conditions [2].
Among various hydrogen storage media, metal-organic frameworks (MOFs), crystalline structures composed of metal nodes and organic linkers, have attracted strong research interest [3]. Their exceptional surface area, tunable pore geometry, and modular chemical structure enable highly selective and reversible hydrogen adsorption under moderate conditions [4]. These properties make MOFs ideal candidates for physisorption-based hydrogen storage under moderate conditions [5].
Conventional approaches to optimize MOF structures, whether via high-throughput simulations or empirical synthesis, are often costly, time-consuming, and sensitive to experimental conditions, limiting their rapid discovery and scale-up [6]. For instance, Avci et al. [7] performed an extensive computational screening of MOF adsorbents and membranes for CO2/H2 separation, uncovering H2/CO2 selectivity ranging from 2.1 × 10−5 to 6.3 and H2 permeabilities between 230 and 1.7 × 106 Barrer under pressure swing adsorption (PSA) and vacuum swing adsorption (VSA) conditions; yet each simulation run required over 104 CPU hours, and key effects such as charge transfer at open metal sites were undervalued due to simplified force-field models. Meanwhile, empirical synthesis and characterization of MOFs often follow iterative, trial-and-error protocols. According to the EU HyStorPor project, developing a novel MOF for hydrogen storage averages 18 months and costs approximately EUR 200,000, underscoring the impracticality of exhaustive laboratory screening [8]. The performance of MOFs is highly sensitive to subtle environmental fluctuations [9]. Minor temperature shifts can alter coordination bonds, modifying pore size and adsorption sites, while pressure variations affect crystal packing density and accessible surface area [10]. These dynamic responses necessitate repeated experiments or simulations under finely controlled conditions, further prolonging the discovery cycle.
Machine learning (ML) has emerged as a powerful tool to predict material properties by uncovering complex, non-linear relationships within existing datasets [11,12,13,14,15]. Yuan et al. demonstrated the potential of tree-based models by training gradient boosting decision trees (GBDTs) on a diverse dataset of porous carbons, achieving an R2 of 0.98 on the training set and 0.84 on held-out test data, demonstrating a strong fit and good generalization ability of the model [16]. They further applied the GBDT classifier to distinguish between regular porous carbon (RPC) and heteroatom-doped porous carbon (HDPC), where the model again excelled. On this basis, Wang et al. [17] conducted a hierarchical screening of approximately 330,000 hypothetical MOFs for methane adsorption, ultimately identifying the four best candidates, which had a calculated working capacity of 145 cm3/cm3 at 5–35 bar and 298 K, close to the U.S. Department of Energy’s 2015 target of 180 cm3/cm3.
In addition to tree-based approaches, further advances have been made in neural networks and hybrid algorithms. Anderson et al. [18] trained a multilayer perceptron (MLP) to predict complete adsorption isotherms, finding that the MLP’s outputs correlated linearly with Grand Canonical Monte Carlo (GCMC) simulations when loading exceeded 50% of saturation, while maintaining clustering near the parity line at low loadings (<25%). Wei et al. [19] combined random forest (RF) with k-nearest neighbors (KNN) to handle high-dimensional synthesis data, revealing key features that generalize across MOF chemistries. Guo et al. [20] compared MLP and long short-term memory (LSTM) networks for gas adsorption prediction, showing that deep learning models outperformed conventional RF methods, especially under low-pressure conditions.
A solid foundation for ML-driven MOF property prediction has been established based on previous studies, yet two critical challenges remain. On the one hand, most studies evaluate either one or more ML techniques in isolation without exploring the relative merits of other regressors. Pardakhti et al. [21] trained a random forest on just 8% of 130,398 hypothetical MOFs and achieved an R2 of 0.98 (training) and a mean absolute percent error of ~7% (testing), completing all predictions in roughly two hours on a single desktop, which was orders of magnitude faster than comparable molecular simulations. To further accelerate screening, Choudhary et al. [22] applied the Atomistic Line Graph Neural Network (ALIGNN) to predict CO2 adsorption across 137,953 hypothetical MOFs, demonstrating orders-of-magnitude speedups over Grand Canonical Monte Carlo simulations while retaining high accuracy. On the other hand, hyperparameter tuning remains a key issue: grid searches can only exhaustively sample a very small fraction of the combined space, typically less than 0.1% of the possible configurations, leading to high computational costs in high dimensions [23,24]. Bayesian optimization addresses this by modeling performance as a Gaussian process and sequentially selecting promising hyperparameter sets, routinely outperforming grid and random search and matching or exceeding expert-tuned results [25,26]. Building on this, metaheuristic methods transfer knowledge from prior tuning tasks to new problems, slashing hyperparameter optimization times from days to under two hours by leveraging surrogate performance predictors trained on historical trials, but such strategies have yet to be systematically applied to MOF discovery [27].
To bridge the identified research gaps, this study introduces an innovative integrated framework that synergizes six regression ML models with metaheuristic optimization algorithms, with the aim of achieving three overarching objectives: (1) compare the predictive accuracy of six ML models, including artificial neuron network (ANN), support vector regression (SVR), random forest (RF), extreme learning machine (ELM), kernel extreme learning machine (KELM), and generalized regression neural network (GRNN), in estimating the hydrogen storage capacities of MOFs; (2) embed a metaheuristic algorithm called artificial lemming algorithm (ALA) to improve hyperparameter optimization, ensuring more rapid convergence and enhanced model generalizability across different datasets; and (3) conduct comprehensive sensitivity analysis to pinpoint the key physicochemical features that govern hydrogen storage in MOFs. The remainder of this paper is structured as follows: Section 2 provides an in-depth description of the methodologies employed, including details on the ML models, the enhanced metaheuristic optimization algorithm, and the sensitivity analysis approach. Section 3 delves into the characteristics and composition of the database used in this paper. Section 4 outlines the development and implementation of prediction models. Section 5 presents and discusses the results obtained from comparative analysis, hyperparameter optimization, and model interpretability. Finally, Section 6 summarizes the key findings, highlights the study’s contributions, and outlines directions for future research.

2. Methodologies

2.1. Artificial Neuron Network (ANN)

ANNs are a class of computational models inspired by biological neural systems, designed to learn and perform complex tasks in pattern recognition, classification, and regression [28]. An ANN typically comprises multiple layers of neurons (also referred to as nodes), which interact through weighted connections. As illustrated in Figure 1a, a standard ANN architecture comprises an input layer, one or more hidden layers, and an output layer. The input layer receives external data. Composed of numerous neurons, the hidden layers perform feature extraction and nonlinear transformation of the input. The output layer produces the final output of the network, which is commonly for classification or regression tasks. In the context of regression prediction using an ANN model, the model output can be mathematically represented as follows:
y = f i = 1 n w i x i + b
where x and y denote the input and output features, respectively. wi represents the weight of the i-th feature, and b represents a bias value. f() is an activation function.

2.2. Support Vector Regression (SVR)

Support vector machines (SVMs) are a type of supervised learning method applicable to both classification and regression tasks. The core concept of SVMs is to identify an optimal hyperplane that maximally separates samples of different classes, thereby enhancing the model’s generalization ability. Originally proposed by Vapnik et al. [29] and grounded in statistical learning theory, SVMs possess a solid mathematical foundation. The primary objective is to find a hyperplane in the feature space that not only correctly classifies the data points but also maximizes the margin between different classes. For regression tasks, the support vector regression (SVR) model adapts the SVM framework by introducing an ε-insensitive loss function, making it suitable for predicting continuous outcomes. The goal in SVR is to identify an optimal hyperplane that fits the data within a specified tolerance level of error (see Figure 1b). To this end, SVR typically employs an ε-insensitive loss function and constraint conditions to formulate the prediction objective equation.
L ( y , f x ) = 0 y f x ε y f x ε y f x > ε
min w , b 1 2 w 2 + C i = 1 n ξ i + ξ i *
where w is the regularization coefficient that controls model complexity. C is a penalty parameter. ξ i and ξ i * are slack variables introduced to allow certain deviations from the ε-insensitive tube. To enhance the model’s applicability to nonlinear data, the kernel method is employed to map the original data into a higher-dimensional feature space, where linear separation becomes feasible. In this paper, the radial basis function (RBF) kernel is selected to further improve the predictive performance of the SVR model.

2.3. Random Forest (RF)

RF is an ensemble learning method widely used for classification and regression tasks. By introducing randomness into the decision tree (DT) construction process, RF improves both the model’s generalization ability and robustness. RF regression is a type of bootstrap aggregation (bagging) technique that builds multitude decision trees and averages their outputs to produce the final prediction. This ensemble approach effectively reduces overfitting while improving model stability and accuracy [30]. The RF training process can be summarized in Figure 1c and involves the following steps: (1) random sampling of training data (with replacement); (2) random sampling of features at each node split; (3) construction of individual DTs; and (4) aggregation of multiple DTs. For regression tasks, the final prediction is obtained by averaging the outputs of all the individual DTs:
y = 1 N n = 1 N D T n
where N represents the maximum number of DTs in an RF model.

2.4. Extreme Learning Machine (ELM)

ELM is a single-hidden-layer feedforward neural network characterized by high computational efficiency. In the ELM model, the weights and biases of the hidden layer neurons are randomly assigned and remain fixed, while the output weights are analytically determined using the least squares method. This approach enables fast learning and training, distinguishing ELM from traditional neural network training methods such as backpropagation, which require iterative updates of weights and biases. By eliminating the need for backpropagation, ELM significantly simplifies the training process and greatly improves training speed [31]. As shown in Figure 1d, the ELM architecture typically consists of three layers: (1) an input layer receives external data; (2) a single hidden layer comprising multiple neurons that perform nonlinear transformations of the input. The weights and biases in this layer are randomly initialized and require no training; and (3) an output layer produces the final output of the network.

2.5. Kernel Extreme Learning Machine (KELM)

KELM is an advanced version of ELM that incorporates kernel methods to better handle nonlinear problems. As shown in Figure 1e, KELM bypasses explicit hidden layer construction by using a kernel function to map input data into a high-dimensional feature space. This allows the model to capture complex patterns more effectively. The output weights are computed analytically through a regularized least squares solution, thus retaining ELM’s fast learning speed and strong generalization performance. In this paper, the RBF kernel is also employed to further improve the predictive performance of the KELM model.

2.6. Generalized Regression Neural Network (GRNN)

GRNN is a type of single-hidden-layer feedforward neural network based on kernel density estimation and conditional expectation theory. GRNN assumes that the input and output variables follow a joint probability distribution. The relationship between inputs and output is modeled using conditional expectation. It estimates the joint probability density function from training samples and computes the conditional expectation to generate predictions. GRNN consists of four layers: an input layer, a pattern layer, a summation layer, and an output layer (see Figure 1f). This structure enables GRNN to achieve fast convergence and high accuracy, particularly in nonlinear regression problems [32].

2.7. Artificial Lemming Algorithm (ALA)

Xiao et al. [33] proposed a novel metaheuristic algorithm, called the Artificial Lemming Algorithm (ALA), designed to address common challenges in high-dimensional search spaces, such as premature convergence, insufficient exploration, and lack of robustness. The algorithm is inspired by the natural behaviors of lemmings, including long-distance migration, burrowing, foraging, and evading predators. In the wild, lemmings exhibit a unique behavioral pattern that provides a rich biological basis for algorithm design: (1) Long-distance migration–This behavior often occurs when population density surpasses environmental carrying capacity and food resources becomes limited. Rather than being a deliberate search for new habitats, migration represents a spontaneous group-level response to ecological stress. As the population increases, some lemmings naturally disperse to other regions in search of better access to food and space. (2) Digging holes–Lemmings are skilled burrowers that use their strong forelimbs and sharp claws to construct intricate underground tunnel systems. These complex burrow networks not only offer shelter from predators and harsh environmental conditions but also serve as storage sites for food and safe spaces for reproduction. (3) Foraging–As herbivores, lemmings primarily consume plant-based resources such as grasses, seeds, fruits, and roots. They forage both above ground and within their burrows, using acute sensory abilities to detect food. During winter, lemmings are even capable of digging beneath snow to access vegetation, and they often stockpile food to survive periods of scarcity. (4) Evading natural predators–Lemmings possess highly developed sensory systems that enable them to detect approaching predators promptly. Upon sensing danger, they emit alarm signals to alert nearby individuals, prompting them to hide or flee to safety. These four core behaviors—migration, digging, foraging, and predator evasion—collectively inspired the development of ALA. This population-inspired metaheuristic algorithm simulates the adaptive and survival strategies of lemmings to enhance global search ability, balance exploration and exploitation, and improve robustness in high-dimensional optimization tasks [33]. The mathematical formulation of ALA during the optimization process is as follows:
(a)
Population initialization
As social animals, lemmings search for food and evade predators within a defined spatial range. Accordingly, the initial positions in the optimization process are defined as
L i = L B + r a n d ( U B L B )
where Li represents the current position of the i-th lemming. LB and UB represent lower and upper boundaries of searching space, respectively. rand is a random number, which is constrained within the range of [0, 1].
(b)
Long-distance migration (exploration)
As food becomes scarce and the population size increases, lemmings must undertake long-distance migration to seek habitats with more abundant resources for the survival of the group. As shown in Figure 2a, this behavior corresponds to the exploration phase of ALA and is mathematically formulated as
L i ( t + 1 ) = L b e s t ( t ) + F B M ( R ( L b e s t ( t ) L i ( t ) ) + ( 1 R ) ( L i ( t ) L a ( t ) ) )
where L i ( t ) represents the current position of the i-th lemming at the t-th iteration. L i ( t + 1 ) represents the current position of the i-th lemming at the t + 1-th iteration. L b e s t ( t ) represents the current optimal position of the lemming at the t-th iteration. L a ( t ) represents the current position of a random individual lemming at the t-th iteration. F is a factor that controls the direction of migration. BM is a random number, which plays a crucial role in the deep exploration of the search space. R represents a factor that controls the movement of the best and random lemmings in the population.
(c)
Digging holes (exploration)
Once a suitable habitat is found, lemmings engage in burrowing behavior to prepare for future food acquisition and storage (see Figure 2b). The mathematical expression for this behavior can be formulated as
L i ( t + 1 ) = L i ( t ) + F I ( L b e s t ( t ) L b ( t ) )
where I is a random number. L b ( t ) represents the current position of a random individual lemming at the t-th iteration.
(d)
Foraging for food (exploitation)
Within the habitat, lemmings explore the entire space to locate the largest quantity of high-quality food and complete the storage for the group’s survival. As shown in Figure 2c, the position of lemmings during this process can be mathematically represented as
L i ( t + 1 ) = L b e s t ( t ) + F s p i r a l r a n d L i ( t )
where spiral represents the shape of searching space.
(e)
Evading natural predators (exploitation)
In the foraging process (see Figure 2d), if lemmings encounter a predator attack, they will quickly execute an avoidance behavior, including retreating to their burrows and utilizing deceptive behaviors to ensure self-protection. The mathematical expression for this behavior can be represented as
L i ( t + 1 ) = L b e s t ( t ) + F G l e v y L b e s t ( t ) L i ( t )
where G represents the escape coefficient of lemmings, which decreases with iteration increases. Levy is the Levy flight function, which is used to simulate the deceptive behaviors during the escape process.

3. Database

Previous studies have identified four key features as primary factors influencing hydrogen storage capacity [33]: surface area, pore volume, pressure, and temperature. Surface area, in particular, is a critical factor affecting the hydrogen storage performance of MOFs. Several techniques are available for determining surface area, among which the Langmuir and Brunauer–Emmett–Teller (BET) methods are the most widely employed. The Langmuir isotherm assumes monolayer adsorption on a homogeneous surface and does not account for multilayer adsorption phenomena [34]. In contrast, the BET approach incorporates the possibility of multilayer adsorption, typically yielding lower surface area values than those derived from the Langmuir model [35]. In this paper, the surface area measured by the BET method was considered as an input feature to predict the hydrogen storage. Furthermore, an increase in pore volume provides additional available space within the MOF structure, enabling it to accommodate a larger number of hydrogen molecules [36]. In addition, pressure and temperature are critical operational parameters for accurately predicting hydrogen storage capacity. For example, physisorption-based adsorbents typically exhibit higher gas storage capacities at cryogenic temperatures. At ambient conditions, weakened van der Waals forces lead to reduced storage capacity [36]. In light of the challenges associated with the experimental determination of hydrogen storage performance of MOFs, this paper utilizes 294 experimental datasets compiled by Salehi et al. [37] to generate the predictive models. In these experiments, hydrogen storage is primarily quantified in two primary forms: total and excess storage capacities. To achieve data consistency and ensure the robustness of the model, total hydrogen storage values were converted into excess hydrogen storage values using Equation (10).
N e x c e s s = N t o t a l d g a s V p o r e
where N t o t a l and N e x c e s s represent the total and excess hydrogen storage, respectively. d g a s and V p o r e represent the density and pore volume of compressed gas, respectively.
The statistical analyses and correlation test results for all datasets are comprehensively presented in Table 1 and Figure 3, providing a detailed overview of the variables’ statistical properties and the relationships among them. As can be seen from Figure 3, the correlation coefficient between BET surface area and pore volume reaches 0.91, which is attributed to the fact that these two features fundamentally reflect the pore characteristics of MOF materials. For instance, a MOF with a high BET surface area usually possesses a large number of pores with a well-ordered and uniform structure, which in turn results in a high pore volume. In addition, the relatively low correlation coefficients among all input features indicate that the effect of each parameter on the target output is largely independent. On the other hand, the correlation coefficients between most of the input features and the output variables are greater than 0.4. According to the Pearson correlation-based feature selection criterion [38], all input features are deemed suitable for inclusion in the predictive modeling process.

4. Development of Prediction Models

In this paper, six ML models were adopted to predict the hydrogen storage performance in MOFs. A novel metaheuristic algorithm (i.e., ALA) was utilized to optimize all models for improving prediction accuracy. To develop high-performance prediction models, as shown in Figure 4, several steps were undertaken:
(1)
Data preparation
As previously mentioned, 294 data samples were collected to construct the prediction models. First, the majority of data samples were allocated to form the training set, enabling the models to learn the underlying relationships between input features and the target variable. Subsequently, the remaining samples were assigned to the test set to evaluate the predictive accuracy of the trained models. Common ratios between the training and test sets include 7:3, 8:2, or 9:1. In this paper, a 7:3 ratio (training set: 206 samples and test set: 88 samples) was adopted based on modeling procedures and predictive performance reported in related prior research by Salehi et al. [37]. Furthermore, all input features were normalized into the range of [−1, 1] to prevent overfitting caused by discrepancies in the magnitudes or units of different variables [39].
(2) Model optimization
The predictive accuracy of each model is closely associated with its hyperparameter selection. For the ANN model, the number of hidden layers (Nh) and the number of neurons per layer (Nn) significantly influence model performance, with ranges of [1–3] and [1–10], respectively. For the SVR model, the penalty parameter C and the kernel parameter (k) are the most influential factors, with ranges of [0.25–252] and [0.25–16], respectively. For the RF model, prediction accuracy is primarily determined by the number of trees (Nt) and the minimum leaf point (Minleafsize), varied within the ranges of [1–100] and [1–10], respectively. In the ELM model, selecting an appropriate Nn in the single hidden layer poses a key challenge, with the range defined as [1–150]. The hyperparameter selection and corresponding ranges for the KELM model are consistent with those used in the SVR model. In the GRNN model, the smoothing factor (Sf) is adjusted within the range of 0–5 to evaluate its impact on model performance. Furthermore, the population size and number of iterations significantly influence the optimization performance of the ALA algorithm. Therefore, four values for population sizes (25, 50, 75, and 100) were tested to select the optimal hyperparameters for all ML models over 200 iterations. During this process, a fitness function was constructed to identify the optimal solution for each model. Generally, the root mean square error (RMSE), which does not require consideration of absolute values, is commonly used in conjunction with cross-validation to define the fitness function [40,41]. In this paper, 10-fold cross-validation was combined with RMSE to calculate fitness values for determining the optimal solutions:
f i t n e s s = 1 K k = 1 K R M S E k
where K represents the maximum number of subsets in the cross validation.
(3)
Model evaluation
After determining the optimal solutions for all models, several indices were utilized to evaluate model performance. In this paper, four statistical indices, including coefficient of determination (R2), RMSE, Willmott’s index (WI), and weighted average percentage error (WAPE), were utilized to evaluate model performance. The definitions of these indices are expressed by Equations (12)–(15). Furthermore, several tools such as regression graphs, error analysis, and Taylor graphs were also employed to determine the optimal prediction model.
R 2 = 1 i = 1 n ( Z i z i ) 2 i = 1 n ( Z i Z ¯ ) 2
RMSE = 1 n i = 1 n Z i z i 2
WI = 1 i = 1 n Z i z i 2 i = 1 n z i Z ¯ + Z i Z ¯ 2
WAPE = i = 1 n ( Z i z i ) 2 i = 1 n ( Z i Z ¯ ) 2
where Z i and z i represent the actual and predicted values of the i-th data points in the dataset, respectively. Z ¯ is the average of actual values. n is the maximum number of data points in the dataset.
(4)
Model interpretability
To further elucidate the influence of input features on the prediction target, this paper adopts the Shapley additive explanations (SHAP) method to quantify the importance of each input feature and illustrate its contribution to hydrogen storage prediction. This approach improves the interpretability of the “black-box” model used in this paper.

5. Results and Discussion

5.1. Hyperparameter Selection

In this paper, the ALA algorithm was utilized to determine the optimal hyperparameter combinations for each prediction model. Figure 5 demonstrates the fitness curves of all models over 200 iterations. As shown in Figure 5a, all ANN models optimized by ALA with different population sizes reached their minimum fitness values before achieving the maximum number of iterations. Among these models, the ALA-ANN model with a population size of 75 achieved the lowest fitness value. Figure 5b illustrates the iteration results for the optimized SVR models using four different population sizes. Significant variations in fitness values are observed across various population sizes, with the lowest value achieved by the ALA-SVR model at a population size of 50. For the ALA-RF and ALA-ELM hybrid models, changes in population size resulted in only minor differences in minimum fitness values. Notably, the minimum fitness values for hybrid models with 50 and 75 population sizes were nearly identical (see Figure 5c,d). As demonstrated in Figure 5e,f, both ALA-KELM and ALA-GRNN models obtained their minimum fitness values at a population size of 75. Table 2 summarizes the results of fitness values for each model and their optimal hyperparameter combinations after optimization.

5.2. Model Performance Evaluation

After obtaining the optimal hyperparameters for all models, the model performance needed to be evaluated to determine the best model for predicting the hydrogen storage. First, the prediction accuracy of the models was verified using the training set. Table 3 presents the results of evaluation indices for all optimized ML models. It can be observed that all models achieved satisfactory prediction performance, as reflected in favorable evaluation indices, with R2 values exceeding 0.9. Among these models, the ALA-RF model demonstrated higher competitiveness in hydrogen storage prediction compared to the other models, with higher values of R2 and WI (0.9845 and 0.9961) and lower values of RMSE and WAPE (0.2719 and 0.0667). Following this model, ALA-ANN, ALA-SVR, ALA-KELM, and ALA-GRNN, but not the ALA-ELM model, also exhibited strong predictive performance during the training phase.
Furthermore, regression graphs were generated to illustrate the prediction performance of all models in the training phase. In a regression graph, the position of each data point is determined by the actual and predicted values. If the predicted values equal actual values, the data points lie on the diagonal. Otherwise, the data points deviate from the diagonal based on the prediction bias. However, perfect predictions are rarely achievable in practical engineering problems. Therefore, a certain degree of prediction error is acceptable. This paper introduces two limitation lines (10%) to assess the predictive accuracy of each model in more detail. Data points falling outside the range of these limitation lines are considered “outliers”. The greater the number of outliers in a model’s predictions, the poorer its performance is deemed to be. As shown in Figure 6, 30 outliers were obtained by the ALA-ELM model, ranking the highest among all models, followed by the ALA-GRNN, ALA-KELM, ALA-SVR, and ALA-ANN models. It is worth noting that the ALA-RF model had zero outliers due to its excellent predictive performance, where all data points were within the boundary line.
However, the performance of a model during the training phase does not necessarily reflect its ultimate applicability in real-world engineering scenarios [42,43]. Therefore, the test set was used to assess the actual predictive performance of the trained models. Table 4 presents the results of evaluation indices for all optimized ML models in the testing phase. It is evident that two models (ALA-ELM and ALA-GRNN) obtained unsatisfactory prediction performance, as indicated by poor evaluation indices, including R2 values lower than 0.9. Among the other models, the ALA-RF model demonstrated the strongest performance in predicting hydrogen storage, achieving the highest values of R2 and WI (0.9840 and 0.9959) and the lowest values of RMSE and WAPE (0.2828 and 0.0714). Following ALA-RF, the ALA-ANN, ALA-SVR, and ALA-KELM also exhibited strong predictive performance during the testing phase.
Figure 7 illustrates the regression graphs of all models in the testing phase. As shown in this figure, ALA-ELM and ALA-RF represent the worst and best-performing models among all models, respectively. For ALA-ELM, 15 outliers are distributed outside the limitation lines, whereas ALA-RF shows no outliers in its prediction results. Following ALA-RF, the number of outliers obtained with the ALA-GRNN, ALA-KELM, ALA-SVR, and ALA-ANN models are 9, 7, 5, and 2, respectively.
Moreover, error analysis was conducted to further compare model performance using the test set. Figure 8 shows the error distribution between actual and predicted values for each model. It is evident that all models exhibit a high proportion of prediction errors less than 1, with only the ALA-RF model maintaining prediction errors within 1 for all test samples. Among these models, both ALA-RF and ALA-ANN display a more uniform cumulative distribution of errors and lower maximum error values compared to the others. In contrast, ALA-ELM and ALA-GRNN demonstrate significantly more instances of large prediction errors. Furthermore, this paper employs the standard deviation of prediction errors as a metric to rank the models’ predictive accuracy. A lower standard deviation indicates that the prediction errors are more concentrated, stable, and less volatile. When considered alongside the error distribution analysis, this metric enables a comprehensive evaluation of model performance [44]. The results indicate that the minimum standard deviation of prediction errors was obtained by the ALA-RF model (0.1751), and the ALA-ELM model obtained the maximum standard deviation of prediction errors (0.7931). Notably, ALA-KELM’s prediction error standard deviation was lower than that of ALA-SVR. This outcome may be attributed to large prediction deviations for certain individual samples in the ALA-SVR model.
Therefore, it was necessary to further validate the performance of each model. In this paper, a Taylor graph tool was employed to determine the final performance of all models. In a Taylor graph, the position of each model is determined by three indices, including standard deviation, RMSE, and correlation coefficient. For the test set point, the reference point has a correlation coefficient of 1 and an RMSE of 0, as it lies on the horizontal axis quantified by the standard deviation. For other models developed in this paper, the position relative to the test points reflects their predictive performance. In principle, a model that is closer to the test point indicates better performance. As shown in Figure 9, the ALA-RF model was closest to the test point, followed by the ALA-ANN model. Notably, the ALA-SVR model was closer to the test point compared to the ALA-KELM model, suggesting that the former demonstrated superior predictive performance.
Additionally, a previous study [37] revealed that the committee machine intelligence system (CMIS) model obtained high performance in predicting hydrogen storage, with an overall R2 of 0.982. To further compare the performance of various models in hydrogen storage prediction, three ALA-RF models were reconstructed under different data splitting ratios (6:4, 8:2, and 9:1). As shown in Table 5, the results of performance evaluations indicated that model accuracy on the training set improved as the number of training samples increased. However, when the training set accounted for more than 70% of the total data, the performance gain became marginal, suggesting diminishing returns with additional training data. Based on the R2 metric, the ALA-RF obtained higher prediction accuracy in both the training (R2 of 0.9845) and testing (R2 of 0.9840) phases based on the same data allocation (training set: test set = 7:3). In conclusion, the ALA-RF model was the best-performing model for predicting hydrogen storage among the developed models.

5.3. Model Interpretability

After identifying the optimal model for predicting hydrogen storage in MOFs, the impact and contribution of the input features to the prediction target were analyzed using the SHAP method. Figure 10 shows the importance score of all features in predicting hydrogen storage through the SHAP analysis. Pressure was found to have the greatest impact on the prediction of hydrogen storage, with the highest importance score of 1.22. This was followed by temperature (0.84), BET surface area (0.30), and pore volume (0.18). Furthermore, the contribution of each feature to hydrogen storage prediction is illustrated in Figure 11. It is evident that pressure, BET surface, and pore volume positively influence the prediction of hydrogen storage, indicating that increases in these features correspond to increases in the predicted values of hydrogen storage. With respect to temperature, a negative correlation with hydrogen storage was observed, suggesting that reducing temperature contributes to an increase in hydrogen storage. These findings are consistent with those reported by Salehi et al. [37], thereby reinforcing the validity of the current analysis.

6. Conclusions

In this paper, six ML models were adopted to predict hydrogen storage in MOFs based on 294 data samples. A novel metaheuristic optimization algorithm, ALA, was employed to further improve model performance. After evaluating the model performance, SHAP analysis was conducted to calculate the feature importance score and impact on hydrogen storage prediction. The main conclusions are summarized as follows:
(1)
The evaluation results demonstrated that the ALA-RF model was the optimal model for predicting hydrogen storage in MOFs, yielding the most satisfactory performance in both the training and testing phases. The values of R2, RMSE, WI, and WAPE were 0.9845, 0.2719, 0.9961, and 0.0667 (training set) and 0.9840, 0.2828, 0.9959, and 0.0714 (test set), respectively. In prediction accuracy, the model provided by this paper outperformed the previous model developed using the same database.
(2)
According to the SHAP analysis, pressure was identified as the most important feature for predicting hydrogen storage in MOFs, with the highest importance score of 1.22 among all input features. Temperature exhibited the most significant negative contribution to the prediction results for hydrogen storage.
However, the proposed model demonstrated strong predictive performance, which was limited by the relatively small and structurally narrow dataset of 294 MOF samples. These factors may restrict the model’s generalizability to unexplored or more complex MOF chemistries. The absence of molecular-level descriptors—such as specific metal nodes, organic linkers, or functional groups—also limits the ability to capture important steric and electronic effects that influence hydrogen adsorption. Furthermore, although pressure and temperature were identified as critical factors, the current work did not extend these insights to virtual screening or structural optimization. Future efforts should integrate machine learning models with high-throughput molecular simulations to explore a broader design space and guide the rational development of novel MOFs. Additionally, developing an automated recommendation system that connects predictive models with operational constraints could enable inverse design and accelerate experimental validation.

Author Contributions

Conceptualization: J.Z. (Jiamin Zhang), and Y.L.; Methodology: Y.L., C.L. and J.Z. (Jian Zhou); Investigation: Y.L. and J.Z. (Jian Zhou); Writing—original draft preparation: J.Z. (Jiamin Zhang) and C.L.; Writing—review and editing: C.L., J.Z. (Jiamin Zhang), X.M. and J.Z. (Jian Zhou); Visualization: Y.L., C.L., J.Z. (Jiamin Zhang), X.M. and J.Z. (Jian Zhou); Funding acquisition: J.Z. (Jian Zhou). All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially supported by the National Science Foundation of China (42177164).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding authors.

Acknowledgments

The authors want to thank all the members who provided us with extensive assistance and cooperation.

Conflicts of Interest

Author Jiamin Zhang was employed by the company SINOPEC Research Institute of Petroleum Engineering. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Farghali, M.; Osman, A.I.; Mohamed, I.M.; Chen, Z.; Chen, L.; Ihara, I.; Yap, P.-S.; Rooney, D.W. Strategies to save energy in the context of the energy crisis: A review. Environ. Chem. Lett. 2023, 21, 2003–2039. [Google Scholar] [CrossRef] [PubMed]
  2. Kamran, M.; Turzyński, M. Exploring hydrogen energy systems: A comprehensive review of technologies, applications, prevailing trends, and associated challenges. J. Energy Storage 2024, 96, 112601. [Google Scholar] [CrossRef]
  3. Hu, Y.H.; Zhang, L. Hydrogen storage in metal–organic frameworks. Adv. Mater. 2010, 22, E117–E130. [Google Scholar] [CrossRef]
  4. Xiong, Q.; Chen, Y.; Yang, D.; Wang, K.; Wang, Y.; Yang, J.; Li, L.; Li, J. Constructing strategies for hierarchically porous MOFs with different pore sizes and applications in adsorption and catalysis. Mater. Chem. Front. 2022, 6, 2944–2967. [Google Scholar] [CrossRef]
  5. Qureshi, F.; Yusuf, M.; Ahmed, S.; Haq, M.; Alraih, A.M.; Hidouri, T.; Kamyab, H.; Vo, D.-V.N.; Ibrahim, H. Advancements in sorption-based materials for hydrogen storage and utilization: A comprehensive review. Energy 2024, 309, 132855. [Google Scholar] [CrossRef]
  6. Kancharlapalli, S.; Snurr, R.Q. High-throughput screening of the CoRE-MOF-2019 database for CO2 capture from wet flue gas: A multi-scale modeling strategy. ACS Appl. Mater. Interfaces 2023, 15, 28084–28092. [Google Scholar] [CrossRef] [PubMed]
  7. Avci, G.; Velioglu, S.; Keskin, S. High-throughput screening of MOF adsorbents and membranes for H2 purification and CO2 capture. ACS Appl. Mater. Interfaces 2018, 10, 33693–33706. [Google Scholar] [CrossRef]
  8. Edlmann, K.; Haszeldine, S.; Heinemann, N.; Hassanpouryouzband, A.; Thaysen, E. Hystorpor: Unlocking the UK’s Hydrogen Storage; HyStorPor Brief; University of Edinburgh: Edinburgh, UK, 2021. [Google Scholar]
  9. Yilmaz, G.; Peh, S.B.; Zhao, D.; Ho, G.W. Atomic-and molecular-level design of functional metal–organic frameworks (MOFs) and derivatives for energy and environmental applications. Adv. Sci. 2019, 6, 1901129. [Google Scholar] [CrossRef]
  10. Burtch, N.C.; Jasuja, H.; Walton, K.S. Water stability and adsorption in metal–organic frameworks. Chem. Rev. 2014, 114, 10575–10612. [Google Scholar] [CrossRef]
  11. Zhang, C.; Xie, Y.; Xie, C.; Dong, H.; Zhang, L.; Lin, J. Accelerated discovery of porous materials for carbon capture by machine learning: A review. MRS Bull. 2022, 47, 432–439. [Google Scholar] [CrossRef]
  12. Mei, X.; Li, J.; Zhang, J.; Cui, Z.; Zhou, J.; Li, C. Predicting energy absorption characteristic of rubber concrete materials. Constr. Build. Mater. 2025, 465, 140248. [Google Scholar] [CrossRef]
  13. Zhang, W.; Wu, C.; Tang, L.; Gu, X.; Wang, L. Efficient time-variant reliability analysis of Bazimen landslide in the Three Gorges Reservoir Area using XGBoost and LightGBM algorithms. Gondwana Res. 2023, 123, 41–53. [Google Scholar] [CrossRef]
  14. Li, C.; Mei, X. Application of SVR models built with AOA and Chaos mapping for predicting tunnel crown displacement induced by blasting excavation. Appl. Soft Comput. 2023, 147, 110808. [Google Scholar] [CrossRef]
  15. Nie, F.; Su, X.; Wang, M.; Ma, X.; Ou, K.; Liu, J.; Lin, H. Debonding of asphalt-aggregate interface under coupled moisture and temperature conditions: An atomistic study. Case Stud. Constr. Mater. 2025, 22, e04554. [Google Scholar] [CrossRef]
  16. Yuan, X.; Suvarna, M.; Low, S.; Dissanayake, P.D.; Lee, K.B.; Li, J.; Wang, X.; Ok, Y.S. Applied machine learning for prediction of CO2 adsorption on biomass waste-derived porous carbons. Environ. Sci. Technol. 2021, 55, 11925–11936. [Google Scholar] [CrossRef]
  17. Wang, R.; Zhong, Y.; Bi, L.; Yang, M.; Xu, D. Accelerating discovery of metal–organic frameworks for methane adsorption with hierarchical screening and deep learning. ACS Appl. Mater. Interfaces 2020, 12, 52797–52807. [Google Scholar] [CrossRef]
  18. Anderson, R.; Biong, A.; Gómez-Gualdrón, D.A. Adsorption isotherm predictions for multiple molecules in MOFs using the same deep learning model. J. Chem. Theory Comput. 2020, 16, 1271–1283. [Google Scholar] [CrossRef]
  19. Wei, F. Synthetic prediction of MOF based on RF and K-nearest neighbor. In Proceedings of the 2024 IEEE 2nd International Conference on Electrical, Automation and Computer Engineering (ICEACE), Changchun, China, 21–31 December 2024; pp. 1005–1011. [Google Scholar]
  20. Guo, W.; Liu, J.; Dong, F.; Chen, R.; Das, J.; Ge, W.; Xu, X.; Hong, H. Deep learning models for predicting gas adsorption capacity of nanomaterials. Nanomaterials 2022, 12, 3376. [Google Scholar] [CrossRef]
  21. Pardakhti, M.; Moharreri, E.; Wanik, D.; Suib, S.L.; Srivastava, R. Machine learning using combined structural and chemical descriptors for prediction of methane adsorption performance of metal organic frameworks (MOFs). ACS Comb. Sci. 2017, 19, 640–645. [Google Scholar] [CrossRef]
  22. Choudhary, K.; Yildirim, T.; Siderius, D.W.; Kusne, A.G.; McDannald, A.; Ortiz-Montalvo, D.L. Graph neural network predictions of metal organic framework CO2 adsorption properties. Comput. Mater. Sci. 2022, 210, 111388. [Google Scholar] [CrossRef]
  23. Liang, H.; Jiang, K.; Yan, T.A.; Chen, G.H. XGBoost: An optimal machine learning model with just structural features to discover MOF adsorbents of Xe/Kr. ACS Omega 2021, 6, 9066–9076. [Google Scholar] [CrossRef] [PubMed]
  24. Bergstra, J.; Bengio, Y. Random search for hyper-parameter optimization. J. Mach. Learn. Res. 2012, 13, 281–305. [Google Scholar]
  25. Chowdhury, C. Bayesian optimization for efficient prediction of gas uptake in nanoporous materials. ChemPhysChem 2024, 25, e202300850. [Google Scholar] [CrossRef]
  26. Du, Y.; Du, D. Bayesian optimization enhanced neural networks for predicting metal-organic framework morphology: A ZIF-8 synthesis case study. Mater. Lett. 2025, 380, 137738. [Google Scholar] [CrossRef]
  27. Sun, Y.; DeJaco, R.F.; Li, Z.; Tang, D.; Glante, S.; Sholl, D.S.; Colina, C.M.; Snurr, R.Q.; Thommes, M.; Hartmann, M.; et al. Fingerprinting diverse nanoporous materials for optimal hydrogen storage conditions using meta-learning. Sci. Adv. 2021, 7, eabg3983. [Google Scholar] [CrossRef] [PubMed]
  28. Agatonovic-Kustrin, S.; Beresford, R. Basic concepts of artificial neural network (ANN) modeling and its application in pharmaceutical research. J. Pharm. Biomed. Anal. 2000, 22, 717–727. [Google Scholar] [CrossRef]
  29. Vapnik, V.; Guyon, I.; Hastie, T. Support vector machines. Mach. Learn 1995, 20, 273–297. [Google Scholar]
  30. Ahmad, M.; Kamiński, P.; Olczak, P.; Alam, M.; Iqbal, M.J.; Ahmad, F.; Sasui, S.; Khan, B.J. Development of prediction models for shear strength of rockfill material using machine learning techniques. Appl. Sci. 2021, 11, 6167. [Google Scholar] [CrossRef]
  31. Li, C.; Zhou, J.; Tao, M.; Du, K.; Wang, S.; Armaghani, D.J.; Mohamad, E.T. Developing hybrid ELM-ALO, ELM-LSO and ELM-SOA models for predicting advance rate of TBM. Transp. Geotech. 2022, 36, 100819. [Google Scholar] [CrossRef]
  32. Zhang, Y.; Niu, J.; Na, S. A novel nonlinear function fitting model based on FOA and GRNN. Math. Probl. Eng. 2019, 2019, 2697317. [Google Scholar] [CrossRef]
  33. Xiao, Y.; Cui, H.; Khurma, R.A.; Castillo, P.A. Artificial lemming algorithm: A novel bionic meta-heuristic technique for solving real-world engineering optimization problems. Artif. Intell. Rev. 2025, 58, 84. [Google Scholar] [CrossRef]
  34. Heraldy, E.; Hidayat, Y.; Firdaus, M. The langmuir isotherm adsorption equation: The monolayer approach. IOP Conf. Ser. Mater. Sci. Eng. 2016, 107, 012067. [Google Scholar]
  35. Walton, K.S.; Snurr, R.Q. Applicability of the BET method for determining surface areas of microporous metal-organic frameworks. J. Am. Chem. Soc. 2007, 129, 8552–8556. [Google Scholar] [CrossRef] [PubMed]
  36. Suh, M.P.; Park, H.J.; Prasad, T.K.; Lim, D.W. Hydrogen storage in metal–organic frameworks. Chem. Rev. 2012, 112, 782–835. [Google Scholar] [CrossRef]
  37. Salehi, K.; Rahmani, M.; Atashrouz, S. Machine learning assisted predictions for hydrogen storage in metal-organic frameworks. Int. J. Hydrogen Energy 2023, 48, 33260–33275. [Google Scholar] [CrossRef]
  38. Zhang, W.G.; Li, H.R.; Wu, C.Z.; Li, Y.Q.; Liu, Z.Q.; Liu, H.L. Soft computing approach for prediction of surface settlement induced by earth pressure balance shield tunneling. Undergr. Space 2021, 6, 353–363. [Google Scholar] [CrossRef]
  39. Li, C.; Zhou, J. Prediction and optimization of adverse responses for a highway tunnel after blasting excavation using a novel hybrid multi-objective intelligent model. Transp. Geotech. 2024, 45, 101228. [Google Scholar] [CrossRef]
  40. Tang, L.; Na, S. Comparison of machine learning methods for ground settlement prediction with different tunneling datasets. J. Rock Mech. Geotech. Eng. 2021, 13, 1274–1289. [Google Scholar] [CrossRef]
  41. Qiu, J.; Huang, R.; Wang, H.; Wang, F.; Zhou, C. Rate-dependent tensile behaviors of jointed rock masses considering geological conditions using a combined BPM-DFN model: Strength, fragmentation and failure modes. Soil Dyn. Earthq. Eng. 2025, 195, 109393. [Google Scholar] [CrossRef]
  42. Zhou, J.; Asteris, P.G.; Armaghani, D.J.; Pham, B.T. Prediction of ground vibration induced by blasting operations through the use of the Bayesian Network and random forest models. Soil Dyn. Earthq. Eng. 2020, 139, 106390. [Google Scholar] [CrossRef]
  43. Zhou, C.; Rui, Y.; Qiu, J.; Wang, Z.; Zhou, T.; Long, X.; Shan, K. The role of fracture in dynamic tensile responses of fractured rock mass: Insight from a particle-based model. Int. J. Coal Sci. Technol. 2025, 12, 39. [Google Scholar] [CrossRef]
  44. Li, J.; Li, C.; Zhang, S. Application of Six Metaheuristic Optimization Algorithms and Random Forest in the uniaxial compressive strength of rock prediction. Appl. Soft Comput. 2022, 131, 109729. [Google Scholar] [CrossRef]
Figure 1. Demonstration of all ML models used in this paper.
Figure 1. Demonstration of all ML models used in this paper.
Materials 18 03122 g001
Figure 2. Demonstration of ALA optimization algorithm used in this paper [33].
Figure 2. Demonstration of ALA optimization algorithm used in this paper [33].
Materials 18 03122 g002
Figure 3. Correlation matrix of all features.
Figure 3. Correlation matrix of all features.
Materials 18 03122 g003
Figure 4. The flowchart of predicting hydrogen storage in MOFs.
Figure 4. The flowchart of predicting hydrogen storage in MOFs.
Materials 18 03122 g004
Figure 5. Iteration curves of all models during optimization process.
Figure 5. Iteration curves of all models during optimization process.
Materials 18 03122 g005
Figure 6. Regression graph of all models based on the training set.
Figure 6. Regression graph of all models based on the training set.
Materials 18 03122 g006
Figure 7. Regression graphs of all models based on the test set.
Figure 7. Regression graphs of all models based on the test set.
Materials 18 03122 g007
Figure 8. Error distribution of all models for predicting hydrogen storage in the testing phase.
Figure 8. Error distribution of all models for predicting hydrogen storage in the testing phase.
Materials 18 03122 g008
Figure 9. Taylor graph of all models based on the test set.
Figure 9. Taylor graph of all models based on the test set.
Materials 18 03122 g009
Figure 10. Importance score of all features in predicting hydrogen storage.
Figure 10. Importance score of all features in predicting hydrogen storage.
Materials 18 03122 g010
Figure 11. SHAP values of all features in predicting hydrogen storage.
Figure 11. SHAP values of all features in predicting hydrogen storage.
Materials 18 03122 g011
Table 1. Statistical analyses of all features with the database.
Table 1. Statistical analyses of all features with the database.
FeaturesUnitStatistical Indices
MedianStandard DeviationMinimumMaximum
BET surface area/2798.501225.16150.006240.00
Pore volumecm3/g1.190.520.043.60
PressureBar18.2436.510.19100.00
TemperatureK77.0074.1030.00300.00
Hydrogen storage%2.752.210.039.95
Table 2. Fitness values and optimal hyperparameter combinations of all models.
Table 2. Fitness values and optimal hyperparameter combinations of all models.
ModelsPopulation SizesHyperparameters
255075100
ALA-ANN0.089490.093280.088690.09088Nh: 1; Nn: 8
ALA-SVR0.107450.100630.101680.10512C: 135.2; k: 1.35
ALA-RF0.089200.081220.081240.08724Nt: 45; Minleafsize: 1
ALA-ELM0.255770.254710.255320.25606Nn: 72
ALA-KELM0.143640.144870.142630.14543C: 114.5; k: 1.87
ALA-GRNN0.224410.226960.223300.22568Sf: 0.4
Table 3. Statistical results of four evaluation indices for six models in training phase.
Table 3. Statistical results of four evaluation indices for six models in training phase.
ModelsEvaluation Indices
R2RMSEWIWAPE
ALA-ANN0.96490.40980.99090.1031
ALA-SVR0.94110.53060.98480.0922
ALA-RF0.98450.27190.99610.0667
ALA-ELM0.90630.66930.97690.1697
ALA-KELM0.93270.56720.98340.1420
ALA-GRNN0.92320.60580.98080.1500
Table 4. Statistical results of four evaluation indices for six models in testing phase.
Table 4. Statistical results of four evaluation indices for six models in testing phase.
ModelsEvaluation Indices
R2RMSEWIWAPE
ALA-ANN0.96490.40980.99090.1031
ALA-SVR0.94110.53060.98480.0922
ALA-RF0.98450.27190.99610.0667
ALA-ELM0.90630.66930.97690.1697
ALA-KELM0.93270.56720.98340.1420
ALA-GRNN0.92320.60580.98080.1500
Table 5. Performance comparison between the previous and present models.
Table 5. Performance comparison between the previous and present models.
ModelsRatio of Training Set to Test SetR2
TrainingTest
ALA-RF6:40.97050.8917
ALA-RF7:30.98450.9840
ALA-RF8:20.98750.9421
ALA-RF9:10.98920.9213
CMIS7:30.98300.9780
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, J.; Li, Y.; Li, C.; Mei, X.; Zhou, J. Application of Soft Computing Represented by Regression Machine Learning Model and Artificial Lemming Algorithm in Predictions for Hydrogen Storage in Metal-Organic Frameworks. Materials 2025, 18, 3122. https://doi.org/10.3390/ma18133122

AMA Style

Zhang J, Li Y, Li C, Mei X, Zhou J. Application of Soft Computing Represented by Regression Machine Learning Model and Artificial Lemming Algorithm in Predictions for Hydrogen Storage in Metal-Organic Frameworks. Materials. 2025; 18(13):3122. https://doi.org/10.3390/ma18133122

Chicago/Turabian Style

Zhang, Jiamin, Yanzhe Li, Chuanqi Li, Xiancheng Mei, and Jian Zhou. 2025. "Application of Soft Computing Represented by Regression Machine Learning Model and Artificial Lemming Algorithm in Predictions for Hydrogen Storage in Metal-Organic Frameworks" Materials 18, no. 13: 3122. https://doi.org/10.3390/ma18133122

APA Style

Zhang, J., Li, Y., Li, C., Mei, X., & Zhou, J. (2025). Application of Soft Computing Represented by Regression Machine Learning Model and Artificial Lemming Algorithm in Predictions for Hydrogen Storage in Metal-Organic Frameworks. Materials, 18(13), 3122. https://doi.org/10.3390/ma18133122

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop