Assessment of Cubic Equations of State: Machine Learning for Rich Carbon-Dioxide Systems

Carbon capture and storage (CCS) has attracted renewed interest in the re-evaluation of the equations of state (EoS) for the prediction of thermodynamic properties. This study also evaluates EoS for Peng–Robinson (PR) and Soave–Redlich–Kwong (SRK) and their capability to predict the thermodynamic properties of CO2-rich mixtures. The investigation was carried out using machine learning such as an artificial neural network (ANN) and a classified learner. A lower average absolute relative deviation (AARD) of 7.46% was obtained for the PR in comparison with SRK (AARD = 15.0%) for three components system of CO2 with N2 and CH4. Moreover, it was found to be 13.5% for PR and 19.50% for SRK in the five components’ (CO2 with N2, CH4, Ar, and O2) case. In addition, applying machine learning provided promise and valuable insight to deal with engineering problems. The implementation of machine learning in conjunction with EoS led to getting lower predictive AARD in contrast to EoS. An of AARD 2.81% was achieved for the three components and 12.2% for the respective five components mixture.


Introduction
The increased global awareness of the effect of CO 2 on the climate has renewed the interest in carbon capture and storage (CCS) technologies for the reduction of CO 2 emissions relative to historic emissions; these innovative technologies aim to achieve a lower relative rise in global mean temperature (GMT) long-term [1]. Though the accurate prediction of the thermodynamic properties of CO 2 -rich mixtures is critical in terms of design calculations for the selectivity, costing, and safety of operations, improper property prediction will have an adverse effect on a project's economic feasibility in regards to overdesign or lack thereof [2,3].

A Brief Insight into Equation of State and Machine Learning
There are a vast number of equations of state (EoS) used for the prediction of thermodynamic properties of gas mixtures. These EoS are used in a fluid package within software such as Aspen HYSYS to predict fundamental properties such as density, heat capacity, viscosity, etc. [4]. Also, it is highly important to combine the existing EoS with other methods to improve the reliability and accuracy of predicting thermodynamic properties. It is because using EoS in the design calculations in software packages such as AspenONE, without access to autonomous property prediction programes such as NIST Thermo Data Engine [5], is not sufficiently accurate. Machine learning itself can be used to predict property data, however, current works are limited to species within fractional components X, Y, etc. Combined with EoS principles, it can be used to predict mixtures' thermody-find the PVT behavior in gas and liquid states by including a repulsive and attraction term [14]. The underlying principle of the Van der Waals' equation is the description of phase behavior through dimensionless reduced variables (a, b, and c constants) relating to reduced fluid properties: Tr, Pr, and Vr which is defined by the Equation (1): These reduced variables yield a universal reduced equation, though this principle is valid for substances that present similar PVT behavior in terms of reduced variables [15]. The theorem of corresponding states for fluid phases is expressed by a compressibility factor (Z) (Equation (2)), which may be used in the ideal gas formulae to define real fluid properties: The nature of cubic EoS, being of cubic nature, allows for three possible roots, where T r being smaller than 1 can enter the vapor-liquid equilibrium regime, though as the reduced temperature is decreased further, minima and maxima are present (for the turning point) in the equation of state shown in where three real roots are present. It has been observed that the low-temperature results in a metastable fluid, this phenomenon is known as "stretched liquid" and causes negative absolute pressure. However, this is not further explored in this work [16]. As a result of the abovementioned Corresponding States Theorem, the root for the basis of prediction for all cubic EoS has empirical correlations to the critical compressibility of a set of mixtures.
Van Der Waals acknowledges that substances behave alike at the same reduced states, that being said, at their respective reduced state the properties of these substances are alike, which is known as the corresponding states principle. Hence, the primary purpose of the use of corresponding states is for the cubic empirical model [17]: The tabulation of the cubic equation of state in Table 1 is found in the literature where Van der Waals, RK, SRK, PR, Harmens-Knapp, and Schmidt-Wenzel EoS were obtained from [13]; Patel-Teja [18], and the modified Nasrifar Moshfeghian EoS [19].  [13,18,19].

Cubic Equation a b c Simplified Equation
Van der Waals a 0 0

Modified Nasrifar Moshfeghian
The parameters for the EoS have an approximate physical meaning, where a is the attraction force between the molecules, b is related to the repulsion and the effective molecule size [20]. The a and b parameters used in this work as presented in Table 2 [16]. The table lists the original works where the Soave-modified Redlich-Kwong alpha function is modified for hydrogen-containing systems due to the conventional extension. This is unable to accurately predict the K values of hydrogen, which may be required in syngas processing [21].
EoS extension to mixtures requires the use of mixing rules to calculate pseudo-reduced properties; the rules applied for mixtures may be as basic as linear mixing, assuming no activity between mixture species. An improvement is an extension to mixtures through the use of Van der Waals mixing for pseudo-reduced attraction and repulsion parameters through interaction parameters. There are two popular mixing rules including linear (Equations (5)- (7)) and Van Der Waals (Equations (6)-(8)) mixing rules [13].
In the current study, in contrast to the semi-empirical correlation, the binary interaction coefficients (k ij ) were extracted from the Aspen HYSYS database for Peng-Robinson and Soave-Redlich-Kwong to solve Equation (8). The binary interaction coefficients for Peng Robinson and Soave Redlich Kwong are provided in Tables 3 and 4.  For the evaluation of EoS and the development of the Artificial Neural Network model, pure component properties, and the properties of a binary mixture such as density and viscosity data were obtained from the literature. For the prediction of binary mixture components, the obtaining of their respective pure component physical properties was undertaken using NIST databanks for properties such as critical pressure, temperature, volume, relative molecular mass, and acentric factor. The pure component properties are given in Table 5. For the development of a single fluid property prediction, the work was done by [21] by measuring densities and thermodynamic properties for three CO 2rich mixtures. Firstly, the evaluation of EoS models including Redlich-Kwong, the Soave-modified Redlich-Kwong, and Peng-Robinson was performed for the prediction of carbon-dioxiderich mixtures for their use in simulation software. The compressibility of the CO 2 mixtures is computed through a three-step process: pseudo-critical properties calculation, mixture attraction-repulsion coefficient calculation, and solving of general EoS. The first step utilizes the linear mixing rule to predict the critical temperature and pressure of the mixture. The second step extends the use of the selected EoS to mixtures; the prediction of cubic equation parameters a, b; and ∝ employs Van der Waals mixing using the VW mix script, where the binary interaction coefficients are inputted from the Aspen HYSYS database, whereas for the three components in the Peng-Robinson EoS, the works by [21] are incorporated for the improved modeling of the VLE behavior. The third step uses the MATLAB symbolic toolbox. The compressibility is solved numerically for V m of the general EoS (in Equation (4)) by using the appropriate variables calculated using Van der Waals mixing for their respective equations: RK, SRK, and PR.
Using the "if" condition to compute the real roots, we classify whether they are in a single-phase region or classify as liquid or vapor by specifying minimum and maximum roots respectively. The molar density may be computed by the reciprocal of the molar volume. The outputs of this function are the compressibility and molar density of the liquid, gas, and single-phase systems.

Machine Learning
The development of machine learning tools for property prediction made use of Knime Analytics and MATLAB, the computation of phase behavior using Weka posed several limitations which in turn led to it not being included in this study due to the inability to have a systematic approach to prediction through a set architecture.
For the prediction of properties of CO 2 mixtures, the logic presented in Figure 1 shows the stages undertaken for the optimization and development of each generation of model envisaged through MATLAB. The sequence of events that lead to the advancement and success in a model lay within the pre-processing data stage, where the manipulation of input data (in this article, the experimental data from literature [22][23][24]) is to extract desired data points and remove irrelevant data, where this data is then used in the calculation of parameters; these techniques for data manipulation are altered where needed to improve predictive accuracy over the preceding model. The manipulated data is then collected for training in the neural network model alongside the specified selected number of layers, neurons per layer, and training algorithm; the trained model is then computed for the AARD.  Within the "Evaluation" stage when the AARD is computed for the neural network model, the model is contrasted against its predecessors for its accuracy and reliability, if it deviates this too much, the alteration of the pre-processing parameters (marked by blue stars below) such as data manipulation method, neuron-layers, and layers-per-neuron is changed to achieve the lowest AARD. Within the "Evaluation" stage when the AARD is computed for the neural network model, the model is contrasted against its predecessors for its accuracy and reliability, if it deviates this too much, the alteration of the pre-processing parameters (marked by blue stars below) such as data manipulation method, neuron-layers, and layers-per-neuron is changed to achieve the lowest AARD.
The approach to achieving the improved models with the enhanced data pre-processing technique is presented in the following sections. It is important to note that each model may require to be re-run, this is due to many solutions available from each unique model owed to the random seed of assigned weights in the initialization of the neural network model, which are then optimized around.

Generation 1 and 2-Knime Analytics
The first and second-generation for property prediction in this work made use of Knime Analytics for the prediction of the density for a single binary mixture at a temperature and pressure range of 273-423 K and 0.95-126.46 MPa, respectively. Generation 1 ( Figure 2) comprised of the prediction of density by training the MLP using temperature, pressure, and composition, however, it was found that around the phase boundary prediction issues arose. ( Figure 2) comprised of the prediction of density by training the MLP using temperature, pressure, and composition, however, it was found that around the phase boundary prediction issues arose. The phase prediction error is reduced in Generation 2 ( Figure 3) where data input is partitioned by phase into their respective categories. This method required more time to compute the data due to the increase in the number of MLP for each phase and offers no solution to property prediction.  The third-generation model developed is aimed at predicting viscosity along with The phase prediction error is reduced in Generation 2 ( Figure 3) where data input is partitioned by phase into their respective categories. This method required more time to compute the data due to the increase in the number of MLP for each phase and offers no solution to property prediction. ( Figure 2) comprised of the prediction of density by training the MLP using temperature, pressure, and composition, however, it was found that around the phase boundary prediction issues arose. The phase prediction error is reduced in Generation 2 ( Figure 3) where data input is partitioned by phase into their respective categories. This method required more time to compute the data due to the increase in the number of MLP for each phase and offers no solution to property prediction.  The third-generation model developed is aimed at predicting viscosity along with density ( Figure 4), however, an improvement was made by using the predicted density as

Generation 3 and 4-Initial MATLAB Models for Predicting Density and Viscosity
The third-generation model developed is aimed at predicting viscosity along with density ( Figure 4), however, an improvement was made by using the predicted density as an additional input using an ANN in series (see Figure 5).

Generation 3 and 4-Initial MATLAB Models for Predicting Density and Viscosity
Generation 5 ( Figure 6) is aimed at improving upon Generation 2, with the prediction of density close to the phase boundary. The prediction of phase uses the MATLAB classification learner through support vector machines (SVM) for the separation of two classes of data through a hyperplane, or a soft margin when necessary. The training set used for training the classifier module defined the phases: gas as "1", liquid as "2", and supercritical as "3"; exporting this model for the prediction phase from the input temperature and pressure for pre-processing and then training the ANN to predict density (see Figure 6).

Generation 3 and 4-Initial MATLAB Models for Predicting Density and Viscosity
Generation 5 ( Figure 6) is aimed at improving upon Generation 2, with the prediction of density close to the phase boundary. The prediction of phase uses the MATLAB classification learner through support vector machines (SVM) for the separation of two classes of data through a hyperplane, or a soft margin when necessary. The training set used for training the classifier module defined the phases: gas as "1", liquid as "2", and supercritical as "3"; exporting this model for the prediction phase from the input temperature and pressure for pre-processing and then training the ANN to predict density (see Figure 6).

Generation 3 and 4-Initial MATLAB Models for Predicting Density and Viscosity
Generation 5 ( Figure 6) is aimed at improving upon Generation 2, with the prediction of density close to the phase boundary. The prediction of phase uses the MATLAB classification learner through support vector machines (SVM) for the separation of two classes of data through a hyperplane, or a soft margin when necessary. The training set used for training the classifier module defined the phases: gas as "1", liquid as "2", and supercritical as "3"; exporting this model for the prediction phase from the input temperature and pressure for pre-processing and then training the ANN to predict density (see Figure 6).

Generation 1 and 2 Multi-Layered Perceptron (Knime Analytics)
Knime Analytics was used for the prediction of density. Table 6 (temperature between 6.71 K and 77.94 K) and Table 7 (temperature between 273.4 K and 323.4 K) show the input, output, and the deviation of the predicted densities from the experimental value. AARDs (%) of 22.27% and 5.95% were obtained for temperatures between 6.71 K and 77.94 K and temperatures between 273.4 K and 323.4 K, respectively. There are two higher absolute relative deviations (ARDs) in Table 6, which make the average higher for the temperature between 6.71 K and 77.94 K. Generally, this method is quite appropriate for the prediction of a mixture density, especially at a higher temperature.

Generation 3 and 4 Neural Network Prediction of Density and Viscosity (MATLAB)
Neural network was used for the prediction of viscosity of a single-phase binary mixture. Table 8 shows an improved prediction of viscosity with lower minimum, maximum, and AARD.

Generation 3 and 4 Neural Network Prediction of Density and Viscosity (MATLAB)
For the prediction of phase through SVM, investigating the accuracy of each method for the number of folds (where data is divided into portions using 1 for test validation and the rest for training repeating through each fold and taking an average error). Generally, it was found that cubic is more accurate than the other ones (Table 9).

Comparison of 3 and 5 Component Binary Mixtures for EoS and Generation 6 ANN with the Association of EoS (MATLAB)
For the computation of Generation 6 ANN, two layers with six neurons each with the Bayesian regularization backpropagation algorithm for the reduction in overfitting of data and consistency in results were used. The following subsections present a graphical comparison of compressibility for the computed EoS and ANN against the literature data for pure CO 2 , CO 2 in a binary mixture with N 2 and CH 4 . As can be seen, ANN has a lower AARD (%) of 2.81. After that, the PR is more accurate than the RK and SRK (see Table 10). Using the full dataset of CO 2 with N 2 , CH 4 , Ar, and O 2 , as there is no available data (for the selection of the number of components) for the binary interaction coefficients for the self-consistent equations, the Van der Waals mixing rule was computed using Aspen interaction parameters [20]. The outcome of this led to ANN being significantly more accurate than the earlier components (Table 11). Generally, in terms of machine learning tools, MATLAB is a better choice for model development than using Knime Analytics due to the increased customization of ANN setup, monitoring of performance, and integration into the property prediction framework. In the case of Weka, though the software is not used to generate results, the analytics package has the opportunity to export the design of neuron layers' respective bias and weight to incorporate the ANN into MATLAB and other code-based mathematical software. The major issue that arose with using MATLAB, which is not presented in Knime Analytics, is the associating of a "seed" to the neural network. This allows for a continually changing framework of the network through different seed converses of the neural network, however, this may be avoided by storing the state of the random number generator before the initialization of the network.
The complete generations of neural network models developed all have merit to property prediction, Generation 3 and 4 for viscosity prediction show a suitable degree of accuracy for prediction, the benefit of using density as an input though has provided a significant reduction in maximum error. In Generation 5, to combat the issue arising phase change, classifying the phase as an independent variable through the Classification Learner app, the best classification method using support vector machines is its cubic variant, this is expected as the phase characterization as a function of temperature and pressure is of cubic behavior.
The use of pseudo-critical properties such as reduced pressure, temperature, and EoS parameters in the Generation 6 model for ANN predictions is able to predict compressibility as a single output because of this, vapor liquid equilibria may not be solved in this version and will have to be undertaken in future works, though the main objective was to achieve an ANN that is able to predict compressibility more accurately than the other methods. However, with the five-component binary mixtures, like with EoS, the predictive accuracy decreased.
In this work, Bayesian regularization backpropagation had the lowest AARD, followed by Levenberg-Marquardt backpropagation. However, in published literature Bayesian regularization has not been considered.

Comparison of 3 and 5 Component Binary Mixtures for EoS and Generation 6 ANN with the Association of EoS (MATLAB)
The major issue arising from the modeling of data is the overfitting of the training data. Even with the use of the Bayesian regularization backpropagation training function, the data still overfitted the training data. This is shown in Figure 8 where the algorithm attempts to find the optimal solution, however, after 600 iterations the MSE of the training set increases significantly with near to no improvement of the actual training set. Though in cases where the number of neurons is lower, such as Figure 9, the optimum solution is found and is not limited by the maximum epoch to prevent overfitting. The current limitation of this work is the prediction of mixture critical properties, as such, a more accurate method for predicting mixture pseudo-critical properties is desired for improving prediction through a better mixing rule for pseudo-properties calculation, where currently linear mixing is used. The optimal ANN setup was found using the three-component mixture. It was obtained that the optimal number of neurons per layer is between two and eight, however, with too many neurons, the AARD deteriorates, overfitting the data and not being able to converge to find the optimal setup. Though the optimal number of layers showed that two neuron layers are the ideal setup for this case; however, it is unable to find optimum weight and bias values to achieve an appropriate fit for the data, unlike with one layer. In this work, it was found that the Bayesian regularization backpropagation has the lowest AARD. Figure 8 shows AARD% as a function of the number of neurons. From Figure 10, increasing from one layer to two layers led to the reduction of AARD% significantly. It may be because the outputs of the ANN model are not a linear function of the input. Moreover, a considerable change in AARD% was not seen with the increase of the number of layers from two to three. Therefore, two layers were selected as the optimal value.

Comparison of 3 and 5 Component Binary Mixtures for EoS and Generation 6 ANN with the Association of EoS (MATLAB)
The major issue arising from the modeling of data is the overfitting of the training data. Even with the use of the Bayesian regularization backpropagation training function, the data still overfitted the training data. This is shown in Figure 8 where the algorithm attempts to find the optimal solution, however, after 600 iterations the MSE of the training set increases significantly with near to no improvement of the actual training set. Though in cases where the number of neurons is lower, such as Figure 9, the optimum solution is found and is not limited by the maximum epoch to prevent overfitting. The current limitation of this work is the prediction of mixture critical properties, as such, a more accurate method for predicting mixture pseudo-critical properties is desired for improving prediction through a better mixing rule for pseudo-properties calculation, where currently linear mixing is used. The optimal ANN setup was found using the three-component mixture. It was obtained that the optimal number of neurons per layer is between two and eight, however, with too many neurons, the AARD deteriorates, overfitting the data and not being able to converge to find the optimal setup. Though the optimal number of layers showed that two neuron layers are the ideal setup for this case; however, it is unable to find optimum weight and bias values to achieve an appropriate fit for the data, unlike with one layer. In this work, it was found that the Bayesian regularization backpropagation has the lowest AARD. Figure 8 shows AARD% as a function of the number of neurons. From Figure 10, increasing from one layer to two layers led to the reduction of AARD% significantly. It may be because the outputs of the ANN model are not a linear function of the input. Moreover, a considerable change in AARD% was not seen with the increase of the number of layers from two to three. Therefore, two layers were selected as the optimal value.

Compressibility Factor
The gas compressibility factor or Z-factor is one of the most vital parameters in the industries such as chemical and petroleum. It should be pointed out that there is a direct relationship between the Z-factor and the density of a gas stream, its flow rate as well as isothermal compressibility [25]. It is challenging to obtain accurate Z-factor values for gas streams because of the fact that mostly there are other gases in the CO2-rich mixture gas and also the system is non-ideal [25]. It is common to use simple empirical correlations for the prediction of the Z-factor, however, the results are usually not accurate and there are calculation convergence difficulties [25]. A combination of different methods can be used to improve the compressibility factor prediction. The work done by Gaganis [26] combined the truncated regularized kernel ridge regression (TR-KRR) algorithm with a simple linear-quadratic interpolation scheme for estimation of the Z-factor. The maximum absolute relative prediction error is around the critical point was obtained around 2%. In the present study, the combination of machine learning with EoS was used for the prediction of the compressibility factor of pure CO2 and CO2-rich mixtures. The compressibility factor as a function of reduced pressure for the pure CO2 models for SRK and Peng-Robinson EoS and ANN at three different temperatures were shown in Figure 11. The results

Compressibility Factor
The gas compressibility factor or Z-factor is one of the most vital parameters in the industries such as chemical and petroleum. It should be pointed out that there is a direct relationship between the Z-factor and the density of a gas stream, its flow rate as well as isothermal compressibility [25]. It is challenging to obtain accurate Z-factor values for gas streams because of the fact that mostly there are other gases in the CO2-rich mixture gas and also the system is non-ideal [25]. It is common to use simple empirical correlations for the prediction of the Z-factor, however, the results are usually not accurate and there are calculation convergence difficulties [25]. A combination of different methods can be used to improve the compressibility factor prediction. The work done by Gaganis [26] combined the truncated regularized kernel ridge regression (TR-KRR) algorithm with a simple linear-quadratic interpolation scheme for estimation of the Z-factor. The maximum absolute relative prediction error is around the critical point was obtained around 2%. In the present study, the combination of machine learning with EoS was used for the prediction of the compressibility factor of pure CO2 and CO2-rich mixtures. The compressibility factor as a function of reduced pressure for the pure CO2 models for SRK and Peng-Robinson EoS and ANN at three different temperatures were shown in Figure 11. The results

Compressibility Factor
The gas compressibility factor or Z-factor is one of the most vital parameters in the industries such as chemical and petroleum. It should be pointed out that there is a direct relationship between the Z-factor and the density of a gas stream, its flow rate as well as isothermal compressibility [25]. It is challenging to obtain accurate Z-factor values for gas streams because of the fact that mostly there are other gases in the CO 2 -rich mixture gas and also the system is non-ideal [25]. It is common to use simple empirical correlations for the prediction of the Z-factor, however, the results are usually not accurate and there are calculation convergence difficulties [25]. A combination of different methods can be used to improve the compressibility factor prediction. The work done by Gaganis [26] combined the truncated regularized kernel ridge regression (TR-KRR) algorithm with a simple linear-quadratic interpolation scheme for estimation of the Z-factor. The maximum absolute relative prediction error is around the critical point was obtained around 2%. In the present study, the combination of machine learning with EoS was used for the prediction of the compressibility factor of pure CO 2 and CO 2 -rich mixtures. The compressibility factor as a function of reduced pressure for the pure CO 2 models for SRK and Peng-Robinson EoS and ANN at three different temperatures were shown in Figure 11. The results showed high accuracy when dealing with the prediction of compressibility factor, however, this was due to the more ideal state of the system. It was assumed that the system is pure CO 2 . showed high accuracy when dealing with the prediction of compressibility factor, however, this was due to the more ideal state of the system. It was assumed that the system is pure CO2. Figure 11. Pure CO2 compressibility chart for all temperatures.
Compressibility as a function of reduced pressure for a binary mixture containing N2 (mole fraction = 0.0442) at three different temperatures is shown in Figure 12. For a binary mixture with the low mole fraction of nitrogen (N2 mole fraction = 0.0442). As it can be seen the SRK, PR, and ANN can predict the compressibility factor relatively accurately at the points far from the critical points. However, the accuracy of the ANN model is higher at the points around critical points. It can be seen as higher deviations for all three EoS at the critical points.  Compressibility as a function of reduced pressure for a binary mixture containing N 2 (mole fraction = 0.0442) at three different temperatures is shown in Figure 12. For a binary mixture with the low mole fraction of nitrogen (N 2 mole fraction = 0.0442). As it can be seen the SRK, PR, and ANN can predict the compressibility factor relatively accurately at the points far from the critical points. However, the accuracy of the ANN model is higher at the points around critical points. It can be seen as higher deviations for all three EoS at the critical points. showed high accuracy when dealing with the prediction of compressibility factor, however, this was due to the more ideal state of the system. It was assumed that the system is pure CO2. Figure 11. Pure CO2 compressibility chart for all temperatures.
Compressibility as a function of reduced pressure for a binary mixture containing N2 (mole fraction = 0.0442) at three different temperatures is shown in Figure 12. For a binary mixture with the low mole fraction of nitrogen (N2 mole fraction = 0.0442). As it can be seen the SRK, PR, and ANN can predict the compressibility factor relatively accurately at the points far from the critical points. However, the accuracy of the ANN model is higher at the points around critical points. It can be seen as higher deviations for all three EoS at the critical points.  The nitrogen content in the gas mixture was increased to evaluate the effect of the higher N 2 content of the gas mixture. The compressibility factor as a function of reduced pressure for a binary mixture containing N 2 (mole fraction = 0.0442 and 0.1442) at three different temperatures is shown in Figures 13 and 14. It was found that there is a lower deviation from experimental data at a lower temperature in the presence of a higher N 2 mole fraction (0.1442) for all three EoS and ANN was higher in the gas mixture. Interestingly, it obtained fairly accurate results even for the RK EoS at lower temperature and higher N 2 content in comparison with pure CO 2 ( Figure 11) and lower N 2 content ( Figure 12) but significant deviation was found for the RK EoS at temperatures of 283 K and 293 K. The deviation from experimental data was increased for the SRK and PR EoS, when temperature increased to 283 K and 293 K, at the points around the critical point. However, there was good agreement between the ANN model values and the experimental data for all three temperatures. The nitrogen content in the gas mixture was increased to evaluate the effect of the higher N2 content of the gas mixture. The compressibility factor as a function of reduced pressure for a binary mixture containing N2 (mole fraction = 0.0442 and 0.1442) at three different temperatures is shown in Figure 13 and Figure 14. It was found that there is a lower deviation from experimental data at a lower temperature in the presence of a higher N2 mole fraction (0.1442) for all three EoS and ANN was higher in the gas mixture. Interestingly, it obtained fairly accurate results even for the RK EoS at lower temperature and higher N2 content in comparison with pure CO2 (Figure 11) and lower N2 content ( Figure  12) but significant deviation was found for the RK EoS at temperatures of 283 K and 293 K. The deviation from experimental data was increased for the SRK and PR EoS, when temperature increased to 283 K and 293 K, at the points around the critical point. However, there was good agreement between the ANN model values and the experimental data for all three temperatures.   The nitrogen content in the gas mixture was increased to evaluate the effect of the higher N2 content of the gas mixture. The compressibility factor as a function of reduced pressure for a binary mixture containing N2 (mole fraction = 0.0442 and 0.1442) at three different temperatures is shown in Figure 13 and Figure 14. It was found that there is a lower deviation from experimental data at a lower temperature in the presence of a higher N2 mole fraction (0.1442) for all three EoS and ANN was higher in the gas mixture. Interestingly, it obtained fairly accurate results even for the RK EoS at lower temperature and higher N2 content in comparison with pure CO2 (Figure 11) and lower N2 content ( Figure  12) but significant deviation was found for the RK EoS at temperatures of 283 K and 293 K. The deviation from experimental data was increased for the SRK and PR EoS, when temperature increased to 283 K and 293 K, at the points around the critical point. However, there was good agreement between the ANN model values and the experimental data for all three temperatures.

Conclusions
The objective of this work was to select an appropriate model for the prediction of thermodynamic properties for CO 2 -rich mixtures. This is clearly of high value to the simulation for the future of process design in CCS and adsorption kinetic in gaswell application. Out of the evaluated EoS that is used in Aspen HSYS, Peng-Robinson more accurately represents CO 2 mixtures. Though there are more accurate EoS developed in the literature, these works are not yet available to incorporate into process simulation using Aspen HYSYS.
The applicability of machine learning techniques applied from Big Data analytics to a broader range of problem-solving capabilities for chemical engineering such as empirical formulation has been successfully considered in this work and may further be used for process control. Prospects of this work aim to incorporate machine learning into process simulations software, where literature or experimental data may be used to accurately describe fluid properties for process design using a pre-set model that is capable of defining a fluid accurately without the need for empirical formulation or choosing a "close enough" approach to using a selection of preset fluid packages.
The importance of data manipulation has been successfully addressed in this article to provide a basis for manipulation of input data to enhance predictive accuracy by extending the desired outputs from being mol fractional inputs of individual components to being a function of pseudo-critical properties. This development may lead to an improved universal model to predict a complex range of mixtures thermodynamic property from their fundamental particle interactions and other underlying mixture interactions not yet explored in this article.
Having explored a range of training algorithms in this work, it has been found that using Bayesian regularization backpropagation has not been previously considered in the published literature [6], which is surprising due to its better classification of data with lower AARD in contrast to conventional lowest AARD, being Levenberg-Marquardt backpropagation. Perhaps this should be considered more throughout the development of ANN for predicting thermodynamic properties in future works.