Next Article in Journal
Shale Oil Occurrence Mechanisms: A Comprehensive Review of the Occurrence State, Occurrence Space, and Movability of Shale Oil
Next Article in Special Issue
Overpressure Generation and Evolution in Deep Longmaxi Formation Shale Reservoir in Southern Sichuan Basin: Influences on Pore Development
Previous Article in Journal
Co-Operative Optimization Framework for Energy Management Considering CVaR Assessment and Game Theory
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Prediction of TOC in Lishui–Jiaojiang Sag Using Geochemical Analysis, Well Logs, and Machine Learning

1
School of Energy Resources, China University of Geosciences, Beijing 100083, China
2
Key Laboratory of Marine Reservoir Evolution and Hydrocarbon Accumulation Mechanism, Ministry of Education, China University of Geosciences, Beijing 100083, China
*
Author to whom correspondence should be addressed.
Energies 2022, 15(24), 9480; https://doi.org/10.3390/en15249480
Submission received: 8 October 2022 / Revised: 9 December 2022 / Accepted: 12 December 2022 / Published: 14 December 2022
(This article belongs to the Special Issue Recent Advances in Shale Oil and Gas Reservoirs)

Abstract

:
Total organic carbon (TOC) is important geochemical data for evaluating the hydrocarbon generation potential of source rocks. TOC is commonly measured experimentally using cutting and core samples. The coring process and experimentation are always expensive and time-consuming. In this study, we evaluated the use of three machine learning (ML) models and two multiple regression models to predict TOC based on well logs. The well logs involved gamma rays (GR), deep resistivity (RT), density (DEN), acoustic waves (AC), and neutrons (CN). The ML models were developed based on random forest (RF), extreme learning machine (ELM), and back propagation neural network (BPNN). The source rock of Paleocene Yueguifeng Formation in Lishui–Jiaojiang Sag was taken as a case study. The number of TOC measurements used for training and testing were 50 and 27. All well logs and selected well logs (including AC, CN, and DEN) were used as inputs, respectively, for comparison. The performance of each model has been evaluated using different factors, including R2, MAE, MSE, and RMSE. The results suggest that using all well logs as input improved the TOC prediction accuracy, and the error was reduced by more than 30%. The accuracy comparison of ML and multiple regression models indicated the BPNN was the best, followed by RF and then multiple regression. The worst performance was observed in the ELM models. Considering the running time, the BPNN model has higher prediction accuracy but longer running time in small-sample regression prediction. The RF model can run faster while ensuring a certain prediction accuracy. This study confirmed the ability of ML models for estimating TOC using well logs data in the study area.

1. Introduction

Resource potential is a basic parameter for evaluating whether a hydrocarbon-bearing block has commercial exploitation value [1]. Evaluation of resource potential requires geochemical test data, which can directly reflect the hydrocarbon generation potential [2,3]. Total organic carbon (TOC) is an important parameter of geochemical data to evaluate resource potential. To obtain accurate TOC, research mainly relies on direct measurement of cutting and core samples. Due to the deep depth of the target development layer, the formation pressure is high, indicating that the difficulty and cost of cutting and core samples collection are generally high [4]. Meanwhile, direct measurement of core samples in the laboratory is expensive. As a result, the difficulty of direct TOC measurement affects exploration and development of oil and gas in deep layers.
Well logs can reflect the physical properties of the rocks in the target interval. Due to the different physical properties, such as gamma radioactivity, resistivity, and density between organic-rich rocks and other surrounding rocks, logging information can effectively distinguish source rocks. TOC prediction methods based on logging parameters are increasingly used. A variety of different TOC prediction methods and empirical formulas have been established by many researchers through fitting actual measured TOC with logging information. Beers [5] first established the relationship between gamma logging and TOC and calculated the TOC of the target interval. Schmoker [6] proposed a TOC prediction technology using density logging, and the technology has been applied to TOC prediction of shale until now. Mendelson and Toksoz [7] first used the multiple regression method to establish the relationship between multiple logging parameters and TOC with a correlation coefficient. Autric and Dumesnil [8] established a prediction relationship between acoustic transit time and TOC and found that physical properties of source rocks affected the prediction accuracy. The accuracy of empirical formulas to predict TOC cannot be guaranteed when applied in other study areas according to their study. Passey et al. [9] proposed the ΔlogR technique, which takes into account the changes in resistivity and porosity of source rocks before and after hydrocarbon generation and realizes effective unification of geological processes and TOC for the first time. However, the ΔlogR technique cannot be applied to TOC prediction of highly mature organic-rich rocks. Kamali and Mirshady [10] extended the ΔlogR technique by using the neuro-fuzzy method so that the technique can also be applied to TOC prediction of gas-producing source rocks. Passey et al. [11] extended the ΔlogR technique to include the TOC of highly mature organic-rich rocks in the prediction range. Hu et al. [12], Wang et al. [13], and Zhao et al. [14] calculated TOC using superposition between resistivity and neutron porosity logs. However, the huge differences in the maturity of source rocks and the background value of TOC content in different regions have a significant effect on the prediction results [13,14]. Since no linear or non-linear relationship occurs between TOC and various logging parameters, there are differences in petrol-physical properties in different study areas [15,16]. Many empirical formulas have different prediction performances in different research areas.
ML is a specific branch of artificial intelligence and is a method of empowering ML to do things that cannot be completed by direct programming [17]. ML is applied with great success in various industries, such as engineering [18,19,20,21,22,23], medicine [24,25,26,27,28,29], economy [30,31,32], and environmental and geospatial modeling [33,34,35]. ML has been applied to macroscopic features of target rocks, such as seismic facies classification [36,37,38,39,40] and logging lithofacies classification [41,42,43]. ML has demonstrated outstanding performance in these areas. An important aspect observed in recent research is that ML can learn and adapt to the dynamics of reservoir conditions, such as formation and depositional environment [44], while making use of geophysical data for lithology identification [45,46], porosity, and permeability [47,48,49,50,51,52,53]. With the development of ML from shallow learning to deep learning, deep learning has also been successfully applied to image-based, geoscience-related issues [54,55], such as seismic facies classification [40], lithology classification [56,57], mineral recognition [58], and carbonate rock recognition [59,60]. Current deep learning is a data-hungry technology. There are many problems in the real world, including in the medical field and oil and gas exploration, which are associated with an insufficient amount of labeled data [61]. The performance of deep learning in the above fields is not good [61]. In recent years, ML has been applied to TOC prediction, and the prediction results are ideal. TOC prediction mainly uses neural networks [16,44,62,63,64,65], extreme learning machine (ELM) [62,66,67], support vector machines [4,68], and decision trees [15,69,70]. Different machine learning algorithms have different operations and kernel functions. Therefore, the results obtained from different algorithms are quite different [63,71,72]. As a result, the choice of algorithm directly affects the accuracy of TOC prediction using ML.
The Lishui–Jiaojiang Sag is an important petroliferous basin in the East China Sea Basin [73,74]. Previous studies have shown that the Lishui–Jiaojiang Sag has excellent migration and storage capacity [75,76,77,78]. Basic research on source rocks is seriously lacking because the cost of core sampling and laboratory analysis of TOC is generally high. The ambiguity of the organic matter abundance in the source rocks severely restricts the oil and gas exploration process in the Lishui Sag. TOC prediction based on well logs is very important for Lishui–Jiaojiang Sag. In this paper, the source rock of Paleocene Yueguifeng Formation in the Lishui–Jiaojiang Sag is taken as the main research object. Three ML algorithms (random forest, extreme learning machine, and back propagation neural network) are used for predicting TOC (target variable). Two well combinations of logging parameters are used as input. This study aims to: (1) establish the multiple regression and ML models of TOC prediction; (2) compare the performance of ML and multiple regression models of TOC prediction; (3) compare the performance of models of TOC prediction using different logging parameters; and (4) determine the best TOC prediction method for the study area.

2. Geological Setting and Stratigraphy

The Lishui–Jiaojiang Sag is located in the western Taipei Depression (see Figure 1a), which is located in the East China Sea Shelf Basin; the sag faces the Yushan Uplift to the north, the Fuzhou Sag and the Minjiang Sag to the east, which is separated by the Yandang Uplift, and the Minjiang Uplift to the west and south (see Figure 1b) [73,75]. Tectonically, the sag lies at the convergence of the Eurasian, Pacific, and Philippine plates. NE-trending segmental faults control the morphology of the basin and have the characteristics of north–south blocks, east–west zoning, and an overall northeast–southwest distribution. The Lishui–Jiaojiang Sag is divided into different structural units (such as the Lishui Western Subsag, the Lingfeng Ridge, and the Lishui Eastern Subsag) by the Lingfeng low bulge [78,79].

2.1. Tectonic

The Lishui–Jiaojiang Sag is a Cenozoic single-fault sag superimposed on a residual Mesozoic basin [78,80]. From the end of the late Cretaceous to the early Tertiary, the Lishui–Jiaojiang Sag experienced four tectonic evolution phases, including a rift stage (late Cretaceous to Paleocene), a post-rift depression stage (early Eocene to late Eocene), an uplift stage (late Eocene to late Miocene), and a regional subsidence stage (late Miocene to Quaternary), which were affected by the changes in the subduction direction of the western Pacific plate and rifting [75,76].

2.2. Stratigraphy

The stratigraphic column of the Lishui–Jiaojiang Sag is divided into late Cretaceous, Paleogene, Neogene, and Quaternary strata from bottom to top. Oil and gas in Lishui Sag have been mainly discovered in Paleocene strata (see Figure 2) [75,78]. The main source rock is the lower Paleocene Yueguifeng Formation (E1y). The main object of this study is the lower Paleocene Yueguifeng (E1y) Formation [73,80]. The lithology of the Paleocene Yueguifeng Formation (E1y) is dominated by gray and dark gray mudstone and silty mudstone interspersed with thin layers of light gray calcium-bearing siltstone, fine sandstone, and a small number of thin layers of fine calcareous sandstone. The Lauguifeng Formation includes two sets of coarse–fine–coarse composite sedimentary cycles. The descending cycle is dominated by dark brown and dark brown mudstone mixed with light gray and off-white fine-medium-grained sandstone. The upper cycle consists of light gray, gray, dark gray, dark gray mudstone, and light gray fine-medium-grained sandstone nearly equal-thickness interbedded, sandwiched by thin layers of light gray, gray siltstone, and two layers of black coal. The overall deposition thickness is 125–400 m.

3. Materials & Methods

In order to test the accuracy of the log multivariate fitting model and the ML models in TOC prediction, only 77 source rock samples were collected in the Lishui–Jiaojiang Sag. These 77 samples were derived from three wells (well A, well B, and well C) in different areas of the Lishui–Jiaojiang Sag. All samples were derived from the lower Paleocene Yueguifeng Formation.

3.1. Geochemical Analysis

Each source rock sample was crushed to 80 meshes and acidified with dilute hydrochloric acid (HCl) to remove inorganic carbon. The TOC measurement was conducted by using the C/S analyzer CS-580A. The pretreatment of samples for pyrolysis analysis is the same as the TOC analysis. Rock-Eval pyrolysis analysis was conducted using a Rock-Eval-II. Before the Rock-Eval pyrolysis analysis, the Rock-Eval-II was calibrated with standard samples from the geochemical laboratory; samples were put into the Rock-Eval analyzer for testing. The measured parameters include TOC, S1, S2, and maximum pyrolysis yield temperature (Tmax). Hydrogen index (HI) was calculated by TOC and S2.

3.2. Log Series Selection and Well Log Models

ML is the process of inferring correlations by continuously learning from input data, which can influence the accuracy of the model. In general, the more feature parameters are supplied, the more thoroughly the objectives are characterized and more accurately the predictions are obtained. If the selected feature parameters are weakly correlated with the target, the system may learn the wrong functional relationship. The prediction may have a larger deviation. Therefore, before performing ML prediction model, the feature parameters should be screened to ensure that these parameters can more accurately describe the characteristics of the target.
Numerous studies have been conducted to clarify the connection between logging settings and TOC. Studies have shown that organic matter has varying degrees of linear correlation with gamma rays (GR), deep resistivity (RT), density (DEN), acoustic waves (AC), and neutrons (CN) [5,6,8,9].
GR logging could record gamma radioactivity during rock formations. The higher the reflectivity of the rock formation, the stronger the GR logging response [5]. Strongly radioactive organic-rich mudstone is abundant in radioactive thorium, potassium, and uranium elements, meaning that the organic-rich mudstones generally have high GR values.
Density logs could reflect the volume density of the formation based on the assumed formation density and drilling fluid density [6,8]. The density of sand and mudstone is generally 2~2.7 g/cm3, and the density of organic-rich stratum is usually lower than that of surrounding rocks. The density of solid organic matter is close to that of water (around 1.0 g/cm3). Organic-rich stratum will produce organic matter pores after hydrocarbon generation and expulsion, and the density will be further reduced. Therefore, density logs can be used to indirectly estimate TOC concentrations.
Neutron logs may identify the degree of scattering of exposed neutrons in the formation, which, afterwards, reflects the formation’s porosity. Organic matter is directly related to the hydrogen atoms and porosity of the rock [14]. Therefore, when the TOC of stratum increases, the values of CN increase. AC logs could measure the response of the rock mass near the borehole to the artificial elastic wave field to achieve the purpose of detecting the properties of the target formation (porosity, longitudinal and transverse wave velocity, etc.).
AC logging is mainly a logging tool for determining lithology, porosity, and fluid type [10,12]. Previous studies have shown that the acoustic transmission time of organic-rich source rocks is significantly shorter. Resistivity logs record the fluid resistivity in the formation and then calculate the true resistivity of the formation [9,11]. The response of TOC is not sensitive to resistivity logs. However, after the mature source rocks generation and expelling hydrocarbons, the resistivity will increase significantly due to the presence of hydrocarbon compounds. Therefore, the TOC content of the source rock section can be indirectly reflected according to the change in the relative resistivity of the source rock section.
Different study areas have different geological conditions, and the relationship between logging parameters and TOC is also different. Figure 3 shows the visual relationship between the logging parameters and the measured TOC. It shows that, as AC, CN, GE, and RT fluctuate positively, TOC also fluctuates in a corresponding positive direction. When DEN fluctuates negatively, TOC also fluctuates positively. This is consistent with the physical principles of logging. However, the visualization law cannot show a quantitative correlation between the individual logging parameters and the TOC. A simple linear regression technique was used to determine the sensitivity of TOC to log data. During this process, a key metric (correction factor) labeled R2 was used to investigate the effect of different logs on the true TOC value. R2 represents the proportion of population variance explained by the model.
Figure 3 shows that the TOC of the Yueguifeng Formation source rock has a better response in DT and CN logging. With the increase in TOC, the logging response increases, showing an obvious positive correlation, with R2 of 0.45 and 0.60, respectively. DEN logging was negatively correlated with TOC to a certain extent, with an R2 of 0.24. However, there was no obvious correlation between TOC and GR and RT. In addition, the Pearson correlation coefficient was used to analyze the sensitivity of TOC to logging data. Table 1 shows that the measured TOC has a good correlation with AC, CN, and DEN, and the R values of the measured TOC value and AC, CN, and DEN are 0.67, 0.77, and −0.49, respectively. There was no obvious correlation between the measured TOC and GR and RT. The linear regression analysis results are basically consistent with the Pearson correlation coefficient analysis results. The sensitivity of the corresponding log and the comprehensiveness of the feature description will affect the performance of the model. Therefore, two feature parameter sets, HXC (including all well logs) and HX (including AC, CN, and DEN), were selected. HXC has high comprehensiveness feature description of TOC. HX and TOC have a high correlation. The accuracy of prediction results with different feature parameter sets has been compared to obtain the optimal combination as input.
It has been pointed out that, if there are two or more uncorrelated independent variables that have a good correlation with the dependent variable, the results of multiple regression are usually better than the results of university analysis, and the multiple regression equation can be based on the correlation of each dataset [7]. Matrix calculation is obtained. The supposed multivariate prediction model is as follows:
TOC = A × AC   +   B × CN   +   C × GR   +   D × RT   +   E DEN +   F
A, B, C, D, E, F, are all constants. Entering the combination of characteristic parameters and TOC into ORIGIN and using regression statistics to obtain two multivariate fitting equations as follows:
TOC = 0.004 × AC   + 15.703 × CN 4.13113 DEN + 1.7546
TOC = 0.03826 × AC   + 12.48936 × CNL   + 0.00327 × GR   + 0.0473 × LLD 7.18502 DEN + 1.85181

3.3. ML Methods

In this study, three ML algorithms were used to predict the TOC of source rocks in the Lishui–Jiaojiang Sag, and 77 log parameter combinations and corresponding measured TOC values were used as input dataset and test set to build and test the prediction model. Three ML algorithms include extreme learning machine (ELM), back propagation neural network (BPNN), random forest (RF).

3.3.1. Method of BPNN

BPNN is the most basic neural network, and its output results are propagated forward, and the error is propagated by back propagation [81]. BPNN models can have one or more hidden layers [82,83]. Taking two hidden layers as an example, the forward propagation operation process is roughly as follows: the input value is xn = i1 = O1, and the output value is y. After linear combination with the weight matrix in the middle, the input i2 of the hidden layer 1 is obtained, and, in the hidden layer 1, the input i2 is activated with the activation function to obtain the output O2, and then it can be known that i2 = W1∙O1. Let f(x) is the activation function, O2 = f(i2). Then, similarly, the weight from hidden layer 1 to 2 is W2, the input of hidden layer 2 is i3 = W2∙O2, and the output is O3 = f(i3). The output y = W3∙O3 obtained from the hidden layer 2 to the output layer can also be written as y = f(w3f(w2f(w1o1))) [10,81,83,84,85]. At this time, the loss function will be used to further process y. Generally, our loss function takes O4 = 1/2(L−y)2, where L is the label of the training set [86]. If the desired output cannot be obtained in the output layer, then turn to back propagation. The error signal is returned along the original connection path, and the weight of each neuron is modified to minimize the error signal [10,44,84]. The weight update of each neuron can be calculated by the formula shown in Equation (5).
Δ W 1 = ( y L ) ( ( W 2 T [ W 2 T f ( i 3 ) ] ) f ( i 2 ) ) O 1 T
Δ W 2 = ( y L ) f ( i 4 ) W 3 T f ( i 3 ) O 2 T
Δ W 3 = ( y L ) f ( i 4 ) O 3 T
Finally, the update of the weight w can be completed by bringing the sample into the continuous forward. BPNN is widely used in pattern recognition, classification, data mining, and other disciplines [10,26,87,88]. The study found that BPNN has excellent performance in nonlinear mapping, and the flexible network structure ensures the accuracy of model prediction [86,89]. However, BPNN has a slow learning speed and is prone to fall into local minima [62,85].

3.3.2. Method of ELM

Extreme learning machine (ELM) was first proposed by Huang Guangbin in 2004 [90]. Extreme learning machine is to improve the back propagation algorithm (backward propagation, BP) to improve learning efficiency and simplify the setting of learning parameters [91,92]. The back propagation algorithm uses the gradient algorithm (back propagation). ELM adopts random selection of input layer weights and hidden layer deviations, and output layer (see Figure 4) weights are calculated and analyzed according to Moore–Penrose generalized inverse matrix theory by minimizing the loss function of the training error term and the output layer weight norm [93] as follows:
y = i = 1 L β i G ( a i x + b i )
X is the input data, β is the output weight, G(x) is the feature map or activation function (activation function) [91,92,94]. The main operation steps of ELM are divided into three steps: (1) randomly assign node parameters; (2) calculate the output matrix of the hidden layer; (3) solve for output weights. The core of the ELM algorithm is to use the least squares method to solve the output weights to minimize the error function [90,92,95]. The weight value can be obtained by the following system of equations.
min β H β T
By solving the least squares problem, the hidden layers’ output weights are implemented as follows:
β = ( H T T ) 1 H T T
The model can be established by using the weight value and the hidden layer deviation and make a prediction of the target value. The study found that ELM can make predictions on small sample data with fewer feature parameters. It can maintain high prediction accuracy set generalization [65].

3.3.3. Method of RF

RF is an ensemble learning algorithm [96,97]. The understanding of random forest can be divided into “random” and “forest”: “random” refers to random sampling of data and random sampling of features, mainly reflected in Bagging; its full name is Bootstrap aggregation [96,98]. It is a kind of sampling method for replacement [99]. The “forest” refers to the use of multiple freely growing CART decision trees to form a “forest” [100]. The core of the RF is the regression tree (ntree: the default value is 500 trees) and the number of branches (mtry: the default value is 1/3 of the total number of features) [101]. The principle of RF operation (see Figure 5) is (1) based on the Bagging ensemble framework, sampling n training samples randomly and with replacement each time from the original samples, performing T rounds of sampling to obtain T training sets, (2) using the T training sets, train T CART decision trees independently and combine T CART decision trees into a strong learner. The core of a single CART decision tree lies in the recursive growth process. Assuming the input dataset Xc, according to the segmentation method of the p-th eigenvalue V as Sp, v, the input formula can be expressed as:
S p , v : X L = { x p > V | x X c } ; X R = { x p V | x X c }
where XL and XR are two subsets of the dichotomy [73,74]. During the learning process, the CART decision tree obtains a nested IFELESE logic chain through stepwise recursive bisection segmentation [73,75]. This logical chain is stored and displayed using a binary tree structure [78]. Then, for the prediction problem, the prediction variance is used to optimize the segmentation method [72,73,79]. The prediction variance (variance) can be written as:
vari ( X ) = Î X ( y ( x ) E ( y | x X ) ) 2 X
After fully recursive optimal growth, the entire input space is divided into L mutually exclusive subspaces. Inputting the sample x’ into the tree structure, its predicted value can be written as:
E ( y | x = x ) = i = 1 n y i × I ( x i X L ) i = 1 n I ( x i X L )
(3) In the end, Bagging will assemble all T CART decision tree prediction results in an equal weight manner to establish an integrated prediction result [96,102,103]. The resulting function can be written as:
H ( x ) = 1 T T = 1 T E m ( x )
To be consistent with the logging prediction model, two kinds of characteristic parameter combinations, HXC and HX, were selected. The data combination of 50 randomly selected well logging data and measured geochemical data were used as a training set for the ML algorithm to build a prediction model. The remaining 27 data combinations were used as the test set to test the performance of the three prediction models. The ML algorithm modeling and TOC prediction in this study were conducted using MATLAB2019 software.

3.4. Evaluation Criteria

In this study, R2, MAE, MSE, and RMSE have been selected to evaluate the performance of prediction models. The mean absolute error (MAE) reflects the average of the absolute errors between the predicted and observed values [69]. Mean squared error (MSE) is the most commonly used regression loss function and is calculated by summing the squares of the distances between the predicted and true values [15]. The root mean square error (RMSE) represents the sample standard deviation of the difference between the predicted and observed values (called residuals) [4]. The RMSE is used to describe the degree of dispersion of the samples.

4. Results and Discussion

4.1. TOC and Source Rock Study from Geochemical Analyses

The TOC concentrations of the study area showed high values, ranging from 0.2% to 3.34%, with an average value of 1.6% (Table 2). There are some differences among the wells. The TOC concentration of well B (Jiaojiang Depression) is the highest, ranging from 1.88% to 3.34%, with an average of 2.63%. The TOC concentration in well C (the Lishui Eastern Subsag) is lower, ranging from 0.2% to 2.89%, with an average of 1.57%. The TOC of well A (the Lishui Western Subsag) is the lowest, ranging from 0.35% to 1.58%, with an average of 0.74%. The TOC of well A (the Lishui Western Subsag) is the lowest, ranging from 0.35% to 1.58%, with an average of 0.74%. The hydrogen index of Yueguifeng showed high values, ranging from 7.14 mg HC/g TOC to 390.64 mg HC/g TOC, with an average of 100.94 mg HC/g TOC. The HI of well B (Jiaojiang Depression) is the highest, ranging from 158.51 mg HC/g TOC to 388.68 mg HC/g TOC, with an average of 323.67 mg HC/g TOC. The HI of well A (the Lishui Western Subsag) is lower, ranging from 34 to 247.5 mg HC/g TOC, with an average of 102.82 mg HC/g TOC. The HI of well C (the Lishui Eastern Subsag) is the lowest, ranging from 17.14 to 100 mg HC/g TOC, with an average of 56.01 mg HC/g TOC. The total distribution of hydrocarbon generation potential (S1 + S2) in Yueguifeng Formation is ranging from 0.05 mg HC/g TOC to 12.51 mg HC/g TOC. The S1 + S2 of well A (the Lishui Western Subsag) is low, ranging from 3.2 mg HC/g TOC to 12.51 mg HC/g TOC, with an average of 8.96 mg HC/g TOC. The S1 + S2 of well B (Jiaojiang Sag) is the highest, ranging from mg HC/g TOC to 3.34 mg HC/g TOC, with an average of 8.96 mg HC/g TOC. The S1 + S2 of well C (the Lishui Eastern Subsag) is next, ranging from 0.05 mg HC/g TOC to 3.98 mg HC/g TOC, with an average of 1.45 mg HC/g TOC (see Table 2).
The type of organic matter has been classified according to the van Krevelen plot of whole rock hydrogen index (S2/TOC × 100) and TMAX shown in Figure 6. The intersection diagram shows that the source rock kerogen of the Yueguifeng Formation is dominated by type II, with a small amount of type I and type III kerogen. Type II kerogen is dominated by type II2, and type II2 kerogen accounts for only about 15%. The type of kerogen in well A is dominated by type II2, with a small amount of type I and type III kerogen. The type of kerogen in well B is dominated by type II1, with a small amount of type II2 kerogen. The kerogen type of well C is dominated by type II2, with a small amount of type I and type III kerogen. At the same time, the high degree of thermal evolution of the source rock in well C results in a low hydrogen index. The distribution of each well reflects the heterogeneity of kerogen. Based on the above geochemical evaluation, the source rocks of the Yueguifeng Formation have objective hydrocarbon generation potential. The source rocks mainly generate oil with a small amount of gas.

4.2. TOC Quantification from Multiple Regression Models

After using two logging multiple regression models to predict the TOC of the Yueguifeng Formation source rock in the Lishui–Jiaojiang Sag (see Table 3), R2, MAE, MSE, and RMSE have been calculated to evaluate the performance of the two models (see Table 4). The calculation results show that the R2 of the multiple regression model using the HXC is 0.63 (see Figure 7). The values of MAE, MSE, and RMSE are 0.39, 0.24, and 0.49, respectively. The R2 of the multiple regression model using the HX is 0.60. The values of MAE, MSE, and RMSE are 0.41, 0.26, and 0.51, respectively. The R2 of the multiple regression model using the HX is 0.60 (see Figure 7). The values of MAE, MSE, and RMSE are 0.41, 0.26, and 0.51, respectively. The R2 of the multiple regression model using the HXC is higher than the R2 of the multiple regression model using the HX. Error values calculated by MAE, MSE, and RMSE also reflect the lower prediction error of the multiple regression model using the HXC. The analysis shows that the accuracy of the multiple regression model using the HXC is significantly higher than that of the multiple regression model using the HX, but the performance is limited, which is also consistent with previous analysis. The relationship between TOC and various logging parameters is not a simple linear or nonlinear relationship, and the internal functional relationship is complex. Simple linear correlation analysis does not confirm whether low correlation parameters are helpful for establishment of predictive models.

4.3. TOC Quantification Using ML Methods

After using three ML test models to predict the TOC of the Yueguifeng Formation source rock in the Lishui–Jiaojiang Sag (see Table 5). R2, MAE, MSE, and RMSE were calculated to evaluate the performance of the models(see Table 4). In terms of the RF model using the HXC, the R2 of the training result is 0.97 (Figure 8). The MAE, MSE, and RMSE of the training results are 0.03, 0.02, and 0.15 (Table 4), respectively. The R2 of the test results is 0.65 (Figure 8). The MAE, MSE, and RMSE of the test results are 0.39, 0.26, and 0.61 (Table 5), respectively. In terms of the RF model using the HX (Figure 9), the R2 of the training result is 0.97. The MAE, MSE, and RMSE of the training results are 0.03, 0.02, and 0.15 (Table 5), respectively. The R2 of the test results is 0.49 (Figure 9). The MAE, MSE, and RMSE of the test results are 0.56, 0.45, and 0.67 (Table 5), respectively.
In terms of the BP model using the HXC, the R2 of the training result is 0.91 (Figure 8). The MAE, MSE, and RMSE of the training results are 0.22, 0.07, and 0.26 (Table 5), respectively. The R2 of the test results is 0.70 (Figure 8). The MAE, MSE, and RMSE of the test results are 0.37, 0.24, and 0.48 (Table 5), respectively. In terms of BP model using the HX, the R2 of the training result is 0.84 (Figure 9). The MAE, MSE, and RMSE of the training results are 0.26, 0.26, and 0.51 (Table 5), respectively. The R2 of the test results is 0.53 (Figure 9). The MAE, MSE, and RMSE of the test results are 0.47, 0.43, and 0.66 (Table 5), respectively.
In terms of the ELM model using the HXC, the R2 of the training result is 0.88 (Figure 8). The MAE, MSE, and RMSE of the training results are 0.23, 0.10, and 0.31 (Table 5), respectively. The R2 of the test results is 0.65 (Figure 8). The MAE, MSE, and RMSE of the test results are 0.42, 0.27, and 0.52 (Table 5), respectively. In terms of the ELM model using the HX, the R2 of the training result is 0.82 (Figure 9). The MAE, MSE, and RMSE of the training results are 0.28, 0.12, and 0.35 (Table 5), respectively. The R2 of the test results is 0.57 (Figure 9). The MAE, MSE, and RMSE of the test results are 0.45, 0.33, and 0.57 (Table 5), respectively.
The error comparison analysis of the training data and test data obtained using three different ML methods has been carried out. It shows that the R2 of the training results of all ML methods is greater than 0.8, regardless of which combination was used. The values of MAE, MSE, and RMSE are ideal. It indicates that the models using the three ML methods have strong stability and are very accurate for re-prediction of input data. However, the three ML methods performed differently in terms of the accuracy of the test results. The prediction accuracy of the ELM model using the HX is the highest, followed by the BPNN model and is the lowest in the RF model. This may be because ELM is different from the back propagation algorithm of the neural network, and the output layer weights are calculated by generalized inverse matrix theory. This allows ELM to ensure a certain prediction accuracy with fewer training parameters. The core of the RF algorithm is random sampling with replacement. The RF model has a poor learning effect on the target because of the lesser feature input using the HX. It leads to the lowest prediction accuracy of the RF model. The BPNN model using the HXC has the highest prediction accuracy, and the ELM and RF perform about the same. This verifies that the difference in the core algorithm affects the accuracy of the prediction again. At the same time, the dimension and nature of the input data and the matching degree of the algorithm will also affect the prediction accuracy. Compared with ELM and RF, BPNN has a simpler structure and simpler operation. This advantage makes BPNN potentially more accurate when dealing with low-latitude data. It also shows that advanced ML algorithms do not always perform better.
The accuracy of ML models using two different parameter combinations has been compared. The average R2 of the ML model using the HXC is 0.67. The average MAE, MSE, and RMSE are 0.4, 0.25, and 0.5, respectively. The average R2 of the ML models using the HX is 0.53. The average MAE, MSE, and RMSE are 0.49, 0.40, and 0.63, respectively. It shows that the accuracy of different ML models using HXC is significantly higher than those using HX, which is related to the completeness of feature information. The HX has the highest correlation with the expected value but provides no sufficient feature information, resulting in a weakening of the information transfer process of the conditional entropy and a decrease in the prediction accuracy. Although the two additional logging parameters are not highly correlated with the expected value, they supplement more characteristic information of the expected value and improve the prediction performance. Therefore, more parameter inputs that can describe the characteristics of the target value are beneficial to the prediction of the target value.
In terms of running time, the RF model has the shortest running time, followed by the ELM model, and the BPNN model has the longest running time (see Figure 10). The running time of the RF model using HXC parameters is 0.24 s. The running time of the RF model using HX parameters is 0.288 s. The running time of the ELM model using HXC parameters is 1.031 s. The running time of the ELM model using HX parameters is 1.022 s. The running time of the BP model using HXC parameters is 2.601 s. The running time of the BP model using the HX parameter is 2.505 s. The ELM model and BP model running time display fewer feature parameters that have faster running times. The less feature parameter input, the less model computation time, but the predictions will be even worse, while the RF model run time shows different characteristics: the more feature parameters, the faster the run time. This may be due to the fact that random sampling of samples is more complicated because there are fewer feature parameters. In the final integration of the prediction results, the RF model using the HX parameter requires more computation time. Considering the prediction accuracy and running time, the BPNN model has higher prediction accuracy but longer running time in small-sample regression prediction. The RF model can run faster while ensuring a certain prediction accuracy. According to the run time comparison, the RF model is much faster than the BPNN model, which indicates that RF should be chosen as the better option if processing speed is important.
This study also found that the prediction accuracy of the training set was significantly higher than that of the test set. All models exhibit varying degrees of overfitting. Possible reasons are: (1) the number of training samples is insufficient, resulting in a gap in the feature extraction and distribution simulation of the expected value by each ML algorithm; (2) since the purpose of this research is to test the universality of different regions, there are a great deal of well logging data. There must be uncontrollable human error, resulting in low prediction accuracy. However, in general, the correlation between the measured value and the predicted value of the BPNN test set still reaches 70%, and it still has a good application prospect. However, the generalized TOC prediction model in the study area seems to need more data supplementation and in-depth research. We believe that the prediction model will perform better with more data.

4.4. Compare between Multiple Regression and ML

The predicted values of each model using the better parameters (HXC) were compared. Figure 11 shows that the ML prediction accuracy is significantly higher than the multiple regression model. The results suggest that the prediction accuracy of the BPNN model is 10% higher than that of the multiple regression model. It also indicates the importance of feature parameters. When there are enough feature parameters, the more comprehensive the description of the target value, the better the prediction performance of the ML model.

5. Conclusions

In this study, the authors established ML models and multiple regression models for TOC prediction with logging parameters and made a comparison of the prediction effect of each model. The datasets including all available well logs and TOC core measurements conducted in Lishui–Jiaojiang Sag served as necessary data for constructing the models. The ML models were developed based on random forest (RF), extreme learning machine (ELM), and back propagation neural network (BPNN). The performance of each model has been evaluated using different factors, including R2, MAE, MSE, and RMSE. Based on the above results, the following conclusions can be drawn:
  • The TOC content of the source rocks in Yueguifeng Formation is relatively high, with an overall distribution of 0.2–3.34%. The S1 + S2 is generally distributed in the range of 0.5 mg HC/g TOC to 12.51 mg HC/g TOC. The type of kerogen is mainly type II. The source rocks of Yueguifeng Formation have good hydrocarbon generation potential.
  • The correlations between each logging parameter and TOC were evaluated through linear regression method and Pearson correlation coefficient analysis. The results indicate that the TOC of Yueguifeng Formation source rock has a better response in DEN, DT, and CN logging. The performance of each model using all well logs and selected well logs shows that each model with all well logs as input performed much better than the models with selected well logs.
  • In terms of accuracy, the results of error analysis show that each ML model with all well logs as input performed much better than the multiple regression models. In addition, it can be concluded that the BPNN model outperforms the other ML models. According to the run time comparison, the RF model is much faster than the BPNN model, which indicates that RF should be chosen as the better option if processing speed is important. This study confirmed the ability of ML models for building an efficient model for estimating TOC from readily available borehole logs data in the study area.

Author Contributions

X.H.: Investigation, Methodology, Writing—review and editing; D.H.: Conceptualization, Methodology, Supervision, Writing—review and editing; X.C. and C.N.: Investigation; Y.L. and S.C.: software. All authors have read and agreed to the published version of the manuscript.

Funding

The National Natural Science Foundation of China under contract No. 41872131.

Data Availability Statement

Not applicable.

Acknowledgments

We would like to thank the CNOOC Shanghai Company for providing the core samples used in this work.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Bechtel, A.; Movsumova, U.; Strobl, S.A.; Sachsenhofer, R.F.; Soliman, A.; Gratzer, R.; Püttmann, W. Organofacies and paleoenvironment of the Oligocene Maikop series of Angeharan (eastern Azerbaijan). Org. Geochem. 2013, 56, 51–67. [Google Scholar] [CrossRef]
  2. Hakimi, M.H.; Abdullah, W.H.; Alqudah, M.; Makeen, Y.M.; Mustapha, K.A. Organic geochemical and petrographic characteristics of the oil shales in the Lajjun area, Central Jordan: Origin of organic matter input and preservation conditions. Fuel 2016, 181, 34–45. [Google Scholar] [CrossRef]
  3. Hakimi, M.H.; Abdullah, W.H.; Makeen, Y.M.; Saeed, S.A.; Al-Hakame, H.; Al-Moliki, T.; Al-Sharabi, K.Q.; Hatem, B.A. Geochemical characterization of the Jurassic Amran deposits from Sharab area (SW Yemen): Origin of organic matter, paleoenvironmental and paleoclimate conditions during deposition. J. Afr. Earth. Sci. 2017, 129, 579–595. [Google Scholar] [CrossRef]
  4. Tan, M.J.; Song, X.D.; Yang, X.; Wu, Q.Z. Support-vector-regression machine technology for total organic carbon content prediction from wireline logs in organic shale: A comparative study. J. Nat. Gas Sci. Eng. 2015, 26, 792–802. [Google Scholar] [CrossRef]
  5. Beers, R.F. Radioactivity and Organic Content of Some Paleozoic Shales1. AAPG Bull. 1945, 29, 1–22. [Google Scholar]
  6. Schmoker, J.W.; Hester, T.C. Organic Carbon in Bakken Formation, United States Portion of Williston Basin1. AAPG Bull. 1983, 67, 2165–2174. [Google Scholar]
  7. Mendelzon, J.D.; Toksoz, M.N. Source Rock Characterization Using Multivariate Analysis of Log Data. In Proceedings of the SPWLA 26th Annual Logging Symposium, Dallas, TX, USA, 17 June 1985; p. SPWLA-1985-UU. [Google Scholar]
  8. Autric, A. Resistivity, Radioactivity And Sonic Transit Time Logs To Evaluate The Organic Content Of Low Permeability Rocks. Log Anal. 1985, 26, SPWLA-1985-vXXVIn3a3. [Google Scholar]
  9. Passey, Q.R.; Creaney, S.; Kulla, J.B.; Moretti, F.J.; Stroud, J.D. A Practical Model for Organic Richness from Porosity and Resistivity Logs1. AAPG Bull. 1990, 74, 1777–1794. [Google Scholar]
  10. Kamali, M.R.; Allah Mirshady, A. Total organic carbon content determined from well logs using ΔLogR and Neuro Fuzzy techniques. J. Pet. Sci. Eng. 2004, 45, 141–148. [Google Scholar] [CrossRef]
  11. Passey, Q.R.; Bohacs, K.M.; Esch, W.L.; Klimentidis, R.; Sinha, S. From Oil-Prone Source Rock to Gas-Producing Shale Reservoir—Geologic and Petrophysical Characterization of Unconventional Shale-Gas Reservoirs. In Proceedings of the International Oil and Gas Conference and Exhibition in China, Beijing, China, 8 June 2010; p. SPE-131350-MS. [Google Scholar]
  12. Hu, H.T.; Lu, S.F.; Liu, C.; Wang, W.M.; Wang, M.; Li, J.J.; Shang, J.H. Models for Calculating Organic Carbon Content from Logging Information: Comparison and Analysis. Acta Sedimentol. Sin. 2011, 29, 1199–1205. [Google Scholar]
  13. Wang, P.; Chen, Z.; Pang, X.; Hu, K.; Sun, M.; Chen, X. Revised models for determining TOC in shale play: Example from Devonian Duvernay Shale, Western Canada Sedimentary Basin. Mar. Pet. Geol. 2016, 70, 304–319. [Google Scholar] [CrossRef]
  14. Zhao, P.; Ma, H.; Rasouli, V.; Liu, W.; Cai, J.; Huang, Z. An improved model for estimating the TOC in shale formations. Mar. Pet. Geol. 2017, 83, 174–183. [Google Scholar] [CrossRef]
  15. Siddig, O.; Ibrahim, A.F.; Elkatatny, S. Application of Various Machine Learning Techniques in Predicting Total Organic Carbon from Well Logs. Comput. Intell. Neurosci. 2021, 2021, 7390055. [Google Scholar] [CrossRef] [PubMed]
  16. Zheng, D.Y.; Wu, S.X.; Hou, M.C. Fully connected deep network: An improved method to predict TOC of shale reservoirs from well logs. Mar. Pet. Geol. 2021, 132, 105205. [Google Scholar] [CrossRef]
  17. Filipič, B.; Junkar, M. Using inductive machine learning to support decision making in machining processes. Comput. Ind. 2000, 43, 31–41. [Google Scholar] [CrossRef]
  18. Kim, D.H.; Kim, T.J.; Wang, X.; Kim, M.; Quan, Y.J.; Oh, J.W.; Min, S.H.; Kim, H.; Bhandari, B.; Yang, I.; et al. Smart Machining Process Using Machine Learning: A Review and Perspective on Machining Industry. Int. J. Precis. Eng. Manuf.-Green Technol. 2018, 5, 555–568. [Google Scholar] [CrossRef]
  19. Rehman, T.U.; Mahmud, M.S.; Chang, Y.K.; Jin, J.; Shin, J. Current and future applications of statistical machine learning algorithms for agricultural machine vision systems. Comput. Electron. Agric. 2019, 156, 585–605. [Google Scholar] [CrossRef]
  20. Shin, S.J.; Kim, Y.M.; Meilanitasari, P. A Holonic-Based Self-Learning Mechanism for Energy-Predictive Planning in Machining Processes. Processes 2019, 7, 739. [Google Scholar] [CrossRef] [Green Version]
  21. Ucar, F.; Cordova, J.; Alcin, O.F.; Dandil, B.; Ata, F.; Arghandeh, R. Bundle Extreme Learning Machine for Power Quality Analysis in Transmission Networks. Energies 2019, 12, 1449. [Google Scholar] [CrossRef] [Green Version]
  22. Xu, Y.; Zhou, Y.; Sekula, P.; Ding, L. Machine learning in construction: From shallow to deep learning. Dev. Built Environ. 2021, 6, 100045. [Google Scholar] [CrossRef]
  23. Wong, L.J.; Michaels, A.J. Transfer Learning for Radio Frequency Machine Learning: A Taxonomy and Survey. Sensors 2022, 22, 1416. [Google Scholar] [CrossRef] [PubMed]
  24. Burbidge, R.; Trotter, M.; Buxton, B.; Holden, S. Drug design by machine learning: Support vector machines for pharmaceutical data analysis. Comput. Chem. 2001, 26, 5–14. [Google Scholar] [CrossRef] [PubMed]
  25. Sample, P.A.; Goldbaum, M.H.; Chan, K.; Boden, C.; Lee, T.W.; Vasile, C.; Boehm, A.G.; Sejnowski, T.; Johnson, C.A.; Weinreb, R.N. Using machine learning classifiers to identify glaucomatous change earlier in standard visual fields. Invest. Ophthalmol. Visual Sci. 2002, 43, 2660–2665. [Google Scholar]
  26. Bax, J.J.; van der Bijl, P.; Delgado, V. Machine Learning for Electrocardiographic Diagnosis of Left Ventricular Early Diastolic Dysfunction∗. J. Am. Coll. Cardiol. 2018, 71, 1661–1662. [Google Scholar] [CrossRef] [PubMed]
  27. Singh, K.; Beam, A.L.; Nallamothu, B.K. Machine Learning in Clinical Journals Moving From Inscrutable to Informative. Circ. Cardiovasc. Qual. Outcomes 2020, 13, e007491. [Google Scholar] [CrossRef]
  28. Mainali, S.; Darsie, M.E.; Smetana, K.S. Machine Learning in Action: Stroke Diagnosis and Outcome Prediction. Front. Neurol. 2021, 12, 2153. [Google Scholar] [CrossRef]
  29. Shuhaiber, J.H.; Conte, J.V. Machine learning in heart valve surgery. Eur. J. Cardio-Thorac. Surg. 2021, 60, 1386–1387. [Google Scholar] [CrossRef]
  30. Tong, S.; Koller, D. Support vector machine active learning with applications to text classification. J. Mach. Learn. Res. 2002, 2, 45–66. [Google Scholar]
  31. Chen, Z.Y.; Khoa LD, V.; Teoh, E.N.; Nazir, A.; Karuppiah, E.K.; Lam, K.S. Machine learning techniques for anti-money laundering (AML) solutions in suspicious transaction detection: A review. Knowl. Inf. Syst. 2018, 57, 245–285. [Google Scholar] [CrossRef]
  32. Andres, A.R.; Otero, A.; Amavilah, V.H. Using deep learning neural networks to predict the knowledge economy index for developing and emerging economies. Expert Syst. Appl. 2021, 184, 115514. [Google Scholar] [CrossRef]
  33. Xie, X.; Sun, S. Multi-view Laplacian twin support vector machines. Appl. Intell. 2014, 41, 1059–1068. [Google Scholar] [CrossRef]
  34. de Bruijne, M. Machine learning approaches in medical image analysis: From detection to diagnosis. Med. Image Anal. 2016, 33, 94–97. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. De Iaco, S.; Hristopulos, D.T.; Lin, G. Special Issue: Geostatistics and Machine Learning. Math. Geosci. 2022, 54, 459–465. [Google Scholar] [CrossRef]
  36. de Matos, M.C.; Osorio PL, M.; Johann, P.R.S. Unsupervised seismic facies analysis using wavelet transform and self-organizing maps. Geophysics 2007, 72, P9–P21. [Google Scholar] [CrossRef]
  37. de Matos, M.C.; Yenugu, M.; Angelo, S.M.; Marfurt, K.J. Integrated seismic texture segmentation and cluster analysis applied to channel delineation and chert reservoir characterization. Geophysics 2011, 76, P11–P21. [Google Scholar] [CrossRef]
  38. Roy, A.; Romero-Pelaez, A.S.; Kwiatkowski, T.J.; Marfurt, K.J. Generative topographic mapping for seismic facies estimation of a carbonate wash, Veracruz Basin, southern Mexico. Interpret.—A J. Subsurf. Charact. 2014, 2, SA31–SA47. [Google Scholar] [CrossRef] [Green Version]
  39. Qi, J.; Lin, T.F.; Zhao, T.; Li, F.Y.; Marfurt, K. Semisupervised multiattribute seismic facies analysis. Interpret.—A J. Subsurf. Charact. 2016, 4, SB91–SB106. [Google Scholar] [CrossRef]
  40. Qian, F.; Yin, M.; Liu, X.Y.; Wang, Y.J.; Lu, C.; Hu, G.M. Unsupervised seismic facies analysis via deep convolutional autoencoders. Geophysics 2018, 83, A39–A43. [Google Scholar] [CrossRef]
  41. Harris, J.R.; Grunsky, E.C. Predictive lithological mapping of Canada’s North using Random Forest classification applied to geophysical and geochemical data. Comput. Geosci. 2015, 80, 9–25. [Google Scholar] [CrossRef]
  42. Sebtosheikh, M.A.; Motafakkerfard, R.; Riahi, M.A.; Moradi, S.; Sabety, N. Support vector machine method, a new technique for lithology prediction in an Iranian heterogeneous carbonate reservoir using petrophysical well logs. Carbonates Evaporites 2015, 30, 59–68. [Google Scholar] [CrossRef]
  43. Ai, X.; Wang, H.Y.; Sun, B.T. Automatic Identification of Sedimentary Facies Based on a Support Vector Machine in the Aryskum Graben, Kazakhstan. Appl. Sci. 2019, 9, 4489. [Google Scholar] [CrossRef] [Green Version]
  44. Mulashani, A.K.; Shen, C.B.; Asante-Okyere, S.; Kerttu, P.N.; Abelly, E.N. Group Method of Data Handling (GMDH) Neural Network for Estimating Total Organic Carbon (TOC) and Hydrocarbon Potential Distribution (S-1, S-2) Using Well Logs. Nat. Resour. Res. 2021, 30, 3605–3622. [Google Scholar] [CrossRef]
  45. Hossain, T.M.; Watada, J.; Aziz, I.A.; Hermana, M. Machine Learning in Electrofacies Classification and Subsurface Lithology Interpretation: A Rough Set Theory Approach. Appl. Sci. 2020, 10, 5940. [Google Scholar] [CrossRef]
  46. Ashraf, U.; Zhang, H.; Anees, A.; Mangi, H.N.; Ali, M.; Zhang, X.; Imraz, M.; Abbasi, S.S.; Abbas, A.; Ullah, Z.; et al. A Core Logging, Machine Learning and Geostatistical Modeling Interactive Approach for Subsurface Imaging of Lenticular Geobodies in a Clastic Depositional System, SE Pakistan. Nat. Resour. Res. 2021, 30, 2807–2830. [Google Scholar] [CrossRef]
  47. Shokir, E.M.E. A novel model for permeability prediction in uncored wells. SPE Reserv. Eval. Eng. 2006, 9, 266–273. [Google Scholar] [CrossRef]
  48. Al-Anazi, A.; Gates, I.D. Support-Vector Regression for Permeability Prediction in a Heterogeneous Reservoir: A Comparative Study. SPE Reserv. Eval. Eng. 2010, 13, 485–495. [Google Scholar] [CrossRef]
  49. Kaydani, H.; Mohebbi, A.; Eftekhari, M. Permeability estimation in heterogeneous oil reservoirs by multi-gene genetic programming algorithm. J. Pet. Sci. Eng. 2014, 123, 201–206. [Google Scholar] [CrossRef]
  50. Zhang, G.Y.; Wang, Z.Z.; Li, H.J.; Sun, Y.A.; Zhang, Q.C.; Chen, W. Permeability prediction of isolated channel sands using machine learning. J. Appl. Geophys. 2018, 159, 605–615. [Google Scholar] [CrossRef]
  51. Zhang, G.Y.; Wang, Z.Z.; Mohaghegh, S.; Lin, C.Y.; Sun, Y.N.; Pei, S.J. Pattern visualization and understanding of machine learning models for permeability prediction in tight sandstone reservoirs. J. Pet. Sci. Eng. 2021, 200, 108142. [Google Scholar] [CrossRef]
  52. Liu, J.Z.; Liu, X.Y. Recognition and Classification for Inter-well Nonlinear Permeability Configuration in Low Permeability Reservoirs Utilizing Machine Learning Methods. Front. Earth Sci. 2022, 10, 218. [Google Scholar] [CrossRef]
  53. Wang, Y.J.; Li, H.Y.; Xu, J.C.; Liu, S.Y.; Wang, X.P. Machine learning assisted relative permeability upscaling for uncertainty quantification. Energy 2022, 245, 123284. [Google Scholar] [CrossRef]
  54. de Lima, R.P.; Surianam, F.; Marfurt, K.J.; Pranter, M.J. Convolutional neural networks as aid in core lithofacies classification. Interpret.—A J. Subsurf. Charact. 2019, 7, SF27–SF40. [Google Scholar]
  55. Koeshidayatullah, A.; Morsilli, M.; Lehrmann, D.J.; Al-Ramadan, K.; Payne, J.L. Fully automated carbonate petrography using deep convolutional neural networks. Mar. Pet. Geol. 2020, 122, 104687. [Google Scholar] [CrossRef]
  56. de Lima, R.P.; Duarte, D.; Nicholson, C.; Slatt, R.; Marfurt, K.J. Petrographic microfacies classification with deep convolutional neural networks. Comput. Geosci. 2020, 142, 104481. [Google Scholar] [CrossRef]
  57. Baraboshkin, E.E.; Ismailova, L.S.; Orlov, D.M.; Zhukovskaya, E.A.; Kalmykov, G.A.; Khotylev, O.V.; Baraboshkin, E.Y.; Koroteev, D.A. Deep convolutions for in-depth automated rock typing. Comput. Geosci. 2020, 135, 104330. [Google Scholar] [CrossRef] [Green Version]
  58. Izadi, H.; Sadri, J.; Bayati, M. An intelligent system for mineral identification in thin sections based on a cascade approach. Comput. Geosci. 2017, 99, 37–49. [Google Scholar] [CrossRef]
  59. Saporetti, C.M.; da Fonseca, L.G.; Pereira, E.; de Oliveira, L.C. Machine learning approaches for petrographic classification of carbonate-siliciclastic rocks using well logs and textural information. J. Appl. Geophys. 2018, 155, 217–225. [Google Scholar] [CrossRef]
  60. Silva, A.A.; Tavares, M.W.; Carrasquilla, A.; Missagia, R.; Ceia, M. Petrofacies classification using machine learning algorithms. Geophysics 2020, 85, WA101–WA113. [Google Scholar] [CrossRef]
  61. Zhu, L.; Zhang, C.; Zhang, C.; Zhang, Z.; Nie, X.; Zhou, X.; Liu, W.; Wang, X. Forming a new small sample deep learning model to predict total organic carbon content by combining unsupervised learning with semisupervised learning. Appl. Soft Comput. 2019, 83, 105596. [Google Scholar] [CrossRef]
  62. Shi, X.; Wang, J.; Liu, G.; Yang, L.; Ge, X.M.; Jiang, S. Application of extreme learning machine and neural networks in total organic carbon content prediction in organic shale with wire line logs. J. Nat. Gas Sci. Eng. 2016, 33, 687–702. [Google Scholar] [CrossRef]
  63. Johnson, L.M.; Rezaee, R.; Kadkhodaie, A.; Smith, G.; Yu, H.Y. Geochemical property modelling of a potential shale reservoir in the Canning Basin (Western Australia), using Artificial Neural Networks and geostatistical tools. Comput. Geosci. 2018, 120, 73–81. [Google Scholar] [CrossRef]
  64. Zhu, L.; Zhang, C.; Zhang, C.; Wei, Y.; Zhou, X.; Cheng, Y.; Huang, Y.; Zhang, L. Prediction of total organic carbon content in shale reservoir based on a new integrated hybrid neural network and conventional well logging curves. J. Geophys. Eng. 2018, 15, 1050–1061. [Google Scholar] [CrossRef] [Green Version]
  65. Wang, H.J.; Wu, W.; Chen, T.; Dong, X.J.; Wang, G.X. An improved neural network for TOC, S-1 and S-2 estimation based on conventional well logs. J. Pet. Sci. Eng. 2019, 176, 664–678. [Google Scholar] [CrossRef]
  66. Zhu, L.Q.; Zhang, C.; Zhang, C.M.; Zhou, X.Q.; Wang, J.; Wang, X. Application of Multiboost-KELM algorithm to alleviate the collinearity of log curves for evaluating the abundance of organic matter in marine mud shale reservoirs: A case study in Sichuan Basin, China. Acta Geophys 2018, 66, 983–1000. [Google Scholar] [CrossRef]
  67. Liu, X.Z.; Tian, Z.; Chen, C. Total Organic Carbon Content Prediction in Lacustrine Shale Using Extreme Gradient Boosting Machine Learning Based on Bayesian Optimization. Geofluids 2021, 2021, 6155663. [Google Scholar] [CrossRef]
  68. Rui, J.W.; Zhang, H.B.; Zhang, D.L.; Han, F.L.; Guo, Q. Total organic carbon content prediction based on support-vector-regression machine with particle swarm optimization. J. Pet. Sci. Eng. 2019, 180, 699–706. [Google Scholar] [CrossRef]
  69. Amosu, A.; Imsalem, M.; Sun, Y.F. Effective machine learning identification of TOC-rich zones in the Eagle Ford Shale. J. Appl. Geophys. 2021, 188, 104311. [Google Scholar] [CrossRef]
  70. Rong, J.; Zheng, Z.; Luo, X.; Li, C.; Li, Y.; Wei, X.; Wei, Q.; Yu, G.; Zhang, L.; Lei, Y. Machine Learning Method for TOC Prediction: Taking Wufeng and Longmaxi Shales in the Sichuan Basin, Southwest China as an Example. Geofluids 2021, 2021, 6794213. [Google Scholar] [CrossRef]
  71. Zhu, L.; Zhang, C.; Zhang, C.; Zhang, Z.; Zhou, X.; Liu, W.; Zhu, B. A new and reliable dual model- and data-driven TOC prediction concept: A TOC logging evaluation method using multiple overlapping methods integrated with semi-supervised deep learning. J. Pet. Sci. Eng. 2020, 188, 106944. [Google Scholar] [CrossRef]
  72. Handhal, A.M.; Al-Abadi, A.M.; Chafeet, H.E.; Ismail, M.J. Prediction of total organic carbon at Rumaila oil field, Southern Iraq using conventional well logs and machine learning algorithms. Mar. Pet. Geol. 2020, 116, 104347. [Google Scholar] [CrossRef]
  73. Ao, S.U.; Honghan, C.H.E.N.; Laisheng, C.A.O.; Mingzhu, L.E.I.; Cunwu, W.A.N.G.; Yanhua, L.I.U.; Peijun, L.I. Genesis, source and charging of oil and gas in Lishui sag, East China Sea Basin. Pet. Explor. Dev. 2014, 41, 574–584. [Google Scholar]
  74. Jiang, Z.; Li, Y.; Du, H.; Zhang, Y. The Cenozoic structural evolution and its influences on gas accumulation in the Lishui Sag, East China Sea Shelf Basin. J. Nat. Gas Sci. Eng. 2015, 22, 107–118. [Google Scholar] [CrossRef]
  75. Zhang, M.; Zhang, J.; Xu, F.; Li, J.; Liu, J.; Hou, G.; Zhang, P. Paleocene sequence stratigraphy and depositional systems in the Lishui Sag, East China Sea Shelf Basin. Mar. Pet. Geol. 2015, 59, 390–405. [Google Scholar] [CrossRef]
  76. Li, Y.; Zhang, J.; Liu, Y.; Shen, W.; Chang, X.; Sun, Z.; Xu, G. Organic geochemistry, distribution and hydrocarbon potential of source rocks in the Paleocene, Lishui Sag, East China Sea Shelf Basin. Mar. Pet. Geol. 2019, 107, 382–396. [Google Scholar] [CrossRef]
  77. Liu, L.; Li, Y.; Dong, H.; Sun, Z. Diagenesis and reservoir quality of Paleocene tight sandstones, Lishui Sag, East China Sea Shelf Basin. J. Pet. Sci. Eng. 2020, 195, 107615. [Google Scholar] [CrossRef]
  78. Lei, C.; Yin, S.; Ye, J.; Wu, J.; Wang, Z.; Gao, B. Characteristics and deposition models of the paleocene source rocks in the Lishui Sag, east China sea shelf basin: Evidences from organic and inorganic geochemistry. J. Pet. Sci. Eng. 2021, 200, 108342. [Google Scholar] [CrossRef]
  79. Huang, Y.; Tarantola, A.; Wang, W.; Caumon, M.C.; Pironon, J.; Lu, W.; Yan, D.; Zhuang, X. Charge history of CO2 in Lishui sag, East China Sea basin: Evidence from quantitative Raman analysis of CO2-bearing fluid inclusions. Mar. Pet. Geol. 2018, 98, 50–65. [Google Scholar] [CrossRef]
  80. Li, Y.; Zhang, J.; Xu, Y.; Chen, T.; Liu, J. Improved understanding of the origin and accumulation of hydrocarbons from multiple source rocks in the Lishui Sag: Insights from statistical methods, gold tube pyrolysis and basin modeling. Mar. Pet. Geol. 2021, 134, 105361. [Google Scholar] [CrossRef]
  81. Nagasaka, Y.; Iwata, A. Performance evaluation of BP and PCA neural networks for ECG data compression. In Proceedings of 1993 International Conference on Neural Networks (IJCNN-93-Nagoya, Japan), Nagoya, Japan, 25–29 October 1993; Volume 1, pp. 1003–1006. [Google Scholar]
  82. Fung, C.C.; Wong, K.W.; Eren, H.; Charlebois, R.; Crocker, H. Modular artificial neural network for prediction of petrophysical properties from well log data. In Proceedings of the Quality Measurement: The Indispensable Bridge between Theory and Reality (No Measurements? No Science! Joint Conference-1996: IEEE Instrumentation and Measurement Technology Conference and IMEKO Tec, Brussels, Belgium, 4–6 June 1996; Volume 2, pp. 1010–1014. [Google Scholar]
  83. Huang, Y.; Wong, P.M.; Gedeon, T.D. An improved fuzzy neural network for permeability estimation from wireline logs in a petroleum reservoir. In Proceedings of the Proceedings of Digital Processing Applications (TENCON ′96), Perth, WA, Australia, 29 November 1996; Volume 2, pp. 912–917. [Google Scholar]
  84. Wong, P.M.; Gedeon, T.D.; Taggart, I.J. Fuzzy ARTMAP: A new tool for lithofacies recognition. Ai Appl. 1996, 10, 29–39. [Google Scholar]
  85. Wong, P.M.; Henderson, D.J.; Brooks, L.J. Permeability determination using neural networks in the Ravva Field, offshore India. SPE Reserv. Eval. Eng. 1998, 1, 99–104. [Google Scholar] [CrossRef]
  86. Sfidari, E.; Kadkhodaie-Ilkhchi, A.; Najjari, S. Comparison of intelligent and statistical clustering approaches to predicting total organic carbon using intelligent systems. J. Pet. Sci. Eng. 2012, 86–87, 190–205. [Google Scholar] [CrossRef]
  87. Hansen, K.B. The virtue of simplicity: On machine learning models in algorithmic trading. Big Data Soc. 2020, 7, 2053951720926558. [Google Scholar] [CrossRef]
  88. Goz, E.; Yuceer, M.; Karadurmus, E. Total Organic Carbon Prediction with Artificial Intelligence Techniques. In Computer Aided Chemical Engineering; Elsevier: Amsterdam, The Netherlands, 2019; pp. 889–894. [Google Scholar]
  89. Amiri Bakhtiar, H.; Telmadarreie, A.; Shayesteh, M.; Heidari Fard, M.H.; Talebi, H.; Shirband, Z. Estimating Total Organic Carbon Content and Source Rock Evaluation, Applying ΔlogR and Neural Network Methods: Ahwaz and Marun Oilfields, SW of Iran. Pet. Sci. Technol. 2011, 29, 1691–1704. [Google Scholar] [CrossRef]
  90. Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: A new learning scheme of feedforward neural networks. In Proceedings of the 2004 IEEE international joint conference on neural networks, Budapest, Hungary, 25–29 July 2004; Volumes 1–4, pp. 985–990. [Google Scholar]
  91. Huang, G.B.; Zhou, H.M.; Ding, X.J.; Zhang, R. Extreme Learning Machine for Regression and Multiclass Classification. IEEE Trans. Syst. Man Cybern. Part B-Cybern. 2012, 42, 513–529. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  92. Peng, X.L.; Lin, P.; Zhang, T.S.; Wang, J. Extreme Learning Machine-Based Classification of ADHD Using Brain Structural MRI Data. PLoS ONE 2013, 8, e79476. [Google Scholar] [CrossRef]
  93. Ahmad, I.; Basheri, M.; Iqbal, M.J.; Rahim, A. Performance Comparison of Support Vector Machine, Random Forest, and Extreme Learning Machine for Intrusion Detection. IEEE Access 2018, 6, 33789–33795. [Google Scholar] [CrossRef]
  94. Zong, W.W.; Huang, G.B.; Chen, Y.Q. Weighted extreme learning machine for imbalance learning. Neurocomputing 2013, 101, 229–242. [Google Scholar] [CrossRef]
  95. Tang, J.X.; Deng, C.W.; Huang, G.B. Extreme Learning Machine for Multilayer Perceptron. IEEE Trans. Neural Networks Learn. Syst. 2016, 27, 809–821. [Google Scholar] [CrossRef]
  96. Lariviere, B.; Van den Poel, D. Predicting customer retention and profitability by using random forests and regression forests techniques. Expert Syst. Appl. 2005, 29, 472–484. [Google Scholar] [CrossRef]
  97. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  98. Breiman, L.; Last, M.; Rice, J. Random forests: Finding quasars. In Statistical Challenges in Astronomy; Springer: New York, NY, USA, 2003; pp. 243–254. [Google Scholar]
  99. Ziegler, A.; Konig, I.R. Mining data with random forests: Current options for real-world applications. Wiley Interdiscip. Rev.-Data Min. Knowl. Discov. 2014, 4, 55–63. [Google Scholar]
  100. Scornet, E. Random Forests and Kernel Methods. IEEE Trans. Inf. Theory 2016, 62, 1485–1500. [Google Scholar] [CrossRef]
  101. Shalaby, M.R.; Malik, O.A.; Lai, D.; Jumat, N.; Islam, M.A. Thermal maturity and TOC prediction using machine learning techniques: Case study from the Cretaceous-Paleocene source rock, Taranaki Basin, New Zealand. J. Pet. Explor. Prod. Technol. 2020, 10, 2175–2193. [Google Scholar]
  102. Biau, G. Analysis of a Random Forests Model. J. Mach. Learn. Res. 2012, 13, 1063–1095. [Google Scholar]
  103. Mutanga, O.; Adam, E.; Cho, M.A. High density biomass estimation for wetland vegetation using WorldView-2 imagery and random forest regression algorithm. Int. J. Appl. Earth Obs. Geoinf. 2012, 18, 399–406. [Google Scholar]
Figure 1. (a) Tectonic location of the East China Sea Shelf Basin (modified after (modified after [51]). (b) Structural units of the Lishui–Jiaojiang Sag (modified after [51]).
Figure 1. (a) Tectonic location of the East China Sea Shelf Basin (modified after (modified after [51]). (b) Structural units of the Lishui–Jiaojiang Sag (modified after [51]).
Energies 15 09480 g001
Figure 2. Stratigraphic units of the Paleocene in the Lishui Sag (modified after [51]).
Figure 2. Stratigraphic units of the Paleocene in the Lishui Sag (modified after [51]).
Energies 15 09480 g002
Figure 3. Cross plots of TOC laboratory measurements and logging data.
Figure 3. Cross plots of TOC laboratory measurements and logging data.
Energies 15 09480 g003
Figure 4. Schematic diagram of ELM.
Figure 4. Schematic diagram of ELM.
Energies 15 09480 g004
Figure 5. Schematic diagram of RF.
Figure 5. Schematic diagram of RF.
Energies 15 09480 g005
Figure 6. Cross plot of hydrogen index (HI) and Tmax.
Figure 6. Cross plot of hydrogen index (HI) and Tmax.
Energies 15 09480 g006
Figure 7. Cross plots between TOC values from multiple regression models and TOC values from core data.
Figure 7. Cross plots between TOC values from multiple regression models and TOC values from core data.
Energies 15 09480 g007
Figure 8. Cross plots between TOC values from ML models using HXC and TOC values from core data.
Figure 8. Cross plots between TOC values from ML models using HXC and TOC values from core data.
Energies 15 09480 g008
Figure 9. Cross plots between TOC values from ML models using HX and TOC values from core data.
Figure 9. Cross plots between TOC values from ML models using HX and TOC values from core data.
Energies 15 09480 g009
Figure 10. Run time of ML models using HXC and HX.
Figure 10. Run time of ML models using HXC and HX.
Energies 15 09480 g010
Figure 11. Error analysis of each model using HXC.
Figure 11. Error analysis of each model using HXC.
Energies 15 09480 g011
Table 1. Data of Pearson’s correlation coefficients.
Table 1. Data of Pearson’s correlation coefficients.
RACCNDENGRRTTOC
AC10.85133−0.68255−0.09409−0.165290.67236
CN 1−0.61136−0.06937−0.082790.77855
DEN 10.044740.1027−0.49961
GR 10.426110.00713
RT 10.1442
TOC 1
Table 2. Results from geochemical Rock-Eval/TOC analysis with calculated parameters from Yueguifeng Formation.
Table 2. Results from geochemical Rock-Eval/TOC analysis with calculated parameters from Yueguifeng Formation.
WellDepth (m)TOCS1S2PgHITmax
(mg/g TOC)(mg/g TOC)S1 + S2S2/TOC°C
(mg/g TOC)
A27010.550.120.881159.42438
27071.090.221.191.41109.17444
27201.580.272.112.38133.54434
27210.40.120.991.11247.5442
2916.60.990.020.480.548454
29170.880.020.30.3234452
29400.730.890.771.66105.48430
29600.70.230.961.19137.14428
30790.350.020.180.234451
30920.470.020.310.3365.96451
30990.370.020.210.2356.76453
B2356.53.340.1311.7211.85350.9438
2362.52.990.3311.6812.01390.64441
2395.51.880.222.983.2158.51438
2401.52.420.237.217.44297.93440
2422.52.590.428.578.99330.89443
2425.52.120.377.217.58340.09440
2434.53.180.1512.3612.51388.68440
2488.52.880.147.487.62259.72440
24952.420.229.29.42380.17438
2506.52.650.349.129.46344.15441
2518.52.510.5188.51318.73440
C35760.970.210.450.6646.39451
3577.51.610.551.281.8379.5458
3586.51.240.330.660.9953.23452
3595.51.710.360.981.3457.31461
3604.51.860.211.161.3762.37458
3607.51.870.381.241.6266.31459
3630.51.480.170.610.7841.22460
36400.680.020.190.2127.94470
3640.050.70.010.120.1317.14479
3640.81.290.130.830.9664.34466
3641.62.41.081.983.0682.5470
3641.70.830.050.210.2625.3478
3642.72.170.111.081.1949.77466
36430.530.030.210.2439.55472
3643.60.20.010.040.0520471
3643.70.640.030.160.1925476
3643.790.580.030.180.2131.03452
3645.81.140.170.60.7752.63465
36461.110.10.620.7255.86469
3646.41.110.110.60.7154.05466
3646.951.690.140.550.6932.54463
36471.040.130.60.7357.69462
3647.091.540.380.831.2153.9461
3647.20.930.10.360.4638.71471
3647.251.740.370.711.0840.8470
3647.421.490.120.60.7240.27468
3647.51.240.10.620.7250462
3647.51.230.110.740.8560.16463
3647.811.570.160.670.8342.68465
3648.160.210.030.210.24100455
3688.51.840.291.541.8383.7448
37020.880.160.320.4836.36458
3712.51.160.150.60.7551.72471
3724.50.570.070.260.3345.61461
3739.51.910.881.472.3576.96455
37451.490.521.211.7381.21456
3751.52.261.271.873.1482.74457
37552.771.322.013.3372.56457
3769.51.991.31.32.665.33455
3772.52.491.382.433.8197.59454
3775.52.212.051.933.9887.33453
3784.52.771.081.682.7660.65446
3793.52.061.251.252.560.68457
3793.52.641.031.612.6460.98451
37951.930.991.222.2163.21453
37952.181.011.252.2657.34455
3796.51.90.941.152.0960.53459
3796.52.551.212.43.6194.12458
38131.40.20.680.8848.57457
38522.730.891.612.558.97464
38832.861.271.833.163.99447
38882.891.522.153.6774.39456
39031.470.130.580.7139.46455
3910.51.480.370.610.9841.22453
3913.50.910.280.430.7147.25455
S1: volatile hydrocarbon (HC) content, mg HC/g rock. S2: remaining HC generative potential, mg HC/g rock. HI: hydrogen index = S2 × 100/TOC, mg HC/g TOC. Pg: potential for hydrocarbon generation = S1 + S2 (mg/g).
Table 3. Results of TOC prediction from multiple regression.
Table 3. Results of TOC prediction from multiple regression.
WellDepth (m)TOCPrediction
%Multiple Regression
HXCHX
A27010.550.770.55
27071.090.920.68
27201.580.890.59
27210.40.930.51
2916.60.990.960.82
29170.880.910.81
29400.730.610.6
29600.70.80.71
30790.350.910.9
30920.470.80.81
30990.371.31.36
B2356.53.343.163.11
2362.52.992.912.96
2395.51.882.442.62
2401.52.421.591.68
2422.52.592.612.79
2425.52.122.52.62
2434.53.182.272.52
2488.52.882.42.84
24952.421.872.16
2506.52.651.912.14
2518.52.511.972.15
C35760.971.351.39
3577.51.611.351.41
3586.51.241.431.44
3595.51.712.192.37
3604.51.861.371.54
3607.51.871.842.01
3630.51.481.161.25
36400.681.211.01
3640.050.71.291.09
3640.81.291.241.06
3641.62.40.941.09
3641.70.830.991.1
3642.72.171.291.48
36430.531.171.34
3643.60.21.251.29
3643.70.641.261.27
3643.790.581.261.27
3645.81.141.431.5
36461.111.161.18
3646.41.111.051.11
3646.951.691.431.54
36471.041.431.54
3647.091.541.351.47
3647.20.931.241.34
3647.251.741.241.34
3647.421.491.181.27
3647.51.241.171.26
3647.51.231.171.26
3647.811.571.211.32
3648.160.211.251.35
3688.51.841.411.59
37020.881.51.71
3712.51.160.890.87
3724.50.571.311.33
3739.51.912.242.18
37451.492.162.15
3751.52.262.372.2
37552.772.72.1
3769.51.992.11.91
3772.52.492.131.8
3775.52.212.061.69
3784.52.772.212.06
3793.52.061.961.8
3793.52.642.021.89
37951.932.352.24
37952.182.22.06
3796.51.92.041.87
3796.52.552.061.89
38131.41.461.27
38522.733.243.1
38832.862.882.14
38882.893.552.25
39031.471.972.19
3910.51.481.190.91
3913.50.911.271.29
Table 4. Mean error values of different methods.
Table 4. Mean error values of different methods.
MethodRFBPNNELMMultiple Regression
ParameterHXCHXHXCHXHXCHXHXCHXHXCHXHXCHXHXCHX
TypeTrainingTestedTrainingTestedTrainingTestedTested
R20.970.970.650.490.910.840.700.530.880.820.660.570.630.60
MAE0.030.030.390.560.220.260.370.470.230.280.420.450.390.41
MSE0.020.020.260.450.070.260.230.430.100.120.270.330.240.26
RMSE0.150.150.510.670.260.510.480.660.310.350.520.570.490.51
Table 5. Results of TOC prediction from ML.
Table 5. Results of TOC prediction from ML.
Core DataTraining Dataset PredictionCore DataTested Dataset Prediction
RFBPNNELMRFBPNNELM
HXCHXHXCHXHXCHXHXCHXHXCHXHXCHX
0.990.990.990.711.11 0.750.821.400.700.701.301.221.021.07
0.880.880.880.750.920.740.782.732.771.882.722.772.732.81
0.470.470.470.650.920.660.422.862.772.772.842.822.732.42
3.343.343.343.243.223.262.792.892.772.653.322.863.022.56
2.992.992.992.832.963.012.901.471.911.871.272.942.122.53
1.881.881.881.642.501.902.681.481.611.160.931.001.020.92
2.592.592.592.763.072.463.010.910.571.840.521.481.051.13
3.183.183.183.123.132.732.840.730.470.470.410.720.360.44
2.882.882.883.162.872.803.150.550.470.470.600.750.220.44
2.652.652.652.302.722.292.441.090.470.470.810.790.580.52
2.512.512.512.242.582.032.361.580.470.470.700.710.360.44
0.970.970.970.801.111.051.340.350.990.990.551.471.010.96
1.611.611.611.511.511.081.390.370.970.970.701.480.901.30
1.241.241.241.341.641.611.472.122.592.593.033.042.772.92
1.861.861.861.361.411.431.442.422.651.242.052.502.292.35
1.871.871.871.532.421.892.332.422.593.181.782.111.621.99
0.680.680.680.550.680.590.811.712.513.182.722.972.302.68
0.70.70.700.761.200.880.891.480.830.880.881.271.491.22
2.42.42.402.231.741.512.331.291.231.110.861.100.800.95
0.830.830.830.981.401.361.050.400.700.470.500.760.080.36
2.172.172.172.072.311.331.390.700.470.470.610.690.210.54
0.5310.5310.530.870.530.530.531.491.912.511.992.582.182.36
0.20.20.200.210.200.210.222.261.911.912.172.822.442.45
0.640.580.640.630.630.621.181.111.110.200.951.310.801.10
0.580.580.640.530.530.561.181.541.042.171.261.551.321.40
1.141.141.141.261.771.301.511.241.231.231.111.271.211.16
1.111.111.110.991.120.890.990.211.490.531.231.361.281.27
1.691.041.041.301.671.411.50
1.041.041.041.301.671.011.50
0.930.930.931.221.391.240.91
1.740.930.931.221.391.241.25
1.491.491.491.151.291.211.17
1.231.231.231.111.271.211.16
1.571.571.571.131.341.301.24
1.841.841.841.391.311.451.47
0.880.880.880.961.400.880.90
1.161.161.160.691.360.890.85
0.570.570.570.771.240.560.58
1.911.911.911.811.951.811.90
2.772.772.772.832.772.622.33
1.991.991.992.092.321.992.03
2.492.492.492.152.172.231.87
2.212.212.212.082.052.041.74
2.772.772.772.132.322.392.21
2.062.062.061.912.292.001.90
2.642.642.642.362.292.042.46
1.931.931.932.161.982.022.41
2.182.182.182.142.322.492.22
1.91.91.902.052.362.181.99
2.552.552.552.092.442.252.03
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Han, X.; Hou, D.; Cheng, X.; Li, Y.; Niu, C.; Chen, S. Prediction of TOC in Lishui–Jiaojiang Sag Using Geochemical Analysis, Well Logs, and Machine Learning. Energies 2022, 15, 9480. https://doi.org/10.3390/en15249480

AMA Style

Han X, Hou D, Cheng X, Li Y, Niu C, Chen S. Prediction of TOC in Lishui–Jiaojiang Sag Using Geochemical Analysis, Well Logs, and Machine Learning. Energies. 2022; 15(24):9480. https://doi.org/10.3390/en15249480

Chicago/Turabian Style

Han, Xu, Dujie Hou, Xiong Cheng, Yan Li, Congkai Niu, and Shuosi Chen. 2022. "Prediction of TOC in Lishui–Jiaojiang Sag Using Geochemical Analysis, Well Logs, and Machine Learning" Energies 15, no. 24: 9480. https://doi.org/10.3390/en15249480

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop