A Novel Approach to Oil Layer Recognition Model Using Whale Optimization Algorithm and Semi-Supervised SVM

The dataset distribution of actual logging is asymmetric, as most logging data are unlabeled. With the traditional classification model, it is hard to predict the oil and gas reservoir accurately. Therefore, a novel approach to the oil layer recognition model using the improved whale swarm algorithm (WOA) and semi-supervised support vector machine (S3VM) is proposed in this paper. At first, in order to overcome the shortcomings of the Whale Optimization Algorithm applied in the parameter-optimization of the S3VM model, such as falling into a local optimization and low convergence precision, an improved WOA was proposed according to the adaptive cloud strategy and the catfish effect. Then, the improved WOA was used to optimize the kernel parameters of S3VM for oil layer recognition. In this paper, the improved WOA is used to test 15 benchmark functions of CEC2005 compared with five other algorithms. The IWOA–S3VM model is used to classify the five kinds of UCI datasets compared with the other two algorithms. Finally, the IWOA–S3VM model is used for oil layer recognition. The result shows that (1) the improved WOA has better convergence speed and optimization ability than the other five algorithms, and (2) the IWOA–S3VM model has better recognition precision when the dataset contains a labeled and unlabeled dataset in oil layer recognition.


Introduction
Oil is the blood of the industry, and China is now the world's largest oil importer. With the increasing dependence on foreign oil, the contradiction between supply and demand appears more evident in China [1]. To alleviate the problem of the conflicting supply and demand of oil, one effective means is the accurate identification of oil reservoirs that can stabilize oil production and increase oil reserves' development.
The logging data curve is a data signal of the change of physical properties' formation with oil well depth. It is also the basis for solving various parameters of oil and gas reservoirs in oil logging recognition. Logging data curves provide valuable information for identifying subsurface sedimentation and analyzing the distribution of subsurface material layers [2,3].
Using the logging data curve to identify the lithology is faster and cheaper than other lithology identification methods. Logging data processing mainly includes data pre-processing, attribute reductions, and classification. Classification is the essential work of the logging data process. In oil logging recognition, labeled data are usually scarce; the traditional model will get unsatisfying results when processing the oil-logging data. So, in this paper, we choose the semi-supervised learning method to process the oil-logging data.
Semi-supervised learning (SSL) [4] is notable for utilizing both labeled and unlabeled samples. Semi-supervised learning breaks through the limitation of traditional methods to consider only one sample type and can tap into the hidden information of a large amount of unlabeled data, aided by a small number of labeled samples for training models. the result shows that the modified WOA is suitable for high-dimension global optimization problems. Zhang et al. [28] proposed a new algorithm based on Lamarckian learning to improve the performance of WOA; the result shows that improvement strategies help WOA improve its ability to solve global optimization problems. Chen [29] introduces levy flight and trigger rules to balance the searchability of WOA, and the simulation results show that it is better than WOA.
The above research of WOA shows that it has a good effect on the application. However, the No Free Lunch theorem [30] shows that no universal optimizer can solve all problems. Therefore, it makes good sense to improve WOA further. There are two problems concerning WOA. One is that the individual predatory behavior of whales is random. It will affect the global optimization ability and search time of the algorithm. So, we introduced the adaptive cloud strategy to help the individual whale to mutate more effectively. The second problem is that the whale's global optimum solution is prone to falling into the local optimum solution in the latter part of the iteration. Therefore, we used the catfish effect to help WOA jump out of the local optimum solution in the latter part of the iteration.
Firstly, to verify the effectiveness of IWOA, it is compared with four other classic algorithms and one improved whale optimization algorithm on 15 classic benchmark functions. Those four classic algorithms are PSO, GWO, WOA, MPA. The improved whale optimization algorithm is HWOA [31]. For all of the experiment data, we use a box line diagram and the Wilcoxon Rank Sum Test to test the stability and competitiveness of IWOA. The experiment results show that IWOA is significantly superior to the other five algorithms. Then, we use IWOA to optimize the selection of kernel hyperparameters in the S3VM for classifying UCI datasets and oil layer recognition. The result shows that the IWOA-S3VM model has higher classification accuracy than other algorithms.
The rest of the paper is structured as follows. Section 2.1 describes the introduction of the whale optimization algorithm (WOA). Sections 2.1 and 2.2 describe the drawbacks of WOA and the proposed improved WOA. Section 2.3 present and analyze the experiment results of IWOA compared with other well-known algorithms. The basic semi-supervised support vector machine algorithm is briefly introduced in Section 3. Section 4 presents the S3VM model's details combined with IWOA, and runs the UCI datasets experiment compared to other algorithms. Section 5 presents the oil layer recognition system's details and uses an IWOA-S3VM model for oil layer recognition.
Humpback whales are able to identify the location of their prey and surround them with the following position update equation for this process: where t is time of iteration, X t is the current position of the whale; X t gbest = (X t gbest1 , X t gbest2 , . . . , X t gbestD ) is the current best position vector of the whale, D is the vector of dimension, A· C·X t gbest − X t is the encirclement step length of the whale, and A and C are coefficient vectors.
The coefficient vectors A and C are calculated by the equations below: where the value of rand lies in the [0,1] and the value of a decreases linearly from 2 to 0. The equation is as follows: where t is number of iterations.
The humpback whale moves toward the prey position and hunts through a spiral motion, the position update equation for this process is as follows: where D = X t gbest − X t , D is the distance between whale and prey. The helix shape is given by the constant b, the range of random number l is [-1,1].
It is worth noting that humpback whales swim around their prey in a shrinking range while spiral hunting and this simultaneous behavior is the whale's bubble-net hunting behavior. Assuming that the probability p of individual whales choosing both the encircling hunting mechanism and the spiral hunting mechanism in the bubble-net hunting behavior is 50%, the behavior can be expressed as follows: where value of p is [0, 1].
In addition to the bubble-net hunting behavior, humpback whales can also perform random hunting based on each other's positions to perform random hunting, and the position update equation for this process is as follows: where X t rand is the whale position vector, and |A| > 1.

The Improved WOA(IWOA) Based on Adaptive Cloud Strategy and Catfish Effect
According to the deficiency of the WOA algorithm, we proposed an adaptive cloud strategy to modify whale hunting behavior to improve the global searchability. Furthermore, using the catfish effect to improve whale population viability can help WOA jump out of the local optimum.

Adaptive Cloud Strategy of IWOA
Traditional WOA does not consider the variability of prey forces on whale guidance during the iterative cycle. However, we can introduce inertia weight to WOA to improve the global searchability. The value of the inertia weight is based on the adaptive cloud strategy; the adaptive cloud strategy is from the cloud theory [32]. The cloud theory is characterized by uncertainty in the expression of things, stability, and simultaneous change, in line with the fundamental laws of population evolution in nature. The detailed step of the cloud theory is shown below: Suppose U is a theoretical domain expressed in terms of an exact value (it could be one-dimensional, two-dimensional or multi-dimensional), and U corresponds to the qualitative concept of B. y = e −(x−Ex) 2 2(En) 2 is called the certainty of concept B for x, and the distribution of x over U is called a cloud. The numerical characteristics of the cloud are given by the expectation E x , entropy E n and super-entropy He. They reflect the whole quantitative characteristics of qualitative concept B.
The algorithm of the normal cloud generator is as follows [33].

1.
Generate a normal random number E n , where E n is the expected value and H e is the standard deviation.

2.
Generate a normal random number x, where E x is the expect value and the absolute value of E n is the standard deviation. x is called a cloud droplet in the theoretical space.

4.
Repeat steps 1 to 3 until N cloud droplets are produced.
Then, we will explain the adaptive cloud strategy. Supposing the population size of WOA is N, and the fitness value of individual x i in the k-th iteration is f i , the mean fitness value of the population is f avg = 1 We can obtain f avg by averaging the fitness value, which is better than f avg , and obtain the f avg by averaging the fitness value, which is less than f avg . Then, we use the cloud theory to divide the population into three sub-populations. We can generate strategies to modify individual whale position of Equation (1) and spiral renewal of Equation (5) using the different inertia weight w, the modified equation is shown below: The details in generating rules of w are shown below: 1.
If f i is better than f avg , it means that the individual is close to the global optimum, so it will use a smaller inertia weight and can accelerate the global convergence speed. The value of w will be 0.9.

2.
If f i is better than f avg but worse than f avg , this group of individuals belongs to the general population, so we introduce the cloud theory to adjust the inertial weight of individual x i . The strategy [34] is as follows: He = E n /c 2 (12) where c 1 and c 2 are learning factors, c 1 decreases linearly with the increasing number of iterations, and c 2 increases linearly. In the early stage of iterative optimization, c 1 is larger and c 2 is smaller, which is good for strengthening the global search ability; in the later stage, c 2 gradually increases and c 1 gradually decreases, which is good for jumping out of the local optimum position to get the global optimum solution.
With the decrease in individual fitness value, it will ensure w ∈ [0.4, 0.9] because of Equation (15). The range of is [0,1] because of Equations (10)- (13). 3. If f i is worse than f avg , it means that this group of individuals is the worst population, so it will set the value of w to 0.4. We use the range [0.4,0.9] for w because this range helps WOA to converge faster than other ranges.

Catfish Effect of IWOA
It is necessary to ensure that the algorithm can jump out of the local optimum when premature convergence of WOA occurs. L.Y. Chuang [35] used the catfish effect of nature introduced into the particle swarm optimization to improve the ability of jumping out of the local optimum; the result of the experiment shows its effectiveness. Since all the individual whales in the population are clustered to the optimum individual position in the later iterations, this will cause the premature convergence of WOA [36]. In order to avoid the premature problem of WOA, it is necessary to ensure that the algorithm can jump out of the local optimum when premature convergence occurs. This paper introduces the catfish effect based on the adaptive cloud strategy improvement of the WOA algorithm. The idea of the catfish effect is to introduce a catfish into the sardine's group in order to help the most sardines to keep their vitality. As explained in the algorithm, supposing the global optimum solution of WOA does not evolve in a certain number of iterations, then the algorithm will automatically reinitialize 10% of the particles with the worst result (no active sardines) to recover the group vitality. The strategy of the catfish effect is to improve the breadth search capability of WOA.
In this section, we propose that when the global optimum solution of WOA does not evolve in a certain number of iteration processes, it will proceed with the mutation equation to re-initialize 10% of the worst particles of the whale population. The mutation equation is shown below: where η is a random variable subject to a Gaussian distribution with range of (−1, 0). The pseudocode of the modified WOA algorithm (IWOA) is shown in Algorithm 1 below, and Figure 1 shows the flow chart of IWOA. In this paper, the experimental environment was a 3.40 GHz Intel(R) Core TM i7-4750QM CPU with RAM of 16GB in the Windows 10 system. Each of the experiments was  else if2(|A| > 1) 6. Select a random search agent (Xrand) 7. Update position of search agent use Equation (7). end if2 else if1 (p > 0.5) 8. Update the position with spiral Equation (9). end if1 9. Check if any search agent goes beyond the search space and amend it. Calculate the fitness of each search agent. 10. Update X gbest if there is a better solution and record X gbest and its iteration.
If3 X gbest did not evolve in five iterations, then rank the search agent fitness from best to worst, and initialize the position of 10% of the worst search agents with Equation (16) end if3 end for t = t + 1 end while return X gbest

Experiment Environment and Parameter Setting
In this paper, the experimental environment was a 3.40 GHz Intel(R) Core TM i7-4750QM CPU with RAM of 16GB in the Windows 10 system. Each of the experiments was repeated 30 times independently on the Matlab 2019.
In order to verify the effectiveness and generalization of the improved algorithm, we compare the IWOA with WOA [14], PSO [11], GWO [12], HWOA [31] and MPA [13] on 15 benchmark functions [37], where F1-F5 are unimodal benchmark functions. From F6 to F10 are multimodal benchmark functions, and F11 to F15 are fixed-dimension multimodal benchmark functions. Those variably shaped test functions can effectively check the optimization performance of algorithms. Since the unimodal benchmark function has only one global optimum, the result of the unimodal benchmark function will reflect the algorithm's exploitation capability. The multimodal benchmark function and fixeddimension multimodal benchmark functions have many local optimums and one global optimum. The result of those benchmark functions will reflect the algorithm's ability to jump out of the local optimum and its exploration capability.
The detail information of benchmark functions is shown in Table 1. Table 1 includes the cost function, range of variation of optimization variables, optimum value f min and the design variable, dim. The details of the algorithm parameter settings are in Table 2.  Table 2. Details of parameter settings. This experiment will compare the speed of convergence and optimization accuracy of six algorithms on the benchmark function at the same population size, number of iterations, and number of runs. So, it sets 30 population sizes for each algorithm, the maximum iteration is 1000, and every algorithm will run independently 30 times, each. We use the mean, standard deviation, and maximum and minimum values of the optimum target value as the basis for performance evaluation in this experiment. The result is shown in Table 3. The convergence curves of the six algorithms show in Figure 2. The box line diagram of six algorithms is shown in Figure 3. The result of the Wilcoxon Rank Sum Test is in Table 4.    From the convergence curve of Figure 2, we observe that the proposed IWOA converges significantly faster than the other five algorithms on F1-F9 and F12-F14. IWOA obtains the global optimum on F1-F4, F7, and F9. The convergence speed of HWOA is faster than WOA on most functions. Furthermore, the convergence speed of IWOA always keeps ahead of HWOA. This means that the improved strategy of WOA is more successful than HWOA. PSO has the best convergence speed and the results compare with other algorithms on F10. Although the performance of IWOA is worse than PSO on F10, IWOA still has a faster convergence speed and better results than HWOA and WOA on F10. The result of F10 shows that WOA still has limitations on the multimodal benchmark function. However, the improvement strategies help IWOA to enhance the ability of WOA to jump out of the local optimum on the multimodal benchmark function. The convergence curve of IWOA proves that the ability of IWOA to jump out of the local optimum is effectively enhanced compared to other algorithms on most benchmark functions. Therefore, the IWOA algorithm has a more vital ability to jump out of the local optimum than the other five algorithms. Table 3 shows the optimization results of 15 benchmark functions. The proposed IWOA algorithm gives the best results of 15 benchmark functions. It can be seen that the IWOA algorithm is significantly better than the other five algorithms in terms of the mean and standard deviation of the optimum solution.

Analysis of Exploitation Capability
Functions F1-F5 are unimodal functions, and they have only one global optimum. The result of Table 3 shows that the IWOA is very competitive compared to the other five algorithms. Functions F6-F10 evaluate the exploration capability of an optimization algorithm. The result shows that the IWOA is better than the other five algorithms, except for F10. Functions F11-F15 are fixed-dimension multimodal functions. IWOA provides the best result on F11, F13, F14, and F15. IWOA is still competitive in F12 compared with HWOA and WOA.
Combining the results of Sections 2.3.2 and 2.3.3 shows that the IWOA outperforms the other five algorithms in convergence speed, exploitation capability, and exploration capability on most functions.

Algorithmic Stability Analysis
To more visually show the stability of IWOA compared with another algorithm, we use a box line diagram to show each algorithm's result on different functions after 30 independent runs. Figure 3 shows the result of the box line diagram.

Wilcoxon Rank Sum Test Analysis
In this section, we use a statistical test to evaluate the performance of IWOA in comparison with other algorithms because too many factors will influence the result of the meta-heuristics algorithm. The most common test is the Wilcoxon Rank Sum Test [38]. In the Wilcoxon Rank Sum Test, the value of the p-value is used as the evaluation criteria. Suppose the p-value is greater than 5%; in that case, it means that IWOA is not statistically different on this function compared with other algorithms. If the p-value is equal to NAN, it means that both algorithms achieve a global optimum. The result of the Wilcoxon Rank Sum Test is shown in Table 4.
In Table 4, it can be seen that HOWA and IWOA have no statistical difference on F1, F2, F3, F4, F7, F8, and F9 because they all achieve the theoretical global optimum. IWOA and WOA have the result of NAN on F9 because they achieve the theoretical optimum. IWOA and GWO have no statistical difference on F9 and F12. IWOA and MPA have the result of NAN on F7 and F9. In general, the Wilcoxon Rank Sum Test results show that IWOA is statistically different from the other five algorithms on most benchmark functions. This means that IWOA has better exploitation and exploration capability compared with the other algorithms.

Analysis of Convergence Behavior
The convergence curves of IWOA, WOA, PSO, MPA, GWO and HWOA are provided in Figure 2 to see the convergence rate of the algorithms.
From the convergence curve of Figure 2, we observe that the proposed IWOA converges significantly faster than the other five algorithms on F1-F9 and F12-F14. IWOA obtains the global optimum on F1-F4, F7, and F9. The convergence speed of HWOA is faster than WOA on most functions. Furthermore, the convergence speed of IWOA always keeps ahead of HWOA. This means that the improved strategy of WOA is more successful than HWOA. PSO has the best convergence speed and the results compare with other algorithms on F10. Although the performance of IWOA is worse than PSO on F10, IWOA still has a faster convergence speed and better results than HWOA and WOA on F10. The result of F10 shows that WOA still has limitations on the multimodal benchmark function. However, the improvement strategies help IWOA to enhance the ability of WOA to jump out of the local optimum on the multimodal benchmark function. The convergence curve of IWOA proves that the ability of IWOA to jump out of the local optimum is effectively enhanced compared to other algorithms on most benchmark functions. Therefore, the IWOA algorithm has a more vital ability to jump out of the local optimum than the other five algorithms. Table 3 shows the optimization results of 15 benchmark functions. The proposed IWOA algorithm gives the best results of 15 benchmark functions. It can be seen that the IWOA algorithm is significantly better than the other five algorithms in terms of the mean and standard deviation of the optimum solution.

Analysis of Exploitation Capability
Functions F1-F5 are unimodal functions, and they have only one global optimum. The result of Table 3 shows that the IWOA is very competitive compared to the other five algorithms. Functions F6-F10 evaluate the exploration capability of an optimization algorithm. The result shows that the IWOA is better than the other five algorithms, except for F10. Functions F11-F15 are fixed-dimension multimodal functions. IWOA provides the best result on F11, F13, F14, and F15. IWOA is still competitive in F12 compared with HWOA and WOA.
Combining the results of Sections 2.3.2 and 2.3.3 shows that the IWOA outperforms the other five algorithms in convergence speed, exploitation capability, and exploration capability on most functions.

Algorithmic Stability Analysis
To more visually show the stability of IWOA compared with another algorithm, we use a box line diagram to show each algorithm's result on different functions after 30 independent runs. Figure 3 shows the result of the box line diagram.
From Figure 3, it can be seen that IWOA is more stable than other algorithms on F1, F2, F3, F4, F6, F7, F8, F9, F10, F11 and F13. Although IWOA is less stable than MPA on F12, F14, and F15, the average fitness of IWOA on F12, F14, and F15 is better than MPA. Overall, IWOA has more stability and better performance than the other algorithms on most functions.

Wilcoxon Rank Sum Test Analysis
In this section, we use a statistical test to evaluate the performance of IWOA in comparison with other algorithms because too many factors will influence the result of the meta-heuristics algorithm. The most common test is the Wilcoxon Rank Sum Test [38]. In the Wilcoxon Rank Sum Test, the value of the p-value is used as the evaluation criteria. Suppose the p-value is greater than 5%; in that case, it means that IWOA is not statistically different on this function compared with other algorithms. If the p-value is equal to NAN, it means that both algorithms achieve a global optimum. The result of the Wilcoxon Rank Sum Test is shown in Table 4.
In Table 4, it can be seen that HOWA and IWOA have no statistical difference on F1, F2, F3, F4, F7, F8, and F9 because they all achieve the theoretical global optimum. IWOA and WOA have the result of NAN on F9 because they achieve the theoretical optimum. IWOA and GWO have no statistical difference on F9 and F12. IWOA and MPA have the result of NAN on F7 and F9. In general, the Wilcoxon Rank Sum Test results show that IWOA is statistically different from the other five algorithms on most benchmark functions. This means that IWOA has better exploitation and exploration capability compared with the other algorithms.

Semi-Supervised Support Vector Machine
The most prominent representative of semi-supervised support vector machines (S3VM) is the Transductive Support Vector Machine (TSVM) [39]. In this paper, TSVM was introduced as S3VM. The key concept is to find appropriate mark assignments for unlabeled samples to optimize the interval after the hyperplane division. S3VM uses a local search technique to solve the problem iteratively. The detail of S3VM is in using a labeled sample collection to train an initial SVM, then using the learner to mark unlabeled samples so that all samples are labeled; based on these labeled samples, the SVM is retrained and then it searches for continuous adjustment of error-prone samples until all unlabeled samples are marked.
S3VM is also a learning method for the two-class problem, just like the regular SVM. S3VM tries to consider different potential unlabeled sample label assignments, such as trying to treat each unlabeled sample as a positive or negative example, and then looking for a dividing hyperplane in all of these results that maximizes the spacing on all samples (including labeled samples and marking the assigned unlabeled sample). Once the divisional hyper-plane is calculated, the unlabeled sample's final marker designation is its labels [40].
There are labeled data D l = {(x 1 , y 1 ), (x 2 , y 2 ), . . . , (x l , y l )} and unlabeled data D u = {(x l+1 , x l+2 , . . . , x l+u }, and y i ∈ {−1, +1}, l ≤ u, l + u = m. The purpose of S3VM is to predict the label where (ω, b) is determined as a divide plane, ξ i is the slack variable, C l and C u are trade-off control parameters, the kernel function of S3VM is radial basis kernel function (RBF), . Kernel parameters γ and regularization parameters C are the parameters of RBF.

IWOA-S3VM Model
In Section 2, we improved WOA with the catfish effect and the adaptive cloud strategy. The results of IWOA showed that the improvement strategy was successful. We explained the basic idea of S3VM in Section 3. Since the kernel parameters of S3VM need to be specified artificially, but the kernel parameters' values affect the classification accuracy of S3VM, we use IWOA to optimize the kernel parameter selection of S3VM. The pseudo code of IWOA-S3VM are shown in Algorithm 2.
Based on ∧ y, D l , D u , C l and C u to solve formula (17) to get (w, b), ζ 8. End while C u = min{2C u , C l } 9. End while. 10. End.

Experiment and Result Analysis of IWOA-S3VM
In order to test the classification effect of the algorithm, it chooses five representative UCI datasets with which to run the experiment. Details of these datasets are shown in Table 5. In the experiment, firstly, each dataset is randomly divided into two parts. As shown below in Table 5, one of the total number of samples are used as the test set T. The remaining part of the data is training set L. In this part, we choose SVM and S3VM compared with IWOA-S3VM; the labeled sample L was 30% and 40% of the total training set. The results are shown in Tables 6 and 7.  Tables 6 and 7 show that our algorithm has better classification accuracy than the other two algorithms. It means that the intelligent algorithm can improve the S3VM model classification accuracy. The result also shows that the accuracy of the IWOA-S3VM model will increase with the increase in labeled samples.

Design Model of Oil Layer Recognition
A block diagram of the oil layer recognition model system based on IWOA-S3VM is shown in Figure 4. The primary step of the oil layer recognition system is shown below.   7 show that our algorithm has better classification accuracy than other two algorithms. It means that the intelligent algorithm can improve the S3VM mo classification accuracy. The result also shows that the accuracy of the IWOA-S3VM mo will increase with the increase in labeled samples.

Design Model of Oil Layer Recognition
A block diagram of the oil layer recognition model system based on IWOA-S3VM shown in Figure 4. The primary step of the oil layer recognition system is shown belo  (1) Sample information collection and pre-possessing: Due to the vast volume of log ging information, it is necessary to pick sample information. The data directly co nected to the oil layer details will be chosen, and the data will be separated randomly into the training dataset and the evaluation dataset. In order to prevent sa ration of the measurement, all data will be normalized.
(2) Attribute generalization: We set the judgment attribute in this step D ={ d }, (1) Sample information collection and pre-possessing: Due to the vast volume of logging information, it is necessary to pick sample information. The data directly connected to the oil layer details will be chosen, and the data will be separated randomly into the training dataset and the evaluation dataset. In order to prevent saturation of the measurement, all data will be normalized.

Practical Applications
In this section, we are using two wells of actual oil layer datasets (W1 and W2) to test the performance of the IWOA-S3VM model. The detailed information of the two oil layers is shown in Table 8. Table 9 shows the original attribution and reduction attribution of two oil layer datasets. Figure 5 shows details of the normalized attributes of W2. (1 represent oil layer).
(3) Attribute reduction: In the well-logging data, there are typically at least ten kinds of well-logging information, but in most of the data, there are redundant and unnecessary attributes. In this article, we use the rough set [41] to reduce the attribute. (4) IWOA-S3VM model: Train the IWOA-S3VM model by inputting the oil layer data. (5) Output the IWOA-S3VM oil layer recognition model output when inputting the entire oil layer test results.

Practical Applications
In this section, we are using two wells of actual oil layer datasets (W1 and W2) to test the performance of the IWOA-S3VM model. The detailed information of the two oil layers is shown in Table 8. Table 9 shows the original attribution and reduction attribution of two oil layer datasets. Figure 5 shows details of the normalized attributes of W2.    We used RMSE and MAE as additional performance metrics to evaluate the effectiveness of the IWOA-S3VM model. The equation is shown below: where f (x) is the predicted value and y is the actual value. RMSE reflects the deviation between the predicted value and the true value. MAE reflects the average value of absolute errors. Smaller values of MAE and RMSE represent the better model performance.
To test the performance of the proposed IWOA-S3VM model on semi-supervised tasks, we used 30% of the labeled data of training datasets to train the classification model. We set the maximum iteration time of IWOA as 100 times. The classification result is shown in Table 10.  Table 10 shows that the intelligent algorithm can increase the S3VM model's accuracy. Compared to the other two algorithms, the IWOA-S3VM has better accuracy on W1 and W2. However, IWOA-S3VM uses more time compared with S3VM and SVM. The reason why IWOA takes more time is because IWOA requires several iterations to find the suitable optimization parameters of the S3VM.

Result of Oil Layer Recognition
In oil layer recognition, IWOA-S3VM can cope with large numbers of unlabeled datasets. The result of Table 10 shows that the S3VM model can be improved successfully by IWOA.
The actual oil layer distribution and oil layer distribution observed by the IWOA-S3VM model with the S3VM model are shown in Figure 6. The result in Figure 6 shows that the IWOA-S3VM model can more easily identify the oil layer distribution correctly when there is a quantity of unlabeled data in the oil layer datasets.
RMSE reflects the deviation between the predicted value and the true value. MAE reflects the average value of absolute errors. Smaller values of MAE and RMSE represent the better model performance.
To test the performance of the proposed IWOA-S3VM model on semi-supervised tasks, we used 30% of the labeled data of training datasets to train the classification model. We set the maximum iteration time of IWOA as 100 times. The classification result is shown in Table 10.  Table 10 shows that the intelligent algorithm can increase the S3VM model's accuracy. Compared to the other two algorithms, the IWOA-S3VM has better accuracy on W1 and W2. However, IWOA-S3VM uses more time compared with S3VM and SVM. The reason why IWOA takes more time is because IWOA requires several iterations to find the suitable optimization parameters of the S3VM.

Result of Oil Layer Recognition
In oil layer recognition, IWOA-S3VM can cope with large numbers of unlabeled datasets. The result of Table 10 shows that the S3VM model can be improved successfully by IWOA.
The actual oil layer distribution and oil layer distribution observed by the IWOA-S3VM model with the S3VM model are shown in Figure 6. The result in Figure 6 shows that the IWOA-S3VM model can more easily identify the oil layer distribution correctly when there is a quantity of unlabeled data in the oil layer datasets.

Conclusions
In this paper, we proposed an IWOA-S3VM model to better utilize a large number of unlabeled samples from oil layer datasets for oil layer recognition. Firstly, IWOA is proposed to address the problems of WOA, which include getting easily trapped in the local optimum, and low convergence accuracy. We used two strategies to improve the performance of WOA; one is to use the adaptive cloud strategy to improve the global search ability of WOA, and the second is to use the catfish effect to help the WOA to jump out of the local optimum and maintain population diversity. Among the 15 benchmark functions, IWOA shows superiority over the other five algorithms. Secondly, to address the problem that the selection of kernel parameters in S3VM affects its classification accuracy, we use IWOA to optimize the selection of its kernel parameters. The classification results of the five UCI datasets show the superiority of the IWOA-S3VM model. Finally, the oil layer identification results show that IWOA-S3VM performs better than the other two algorithms when using a small number of labeled data. For future works, we will try to use other semi-supervised classification models for oil layer identification.