1. Introduction
In recent years, the global manufacturing industry has been undergoing a profound transformation with the advent of Industry 5.0. Unlike its predecessor, Industry 4.0, which focused primarily on automation, digitalization, and the integration of cyber–physical systems, Industry 5.0 emphasizes the collaboration between humans and intelligent systems, aiming to create more resilient, sustainable, and human-centric manufacturing processes. This new paradigm leverages advanced technologies, such as artificial intelligence (AI), big data analytics, the Internet of Things (IoT), and robotics, not only to enhance productivity and efficiency but also to enable flexible, adaptive, and personalized solutions across various sectors of the manufacturing industry Van Erp, T et al. (2024) [
1].
As a critical component of the manufacturing sector, the oil and gas industry is also embracing the principles of Industry 5.0 to address the increasing complexity and challenges in reservoir management and production optimization. By integrating AI-driven approaches with traditional engineering expertise, the industry is moving towards intelligent decision-making and autonomous operations, which are essential for achieving higher efficiency, sustainability, and competitiveness in the era of smart manufacturing Zhang G.Y. et al. (2024) [
2]; Zhao B. et al. (2025) [
3]. In this context, integrating artificial intelligence and deep learning methods has become a highly promising direction in the field of reservoir engineering. These alternative models and intelligent optimization algorithms developed based on artificial intelligence technology can quickly and accurately predict the production dynamics of oil and gas reservoirs, enhance intelligent decision-making, and thereby achieve rapid improvements in productivity and production efficiency. This will accelerate the progress of industrialization 5.0 in the petroleum industry.
In the development of a surrogate model, Zhang X.D. et al. (2018) [
4] generated artificial logging curves based on existing logging curves from a machine learning method, which achieved good application results. However, these purely data-driven methods have inherent limitations in terms of physical consistency and extrapolation. Subsequently, the research moved towards more complex dynamic predictions and geological characterizations. Zhu Y. and Zabaras N. (2018) [
5] constructed a model for predicting two-dimensional fluid pressure and velocity using Bayesian and Dense Block methods. Mo et al. (2019) [
6] used the Dense Block model to solve the problem of pressure and saturation changes when CO
2 was stored underground, but this model has a poor extrapolation effect on time steps. In order to enhance the accuracy of geological characterization, Liu Y. et al. (2019, 2010) [
7,
8] proposed the CNN-PCA model to reduce the geological dimension. By reducing the dimension of input geological maps with a convolutional neural network, the model overcomes the shortcoming that PCA cannot fully reflect geological characteristics. However, this method was only tested in river flows and lacks an analysis of its actual effectiveness. Zhu Y. et al. (2019) [
9] solved the steady-state flow problem in random homogeneous media with a convolutional neural network. Based on big data and machine learning, Jia D. et al. (2020) [
10] put forward a set of optimization methods for a fine water injection scheme driven by big data, which made cumulative oil production increase by 7.2%. Shahkarami A. and Mohaghegh S. (2020) [
11] used artificial intelligence and machine learning techniques to develop an intelligent surrogate model for reservoir history fitting, sensitivity analysis, and uncertainty assessment. This model has great advantages in computing speed, time consumption, and cost. Kim J. et al. (2020) [
12] sought the optimal well location design with convolution. These efforts collectively demonstrate the significant advantages of AI methods in enhancing computational efficiency and decision-making speed. However, the models employed are relatively simple, and their practical application value is limited.
With the development of deep learning technology, the research focus has shifted to utilizing more powerful network structures to achieve efficient end-to-end simulation and prediction. Tang M. et al. (2020) [
13] simulated the changes in pressure and saturation distribution of reservoirs of different permeabilities over time with the R-U-Net model and calculated well responses using the Peaceman formula. The R-U-Net model takes 2 s to predict the state of a reservoir at different moments. However, its accuracy is highly dependent on the completeness of the training data. Wang H.L. et al. (2020) [
14] predicted the production of some oilfields in the ultra-high water cut period by using a deep learning method effectively. Zhong Z. et al. (2020) [
15] used a generative adversarial neural network to simulate the change in oil field saturation over time in waterflood development. But this model shows significant errors in capturing the abrupt change in water production wells, reflecting the difficulties of generative models in restoring dynamic details. Gao Z.K. et al. (2021) [
16] used a deep learning method based on complex convolutional networks to characterize the gas–liquid flow. Kuang L. et al. (2021) [
17] summarized and prospectively reviewed the research of artificial intelligence in the field of oil and gas exploration. Ma C.J. (2021) [
18] used an LSTM model to simulate oil production and obtained good results. Chaki S. et al. (2021) [
19] predicted variations in field information (pressure or saturation) and the production of oil reservoirs with a U-Net network, saving time and costs. Cornelio J. et al. (2021) [
20] transferred a model trained on one oilfield’s data to another oilfield through transfer learning, thus reducing the data requirements and the workload of building new models. Based on the U-Net, Zhong Z. et al. (2021) [
21] proposed generating an antagonistic neural network (Co-Gan) by coupling to predict the changes in reservoir pressure and fluid saturation in the process of water injection in heterogeneous reservoirs. The inputs of the model include the reservoir static properties (permeability), injection rate, and time to be predicted, and the outputs include the dynamic state of the reservoir corresponding to the predicted time. Alakeely A. and Horne R. (2022) [
22] constructed surrogate models with historical production data using the generative deep learning (GDL) method to predict the production of new wells. Relying on more than 40 million sets of historical dynamic monitoring data covering different types of oil wells, Wang X. et al. (2022) [
23] prepared oil well condition diagnosis sample sets covering 37 types of working conditions in 5 major categories. On this basis, the convolutional neural network algorithm was selected, and the intelligent diagnosis technology was constructed by using the new generation of artificial intelligence technology of “big data + artificial intelligence”. Using this technology, they completed condition diagnosis 5 million times in the field, with an accuracy of 90% and timely alarm pushing. Yang Y. (2022) [
24] combined GNN with seepage physical process information to establish a graph neural network model suitable for the injection and production balance training of a well pattern. Huang H. et al. (2023) [
25] also used GNN for reservoir simulation and optimization with varying well controls.
In summary, the intelligent surrogate models examined in current research exhibit a pronounced path dependence in their development: their core architectures are largely derived from computer vision (e.g., CNNs, U-Net, GANs), emphasizing the learning of spatiotemporal patterns from data. Yet they generally fail to rigorously embed the reservoir governing equations—capturing the essence of fluid flow—as hard physical constraints, and consequently their predictions often lack robust physical consistency and display a limited extrapolative capability. Meanwhile, most reported advances are grounded in simplified, small-scale theoretical models, and their efficiency and accuracy remain insufficiently validated in real industrial contexts characterized by complex geology, long-term production, large-scale well grids, and comprehensive oilfield management. This highlights a substantial disconnect between theoretical research and large-scale practical application.
In the application of the flow equation, many processing methods based on deep learning models are quite different from those of reduced order models (ROMs) based on partial differential equations and numerical discretization. In the study of reduced-order models, Watter M. et al. (2015) [
26] proposed an embedded control framework (E2C) to predict the evolution of the system state with sensor data (images) and control conditions. Jin Z.L. et al. (2020) [
27] constructed an end-to-end model based on E2C to realize the prediction of production with the changing control conditions of wells in the reservoir. Firstly, the model uses an AE encoder to reduce the dimension of the input data and extract features; then, according to the control conditions of wells, the fully connected converter is used to obtain new features; finally, the distributions of saturation and pressure at the new time are obtained by the decoder. Although the E2C model proposed by Jin et al. works well in simple reservoirs, it is poor in accuracy in large-scale reservoirs and cannot give production information directly. Qin F. et al. (2022) [
28] used the E2C model to simulate the gas reservoir problems with small wells. Huang H. et al. (2023) [
29] optimized the E2C model proposed by Jin to improve its performance and added the prediction of production evolution to it. In the optimized model, skip connection is used to improve the accuracy of the decoder, while the converter is modified to directly calculate production information. Although the model has been verified in some reservoirs, the reservoirs used in the verification were small, short in production period, and simple in well production scheme.
In this work, by leveraging actual production data from a complex oil field, we utilized an Embed-to-Control (E2C) model to predict reservoir pressure, saturation, and well response dynamics. Subsequently, we employ the trained surrogate model to optimize water injection volumes for enhanced oil production. For the first time, a fully integrated deep learning workflow combining model construction, training, and optimization decision-making is used within a realistic, complex reservoir setting that contains frequent injection-to-production conversions. Furthermore, we innovatively enhanced the E2C architecture by optimizing the production calculation network using residual connections, specifically overcoming original network limitations in accurately simulating intricate well control dynamics like injection–production conversion. The application of this optimized surrogate model integrated with optimization algorithms demonstrably improved daily oil production by 13.84%, significantly boosting operational efficiency. Finally, this work exemplifies the surrogate model’s high practical value in Industry 5.0, providing a validated pathway for AI-driven intelligent decision-making and efficiency gains within industrial production systems.
2. Theoretical Basis
Based on the mass conservation law of each component, and combined with Darcy’s law, the equation governing immiscible oil–water flow is derived as follows:
In Equation (1), the subscript
(
for oil and water, respectively) denotes the phase of the fluid,
and
denote porosity and permeability, respectively,
denotes the interaction between the rock and fluid,
and
represent pressure and saturation, respectively,
represents density,
is the source/sink terms for the well
,
represents gravitational acceleration, and
represents depth. The variable
denotes the state vector of the calculated area at a specific time step
,
is the number of grids in the model,
denotes the control conditions of the wells from time
to time t + 1, and
indicates the number of all wells in the system. Solving the governing Equation (1) with a fully implicit finite volume method, the following equation is obtained:
where
is a function, the subscript t indicates the current moment, and t + 1 indicates the next moment. The full-order discrete nonlinear system defined by Equation (2) is usually solved using the Newton method, so each iteration needs to construct a sparse Jacobian matrix of
dimensions, and the solution of these matrices is highly time-consuming, so how to reduce the dimensions of high-dimensional systems has become a major research direction.
He J. and Durlofsky L.J. (2014, 2015) [
30,
31] proposed a method combining orthogonal decomposition and trajectory piecewise linearity (POD-TPWL) to solve the above equation. In this method, Equation (2) is first Taylor expanded to give the following:
The subscript i + 1 represents some known solutions (which can be regarded as a training dataset used to determine some parameters), and t + 1 represents a new example to be calculated. For details on the POD-TPWL’s implementation, see He J. and Durlofsky L.J. (2014, 2015) [
30,
31].
By means of orthogonal decomposition, the high-dimensional state matrix
can be represented by the low-dimensional variable
, namely
where
denotes the dimension of the lower dimensional space, satisfying
. Because
is orthogonal, it is deduced that
, and thus the mutual transformation between high-dimensional space and low-dimensional space is realized.
In the POD-TPWL method, by using Newton iteration, Equation (3) can be expressed as follows:
where
,
, and
. Substituting Equation (4) into Equation (5), we obtain
where
,
, and
is a constraint matrix. Since
is also unknown in the iterative calculation, it is replaced by
approximately, and the final result can be expressed as follows:
Suppose
,
,
, and
; then, Equation (7) can be expressed as follows:
where
represents the characteristic of the low-dimensional space at the current moment,
represents the characteristic of the low-dimensional space at the new moment, and
represents the control conditions of the well under the current production state;
and
are two matrixes related to the current characteristic and time interval. Based on Equation (8), the features at different moments in the low-dimensional space can be iterated, and then the solution in the high-dimensional space can be obtained through the inverse transformation of Equation (4).
It can be seen from the above introduction that POD-TPW generally includes three steps: the projection from a high-dimensional space to a low-dimensional subspace; a dynamic linear approximation in the low-dimensional subspace; and mapping from the low-dimensional space to the high-dimensional space. Each of these steps involves lots of matrix calculations. In order to simplify the calculation method and improve the calculation rate, deep learning is used to replace the matrix solution in each step, and finally the E2C model is generated.
3. Scheme Design
The purpose of a surrogate model is to quickly predict production dynamics based on given inputs. In the construction of the surrogate model, we first need to establish several physical models under different input conditions and then calculate these physical models using a numerical simulation method to extract the required production dynamic data; finally, based on these datasets, the deep learning method is used to obtain the mapping relationship between the input parameters and the production dynamics. The deep learning model can automatically detect useful features from the data and obtain the mapping from input to output. The calculation process can be expressed as follows:
where
is the surrogate model,
is the input variable,
is the output (pressure, saturation, well production, etc.), and
is the deep neural network parameter, which needs to be determined in training.
3.1. Model Design
Based on deep learning, the embedded control framework (E2C) was used to predict the pressure, saturation, and well production information in the process of reservoir development in this study. The model was first proposed by Watter et al. and successfully applied to the robot control system. Jin et al. extended the E2C model in 2020 and applied it to reservoir simulation. However, the accuracy of the model is low in actual oilfields. Huang et al. modified the model in 2023 and improved its accuracy in actual oilfields. But the oilfield used to verify the model then had a short production duration and a small number of wells, and even these wells occupied a large proportion of the video memory. In this study, the E2C model proposed by Huang et al. was optimized so that it can complete the development prediction of large-scale complex oilfields with a smaller occupancy rate of the video memory.
The E2C model mainly uses the convolutional neural network to extract features and calculate fields. The convolutional neural network is composed of one or more convolutional layers and active layers, and the specific calculation method is as follows:
where
represents the activation function,
represents the output,
represents the input, and
represents the layers of the neural network.
and
denote weights in the convolution calculation and need to be determined using an optimization algorithm in the training process.
In the proper orthogonal decomposition–piecewise linearization of trajectory (POD-TPWL) principle, solving the reservoir control equation under different well control conditions mainly includes three steps: down-sampling the original space through orthogonal decomposition, transforming the down-sampled features into features at a new moment through linear transformation, and transforming the features at the new moment into the original space through the inverse transformation of orthogonal decomposition. Similarly to the solving process of POD-TPWL, the E2C model is mainly divided into three parts: (1) an encoder that projects data from a high-dimensional space to a low-dimensional space; (2) a converter that performs feature conversion in the low-dimensional space; and (3) a decoder that performs well production performance prediction and pressure and saturation calculation from the converted low-dimensional space. The specific structure is shown in
Figure 1. In the surrogate model, the encoder is used to down-sample the current feature, the converter is used to evolve the time, and the decoder is used to recover a new feature in the spatial dimension.
It can be seen from
Figure 1 that the inputs of the surrogate model are the pressure distribution, saturation distribution, and well control conditions at the current moment (
), and the outputs are the pressure, saturation, and well production information at the next moment (
). The figure uses color-coding to differentiate module types: pink denotes the convolutional block (Conv Block), light green the dense block (Dense Block), cyan-blue the convolution operator (Conv), light purple the fully connected layer (FC), yellow the fully connected block (FC Block), red the latent features (e.g.,
and
), dark purple the deconvolutional block (DeConv Block), and gray the well-response output (Well Response). The long red lines represent skip connections from the encoder to the decoder, which inject high-resolution detail and thereby enhance reconstruction quality.
In the encoder, three convolutional blocks—each comprising a convolutional layer, batch normalization, and PReLU (denoted as a Conv Block in
Figure 1)—are first applied with a stride of 2 and a 3 × 3 × 3 kernel to extract features from the input information. Then, the Dense Block is used to extract features in the low-dimensional space. The Dense Block is a special convolution structure which still uses convolution operation, but the output is composed of the result of this calculation and the input. Please see the paper of Huang G. et al. [
32] in 2017 for more details of the calculation of the Dense Block. Then, a Conv operation is used to fuse the channels to obtain new features,
. Finally, global average pooling (GAP) and full connection (FC) are used to obtain the features,
, for subsequent calculations.
In the converter (from
to
), the feature is updated in the same manner as in Equation (8), where
The variables are the weights to be trained, and denotes the converter. According to Equations (11) and (12), in the converter, the variable is first calculated from the feature using three fully connected blocks (FC Blocks). Similarly to the convolution block, all fully connected blocks consist of “FC + BN + PReLU”. The matrices and are then obtained using two different FCs, respectively. Finally, according to Equation (8), the characteristic of the next moment is generated with the control conditions and characteristic of the well as inputs.
In the decoder, two groups of network structures are used to obtain information on the distributions of pressure and saturation and well production, respectively. First, on the basis of
, the production information of the well is calculated using one FC, one FC residual block, one FC Block, and C (the bottom part of the decoder,
in
Figure 1). As the oil field in this study has wells larger in number and complexity, three fully connected layers and residuals are used in the production calculation to increase the nonlinearity of the well production evolution. Then, based on the same
, the pressure and saturation distributions at the new moment are obtained through a series of operations (the upper right part of
in
Figure 1). In the evolution of pressure and saturation, FC is used first to expand the features and rearrange them according to the dimension of
Since the spatial information is obtained through feature rearrangement, the accuracy is not that high. Therefore, the skip connection is used to rearrange features to improve their accuracy (red arrows in
Figure 1). One Conv and two Dense Blocks are then used to compute the features, and three DeConv blocks are used to up-sample the spatial dimensions. Finally, the pressure and saturation distributions at the new moment are obtained by reorganizing the channels through a Conv.
In the above structure, the parameter correction activation function (ParametricRectified Linear Unit, PReLU) is used to make nonlinear changes to the model, and the calculation method is as follows:
It can be seen from the above calculation that, when the input variable , the gradient of PReLU is , thus avoiding the phenomenon that the gradient is 0 on a negative number and solving the problem that the gradient is not updated. Slope can also be learned in model training to reduce the trouble caused by artificial settings.
3.2. Design of Loss Function and Training Hyperparameters
In the model training, the total loss function is composed of pressure loss, saturation loss, and well production loss. The calculation formulas of the losses are Equations (14)–(16).
In the above equations, the variables with a “hat” represent predicted values, and the variables without a “hat” represent true values, where represents the saturation and pressure, respectively, and represents the oil production rate and water production rate of the well, respectively. The variable indicates the number of all grids and indicates the total number of production wells. The variable represents the final loss and represents the weight.
4. Test and Analysis of the Model
Taking a typical low-permeability water drive reservoir in Northwest China as an example, the performance of the E2C model has been analyzed using three aspects: the example background, the analysis of model results, and optimization and tests.
4.1. Design of the Test Example
The reservoir model has 1,312,830 (503 × 435 × 6) grids and 478 wells. The permeability distribution in the X direction and some properties of the reservoir are shown in
Figure 2 and
Table 1.
The development period of the block is from 1 December 2004 to 1 April 2019, and the control conditions of the well were revised once a month. The control condition of the injection wells was a constant daily water injection rate, and the condition of the production wells was a constant daily liquid production rate. As time went on, new wells were added, and the number of production wells changed accordingly, so the flow in the reservoir became more and more complex, adding difficulty to the simulation of the surrogate model.
For this reservoir, 300 training datasets were first generated by the simulator, of which 270 were used to train the model and 30 were used to test the model. In the prediction stage, the trained model was loaded first. Then, the pressure and saturation of the zero month (initial time) were taken as inputs, and the distributions of pressure, saturation, and production of the first month were obtained according to the control conditions of the first month. Then, the pressure and saturation of the first month obtained from E2C were used as inputs, and the pressure, saturation, and dynamic response of the wells in the second month were obtained according to the control conditions of the second month. Finally, the pressure, saturation, and well production-related information at all the time steps were obtained by continuously calling the model according to the above method. The specific calculation process is shown in
Figure 3.
Based on the loss obtained from Equation (16), the optimal hyperparameters determined using the random grid search algorithm were as follows: a learning rate of 2 × 10−4, the optimizer Adam, and the learning rate decaying once every 10 epochs at a decay coefficient of 0.6, pressure loss weight , saturation loss weight , weight , and weight of water production loss . The model was trained by setting it to the above hyperparameters.
4.2. Analysis of Model Test Results
Firstly, based on the pressure and saturation on 1 December 2004, changes in the reservoir at each month in the 172 months were predicted in turn according to the control conditions in the corresponding months. Then, the performance of the E2C model was analyzed.
Based on the examples randomly selected from the test set, the pressure and saturation changes in the 30th, 60th, 90th, 120th, 150th, and 170th months were simulated. The results are shown in
Figure 4 and
Figure 5.
Figure 4 and
Figure 5 show the simulation results of the E2C model and the numerical simulator for the randomly selected examples, respectively, where the E2C simulation results are in the first row and the simulation results of the numerical simulator are in the second row. It can be seen from
Figure 4 and
Figure 5 that the results of the E2C model are basically consistent with the results obtained using the numerical simulator in terms of the predicted pressure and saturation.
In order to quantitatively analyze the accuracy of the pressure and saturation prediction of the E2C model from a numerical point of view, the concept of average relative error is introduced, and the specific calculation process is shown in Equation (17):
where
represents the pressure,
represents the saturation, superscripts
represent the coordinates of the grid points,
represent the dimensions of the grid,
represents the number of test datasets, and the superscript
represents error. The subscript
indicates time, and
is used to avoid the case where the denominator is 0. The pressure and saturation errors calculated using Equation (17) are shown in
Figure 6. It can be seen from
Figure 6 that the saturation and pressure predicted by the E2C model have maximum relative errors of less than 1% and less than 2%, respectively. Apparently, the E2C model has a higher accuracy in predicting pressure and saturation.
Next, four wells were randomly selected and their daily production rates over time were analyzed, and the results are shown in
Figure 7 and
Figure 8. It can be seen from them that the production obtained using the E2C model is basically consistent with the simulation result of the numerical simulator. For well B151, although the daily oil and water productions predicted using E2C at some moments are smaller than the results of the numerical simulator (the peak value of the red line of B151 is lower than the blue line in the figure), with a very low production, this well has little impact on the total production of the reservoir.
To analyze the accuracy of the daily production of the reservoir, the daily production of the reservoir at different moments was calculated using Equation (18). The specific calculation method is as follows:
where
represents oil and water, respectively,
represents the moment,
represents the well,
represents the total number of wells,
represents the daily production of each well, and
represents the daily production of the reservoir. According to the above equation, the daily reservoir productions at different moments were obtained using the simulator and E2C, and the results are shown in
Figure 9 and
Figure 10. It can be seen from the two figures that all data points are distributed around the 45° line, which proves that the E2C model has a higher consistency and reliable accuracy in predicting the daily production of the reservoir compared with the numerical simulator.
In order to avoid the contingency of random sampling, the relative errors of all test sets were counted and analyzed. As the daily production of some wells is 0 at some moments, in order to reduce the impact of the productions with a denominator of 0 on the error, the error calculation method used in this paper is as follows:
In Equations (19)–(21), the superscripts represent oil and water, respectively, represents the moment, represents the number of all time steps, represents the well, represents the number of wells, represents the daily production of the well, represents the absolute error of the average daily production in each example, represents the average daily production in each example, and represents the relative error of the average daily production in each example.
According to the above equation, the relative error distributions of all test examples were counted, and the results are shown in
Figure 11. It can be seen from this figure that the relative error of the average daily production of the reservoir is about 13%. This is mainly because the daily production of some test wells is low, and the denominator is too small.
Finally, the running time of the numerical simulator was compared with that of E2C. The test cases were simulated using the simulator on an Intel Xeon (R) W-3275M dual CPU (50 cores) node, and each took approximately 4080 s. In comparison, on the RTX1080 GPU node with 11 GB video memory, the E2C model completed the calculation of 30 test cases in 521 s, which was 17.3 s per case on average. Clearly, compared with the traditional simulator, the E2C model is about 200 times faster in computing speed. Although there is a difference between the GPU and CPU, the simulation speed of the simulator did not show any significant improvement using the GPU.
To better demonstrate the effectiveness of the model proposed in this paper, we used the same data to conduct a comparison with the E2C model proposed by Huang H et al. (2023) [
29] The results are presented in
Appendix A.
4.3. Analysis of Model Optimization Results
The E2C model trained in 3.2 was used to optimize the water injection conditions of the reservoir. In the process of model construction, some historical data were needed. In order to use as much data as possible and illustrate the use of a surrogate model in optimization, only the last month was analyzed in this work. The goal of optimization in this part was to improve the daily oil production of the reservoir in the last month by adjusting the daily water injection rate of the water injection wells. According to the production history, a total of 97 injection wells were put into operation in the last month, so the number of independent variables to be optimized was
. The initial water injection rates of the 97 injection wells are shown in
Figure 12. In this study, the liquid production rate of each production well was fixed at the initial rate, the daily water injection volume of the reservoir was 2136.63 m
3, and the daily injection volume of each well ranged from 0 to 28 m
3.
The particle swarm optimization (PSO) algorithm is usually used to update independent decision variables and find the optimal solution. The quality of the optimal solution found using PSO usually depends on the number of particles. Making use of the speedup provided by the E2C model, the PSO algorithm used in this study contained 200 particles and 100 iteration steps to find the optimal solution within a wider distribution. The relationship between the daily oil production of the reservoir and the iteration number is shown in
Figure 13. It can be seen from the figure that, in the first 30 iterations, the daily oil production of the reservoir increases rapidly and then tends to be stable.
Figure 14 shows the difference between the optimal daily water injection volume and the initial daily water injection volume, where the black dotted line is the 45° line, representing that the results are completely consistent. It can be seen from the figure that the optimal daily water injection volume of each well deviates from the 45° line, and in some wells the water injection volume can reach the set maximum boundary. Therefore, there is a big difference between the optimal daily water injection volume and the initial daily water injection volume, which also sets higher requirements on the generalization ability of the surrogate model.
Figure 15 shows the change in daily oil production of the reservoir before and after the optimization. It can be seen from the figure that the daily oil production of the reservoir at the last time step is 393.09 m
3 under the initial daily water injection volume, while the optimal daily oil production of the reservoir obtained using E2C + PSO is 477.78 m
3. Through optimization, the daily oil production of the reservoir increases by 13.84%. In addition, due to the high calculation efficiency of the E2C model, it only took about 5 h to complete 100 iterations of 200 particles. Apparently, the model proposed in this study can quickly complete the injection–production optimization analysis of the reservoir.
In order to better illustrate the reliability of the above optimization results, the optimized daily water injection volume was input into the simulator, the daily oil production of all wells at the last time step was simulated, and the results are shown in
Figure 16. It can be seen from the figure that the daily oil production of most wells is near the 45° line, so the above optimization results are credible.
This study serves as a proof-of-concept, aimed at demonstrating the feasibility and scalability of the E2C surrogate model under high-dimensional controls and complex, coupled physics.
In practice, optimization is a multi-objective, multi-constraint problem that must simultaneously balance cumulative production, economic performance, energy consumption, and operating costs; incorporate operational windows, facility and pipeline capacity, progressive well pattern expansion, injection–production coordination, the duration of stable production periods, and shut-in/restart scheduling; and account for risks arising from geological uncertainty and data noise. These factors substantially increase the problem dimensionality and algorithmic complexity, which lie beyond the scope of this paper and constitute a distinct research area.
To address these challenges, our future work will prioritize (1) developing a graph neural network (GNN)-based surrogate model that accommodates dynamic well locations, enabling the reuse and transfer of learned representations under varying well counts and evolving well patterns; (2) integrating uncertainty quantification and robust/stochastic optimization to enhance stability and reliability under field conditions; (3) establishing a multi-stage rolling optimization framework and coupling it with surface facility models for long-horizon validation; and (4) conducting systematic evaluation on synthetic and field datasets, reporting standardized error metrics and ablation studies. These directions will be central to our subsequent research, enabling an assessment of the method’s practicality under conditions that more closely reflect engineering practice.