Abstract
To address the challenges of low production efficiency, high energy consumption, and frequent equipment failures in low-producing wells, this study proposes an intelligent production parameter optimization method based on deep learning and multi-indicator fusion. First, a Long short-term memory (LSTM)-based prediction model was developed using the displacement–load characteristics of pumping-unit dynamometer cards. The results show that the model achieves an average prediction metric Q = 0.08, outperforming Back Propagation and Recurrent Neutral Network (BP and RNN) models. Second, a Convolutional Neutral Network (CNN) was employed to extract the fluid supply capability features, achieving a recognition accuracy exceeding 98%, thereby validating the model’s effectiveness. Combined with three types of time series data (liquid supply degree, dynamic liquid level, and liquid production rate), a multi-index fusion parameter optimization method is proposed. A comprehensive decision-making model is constructed based on the Analytic Hierarchy Process (AHP), which takes long-term, short-term, and overall feature changes as the basis, and forms a hierarchical framework by decomposing objectives and quantifying feature weights. The consistency ratio (CR) of the model is less than 0.1, meeting the requirement of logical consistency and enabling the output of standardized regulation suggestions. Consequently, a closed-loop system of “data preprocessing—condition prediction—state identification—parameter optimization” was constructed, enabling early dynamometer card prediction, accurate fluid supply fluctuation identification, and automatic generation of optimization schemes. This system effectively enhances production efficiency and equipment stability in low-producing wells, providing a technical foundation for intelligent oilfield development.
1. Introduction
A low-production oil well refers to an oil well with a daily output below the specific industry standard. In China’s oil and gas development field, there is a huge stock of low-production oil wells, and their efficient development is of great practical significance for ensuring stable oil and gas production, improving efficiency, and enhancing the comprehensive utilization rate of resources. According to statistical data from PetroChina, the average daily output of low-production oil wells in China is approximately 2–3 tons, which generally face the problems of low production efficiency and high unit development cost. Given the wide distribution of oil and gas resources and the complex geological conditions [], pumping units are widely used in oil and gas extraction because of their strong adaptability, ease of operation, and low cost. As a core monitoring method for the operating state of the pumping system, the dynamometer card intuitively reflects the relationship between load and displacement over a single stroke; by analyzing its morphological features and variation trends, the operating status of sucker rods, pumps, and tubing can be identified, providing a basis for condition assessment []. Under complex operating conditions, sucker-rod pump systems are prone to faults such as sand or wax sticking, gas locking, pump leakage, and rod breakage or separation, which not only reduce production but also increase equipment risk []. Predictive analysis based on dynamometer cards can extract features and identify faults to support fault localization and production parameter optimization, enabling early warning and dynamic control, thereby significantly improving system safety, stability, and efficiency []. In recent years, dynamometer card recognition methods have developed rapidly. Traditional methods rely on manual feature extraction and rule matching—such as shape analysis, template matching, and statistical pattern recognition—which are limited by expert experience and have poor generalization; theory-based methods simulate faults by building the working principles and mathematical models of the sucker-rod system []; overall, traditional approaches that depend on manual analysis or template matching have low accuracy and poor flexibility in complex scenarios [].
With the development of machine learning and deep learning, methods such as Support Vector Machine (SVM) and random forest have been applied to dynamometer card classification to improve recognition accuracy []. Recently, Convolutional Neural Network (CNN) has shown outstanding performance in extracting spatial features from image-like dynamometer cards, while Recurrent Neutral Network (RNN) and its variant Long short-term memory (LSTM) are adept at processing time series data; attention mechanisms and transfer learning offer new solutions for small-sample and cross-field recognition scenarios [,,]. Zhang, K. et al. [] used deep autoencoders for an intelligent diagnosis of rotating machinery. Liu J. et al. [] proposed optimizing the Chicken Swarm Optimization (CSO) algorithm with differential mutation and an adaptive inertia strategy (DACSO) to optimize an SVM model, achieving low production fault diagnosis via dynamometer card classification; Yu, D. et al. [] extracted key features from progressive cavity pump operation data to build an SVM-based operating-state recognition model and validated it with experimental simulation platform measurements, showing good diagnostic performance. Liu, S. et al. [] proposed a method for early fault detection in sucker-rod systems by monitoring operating parameters and applying the AdaBNet algorithm to improve accuracy, which can enhance asset monitoring, extend well life, reduce maintenance costs, and enable remote monitoring and prediction of potential issues, though the model has poor interpretability and requires large amounts of training data. In recent years, deep-learning-based fault diagnosis methods, particularly convolutional neural networks, have been used for pumping unit state recognition. Compared with methods such as random forest (RF), k-nearest neighbors (k-NN), spectral clustering, and density peak clustering (DPC), deep learning techniques demonstrate much stronger pattern-mining capability in scenarios with many features, low signal sensitivity, and complex data types, and have been widely applied in object detection, image recognition, behavior prediction, 3D reconstruction, and visual relocalization [,,,,,].
Furthermore, deep learning methods can automatically extract image features from dynamometer cards using deep neural networks []; Moghimihanjani, M. et al. [] and Dong, L. et al. [] applied CNNs for bearing fault diagnosis, showing that CNNs have become leading architectures and perform well across many tests. Peng, Y. et al. [] proposed a CNN-based method for dynamometer card recognition, classifying cards under different conditions with good results. He, Y.-P. et al. [] proposed a capsule-network-based method for pump dynamometer card recognition; capsule networks are a novel deep learning model able to better capture feature relationships in graphics. Shan, L. et al. [] used an improved CNN-BiLSTM model for well production forecasting; this model combines convolutional neural networks and bidirectional LSTM to effectively predict well performance. [] proposed both image-based and data-based CNN methods, confirming that CNN-based approaches yield higher recognition accuracy than machine learning algorithms like random forest and k-NN. Daolun, L.I. et al. [] also validated the effectiveness of image-based CNNs in well fault classification. Although BP neural networks are suitable for static data tasks, they lack temporal memory and the ability to capture time-dependent relationships due to structural limitations, failing to adapt to the dynamic characteristics of dynamometer cards and resulting in insufficient prediction accuracy. Traditional RNNs, while specifically designed for temporal data, struggle to maintain long-term memory due to gradient vanishing and gradient exploding issues, being only applicable to short-time sequences and exhibiting limited performance when dealing with complex long-time sequences. In contrast, LSTM networks, endowed with excellent long-term dependency processing capabilities by their gating mechanisms, can accurately capture key temporal features and adapt to complex dynamic patterns of data. They demonstrate significant advantages in dynamometer card prediction for oil wells, providing an ideal model choice for the accurate prediction of full-cycle dynamometer cards.
The degree of liquid supply in a well reflects the ratio of actual fluid output to the theoretical maximum output and can be preliminarily judged by indicators such as production and water cut; it is important for improving production efficiency [,,], reducing energy consumption, and optimizing field development. Traditional measurement methods are complex, have poor real-time performance, and are easily disturbed. Andrade, G.M.P. et al. [] analyzed the principles and pros and cons of various flowmeters and downhole test tools and proposed improvements, laying a foundation for modern measurement technologies and AI development; Zhong, Z. et al. [] designed a monitoring system based on surface sensors that judges the liquid supply degree of low-producing wells from wellhead pressure, temperature, and other parameters—this system is simple to install and maintain and has been field-validated. Han, D. et al. [] built a liquid supply recognition model based on machine learning that predicts low-producing well supply conditions from historical production and real-time monitoring data, showing high predictive accuracy and generalization. The above methods generally suffer from model complexity and the need for large sets of operating parameters, highlighting the potential of deep learning for predicting liquid supply in low-producing wells.
Research on production parameter optimization for low-producing wells is mainly divided into mechanism-based modeling and data-driven approaches. Weiwei, T. et al. [] proposed a mechanism-simulation-based parameter optimization method for sucker-rod pump lifting systems, using big data analysis to increase production and reduce energy consumption, and combined simulation with field data to build deep learning models that recommend efficient schemes; however, adaptability across different geological conditions is limited. Wang, H. et al. [] used multiple machine-learning algorithms to build models relating production parameters and output and improved well production through parameter optimization. Dong, Y. et al. [] proposed a curve-adaptive and simulated annealing particle swarm optimization algorithm that takes production time, injection volume, and bottom hole flow pressure as inputs and optimizes injection production parameters to maximize net present value. Data-driven methods integrate multi-source data and continuously update models to achieve ongoing production parameter optimization and significantly improve well production efficiency.
In summary, this paper constructs an AI-based research approach to address the full process needs of low-producing wells-“condition diagnosis liquid supply recognition parameter optimization.” In this study, LSTM neural networks are used to model the time series features of dynamometer cards to improve prediction accuracy; a high-precision liquid supply recognition model is established by combining CNNs with an iterative labeling method; and the Analytic Hierarchy Process (AHP) is introduced to integrate multi-indicator time series features, constructing a long- and short-term fused comprehensive decision model to achieve quantitative optimization of production parameters. This method aims to overcome current technical shortcomings in adaptation to complex operating conditions and multi-indicator coordinated control, and it has important research value and application potential for intelligent operation and maintenance of low-producing wells.
2. Materials and Methods
2.1. Backpropagation Neural Network (BP)
Training of a BP neural network alternates between forward propagation and backpropagation: forward propagation computes outputs layer by layer from the inputs, while backpropagation updates weights based on the error between predictions and target values to achieve learning []. For the same training set, model performance can be tuned by selecting different activation functions and parameters. The BP principle is shown in Figure 1.
Figure 1.
BP neural network schematic diagram.
The error function is:
Randomly select the K-th input sample and the corresponding desired output:
For each neuron in a hidden layer, compute its input and output:
Using the network’s expected outputs and actual outputs, compute the partial derivatives of the error function with respect to the output layer neurons, applying the chain rule:
The output layer neuron and the hidden layer neuron output are used to modify the connection weights, then:
In the formula, and represent learning rates; by adjusting the step size, one can avoid divergence caused by overly large updates and approach the optimal solution.
Use the deltas of the hidden layer neurons and the inputs of the input layer neurons to correct the connection weights:
Compute the global error:
Evaluate the network error; the algorithm terminates when this error reaches a preset accuracy or exceeds a preset maximum number of iterations N.
2.2. Recurrent Neural Network (RNN)
A recurrent neural network (RNN), proposed by Elman [], is a neural network for processing sequential data and is suitable for sequences such as natural language, audio, and stock trends. Its core is that the hidden layer input fuses the current input with the previous time step’s hidden layer output, enabling the capture of temporal information and learning of sequence nonlinearities. The RNN principle is shown in Figure 2.
Figure 2.
Recurrent Neural Network workflow.
Figure 3: Schematic diagram of a Recurrent Neural Network (RNN) recurrent cell. As shown in the figure, the core unit of an RNN is the recurrent cell, whose structure is similar to a fully connected layer. Although the weight matrices for the current input () and the previous hidden output () are shown separately as and , this is equivalent to concatenating and and then applying a single parameter matrix. Key components are defined as follows: (a) : Previous hidden state, representing the memory information passed from the last time step; (b) : Input data at the current time step; (c) tanh: Activation function that introduces non-linearity to transform the combined input and hidden state; (d) : Hidden state at the current moment, the output of the RNN unit carrying both current input and historical information, which is also passed to the next time step. Abbreviations: RNN = Recurrent Neural Network. The mathematical expression is:
which can also be written as:
Figure 3.
Recurrent Neural Network cycle body diagram.
In the formula, denotes vector concatenation.
The components are labeled as follows: (a) : Hidden state at the previous time step, representing the memory information passed from the last moment. (b) : Input data at the current time step. (c) : Weight matrix for the hidden state input. (d) : Weight matrix for the current input data. (e) σ: Sigmoid activation function, which introduces non-linearity to transform the combined input and hidden state. (f) : Output at the current time step, generated by the activation function processing. (g) : Hidden state at the current time step, carrying both current input and historical information, and passed to the next time step.
2.3. Long Short-Term Memory (LSTM)
The Long Short-Term Memory (LSTM) neural network is a special form of Recurrent Neural Network (RNN). Built on the foundation of multi-layer feedforward networks, RNN achieves memory functionality by transmitting information from the previous time step through lateral connections of hidden layers and weight matrices []. Through the synergistic design of three core gating units (Forget Gate, Input Gate, Output Gate) and a Cell State, LSTM effectively addresses the vanishing gradient problem of traditional RNNs. It can accurately capture and retain key information as well as temporal dependencies in long time series, dynamically adapt to complex fluctuations of data while preserving long-term critical features, and break through the inherent limitation of RNNs that only support short-term memory, ultimately delivering more accurate and stable prediction results; The LSTM workflow is shown in Figure 4.
Figure 4.
The Long Short-Term Memory workflow.
By designing three core gating units (Forget Gate, Input Gate, Output Gate) and a Cell State, the network effectively alleviates the vanishing gradient problem faced by traditional Recurrent Neural Networks (RNNs) in long-sequence modeling, enabling accurate storage and transmission of long-term dependent information: (a) Forget Gate, composed of a Sigmoid activation function layer, whose output vector acts on the previous cell state () through element-wise multiplication to adaptively filter out redundant historical information; (b) Input Gate, synergistically constituted by a Sigmoid layer and a tanh layer: the Sigmoid layer outputs an update weight vector for selecting key information to be integrated into the cell state; the tanh layer generates a candidate cell state () with a value range of [−1, 1], and the two complete the cell state update through element-wise multiplication and addition operations; (c) Output Gate, composed of a Sigmoid layer and a tanh layer: the Sigmoid layer outputs a filtering vector to filter information from the current cell state (); the tanh layer normalizes the filtered cell state to the [−1, 1] interval, ultimately generating the current hidden state () while transmitting the updated cell state () to the LSTM unit at the next time step.
Abbreviations: LSTM = Long Short-Term Memory; RNN = Recurrent Neural Network; = Previous Cell State; = Current Cell State; = Current Input Feature Vector; = Candidate Cell State; = Current Hidden State.
Forget gate (selectively forgets information in the cell state):
Input gate (selectively records new information into the cell state):
Cell state update:
Output gate (controls the output based on input and cell state, and updates the next hidden state):
In the formula, is the input information of the current moment; , , , are the weights of forgetting gate, input gate, memory unit and output gate, respectively; , , , for the corresponding bias.
2.4. Convolutional Neural Network (CNN)
A convolutional neural network (CNN) is a deep learning architecture specifically designed for data with a known grid structure. It can excavate complex structures in the data, which gives it high robustness to small perturbations in the input []. The CNN principle is shown in Figure 5.
Figure 5.
Convolutional Neural Network (CNN) workflow.
The convolutional neural network (CNN) consists of an input layer (image), convolution + ReLU layers, a max-pooling layer, a fully connected layer, and an output layer. The image recognition process, the data flow between the layers (indicated by arrows in the diagram), and the meaning of the colors are as follows:
The input layer receives the raw image data (represented in purple in the diagram, corresponding to the original image block), which is then passed through an arrow to the convolution + ReLU layer. The feature maps generated by this layer are shown in cyan, with the light green patches within the cyan areas representing the local feature regions extracted via convolution. In this layer, convolution operations extract local features, and the ReLU activation function introduces nonlinearity to enhance the feature representation capability, ultimately generating multiple feature maps.
These cyan feature maps are then passed through an arrow to the max-pooling layer, where the output feature maps are displayed in orange. The pink patches within the orange areas correspond to the key features retained after down-sampling. This layer reduces dimensionality via down-sampling while preserving important features and enhancing the network’s translation invariance, resulting in smaller feature maps.
The pooled orange feature maps are then "flattened" and passed (via an arrow) to the fully connected layer, where the nodes are shown in dark green (with gradient effects to represent the process of feature integration). In this layer, the features are integrated, establishing a mapping between the features and output classes for classification computation.
Finally, the classification result computed by the fully connected layer is passed through an arrow to the output layer (represented in light blue, with yellow circular nodes indicating the output classification nodes). The output layer then outputs the classification result, completing the image recognition task.
- (1)
- Convolutional layer
The convolutional layer is the core of a CNN and is used for feature extraction. It applies learnable filters that slide over the input and perform dot product operations on local regions to produce two-dimensional activation maps (feature maps) [].
The convolution operation is:
In the formula, is the value of the input image at position ; is the filter value at position ; and are the filter height and width; and is the output value at position .
- (2)
- Pooling layer
The pooling layer reduces the spatial dimensions of feature maps, lowers computational cost, and increases robustness to feature position shifts:
In the formula, is the input feature map; is the pooled output feature map; is the stride; indexes positions inside the pooling window.
- (3)
- Fully connected layer
After several convolutional and pooling layers, one or more fully connected layers are used for classification or other tasks. In these layers, the network integrates all previously extracted features to produce the final outputs.
- (4)
- Output layer
The output from the last fully connected layer is fed to the output layer, typically a softmax layer, which yields class probabilities; the class with the highest probability is taken as the prediction.
2.5. Analytic Hierarchy Process (AHP)
The Analytic Hierarchy Process (AHP) [], developed by Thomas L. Saaty, is an important multi-criteria decision-making method. It decomposes a complex problem into hierarchical levels and factors, builds a judgment matrix by pairwise comparisons, computes the eigenvector to determine alternative weights, and thereby provides a basis for optimal decision-making. AHP is widely used in management, project evaluation, resource allocation, and risk assessment. The hierarchy model is shown in Figure 6.
Figure 6.
Analytic hierarchy model structure diagram.
The analytic hierarchy process (AHP) model is constructed as a three-level architecture from top to bottom, including the Decision-Making Objectives layer, Pre-test Items layer (Pre-test 1~i), and Decision Alternatives layer (Decision 1~j): By decomposing complex decision-making objectives into several pre-test items level by level, systematic analysis is conducted based on the correlation between pre-test items and decision alternatives. Through pairwise comparisons, a judgment matrix is constructed, and eigenvectors are calculated to determine the weight of each alternative, ultimately achieving hierarchical and quantitative decision-making for complex problems.
- (1)
- Constructing the judgment matrix
For n elements at a given level , pairwise comparisons produce an n × n matrix A, i.e., the judgment matrix:
In the formula, denotes the importance scale of element relative to . This matrix must satisfy two conditions to form a positive reciprocal matrix.
- (2)
- Calculating the weight vector
Geometric mean method (root method):
Arithmetic mean method (sum method):
3. Results and Discussion
3.1. Data Preprocessing and Selection of Dynamometer Card Prediction Method
3.1.1. Data Preprocessing
This study uses 500 measured dynamometer cards continuously collected from a production plant in Oilfield X as the original time series data. The acquisition parameters were set to 200 sampling points per dynamometer and a sampling interval of 20 min. With extended pumping time, the “handle” shape of low-producing wells’ dynamometer cards gradually elongates, showing a temporal trend that is valuable for prediction studies.
A dynamometer card consists of the polished rod (suspension point) displacement and the load; both are sampled at fixed time intervals within a single stroke, but displacement data vary between cards. Predicting displacement directly would increase model complexity, so interpolation is used to fix displacement sampling points and predict only the load.
The specific processing steps are as follows: first, take the maximum displacement as the midpoint and divide a single stroke into upstroke and downstroke; then set the number of interpolation points to 100 and compute the interpolated displacement points sequentially according to Equation (28); finally, apply cubic spline interpolation to complete the data processing.
In the formula, denotes the interpolated sequence and is the maximum displacement.
In the formula, is the segment spline, each segment spline is defined in [, ], and its coefficient , , , is determined by the following constraints:
- (1)
- Interpolation condition: the spline must pass through the end-point data;
- (2)
- Continuity condition: function values and first derivatives are continuous at joints;
- (3)
- Smoothness condition: second derivatives are continuous at joints;
- (4)
- Boundary conditions: endpoints X0 and Xn must have specified boundary conditions. This yields dynamometer card data with fixed displacement and variable load, which are then normalized.
Because different physical quantities have widely different scales, direct modeling can cause biased regression coefficients, reduced prediction accuracy, and poor generalization. To improve prediction accuracy, data must be standardized before the correlation analysis. Max–Min normalization was applied to rescale the load values to the [0,1] interval, thereby improving the efficiency and stability of subsequent network training []. The mathematical expression is given in Equation (30). The preprocessed upstroke and downstroke dynamometer cards are shown in Figure 7 and Figure 8, respectively.
Figure 7.
The stroke diagram on the indicator diagram after pretreatment.
Figure 8.
The stroke diagram under the indicator diagram after pretreatment.
In the formula, X1 is the normalized data, X is the original data, Xmin is the minimum of the original data, and Xmax is the maximum of the original data.
Figure 7 and Figure 8, respectively, depict the dynamic variation relationship between payload and displacement during the upstroke and downstroke processes: the upstroke curve shows a rapid upward trend of payload in the initial stage of displacement (within a small range) and then enters a fluctuating stable phase, reflecting the evolutionary characteristic of the mechanical response transitioning from drastic change to stability at this stage; the downstroke curve first exhibits slight fluctuations, and when the displacement approaches 1.0, the payload presents a sharp upward tendency, which reveals the non-linear response law of the payload-displacement during the downstroke phase. Together, these two curves clearly demonstrate the phase-dependent variation characteristics of the payload-displacement relationship throughout the full stroke, providing robust support for the dynamic response analysis of pumping units.
3.1.2. Dynamometer Card Prediction Methods and Evaluation Metrics
- (1)
- Evaluation metrics
To evaluate the model performance, mean squared error (MSE), mean absolute error (MAE), and the coefficient of determination (R2) are used. MAE reflects model robustness, MSE characterizes the overall error level, and R2 measures goodness of fit. In addition, a metric Q is defined as the mean of the squared errors between the predicted load sequence (normalized values) and the true values—a lower Q indicates higher prediction accuracy. The metrics are computed as:
In the formula, , , are true value, predicted value and average true value, respectively, and n is the number of samples.
In the formula, is the length of the model output sequence, that is, the number of predicted indicator diagrams; is the first load value of the first real indicator diagram; is the first load value in the corresponding predicted indicator diagram.
- (2)
- Optimization of prediction methods
According to the same input and output sequence, BP neural network, RNN and LSTM are used to predict the oil well indicator diagram, respectively. The prediction comparison results are shown in Table 1 and Figure 9.
Table 1.
Average performance evaluation of each prediction model.
Figure 9.
Comparison between normalized predicted indicator diagram and real indicator diagram at different time steps. (a) After 1T; (b) After 2T; (c) After 3T; (d) After 4T; (e) After 5T.
From Table 1 and Figure 9, it can be seen that the LSTM model outperforms the RNN and BP neural networks in mean squared error (MSE = 0.1586) and the prediction metric Q (Q = 0.018). The comparison shows that LSTM’s predictions are closer to the measured values, exhibiting the smallest errors and demonstrating its superiority in dynamometer card prediction.
As shown in Figure 10, the loss trend during model training further validates the model’s high accuracy. Moreover, LSTM’s gating mechanisms enable it to effectively capture long-term dependencies in the data, outperforming RNNs that are prone to vanishing gradients and BP networks that lack temporal memory. Considering these factors, the LSTM model was ultimately chosen for dynamometer card prediction.
Figure 10.
The loss change curve of the indicator diagram prediction model training process.
3.2. LSTM-Based Dynamometer Card Prediction Model
3.2.1. Model Structure Design
The dynamometer card prediction workflow consists mainly of model training and forecasting.
Step 1—training set construction: process the collected dynamometer cards from low-producing wells with sampling interval T = 20 min; use k = 10 dynamometer cards as the time window to predict the next t = 5 dynamometer cards. A sliding window method with stride 1 is employed, producing 486 samples from N = 500 original dynamometer cards.
Step 2—interpolation and normalization: since dynamometer cards evolve over time, normalizing is performed using the maximum displacement and maximum load of the first card in each group, and cubic spline interpolation is applied to unify sampling.
Step 3—LSTM-based sequence-to-sequence prediction: the sequence-to-sequence model built with LSTM is shown in Figure 11. It predicts load data Y and combines them with fixed displacements to form predicted dynamometer cards. The model takes current and historical load sequences Z as input and future load sequences Y as output; the input length is controlled by k = 10 (10 normalized, interpolated load traces, approximately 2000 values) and the output length by I = 5 (5 normalized, interpolated load traces, approximately 1000 values). Considering the comprehensive input–output scale (2000→1000 values) and the temporal characteristics of the load sequence, the optimal number of LSTM layers is one. This structure can effectively extract temporal features and possesses a robust long-term prediction capability.
Figure 11.
Structure diagram of sequence-to-sequence indicator diagram prediction model based on LSTM.
Both inputs and outputs are load sequences of dynamometer cards: each sample’s input is the preceding k = 10 load sequences and the output label is the following I = 5 load sequences. The loss function is mean squared error (MSE). Hyperparameter settings, which significantly affect robustness and performance, are listed in Table 2.
Table 2.
Prediction model hyperparameters.
3.2.2. Sensitivity Analysis of Prediction Model Parameters
Conducting sensitivity analysis on the LSTM-based dynamometer card prediction model helps determine the optimal input and output sequence lengths, thereby improving prediction accuracy and computational efficiency. The analysis can identify information redundancy introduced by overly long input sequences and thereby simplify the model architecture and reduce unnecessary computational overhead; it also helps detect and avoid overfitting caused by excessively long inputs, thus safeguarding the model’s generalization to unseen data. Properly optimizing the input sequence length can substantially reduce inference time and speed up predictions, which is especially beneficial for real-time or near-real-time applications.
In addition, sensitivity analysis can reveal potential prediction instability when the output sequence is too long, ensuring the reasonableness and credibility of model outputs. Based on these considerations, sensitivity analyses were performed along two dimensions—the input sequence length k and the output sequence length I—to determine k and I values suitable for on-site use in low producing wells.
- (1)
- Sensitivity analysis of input load-sequence length
To improve the accuracy of the dynamometer card prediction model for low-producing wells, an initial input load sequence length was selected empirically and then refined via sensitivity analysis. During the analysis, the output sequence length, model structure, training parameters, and sampling interval were held constant. Considering that the dynamometer card of low-producing wells changes rapidly, the value range of the input load sequence length k is set to 1–10 with a step size of 1; the output load sequence length l is fixed at 5. The resulting variation in the model’s Q metric with k is shown in Figure 12.
Figure 12.
Schematic prediction of model Q with input sequence length k.
As the input load sequence length K increases, the prediction performance metric Q generally decreases, with a pronounced inflection at K = 4. Combined with the variation in Q with k shown in Table 3 for the dynamometer chart prediction model, the optimal input sequence length under the conditions of this study is K = 5. For smaller K values the model is highly sensitive to changes in K and exhibits large performance fluctuations, which is consistent with the rapid evolution characteristics of dynamometer charts from low-producing wells; once K exceeds a certain threshold, Q tends to stabilize. The Q index essentially reflects the relative deviation between the predicted oil well production and the on-site actual production. It serves as a core engineering indicator for quantifying the prediction accuracy of the model, with smaller values indicating that the predicted results are closer to the actual production and the model is more reliable.
Table 3.
The variation in Q value of indicator diagram prediction model with K value.
In summary, an input sequence that is too short cannot fully capture the trend information of the dynamometer chart, whereas an excessively long sequence tends to introduce data redundancy and provides limited improvement in prediction accuracy.
- (2)
- Sensitivity analysis of output load sequence length
In the experiment, the LSTM-based dynamometer chart prediction model keeps the input sequence length fixed at k = 5, while the output sequence length I is varied over I ∈ [1,10] with a step size of 1; the prediction metric Q is recorded for each I (Figure 13).
Figure 13.
Variation in model Q with output sequence length I in dynamometer card prediction.
The results show that Q increases with I and that the growth rate accelerates; pronounced inflection points occur at I = 5 and I = 8, after which the curve’s slope increases markedly. Combining the Q values in Table 4 and considering overall predictive performance and prediction horizon, I = 5 is chosen as the optimal output sequence length, as this setting ensures a relatively long prediction period while maintaining low prediction error.
Table 4.
Variation in the dynamometer-chart prediction model’s Q value with I.
3.2.3. Model Prediction Results Validation
Using actual dynamometer charts as examples, the prediction performance is compared and analyzed. Five consecutive historical dynamometer charts ending at time T were selected, normalized, and resampled to a unified sampling interval (30 min) using cubic spline interpolation. The superimposed normalized continuous dynamometer charts are shown in Figure 14.
Figure 14.
Overlay diagram of continuous historical indicator diagram.
In the figure, “T” denotes the current dynamometer chart, “1T prior” denotes the previous sampling interval (30 min), and so on up to “4T prior.” As shown in Figure 14, the closed area in the lower left of the dynamometer chart gradually decreases over time, indicating a reduced degree of pump fill-up and a marked weakening of the well’s fluid supply capacity. This trend directly reflects the impact of formation energy depletion on the pumping system’s fluid delivery condition.
Based on the preceding sensitivity analysis, both the input and output sequence lengths of the LSTM model were set to k = 5 and l = 5. The sequence of consecutive historical dynamometer charts was fed into the model to predict dynamometer chart evolution over the next 5T. Figure 15 and Figure 16 presents the comparison between predicted and measured dynamometer charts. Although the predicted curves show minor fluctuations at some time points relative to the measured curves, the overall shape features remain highly consistent, and key evolving features such as the “knife-handle” are effectively captured.
Figure 15.
The prediction result diagram of normalized indicator diagram of LSTM model.
Figure 16.
Normalized measured indicator diagram.
A smaller Q value indicates that the predicted results are closer to the actual production, and the model exhibits higher reliability. The model’s performance metric Q is below 0.2 at all time points, with a mean value of 0.08 (Table 5), indicating good accuracy and stability. The method not only meets operational monitoring requirements but also provides a reliable basis for early intervention and intelligent control of low-producing wells.
Table 5.
Errors between predicted and observed dynamometer charts.
3.3. CNN-Based Identification and Analysis of Fluid-Supply Degree
3.3.1. Workflow and Method for Fluid-Supply Degree Identification
Dynamometer chart data were collected and converted, followed by image preprocessing. An AlexNet-based model was then constructed from the dynamometer chart features, and the network was trained and validated.
Summary of Core Parameter Settings for the CNN Model:
Network Structure: A simplified architecture is adopted, consisting of four convolutional layers, three pooling layers, and two fully connected layers. Through structural streamlining and optimization, the model training time is significantly shortened, and network convergence is accelerated.
Kernel Size Optimization: The convolutional kernel size is set to 3 × 3, and the pooling kernel size is optimized to 2 × 2.
Normalization Technique: Batch Normalization (BN) is employed to replace Local Response Normalization (LRN). By standardizing the batch data input to each layer, BN not only accelerates convergence and reduces training time but also improves the overall efficiency of the model.
The overall intelligent identification workflow for the fluid supply degree of low-producing wells is shown in Figure 17, covering data collection, image processing, model construction, and training/validation, thereby enabling an effective transition from raw data to application.
Figure 17.
The total flow chart of intelligent identification of liquid supply degree of low-yield oil wells.
The identification workflow comprises the following: generating grayscale dynamometer chart images, building a sample library and splitting it into training and test sets at an 8:2 ratio; training and validating models and saving the highest-accuracy model; and applying the model in the field—correctly classified charts are output directly, while misclassified results are corrected manually and added back to the sample library. Continuous iteration and augmentation of the sample set enable progressive improvement of model performance.
As shown in Table 6, typical low-producing well dynamometer charts exhibit a “knife-handle” shape with parallel unload and load lines; when the minimum unload line decreases, the “knife-handle” lengthens, indicating a reduction in fluid supply. In sample preparation, the fluid supply degree was divided into 10 levels. Because labeling quality directly affects deep learning performance and manual labeling is time-consuming and poorly scalable, we adopted an iterative labeling strategy: experts provide initial labels for a subset of samples and train a CNN; if accuracy is insufficient, additional samples are labeled and the model retrained; once satisfactory accuracy is reached, the trained model automatically labels the remaining samples and experts only review and correct, which reduces manual cost and improves labeling efficiency.
Table 6.
Sample set of liquid supply degree.
3.3.2. Identification Results and Analysis
We constructed a dataset of 15,000 dynamometer chart images for low-producing wells, with 1500 images per class. To improve training efficiency, GPU-accelerated deep learning was used and the Adam optimizer was selected. Compared with traditional Stochastic Gradient Descent (SGD), Adam adaptively adjusts learning rates, is suitable for large datasets and high-dimensional parameter spaces and is less sensitive to gradient sparsity.
The loss function uses cross-entropy loss, which is often used in multi-classification tasks to measure the difference between the model prediction distribution and the real label distribution, so as to guide the model optimization. The calculation formula is shown in Equation (35).
In the formula, L is the loss value, C is the number of classes, is the ground truth indicator for class i (1 if the sample belongs to class i, otherwise 0), and is the model’s predicted probability that the sample belongs to class .
As shown in Figure 18, the number of iterations was set to 200 in the experiment. The training accuracy exhibited a rapid upward trend in the early stage and stabilized above 98.5% when the iteration reached 100. Figure 19 presents the evolutionary characteristics of the training loss, which decreased rapidly in the initial stage and then tended to be stable, eventually converging to approximately 0.05, indicating that the model achieved excellent training performance.
Figure 18.
CNN model training and verification accuracy change curve.
Figure 19.
CNN model training and verification loss change curve.
Further analysis of the experimental results in Figure 18 and Figure 19 reveals that the validation accuracy remained consistently above 98%, while the validation loss stabilized at around 0.08. During the training and validation processes, the accuracy and loss values showed a consistent variation trend with the increase in the number of iterations without significant fluctuations. This result verifies that the proposed model possesses favorable stability and efficient learning performance in both the training and validation stages.
Using a pre-trained neural network model, the liquid supply level of actual dynamometer cards was identified; the dynamometer sampling interval was 30 min. To evaluate model performance, dynamometer card samples covering a randomly selected continuous 5 h period were analyzed. The quantitative assessment results of the liquid supply level for the low-production oil well are shown in Figure 20 and Figure 21.

Figure 20.
Indicator diagram sample sets at different time points within 5 h (a total of 12 groups).
Figure 21.
Dynamic change diagram of fluid supply degree in low production wells.
Combining Figure 20 and Figure 21, the CNN-based model for identifying the liquid supply level of the low-production well can effectively reflect the supply level and its temporal variation, achieving accurate quantification of this parameter and providing a basis for optimizing production parameters. To further compare the effectiveness of four indicators—liquid supply level, daily liquid production, dynamic fluid level, and pump efficiency—in reflecting well changes, these were collected and analyzed; the results are shown in Table 7.
Table 7.
Comparison of different indicators for the low-production well.
As shown in Table 7, between the adjacent months, the liquid supply level, daily liquid production, and pump efficiency all show declining trends, indicating a decrease in the well’s output. The magnitudes of change in the liquid supply level and pump efficiency are similar, suggesting that the liquid supply level is a more accurate performance evaluation indicator than the other metrics.
3.4. Temporal Feature Analysis and Production Parameter Optimization for Low-Production Wells
3.4.1. Extraction and Analysis of Time Series Data Features for Low-Production Wells
- (1)
- Level feature extraction
The level features of the low-production well time series data reflect the average state of the data over a period and are used to describe its overall stability level. The median, mode and mean are extracted as the level features of the liquid supply level. The median reflects the central value of the data but is less accurate when the distribution is uneven and its segment often coincides with the mode; by contrast, the mean varies with all data points and more comprehensively reflects the data level.
The results of the long-term level feature analysis for the liquid supply level are shown in Figure 22. When extracting short-term and long-term level features for the low-production well time series data, the mean is used comprehensively. Taking the liquid supply level as an example: when the mean is above the normal level, it indicates higher formation energy; when equal to the normal level, it indicates normal formation energy; when below the normal level, it indicates relatively low formation energy.
Figure 22.
Analysis chart of fluid supply level characteristics of long-term low-yield oil wells.
- (2)
- Trend feature extraction
The trend features of the low-production well time series data reflect the direction of change over a period and are an important basis for forecasting and planning in a time series analysis. Because the liquid supply level, dynamic fluid level, and production rate exhibit different trends at different time scales, short-term and long-term trend features must be extracted separately. The short-term trend can be characterized by the relative change rate, whose mathematical expression is as follows:
In the formula, is the relative change rate; is the final value; is the initial value.
When the relative change rate is positive, the data show an upward trend, indicating high formation energy; when zero, the data remain unchanged, indicating normal energy; when negative, the data decline, indicating low formation energy. Because downhole conditions are complex, long-term time series data usually show irregular changes and it is necessary to compare different trend extraction methods; among them, the moving average method and the linear regression method are two commonly used methods. Using the liquid supply level data from Low-production Well-1 and Low-production Well-2 as examples, trend features were extracted, respectively, by the moving-average method (Formulas (37) and (38)) and the linear regression method (Formulas (39) and (40)).
In the formulas: is the liquid supply level; is time.
Table 8 summarizes the moving average method and the linear regression method for trend feature extraction and provides a comparative illustration of their core characteristics.
Table 8.
Trend feature extraction methods.
As shown in Figure 23 and Figure 24, the black solid line represents the Fluid Supply Level, the blue dashed line denotes the Moving Average (Linear), the green dash-dotted line indicates the Moving Average, and the red dash-dotted line stands for the Fluid Supply Level (Linear). The vertical axis represents the Fluid Supply Level (%), and the horizontal axis indicates Time, presenting the changes in the fluid supply level and the results of different trend extraction methods from 10:10:00 to 12:00:00. The moving average and linear regression methods for Low-production Well-1 (Figure 23) are negative, accurately reflecting the declining trend in the liquid supply level and indicating low formation energy; for Low-production Well-2 (Figure 24), the decline in the liquid supply level is larger, and the linear regression slope (−3.4997) is significantly greater in magnitude than the moving average slope (−0.1005), indicating that linear regression is more precise for trend feature extraction.
Figure 23.
Low-Production Oil Well-1 liquid supply degree trend characteristics analysis.
Figure 24.
Low-Production Oil Well-2 liquid supply degree trend characteristics analysis.
In summary, by comparing the moving average and linear regression methods and considering the long-term characteristics of the low-production well time series data, linear regression is preferred to extract the three features. Data change is reflected by the curve slope: a positive slope indicates high formation energy, zero slope indicates normal formation energy, and a negative slope indicates low formation energy.
- (3)
- Fluctuation feature extraction
The fluctuation features of the low-production well time series data reflect the amplitude and frequency of changes, revealing the instability and volatility of the data. Table 9 lists three extraction methods: standard deviation, variance and range. Taking the liquid supply level time series data from Low-production Wells-1, -2 and -3 as examples, the fluctuation features for the three wells are shown in Figure 25.
Table 9.
Different fluctuation feature values for the liquid supply level of three low-production wells.
Figure 25.
Analysis of fluctuation characteristics of fluid supply degree in three low-yield oil wells.
Figure 25 illustrates the fluid supply level trends of three wells over time: the red dotted line represents Well-3, the purple solid line denotes Well-1, and the green dash-dotted line indicates Well-2. The vertical axis stands for Fluid Supply Level (%), and the horizontal axis represents Time, presenting the dynamic fluid supply changes and capacity patterns of the three wells from 10:10:00 to 12:00:00, with the corresponding data shown in Table 9, the standard deviation, variance and range were calculated for the liquid supply level of Wells-1, -2 and -3. If the range is used, Well-3 shows the largest fluctuation while Wells-1 and -2 show the same fluctuation, which contradicts Figure 25, because the range only reflects the maximum and minimum values and ignores intermediate changes. Therefore, both standard deviation and variance can accurately reflect fluctuation, with the standard deviation being more intuitive and suitable as the fluctuation feature for liquid supply level.
Because the short-term window is brief, fluctuation features do not need to be extracted for the short term; only long-term fluctuation features are extracted for the low-production well time series data. When the standard deviation is above the normal fluctuation range, formation energy is relatively high; when within the normal range, formation energy is normal; when below the normal range, formation energy is relatively low.
- (4)
- Example of feature extraction
Time series data of the liquid supply level from an actual field are selected as the study index; the liquid supply level time series for the low-production well is shown in Figure 26. The normal level for this well is 61%, the normal trend is 0.018, and the normal fluctuation range is [6.68, 10.08].
Figure 26.
Analysis chart of fluid supply degree based on time series data of low-production wells.
As shown in Table 10, in the short-term time window, the well’s level is 46.11% with a trend change of −1.33; in the long-term window, the level is 43.55% with a trend change of −3.20 and a fluctuation of 8.77. The comprehensive feature analysis indicates that the well’s formation energy is relatively low and production parameter optimization is required.
Table 10.
Feature extraction for the liquid supply level time series of the low-production well.
3.4.2. Production Parameter Optimization for Low-Production Wells Based on the Analytic Hierarchy Process (AHP)
Most existing production parameter optimization methods for low-production oil wells rely on single-factor decision-making based on the fluid supply degree derived from dynamometer cards, making it difficult to comprehensively characterize the integrated dynamic changes in formation energy. To address this limitation, this study proposes a production parameter optimization method for low-production oil wells based on the Analytic Hierarchy Process (AHP).
This method constructs a weighted average comprehensive evaluation function to systematically integrate multi-dimensional indicators characterizing formation energy, including long-term and short-term fluid supply degrees, liquid production rate, and dynamic liquid level. It thereby realizes multi-index collaborative decision-making for production parameters and ultimately determines the optimal production parameter scheme based on the comprehensive decision results.
The optimization objective function is constructed with the core goal of “production improvement + energy consumption reduction”. Considering the production characteristics of low-production oil wells, such as limited fluid supply capacity and a relatively high proportion of energy consumption, this optimization abandons the single production-oriented decision-making logic and establishes a “production energy consumption” dual-objective optimization system. This system ensures an effective improvement in production capacity while achieving a significant reduction in energy consumption, with the mathematical expressions as follows:
Among them, q represents the displacement of the oil pump, s denotes the stroke length, and n is the pumping frequency. The core optimization logic is “maximizing daily oil production while minimizing the energy consumption per unit oil production under the premise of stable fluid supply”—which not only avoids problems such as equipment idle pumping and a sharp increase in energy consumption caused by blind production increase, but also solves the industry pain point of “high production without high efficiency” in low-production oil wells.
- (1)
- Comprehensive decision-making for long-term feature changes
To optimize production parameters for low-production wells, level, trend and fluctuation features of liquid supply level, dynamic fluid level and production rate are first extracted from a long-term time window to provide the basis for optimization. Based on these features, a multi-level structural model for optimizing production parameters of low-production wells is established. The hierarchy for long-term production parameter optimization of low-production wells is shown in Figure 27.
Figure 27.
Long-term low-yield oil well production parameter optimization hierarchical structure diagram.
The judgment matrix is created by pairwise comparisons of each element, typically using Saaty’s 1–9 scale; the specific comparison scale values are shown in Table 11.
Table 11.
Saaty’s 1–9 scale.
Construct the judgment matrices and perform pairwise comparisons of the features using Saaty’s 1–9 scale (Table 11) to determine relative importance and then compute the weights of each time series feature with respect to the decision objective (Table 12 and Table 13).
Table 12.
Long-term criterion-level judgment matrix.
Table 13.
Long-term Decision-Making Layer Judgment Matrix.
Next, single-level ranking and consistency checks are carried out, yielding the ranking weight vector , ; the consistency check produces: , , , and the judgment matrix passes the consistency test. Using the above steps, the weight vector:
. Then, an overall hierarchical ranking and consistency check are performed: based on the already determined weights of the criterion layer relative to the goal and the weights of the decision layer relative to the criterion layer, the overall hierarchical ranking
, is obtained, and the judgment matrices pass the consistency checks.
Finally, compute the comprehensive decision result from the overall hierarchical ranking and consistency checks using the criterion-to-goal and decision-to-criterion weights, thereby constructing the long-term feature comprehensive decision model. The corresponding results are detailed in the overall summary table in the subsequent section.
- (2)
- Comprehensive decision-making for short-term feature changes
To optimize production parameters for a low-production well, level, trend and fluctuation features of the liquid supply level, dynamic fluid level and production rate are first extracted from a short-time window to provide the basis for optimization. Based on these features, a multi-level structural model for optimizing production parameters of low-production wells is established. The hierarchical structure for short-term optimization is shown in Figure 28.
Figure 28.
Short-term low-yield oil well production parameter optimization hierarchy diagram.
Construct judgment matrices and perform pairwise comparisons of the features using Saaty’s 1–9 scale (Table 11) to determine relative importance, and from these, compute the weights of each time series feature with respect to the decision objective (Table 14).
Table 14.
Short-term decision-level judgment matrix trend feature.
Next, single-level ranking and consistency checks are performed, yielding the ranking weight vector , . The consistency check gives , and the judgment matrix passes the consistency test.
Using the above steps the weight vector is calculated. Then perform overall hierarchical ranking and consistency checks: based on the already determined weights of the criterion layer relative to the goal and the decision layer relative to the criterion layer, compute the overall weights: . The consistency yields , indicating the pairwise comparison matrix passes the consistency test. Finally, construct the short-term feature comprehensive decision model; the results are shown in Table 15.
Table 15.
Comprehensive Decision-Making for Long-Term, Short-Term, and Overall Feature Changes Based on the Analytic Hierarchy Process (AHP).
- (3)
- Comprehensive decision-making for combined long- and short-term features
Long-term feature changes reflect the long-term rising or falling trend of low-production well time series data, while short-term feature changes reflect short-term sudden changes and instability; the two are related in the data. Therefore, in production parameter optimization for low-production wells, short-term comprehensive decisions are integrated with long-term trend comprehensive decisions to form a combined long- and short-term decision. The calculation formula is as follows:
In the formula, is the combined decision factor for long- and short-term features; is the time weight coefficient ( = 0.5); is the long-term feature decision function; is the short-term feature decision function; is the level feature; is the trend feature; is the fluctuation feature.
The AHP-based overall feature comprehensive decision results are shown in Table 15.
According to the earlier single-factor decision example analysis, single-factor decisions have limitations in consistency and scientific validity: some decision factors need adjustment while others do not. Therefore, to achieve a scientifically reasonable optimization of production parameters for low-production wells, a comprehensive decision method is necessary.
From the above calculations, the combined weight of long-term features was determined to be , with the long-term decision factor equal to −0.613.
Meanwhile, the short-term feature weight is . Using the formula for combined long- and short-term decision-making, the resulting comprehensive decision factor is .
The comprehensive decision result falls within the negative adjustment interval , clearly indicating that a negative regulation of the pumping frequency of the low-production oil well should be implemented in the current time period. The core of this decision-making logic lies in moderately reducing the pumping frequency to dynamically match the pumping parameters with the formation fluid supply capacity, thereby avoiding idle pumping losses caused by blind production increase and ultimately achieving the coordinated optimization of the oil well’s production efficiency and energy consumption economy.
4. Conclusions
- (1)
- Aiming at the problem that the LSTM network structure does not match the data characteristics of the low-yield well indicator card, the LSTM prediction model of the low-yield well indicator card is constructed based on the actual well conditions and LSTM characteristics. Compared with the commonly used BP neural network and RNN model, the model performs better in prediction performance, and the average prediction performance index Q is 0.08.
- (2)
- A CNN was used to extract and analyze the liquid supply level data from dynamometer cards of low-production wells, verifying the applicability of CNNs for identifying liquid supply capacity from dynamometer cards. Case results show identification accuracies exceeding 98%, enabling effective prediction-matching of dynamometer cards and accurate extraction of the liquid supply level information.
- (3)
- For the representation relationship among three types of time series data—the liquid supply level, dynamic fluid level, and production rate—and formation energy, a multi-indicator fusion method for production parameter optimization was proposed. Using liquid supply level, dynamic fluid level and production rate as core indices, level, trend and fluctuation features were extracted and combined with AHP to construct a comprehensive decision model (CR < 0.1), achieving standardized control of production parameters.
- (4)
- Based on the representation relationships among liquid supply level, dynamic fluid level and production rate and formation energy, a closed-loop system of “data preprocessing—condition prediction—state identification—parameter optimization” was constructed. This system integrates LSTM, CNN and AHP modules and ensures data quality through standardized processing. Field application results show the system can predict dynamometer cards in advance, accurately identify liquid supply fluctuations, and output optimization schemes, effectively achieving intelligent operation and maintenance of low-production wells.
- (5)
- This study has two limitations: first, the model training data are derived from a single block and have not been validated in oilfields with different geological conditions, so its generalization performance remains to be tested; second, the current parameter optimization only targets pumpjack frequency and does not involve the coordinated optimization of multiple parameters such as stroke length and stroke rate. Future research can be expanded in two aspects: first, introducing key dynamic formation parameters such as real-time reservoir pressure and dynamic fluid supply capacity to enhance the dynamic adaptability between production parameters and formation conditions; second, integrating intelligent algorithms such as reinforcement learning to construct a closed-loop dynamic optimization model based on real-time feedback, realizing the adaptive iterative adjustment of production parameters.
Author Contributions
Conceptualization, L.J. and G.S.; methodology, N.J.; software, S.H.; validation, G.L. and Z.Z. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by Tianshan Elite Talent Cultivation Program, grant number 2022TSYCJC0032.
Data Availability Statement
The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.
Conflicts of Interest
Authors Lu Jia, Guowei Shi, Shunquan Hu, Guangya Li and Zhipeng Zhang were employed by Digital Intelligent Technology Company, Xinjiang Oilfield Branch. Author Nengji Jiang was employed by PetroChina Xinjiang Oilfield Company.
References
- Wenzhi, Z.; Suyun, H.U.; Lianhua, H.O.U.; Tao, Y.; Xin, L.; Bincheng, G.U.O.; Zhi, Y. Types and Resource Potential of Continental Shale Oil in China and Its Boundary with Tight Oil. Pet. Explor. Dev. 2020, 47, 1–11. [Google Scholar] [CrossRef]
- Tian, H.; Deng, S.; Wang, C.; Ni, X.; Wang, H.; Liu, Y.; Ma, M.; Wei, Y.; Li, X. A Novel Method for Prediction of Paraffin Deposit in Sucker Rod Pumping System Based on CNN Indicator Diagram Feature Deep Learning. J. Petrol. Sci. Eng. 2021, 206, 108986. [Google Scholar] [CrossRef]
- Feng, K.; Jiang, Z.; He, W.; Ma, B. A Recognition and Novelty Detection Approach Based on Curvelet Transform, Nonlinear PCA and SVM with Application to Indicator Diagram Diagnosis. Expert Syst. Appl. 2011, 38, 12721–12729. [Google Scholar] [CrossRef]
- Liu, J.; Feng, J.; Xiao, Q.; Liu, S.; Yang, F.; Lu, S. Fault Diagnosis of Rod Pump Oil Well Based on Support Vector Machine Using Preprocessed Indicator Diagram. In Proceedings of the 2021 IEEE 10th Data Driven Control and Learning Systems Conference (DDCLS), Suzhou, China, 14 May 2021; pp. 120–126. [Google Scholar]
- Tecle, S.I.; Ziuzev, A. A Review on Sucker Rod Pump Monitoring and Diagnostic System. In Proceedings of the 2019 IEEE Russian Workshop on Power Engineering and Automation of Metallurgy Industry: Research & Practice (PEAMI), Magnitogorsk, Russia, 4–5 October 2019; pp. 85–88. [Google Scholar]
- Ma, R.; Tian, H.; Cheng, X.; Xiao, Y.; Xu, Q.; Yu, X. Oil-Net: A Learning-Based Framework for Working Conditions Diagnosis of Oil Well through Dynamometer Cards Identification. IEEE Sens. J. 2023, 23, 14406–14417. [Google Scholar] [CrossRef]
- Shafiei, A.; Tatar, A.; Rayhani, M.; Kairat, M.; Askarova, I. Artificial Neural Network, Support Vector Machine, Decision Tree, Random Forest, and Committee Machine Intelligent System Help to Improve Performance Prediction of Low Salinity Water Injection in Carbonate Oil Reservoirs. J. Petrol. Sci. Eng. 2022, 219, 111046. [Google Scholar] [CrossRef]
- Sheikhoushaghi, A.; Gharaei, N.Y.; Nikoofard, A. Application of Rough Neural Network to Forecast Oil Production Rate of an Oil Field in a Comparative Study. J. Petrol. Sci. Eng. 2022, 209, 109935. [Google Scholar] [CrossRef]
- Huang, R.; Wei, C.; Wang, B.; Yang, J.; Xu, X.; Wu, S.; Huang, S. Well Performance Prediction Based on Long Short-Term Memory (LSTM) Neural Network. J. Petrol. Sci. Eng. 2022, 208, 109686. [Google Scholar] [CrossRef]
- Lee, K.; Lim, J.; Yoon, D.; Jung, H. Prediction of Shale-Gas Production at Duvernay Formation Using Deep-Learning Algorithm. Spe J. 2019, 24, 2423–2437. [Google Scholar] [CrossRef]
- Zhang, K.; Zhang, J.; Ma, X.; Yao, C.; Zhang, L.; Yang, Y.; Wang, J.; Yao, J.; Zhao, H. History Matching of Naturally Fractured Reservoirs Using a Deep Sparse Autoencoder. Spe J. 2021, 26, 1700–1721. [Google Scholar] [CrossRef]
- Liu, J.; Feng, J.; Gao, X. Fault Diagnosis of Rod Pumping Wells Based on Support Vector Machine Optimized by Improved Chicken Swarm Optimization. IEEE Access 2019, 7, 171598–171608. [Google Scholar] [CrossRef]
- Yu, D.; Zhang, Y.; Bian, H.; Wang, X.; Qi, W. A New Diagnostic Method for Identifying Working Conditions of Submersible Reciprocating Pumping Systems. Pet. Sci. 2013, 10, 81–90. [Google Scholar] [CrossRef]
- Liu, S.; Raghavendra, C.S.; Liu, Y.; Yao, K.; Balogun, O.; Olabinjo, L.; Soma, R.; Ivanhoe, J.; Smith, B.; Seren, B. Automatic Early Fault Detection for Rod Pump Systems. In Proceedings of the SPE Annual Technical Conference and Exhibition? Denver, CO, USA, 30 October–2 November 2011; p. SPE-146038. [Google Scholar]
- Lei, L.; Xie, S.; Chen, Z.; Carranza, E.J.M.; Bao, Z.; Cheng, Q.; Yang, F. Distribution Patterns of Petroleum Indices Based on Multifractal and Spatial PCA. J. Petrol. Sci. Eng. 2018, 171, 714–723. [Google Scholar] [CrossRef]
- Tian, M.; Tan, M.; Wang, M.; Li, C.; Zhang, L. Evaluation of Oil Saturation in Shale from an NMR T1–T2 Map Based on Density Peak Clustering and Gaussian Mixture Model. J. Pet. Explor. Prod. Technol. 2025, 15, 103. [Google Scholar] [CrossRef]
- Efimov, I.; Gabdulkhakov, R.R.; Rudko, V.A. Fine-Tuned Convolutional Neural Network as a Tool for Automatic Microstructure Analysis of Petroleum and Pitch Cokes. Fuel 2024, 376, 132725. [Google Scholar] [CrossRef]
- Alizadeh, S.M.; Khodabakhshi, A.; Abaei Hassani, P.; Vaferi, B. Smart Identification of Petroleum Reservoir Well Testing Models Using Deep Convolutional Neural Networks (GoogleNet). J. Energy Res. Technol. 2021, 143, 073008. [Google Scholar] [CrossRef]
- Valentín, M.B.; Bom, C.R.; Coelho, J.M.; Correia, M.D.; De Albuquerque, M.P.; de Albuquerque, M.P.; Faria, E.L. A Deep Residual Convolutional Neural Network for Automatic Lithological Facies Identification in Brazilian Pre-Salt Oilfield Wellbore Image Logs. J. Petrol. Sci. Eng. 2019, 179, 474–503. [Google Scholar] [CrossRef]
- Dias, L.O.; Bom, C.R.; Faria, E.L.; Valentín, M.B.; Correia, M.D.; De Albuquerque, M.P.; De Albuquerque, M.P.; Coelho, J.M. Automatic Detection of Fractures and Breakouts Patterns in Acoustic Borehole Image Logs Using Fast-Region Convolutional Neural Networks. J. Petrol. Sci. Eng. 2020, 191, 107099. [Google Scholar] [CrossRef]
- Cheng, Y.; Fu, L.-Y. Nonlinear Seismic Inversion by Physics-Informed Caianiello Convolutional Neural Networks for Overpressure Prediction of Source Rocks in the Offshore Xihu Depression, East China. J. Petrol. Sci. Eng. 2022, 215, 110654. [Google Scholar] [CrossRef]
- Moghimihanjani, M.; Vaferi, B. A Combined Wavelet Transform and Recurrent Neural Networks Scheme for Identification of Hydrocarbon Reservoir Systems from Well Testing Signals. J. Energy Res. Technol. 2021, 143, 013001. [Google Scholar] [CrossRef]
- Dong, L.; Xiao, Q.; Jia, Y.; Fang, T. Review of Research on Intelligent Diagnosis of Oil Transfer Pump Malfunction. Petroleum 2023, 9, 135–142. [Google Scholar] [CrossRef]
- Peng, Y.; Zhao, R.; Zhang, X.; Shi, J.; Chen, S.; Gan, Q.; Li, G.; Zhen, X.; Han, T. Innovative Convolutional Neural Networks Applied in Dynamometer Cards Generation. In Proceedings of the SPE Western Regional Meeting, San Jose, CA, USA, 23–26 April 2019; p. D031S010R002. [Google Scholar]
- He, Y.-P.; Zang, C.-Z.; Zeng, P.; Wang, M.-X.; Dong, Q.-W.; Wan, G.-X.; Dong, X.-T. Few-Shot Working Condition Recognition of a Sucker-Rod Pumping System Based on a 4-Dimensional Time-Frequency Signature and Meta-Learning Convolutional Shrinkage Neural Network. Pet. Sci. 2023, 20, 1142–1154. [Google Scholar] [CrossRef]
- Shan, L.; Liu, Y.; Tang, M.; Yang, M.; Bai, X. CNN-BiLSTM Hybrid Neural Networks with Attention Mechanism for Well Log Prediction. J. Petrol. Sci. Eng. 2021, 205, 108838. [Google Scholar] [CrossRef]
- Chen, Z.; Yu, W.; Liang, J.-T.; Wang, S.; Liang, H. Application of Statistical Machine Learning Clustering Algorithms to Improve EUR Predictions Using Decline Curve Analysis in Shale-Gas Reservoirs. J. Petrol. Sci. Eng. 2022, 208, 109216. [Google Scholar] [CrossRef]
- Daolun, L.I.; Xuliang, L.I.U.; Wenshu, Z.H.A.; Jinghai, Y.; Detang, L.U. Automatic Well Test Interpretation Based on Convolutional Neural Network for a Radial Composite Reservoir. Pet. Explor. Dev. 2020, 47, 623–631. [Google Scholar] [CrossRef]
- Cao, L.; Lv, M.; Li, C.; Sun, Q.; Wu, M.; Xu, C.; Dou, J. Effects of Crosslinking Agents and Reservoir Conditions on the Propagation of Fractures in Coal Reservoirs during Hydraulic Fracturing. Reserv. Sci. 2025, 1, 36–51. [Google Scholar] [CrossRef]
- Li, M.; Liu, J.; Xia, Y. Risk Prediction of Gas Hydrate Formation in the Wellbore and Subsea Gathering System of Deep-Water Turbidite Reservoirs: Case Analysis from the South China Sea. Reserv. Sci. 2025, 1, 52–72. [Google Scholar] [CrossRef]
- Wang, F.; Kobina, F. The Influence of Geological Factors and Transmission Fluids on the Exploitation of Reservoir Geothermal Resources: Factor Discussion and Mechanism Analysis. Reserv. Sci. 2025, 1, 3–18. [Google Scholar] [CrossRef]
- Andrade, G.M.; de Menezes, D.Q.; Soares, R.M.; Lemos, T.S.; Teixeira, A.F.; Ribeiro, L.D.; Vieira, B.F.; Pinto, J.C. Virtual Flow Metering of Production Flow Rates of Individual Wells in Oil and Gas Platforms through Data Reconciliation. J. Petrol. Sci. Eng. 2022, 208, 109772. [Google Scholar] [CrossRef]
- Zhong, Z.Y.; Zhi, X.L.; Yi, W.J. Oil Well Real-Time Monitoring with Downhole Permanent FBG Sensor Network. In Proceedings of the 2007 IEEE International Conference on Control and Automation, Guangzhou, China, 30 May–1 June 2007; pp. 2591–2594. [Google Scholar]
- Han, D.; Kwon, S. Application of Machine Learning Method of Data-Driven Deep Learning Model to Predict Well Production Rate in the Shale Gas Reservoirs. Energies 2021, 14, 3629. [Google Scholar] [CrossRef]
- Weiwei, T.; Jie, L.I.; Wenyuan, M.A.; Hui, Z.; Gang, F.; Chaodong, T.A.N.; Wenhao, Z. Optimization Design of Production Parameters of Rod Pump Lifting System of Pumping Unit Based on Mechanism Simulation. Drill. Prod. Technol. 2024, 47, 108. [Google Scholar]
- Wang, H.; Qiao, L.; Lu, S.; Chen, F.; Fang, Z.; He, X.; Zhang, J.; He, T. A Novel Shale Gas Production Prediction Model Based on Machine Learning and Its Application in Optimization of Multistage Fractured Horizontal Wells. Front. Earth Sci. 2021, 9, 726537. [Google Scholar] [CrossRef]
- Dong, Y.; Zhang, Y.; Liu, F.; Zhu, Z. Research on an Optimization Method for Injection-Production Parameters Based on an Improved Particle Swarm Optimization Algorithm. Energies 2022, 15, 2889. [Google Scholar] [CrossRef]
- Zhang, X.-Q.; Cheng, Q.-L.; Sun, W.; Zhao, Y.; Li, Z.-M. Research on a TOPSIS Energy Efficiency Evaluation System for Crude Oil Gathering and Transportation Systems Based on a GA-BP Neural Network. Pet. Sci. 2024, 21, 621–640. [Google Scholar] [CrossRef]
- Elman, J.L. Finding Structure in Time. Cogn. Sci. 1990, 14, 179–211. [Google Scholar] [CrossRef]
- Connor, J.T.; Martin, R.D.; Atlas, L.E. Recurrent Neural Networks and Robust Time Series Prediction. IEEE Trans. Neural Netw. 1994, 5, 240–254. [Google Scholar] [CrossRef]
- Yang, L.; Sun, S.Z. Seismic Horizon Tracking Using a Deep Convolutional Neural Network. J. Petrol. Sci. Eng. 2020, 187, 106709. [Google Scholar] [CrossRef]
- Haffner, F.; Lacoue-Negre, M.; Pirayre, A.; Gonçalves, D.; Gornay, J.; Moreaud, M. IPA: A Deep CNN Based on Inception for Petroleum Analysis. Fuel 2025, 379, 133016. [Google Scholar] [CrossRef]
- Dawotola, A.W.; Van Gelder, P.; Vrijling, J.K. Risk Assessment of Petroleum Pipelines Using a Com-Bined Analytical Hierarchy Process-Fault Tree Analysis (AHP-FTA). In Proceedings of the 7th International Probabilistic Workshop, Delft, The Netherlands, 5–6 November 2009. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).









