Next Article in Journal
Fixed-Time Event-Triggered Control for Distributed Unmanned Underwater Vehicles
Next Article in Special Issue
A Deep Learning-Integrated Framework for Operational Rip Current Warning
Previous Article in Journal
Spatiotemporal Characteristics and Long-Term Variability of Large-Wave Frequency in the Northwest Pacific
Previous Article in Special Issue
Research on the Detection of Ocean Internal Waves Based on the Improved Faster R-CNN in SAR Images
 
 
Due to scheduled maintenance work on our servers, there may be short service disruptions on this website between 11:00 and 12:00 CEST on March 28th.
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Physics-Informed Fine-Tuned Neural Operator for Flow Field Modeling

1
College of Environmental and Resource Sciences, Zhejiang University, Hangzhou 310027, China
2
Key Laboratory of Coastal Environment and Resources of Zhejiang Province, School of Engineering, Westlake University, Hangzhou 310030, China
*
Authors to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2026, 14(2), 201; https://doi.org/10.3390/jmse14020201
Submission received: 13 December 2025 / Revised: 6 January 2026 / Accepted: 14 January 2026 / Published: 19 January 2026
(This article belongs to the Special Issue Artificial Intelligence and Its Application in Ocean Engineering)

Abstract

Modeling flow field evolution accurately is important for numerous natural and engineering applications, such as pollutant dispersion in the ocean and atmosphere, yet remains challenging because of the highly nonlinear, multi-physics, and high-dimensional features of flow systems. While traditional equation-based numerical methods suffer from high computational costs, data-driven neural networks struggle with insufficient data and lack physical explainability. The physics-informed neural operator (PINO) addresses this by combining physics and data losses but faces a fundamental gradient imbalance problem. This work proposes a physics-informed fine-tuned neural operator for high-dimensional flow field modeling that decouples the optimization of physics and data losses. Our method first trains the neural network using data loss and then fine-tunes it with physics loss before inference, enabling the model to adapt to evaluation data while respecting physical constraints. This strategy requires no additional training data and can be applied to fit out-of-distribution (OOD) inputs faced during inference. We validate our method using the shallow water equation and advection–diffusion equation using a convolutional neural operator (CNO) as the base architecture. Experimental results show a 26.4% improvement in single-step prediction accuracy and a reduction in error accumulation for multi-step predictions.

1. Introduction

Flow fields represent fundamental physical phenomena that govern numerous natural and engineering processes [1,2,3,4]. Natural processes encompass critical environmental phenomena such as pollutant dispersion in oceans and the atmosphere [5], where understanding flow field evolution is essential for predicting contamination spread and assessing environmental impacts; ocean circulation systems, which transport heat, nutrients, and carbon across global scales and play a crucial role in climate regulation [6]; and atmospheric flows, including weather patterns and air quality dynamics that directly affect human health and agricultural productivity [7]. Engineering applications include chemical reactors, aerospace engineering, and energy systems [8]. The accurate modeling of flow field evolution is crucial for understanding, predicting, and controlling these complex systems [9]. However, flow fields exhibit inherent challenges: They are nonlinear, multi-physics coupled, and high-dimensional systems, making their accurate and efficient modeling a significant challenge [9].
Traditional approaches to flow field modeling rely on numerical methods based on Partial Differential Equations (PDEs). The types of numerical methods mainly include finite difference methods, finite element methods, and finite volume methods [10,11]. While these methods have been extensively developed and validated, they suffer from substantial computational costs, particularly for high-resolution simulations and multi-physics coupling. Moreover, these numerical schemes may face stability issues and divergence failures, especially in complex flow regimes [12].
The development of deep learning has opened significant potential avenues for flow field modeling through data-driven neural network approaches [13,14,15]. These methods can learn complex nonlinear mappings from input initial conditions, boundary conditions, and physical parameters to future states directly from data, potentially offering significant computational advantages over traditional numerical methods [16]. However, data-driven approaches face critical limitations: They exhibit poor accuracy when training data is insufficient, and their black-box nature leads to violations of physical principles and a lack of explainability [17,18].
To address these limitations, researchers have developed the physics-informed neural operator (PINO), physics-informed operator learning, and their variants. They incorporate physics by integrating physics loss terms with the data loss [19,20,21]. These methods jointly optimize neural network parameters using a combined loss function: L t o t a l = L d a t a + λ · L p h y s i c s , where λ is a weighting parameter. While promising, this approach faces a fundamental challenge: The gradients of physics loss and data loss differ significantly in magnitude and direction. When λ is small, the physics loss provides minimal constraint, resulting in small accuracy improvements. Conversely, when λ is large, the data loss becomes dominated, causing the optimization to violate the data distribution and potentially increase prediction errors. This gradient imbalance makes it difficult to optimize neural networks in many systems and experimental settings [22,23].
In this work, we propose a physics-informed fine-tuned neural operator for high-dimensional flow field modeling that addresses the gradient imbalance problem and effectively utilizes the physical information by decoupling the optimization of physics loss and data loss. Our approach consists of two distinct phases: (1) Data-Driven Pretraining phase: The deep model is first trained using the training set with only data loss, allowing it to learn the data distribution effectively. (2) Physics-Informed Fine-tuning phase: Before inference on evaluation data, we compute physics loss from the model’s input of evaluation data and output predictions and perform additional optimization steps. This enables the model to transfer to the distribution of specific evaluation data using physical information. There are two advantages of the proposed method. First, it requires no additional training data beyond the original dataset yet improves accuracy through the incorporation of physical information. Second, by separating the optimization phases, we avoid the gradient imbalance problem that plagues joint optimization approaches in physics-informed machine learning for flow field modeling.
To validate our method, we conduct experiments on two distinct flow field systems: (1) Shallow water equation, which models the height field and two-directional velocity fields. This equation is fundamental for understanding wave propagation, tsunami forecasting, and coastal hydrodynamics, playing a critical role in oceanography and atmospheric science for predicting storm surges and flood dynamics [24]. (2) Advection–diffusion equations, which describe the evolution of the concentration field, two-directional velocity fields, and pressure fields. These equations are essential for modeling pollutant transport in environmental systems, chemical reaction processes in industrial applications, and heat and mass transfer phenomena, making them crucial for environmental protection and process optimization [25]. For both systems, we employ a convolutional neural operator (CNO) [26] as the base model architecture, which is suited for 2D spatial–temporal fields. We integrate our physics-informed fine-tuning method with the CNO (PFT-CNO) and evaluate its performance against baseline techniques.
Results of our experiments show the effectiveness of the proposed PFT-CNO. After physics-informed fine-tuning, the model achieves an 8.3% and 26.4% improvement in single-step prediction accuracy compared to the baseline for two systems, respectively. Moreover, in long-term multi-step predictions, the method reduces error accumulation while maintaining relatively better accuracy throughout the prediction horizon.
The main contributions of this work are summarized as follows:
1.
We present a physics-informed fine-tuning method for high-dimensional flow field modeling, which separates the optimization of physics loss and data loss into distinct phases. By fine-tuning using inference data input via physics loss, it addresses the gradient imbalance problem that plagues joint optimization approaches in physics-informed neural operators.
2.
We apply the physics-informed fine-tuning method to the CNO and propose PFT-CNO, which combines the ability of the CNO to handle high-dimensional spatio-temporal flow fields with the advantages of physics-informed fine-tuning, requiring no additional training data yet significantly improving prediction accuracy through physics-informed fine-tuning during inference.
3.
We demonstrate the effectiveness of our method on two distinct flow field systems (shallow water equation and advection–diffusion equation), achieving substantial improvements in both single-step and multi-step predictions, validating the practical utility of PFT-CNO.

2. Related Works

The emergence of deep learning has revolutionized flow field modeling through data-driven approaches. Convolutional neural networks (CNNs) and U-Net have been applied to learn spatial patterns in flow fields [27], while Recurrent Neural Networks (RNNs) and Long Short-Term Memory (LSTM) networks capture temporal dependencies [28,29]. Graph Neural Networks (GNNs) have been employed for irregular mesh-based flow simulations [30,31]. However, these methods discretize the variable-to-variable mapping of the flow field, ignoring the continuity characteristics of the flow field.
Neural operators represent a significant advancement over traditional neural networks by learning mappings between function spaces, enabling generalization to different resolutions and domains. The Fourier Neural Operator (FNO) leverages spectral methods to efficiently process spatial information [32]. DeepONet uses branch and trunk networks to approximate nonlinear operators [33]. The Convolutional Neural Operator (CNO) employs filtered convolutions to maintain spatial locality while learning operator mappings [26]. Transolver introduces attention-based learning of flow fields, enabling the neural network to achieve intrinsic geometry-independent ability [34]. The Wavelet Diffusion Neural Operator (WDNO) introduces the wavelet transform into the diffusion model, enabling more accurate predictions for complex high-dimensional fluid dynamics [35]. These methods have shown promise in learning PDE solution operators, but they still heavily rely on data, have poor accuracy with limited training data, and violate physical laws.
Physics-Informed Neural Networks (PINNs) incorporate physical knowledge by adding physics loss terms to the training objective [36]. The physics loss is computed by evaluating the residual of the governing PDEs at collocation points, ensuring the neural network solution satisfies the underlying physical laws. PINNs have been extended to various applications, including heat transfer [37] and wave propagation [38]. However, PINNs typically solve specific instances of PDEs rather than learning general solution operators. To combine the operator learning capability of neural operators with the physics constraints of PINNs, researchers have developed physics-informed neural operators (PINOs), physics-informed operating learning (PI-DeepONet) [19,39], and others [40,41]. These methods incorporate physics loss into the training of neural operators, enabling them to learn solution operators while respecting physical laws. However, these methods face a fundamental challenge: The joint optimization of data loss and physics loss suffers from gradient imbalance, where the gradients of these two loss components differ significantly in magnitude and direction. This imbalance makes it difficult to leverage physics loss in many cases, often resulting in either insufficient physics constraints or violation of data distributions [22]. Although the physics-informed fine-tuning method has also been introduced in previous works [15,42,43,44], these only studied modeling on partial observations or simple 1D equations [42] or studied transfer learning of PINNs instead of neural operators with generalization ability. We propose PFT-CNO for the first time in this work, thus extending the physics-informed fine-tuned neural operator to 2D high-dimensional flow field modeling.

3. Materials and Methods

3.1. Problem Setup

We consider the problem of learning the solution operator for a family of parametric PDEs. Neural operators extend traditional neural networks from finite-dimensional vector mappings to infinite-dimensional functional mappings, naturally accommodating different mesh resolutions and discretization schemes. To formalize this problem, we consider a parametric PDE defined on a spatial domain D and temporal domain ( 0 , ) :
u t = R ( u ) , in D × ( 0 , ) , u = g , in D × ( 0 , ) , u = u ( 0 ) , in D × { 0 } ,
where R is a possibly nonlinear partial differential operator associated with Banach spaces U and A , g is the known boundary conditions, and u ( 0 ) denotes the initial condition. The unknown function u ( t ) U , where t > 0 , represents the solution we aim to approximate. We assume that u exists and is bounded for all time and for every initial condition in the appropriate function space.
The definition leads to an operator of solution G : A U that maps input functions a A (initial conditions, boundary conditions, and coefficient functions) to output solution functions u U that satisfy the governing PDE. Prototypical examples of such problems include the shallow water equation, advection–diffusion equation, and Navier–Stokes equation, which are fundamental in fluid dynamics and exhibit complex nonlinear behavior.

3.2. Neural Operator

To approximate the solution operator G , we employ neural operators, which are generalizations of standard deep neural networks to the operator setting. Neural operators compose linear integral operators with nonlinear activation functions to approximate highly nonlinear operators. Formally, we define the neural operator model as follows:
G θ : = Q σ ( W L + K L ) σ ( W 1 + K 1 ) P ,
where θ denotes all learnable parameters. This architecture consists of three key components:
  • Lifting operator P: P : R d a R d 1 lifts the lower-dimensional input function a A (with co-dimension d a ) into a higher-dimensional latent space.
  • Iterative layers: The model stacks L layers of σ ( W l + K l ) , where W l R d l + 1 × d l are local linear operators, K l : { D R d l } { D R d l + 1 } are integral kernel operators that capture non-local dependencies in function spaces, and σ denotes pointwise activation functions.
  • Projection operator Q: Q : R d L R d u projects the higher-dimensional function back to the output space (with co-dimension d u ).
The neural operator G θ learns to approximate the above-mentioned operator G by combining local operations with global integral operations, enabling it to capture both fine-scale local features and long-range dependencies in the solution space. This framework provides the foundation for learning mappings between function spaces in operator learning problems.

3.3. Convolutional Neural Operator

As a specific instantiation of the neural operator, we introduce the convolutional neural operator (CNO) [26], which specializes the integral kernel operators K l to convolutional operators. The CNO is designed to operate on the space of bandlimited functions B w ( D ) , where D is the spatial domain and w denotes the bandwidth. Under this framework, the CNO maintains the general form G θ = Q σ ( W L + K L ) σ ( W 1 + K 1 ) P , with the following specializations:
  • Lifting operator P: The lifting operator is instantiated as a convolutional layer that maps the input function u B w ( D , R d x ) to the latent space v 0 B w ( D , R d 0 ) , where d 0 > d x is the number of latent channels.
  • Iterative layers: Each layer l implements the transformation v l + 1 = σ ( W l ( v l ) + K l ( v l ) ) , where
    -
    K l is a convolutional operator with learnable kernels that captures spatial dependencies;
    -
    W l represents skip connections or residual mappings;
    -
    σ applies pointwise nonlinear activations.
    Additionally, resolution adjustment operators (upsampling or downsampling) may be incorporated within layers to enable multi-scale feature extraction, inspired by architectures for image generation.
  • Projection operator Q: The projection operator is realized as a convolutional layer that maps the final latent function v L B w ( D , R d L ) to the output space u B w ( D , R d y ) .
The core computational unit of the CNO is the convolution operator K, which employs learnable kernels of size k × k :
K ( i , j ) = m = 0 k 1 n = 0 k 1 w ( m , n ) · v ( i + m , j + n ) ,
where w ( m , n ) represents trainable kernel weights. A distinguishing feature of the CNO compared to spectral methods such as the FNO is its direct operation in the spatial domain through “real-space” transformations, rather than relying on frequency-domain convolutions via Fourier transforms. This spatial parameterization enhances locality, naturally accommodates non-periodic boundary conditions, and extends readily to irregular geometries. To enable multi-scale feature extraction, it introduces resolution-changing operators, including a downsampling block, upsampling block, residual block, etc. [26].
To effectively capture multi-scale features inherent in flow fields, a modified U-Net architecture is employed for the CNO implementation. The U-Net [45], originally developed for image segmentation tasks, features a symmetric encoder–decoder structure with skip connections that directly link corresponding resolution levels. This design is particularly well-suited for the CNO in several aspects. First, the encoder progressively reduces spatial resolution by downsampling while expanding channel capacity, enabling the extraction of increasingly abstract and global features. Second, the decoder reconstructs fine-scale details by gradually upsampling while reducing channel dimensions. Third, the skip connections through residual blocks bypass the bottleneck to transfer high-frequency information directly from encoder to decoder at matching resolutions, preventing the loss of fine-scale features during the encoding process.
A key property of the CNO is its structure-preserving nature: bandlimited input functions are mapped to bandlimited output functions, respecting the continuous–discrete equivalence. This property enables the operator to generalize across various discretizations while maintaining underlying continuous structures of solution spaces. Moreover, the universal approximation of the CNO is proved by [26]. By leveraging efficient convolutional operations combined with multi-scale processing, the CNO is particularly well-suited for learning operators on high-dimensional spatio-temporal flow fields.

3.4. Physics-Informed Fine-Tuned CNO

In this work, we utilize the CNO as the base model and integrate the physics-informed fine-tuning strategy and then propose the physics-informed fine-tuned convolutional neural operator (PFT-CNO) to adapt to high-dimensional flow field modeling. While the neural network architecture of PFT-CNO is identical to the standard CNO model described in Section 3.3, we employ a novel two-phase training strategy as illustrated in Figure 1 to enhance generalization and mitigate out-of-distribution (OOD) performance degradation. The pseudocode and hyperparameters are shown in Appendix A and Appendix B.

Two-Phase Training Strategy

Phase I: Data-Driven Pretraining. The first phase follows the conventional supervised learning paradigm. We construct the training dataset D = { ( u t ( i ) , u t + 1 ( i ) ) } i = 1 N t r a i n by solving Equation (1) using numerical methods on discrete temporal and spatial domains. Each sample consists of an input–output pair ( u t , u t + 1 ) , where u t represents the system state at time t and u t + 1 represents the corresponding state at time t + 1 . The pretraining objective minimizes the data loss, defined as follows:
L d a t a = 1 N t r a i n i = 1 N t r a i n u ^ t + 1 ( i ) u t + 1 ( i ) 2 u t + 1 ( i ) 2 ,
where u ^ t + 1 ( i ) = G θ ( u t ( i ) ) and · 2 represents the L 2 norm. This phase enables the model to learn the general mapping from the training data distribution.
Phase II: Physics-Informed Fine-Tuning. After pretraining, conventional approaches would directly test the model on the evaluation dataset. However, we introduce an intermediate physics-informed fine-tuning phase to adapt the model to the distribution of evaluation data. The evaluation dataset B = { u t ( j ) , u t + 1 ( j ) } j = 1 N e v a l is generated using the same numerical solver as D , but with different initial conditions and physical parameters (e.g., viscosity coefficient, diffusion coefficient, and Reynolds number).
Crucially, during fine-tuning, we only have access to the input states u t ( j ) from B , not the ground truth outputs u t + 1 ( j ) , as these represent the quantities to be predicted. To overcome this challenge, we leverage the underlying physical information to construct a physics loss function. After the pretraining, data loss has initialized the model parameters to a constrained range. Here, the goal of using physics loss is to start from the point after data loss optimization and further constrain the model parameters to the optimal solution that fits the distribution of the evaluation dataset, avoiding the imbalance problem caused by synchronous competition between the two loss functions. Specifically, the physics loss is computed using the predicted output u ^ t + 1 ( j ) = G θ ( u t ( j ) ) :
L p h y s i c s = 1 N e v a l j = 1 N e v a l | u ^ t + 1 ( j ) t R ( u ) | 2 .
Unlike methods that employ automatic differentiation (autograd) to compute spatial and temporal derivatives, we utilize finite difference approximations following [15,46], which significantly improves computational efficiency and reduces memory consumption. This design choice is particularly advantageous for high-resolution spatial domains where autograd-based approaches become prohibitively expensive.
It is worth noting that although in our work the CNO is employed as the base architecture and relative L 2 error as the data loss, with the finite difference method and MSE as the physics loss, our method can also be extended to other base models, and data loss and physics loss functions.

4. Results

We validate the performance of our proposed PFT-CNO method on two distinct PDE systems: The shallow water equation (SWE) and the advection–diffusion equation (ADE). Our method is compared against several baseline approaches to demonstrate the improvements achieved through physics-informed fine-tuning. The experimental results show that PFT-CNO consistently outperforms existing methods in both single-step and multi-step predictions, demonstrating the effectiveness of decoupling physics loss and data loss optimization.

4.1. Baseline Methods

To examine the modeling accuracy of PFT-CNO, we compare PFT-CNO with the following baseline approaches.
  • DeepONet [33]: This is a data-driven neural operator that uses branch and trunk networks to approximate nonlinear operators. This method serves as a representative baseline for purely data-driven operator learning approaches.
  • PFT-DeepONet [42]: The physics-informed fine-tuning method is applied to DeepONet and validated on simple 1D problems. It serves as a baseline to compare the advantages of our proposed PFT-CNO over existing models for more complex systems.
  • Transolver [34]: This is a data-driven transformer-based neural operator that leverages attention mechanisms to achieve geometry-independent learning of flow fields. This method demonstrates the capability of transformer architectures in operator learning.
  • CNO [26]: It serves as our base model architecture. This purely data-driven method provides the foundation for our physics-informed fine-tuning approach and allows us to isolate the contribution of the fine-tuning strategy.
  • PD-CNO: The Physics-Driven CNO is a variant of the CNO trained exclusively using physics loss without any data loss. This baseline demonstrates the performance when relying solely on physical constraints, highlighting the importance of data-driven initialization.
  • PI-CNO [40]: This is the Physics-Informed CNO, which jointly optimizes both data loss and physics loss during training using the combined objective L t o t a l = L d a t a + λ · L p h y s i c s . This method represents the conventional approach to incorporating physical information into neural operators and allows us to compare against joint optimization strategies.

4.2. Evaluation Metrics

We employ five metrics to evaluate the performance of all methods.
  • Relative L 2 Error: This metric measures the normalized prediction error relative to the ground truth magnitude, defined as follows:
    Rel L 2 Error = 1 N i = 1 N u ^ u 2 u 2 ,
    where N is the total number of spatial points, u represents the ground truth, and u ^ denotes the predicted solution, which is the output of neural operators G θ defined in Section 3.2. As we mentioned in Section 3.1, G θ maps input functions a A (e.g., initial conditions, boundary conditions, or coefficient functions) to output solution functions u U that satisfy the governing PDE. This metric provides a scale-invariant measure of prediction accuracy.
  • Root Mean Square Error (RMSE): The RMSE can be expressed in terms of the standard deviations and correlation coefficient:
    RMSE = σ o 2 + σ p 2 2 σ o σ p R ,
    where σ o and σ p represent ground truth and predicted standard deviations, respectively, and R represents the Pearson correlation coefficient between them. This metric penalizes variance differences and insufficient correlation.
  • Mean Absolute Error (MAE): The MAE quantifies the average absolute deviation between predictions and the ground truth:
    MAE = 1 N i = 1 N | u ^ i u i | .
    This metric offers an intuitive interpretation of prediction errors in the original units.
  • Coefficient of Determination ( R 2 ): The R 2 score measures the proportion of variance in the ground truth explained by the predictions:
    R 2 = 1 i = 1 N ( u i u ^ i ) 2 i = 1 N ( u i u ¯ ) 2 ,
    where u ¯ = 1 N i = 1 N u i is the mean of the ground truth. An R 2 value close to 1 indicates excellent predictive performance, while negative values suggest predictions worse than the mean.
  • Inference Time: This metric measures the computational efficiency by recording the average time required to perform a single forward prediction. Lower inference times indicate better computational efficiency for real-time applications. We measure the inference time of all methods using an NVIDIA H800 GPU, 64 cores, and a batch size of 64.
We also use additional metrics to evaluate the statistical significance of model performance differences following the previous work [47].
  • ANOVA: This assesses whether significant differences exist among multiple models using the F-statistic ( p < 0.05 indicates significance).
    F = i = 1 k n i ( u ¯ i u ¯ ) 2 / ( k 1 ) i = 1 k j = 1 n i ( u i j u ¯ i ) 2 / ( N k ) ,
    where k is the number of models, n i is the sample size of the i-th model, N is the total sample size, u ¯ i is the mean of the i-th model, and x ¯ is the overall mean.
  • T-test: The test compares performance differences between two models on the same data.
    t = d ¯ σ d / n ,
    where d ¯ is the mean of paired differences (MSE1 − MSE2), σ d is the standard deviation of paired differences, and n is the sample size.
  • Cohen’s d: It quantifies the magnitude of the results between two models. For paired samples, it is calculated as follows:
    d = d ¯ σ d .
    d indicates which model performs better: d > 0 suggests that Model 2 has better performance, while d < 0 indicates that Model 1 has better performance.

4.3. Shallow Water Equation

The SWE is a fundamental system in geophysical fluid dynamics [24], modeling the evolution of fluid height and horizontal velocity fields. This system provides a benchmark for evaluating our method on multi-field, nonlinear PDE systems.

4.3.1. Data Collection

The SWE in two dimensions is given by
h t + ( h u ) x + ( h v ) y = 0 , ( h u ) t + x h u 2 + 1 2 g h 2 + ( h u v ) y f h v = ν 2 ( h u ) x 2 + 2 ( h u ) y 2 , ( h v ) t + ( h u v ) x + y h v 2 + 1 2 g h 2 + f h u = ν 2 ( h v ) x 2 + 2 ( h v ) y 2 ,
where h ( x , y , t ) represents the fluid height, u ( x , y , t ) and v ( x , y , t ) denote the velocity field components in the x and y directions, respectively, g = 9.81 m/s2 is the gravitational acceleration, f is the Coriolis parameter, and ν is the viscosity coefficient.
We solve these equations numerically using central difference schemes for spatial derivatives and the explicit fourth-order Runge–Kutta integrator for temporal discretization. The simulations are conducted on a spatial domain discretized with a resolution of 128 × 128 grid points, employing periodic boundary conditions. The temporal discretization uses a time step of Δ t = 0.001 s, and we integrate the system for a total duration of T = 1 s, resulting in 1000 time steps per numerical simulation. Then, we downsample 1000 time steps to 100 time steps for training and evaluation to ensure distinctions between adjacent steps.
To ensure clear evaluation across different flow regimes, we generate 10 distinct trajectories by varying initial conditions and physical parameters. We initialize all three state variables (fluid height h and velocity components u and v) using the Gaussian Random Field (GRF) generated via a spectral method in Fourier space following previous works [19,21]. For each numerical simulation, we sample the GRF parameters from uniform distributions to ensure diverse initial conditions that capture a wide range of flow regimes while maintaining physically reasonable perturbation amplitudes. Moreover, the Coriolis parameter f is randomly sampled from a uniform distribution over the range [ 0 , 1.5 × 10 4 ] to control rotational effects, where this range is selected based on realistic Earth Coriolis parameter values. The viscosity coefficient ν is randomly sampled from a uniform distribution over the range [ 10 5 , 10 3 ] to govern the dissipative behavior. Among these 10 trajectories, 5 are allocated for training the neural operators, while the remaining 5 are reserved for evaluation, enabling assessment of generalization to unseen parameter configurations.

4.3.2. Single-Step Prediction Performance

Table 1 presents the quantitative results for single-step prediction on the evaluation set of the SWE. The results demonstrate the superior performance of our proposed PFT-CNO across all evaluation metrics.
The results reveal several key observations. First, the purely data-driven methods DeepONet and Transolver exhibit higher errors (relative L 2 errors of 0.16331 and 0.16255, respectively) and negative R 2 scores. This demonstrates the challenge of learning high-dimensional flow field dynamics with limited training data when relying solely on data-driven approaches. Moreover, the base model CNO training in a data-driven manner achieves better performance with a relative L 2 error of 0.01284 and an R 2 score of 0.99304, demonstrating the effectiveness of the convolutional architecture for the 2D flow field.
Second, PI-CNO, which jointly optimizes data and physics losses, achieves a slight improvement over CNO (2.18% improvement), demonstrating the potential benefits of incorporating physical information. However, the improvement is modest, likely due to the gradient imbalance problem discussed in the Introduction.
Most importantly, our proposed PFT-CNO shows the best accuracy, with a relative L 2 error of 0.01178 (8.26% improvement over CNO and 6.21% improvement over PI-CNO) and an R 2 score of 0.99394. These results validate the effectiveness of decoupling physics loss and data loss optimization, allowing the model to leverage physical constraints during inference, avoiding the gradient imbalance issues that plague joint optimization approaches. Compared with the existing PFT-DeepONet method, PFT-CNO has significant advantages in spatial feature extraction ability due to its convolutional neural network architecture, which can better capture local and global correlations in high-dimensional space. The inference time of PFT-CNO is also comparable with that of CNO because they have the same architectures. Compared to the computation time of numerical solvers (0.32 s), the speed of neural operators has been greatly improved (at least 43 times), which is an important advantage over numerical solvers. This is also the core motivation for developing fluid solvers based on neural operators, which is to use the high efficiency of neural operators to accelerate simulations [48]. On the other hand, using physics loss directly to train the model instead of fine-tuning (PD-CNO) results in greater errors, as it ignores the data distribution and overfits on the physical equation. This highlights the importance of initialization via a data-driven method and the limitations of physics-only training.
Figure 2 provides visual comparisons of the predictions from all methods, further illustrating that PFT-CNO captures the complex flow fields more accurately. More visualizations are shown in Appendix C.

4.3.3. Multi-Step Prediction Performance

To evaluate the long-term predictive capability and error accumulation behavior, we perform autoregressive multi-step predictions by iteratively applying each model to predict 10 consecutive time steps. At each step, the model takes the previous prediction as input to generate the next state, simulating real-world forecasting scenarios where the ground truth is unavailable.
Figure 3a shows the evolution of the relative L 2 error over the 10-step prediction horizon for all methods. The average relative L 2 error is shown in Table 2. The results reveal critical differences in error accumulation behavior.
The purely data-driven CNO method exhibits substantial error accumulation, with the relative L 2 error increasing rapidly over the prediction steps. This behavior is characteristic of autoregressive models where small errors at each step compound over time, leading to significant deviations from the true solution of the simulation.
In contrast, PFT-CNO demonstrates stability, maintaining consistently low errors throughout the prediction horizon. The physics-informed fine-tuning process enables the model to adapt to the evolving prediction distribution at each step, constraining the predictions to remain physically consistent even as the system state deviates from the training data. The superior multi-step performance of PFT-CNO further validates the effectiveness of our decoupled optimization strategy, demonstrating that physics-informed fine-tuning mitigates error accumulation and maintains prediction accuracy over extended time horizons.

4.4. Advection–Diffusion Equation

The ADE is essential for modeling pollutant transport both in the ocean and atmosphere [25]. The ADE represents a more challenging benchmark due to its inherent complexity, coupling diffusion dynamics with the Navier–Stokes equations to simultaneously govern concentration transport, velocity fields, and pressure distributions. This multi-physics system poses substantial difficulties for learning-based methods, particularly in low-data regimes.

4.4.1. Data Collection

The ADE in two dimensions, coupled with the incompressible Navier–Stokes equations for the velocity and pressure fields, is given by
c t + u c x + v c y = D 2 c x 2 + 2 c y 2 + S , u t + u u x + v u y = 1 ρ p x + ν 2 u x 2 + 2 u y 2 , v t + u v x + v v y = 1 ρ p y + ν 2 v x 2 + 2 v y 2 , u x + v y = 0 ,
where c ( x , y , t ) represents the concentration field, u ( x , y , t ) and v ( x , y , t ) denote the velocity components, p ( x , y , t ) is the pressure field, S ( x , y , t ) is the source term, D is the diffusion coefficient, ν represents the viscosity coefficient, and ρ demotes the fluid density.
We employ the same numerical discretization strategy as for the SWE, where the central difference scheme is used for spatial derivatives and the RK4 method is used for temporal integration. The simulations are conducted on a 128 × 128 spatial grid with periodic boundary conditions, using a time step of Δ t = 0.0005 s and integrating for a total duration of T = 0.6 s, resulting in 1200 time steps per numerical simulation. To ensure that the system has achieved a developed flow state, we utilize the last 1000 time steps from each numerical simulation and downsample to 100 steps for training and evaluation. This is due to the small range of point source diffusion in the initial time step, leading to little difference between trajectories.
The initial conditions employ a two-stage sampling. The velocity fields are generated using the GRF, while the concentration field starts from zero and evolves through a continuous Gaussian point source. First, we generate the initial velocity field ( u 0 , v 0 ) using the same Fourier-based spectral method as described above, but with the incompressibility constraint · u = 0 strictly enforced. Second, the concentration field c 0 is initialized as zero throughout the domain. The initial concentration distribution emerges through a continuous Gaussian point source term S ( x , y ) in the ADE:
S ( x , y ) = Q 2 π σ 2 exp ( x x s ) 2 + ( y y s ) 2 2 σ 2 ,
where Q = 50.0 is the source strength, σ = 0.1 is the standard deviation controlling the spatial extent of the source, and ( x s , y s ) is the source location (set at the domain center). The concentration field evolves from this source term through the coupled advection–diffusion dynamics, naturally generating diverse plume patterns that depend on the underlying velocity field. Moreover, the diffusion coefficient D is randomly sampled from a uniform distribution over the range [ 10 3 , 10 2 ] to control the rate of concentration spreading. The viscosity coefficient ν is randomly sampled from a uniform distribution over the range [ 10 5 , 10 3 ] to govern the momentum diffusion. The dataset is similarly split into five training trajectories and five evaluation trajectories.

4.4.2. Single-Step Prediction Performance

From Table 3, we observe that purely data-driven approaches struggle considerably. DeepONet achieves only a relative L 2 error of 0.25895 with an R 2 of 0.46684, indicating poor predictive capability, while Transolver, despite its attention mechanism, still yields a relatively high error of 0.15784. Even the CNO baseline, which incorporates convolutional neural operators, achieves a relative L 2 error of 0.06477, suggesting that the limited training data are insufficient for the model to fully learn the intricate interactions between flow and diffusion. The same observation as for the SWE is that when only training the model with physics loss (PD-CNO), the accuracy decreases due to the inability to obtain information about the data distribution. PI-CNO, optimized by combining data and physics loss, only brings a small improvement (about 2.8%) compared to the data-driven CNO.
In stark contrast, our proposed PFT-CNO method demonstrates the significant improvement of physics-informed fine-tuning for this coupled system, achieving a relative L 2 error of 0.04767 and R 2 score of 0.99093. This represents a 26.4% improvement over the CNO baseline and a 24.3% improvement over PI-CNO, significantly outperforming the 8.3% gain observed for the SWE. This improvement can be attributed to the physics and data decoupled optimization strategy, which allows the model to effectively leverage physical constraints from both the advection–diffusion equation and the Navier–Stokes equations during fine-tuning without suffering from gradient conflicts. By separating data-driven pretraining from physics-based refinement, PFT-CNO successfully mitigates the data scarcity issue while guiding consistency with the governing equations. Similarly to the SWE experiment, PFT-CNO learned spatial and multi-physics field information better than PFT-DeepONet, resulting in the smaller error. Compared to the computation time of numerical solvers (0.89 s), the speed of neural operators has been significantly improved (at least 120 times).
Figure 4 provides visual comparisons of the predictions from all methods, further illustrating the relatively better accuracy of PFT-CNO in capturing the complex flow field dynamics. More visualizations are shown in Appendix C.

4.4.3. Multi-Step Prediction Performance

We also perform autoregressive multi-step predictions to predict 10 time steps for evaluating the long-term prediction and error accumulation. The average relative L 2 error is shown in Table 2.
Figure 3b demonstrates how prediction errors evolve during the rollout for both methods. By incorporating physical constraints through fine-tuning, PFT-CNO achieves superior long-term accuracy with slower error accumulation. The relative L 2 error of PFT-CNO consistently outperforms that of CNO by a significant margin throughout the prediction sequence. This robust behavior stems from the physics-constrained refinement process, which guides compliance with governing equations and prevents predictions from violating fundamental conservation laws.

4.5. ANOVA and T-Test

Table 4 shows the statistical comparison of different models for the SWE and ADE. Through a paired T-test, ANOVA, and Cohen’s d effect size, the results consistently indicate that PFT-CNO performs well in all comparisons.
For the SWE benchmark, PFT-CNO significantly outperforms the data-driven CNO with a large effect size (Cohen’s d = 5.04, p-Value < 0.001), demonstrating that physics-informed fine-tuning substantially enhances predictive accuracy. Compared to PI-CNO, PFT-CNO shows more effective performance than direct physics-informed training. Against PFT-DeepONet, PFT-CNO shows a large effect size (d = 2.59, p-Value < 0.001), confirming that the convolutional architecture captures spatial features more effectively, which is suitable for high-dimensional flow field modeling.
Similar conclusions emerge for the ADE benchmark. PFT-CNO achieves large effect sizes compared to data-driven CNO (Cohen’s d = 1.58), PI-CNO (Cohen’s d = 0.93), and PFT-DeepONet (Cohen’s d = 2.20), with all comparisons showing p-Value < 0.001. The consistent improvement in these two tests demonstrates the robustness and generalizability of PFT-CNO to spatio-temporal flow field problems.

5. Conclusions

In this work, we propose PFT-CNO with physics-informed fine-tuning for high-dimensional flow field modeling, which addresses the gradient imbalance problem in physics-informed neural operators through a decoupled optimization strategy. By separating the data-driven pretraining phase from the physics-informed fine-tuning phase, our method enables effective integration of physical information without the competing gradient conflicts that plague joint optimization approaches. We validate PFT-CNO on two distinct flow systems: The shallow water equation and the advection–diffusion equation coupled with the Navier–Stokes equation. The experimental results demonstrate improvements over the baseline methods, achieving accuracy gains of 8.3% and 26.4% for the two systems, respectively. More importantly, PFT-CNO exhibits better stability in long-term multi-step predictions, effectively mitigating error accumulation. The greater improvement on the advection–diffusion equation highlights the particular effectiveness of our method for multi-physics systems, where limited training data poses significant challenges for purely data-driven approaches. In addition, PFT-CNO demonstrates significant performance in statistical validation. A key advantage of our method is that it requires no additional training data beyond the original dataset yet improves prediction accuracy by leveraging governing equations and the inputs of evaluation data. This capability makes PFT-CNO valuable in real-world cases where data acquisition is expensive or limited but the underlying physical laws are established.

6. Limitations and Future Perspectives

While our proposed PFT-CNO method demonstrates significant improvements, several limitations warrant discussion and suggest directions for future research. First, the current experimental setup employs training and evaluation datasets generated from the same numerical solver with variations only in initial conditions and physical parameters (e.g., viscosity and diffusion coefficients). Although these variations introduce distinctions, the training and evaluation distributions remain relatively similar, which may not fully demonstrate the potential advantages of physics-informed fine-tuning. We anticipate that when the model encounters more substantial distribution shifts, such as flow fields governed by different but related PDEs or significantly different parameter regimes, the benefits of physics-informed fine-tuning will be more pronounced. Moreover, a particularly promising application lies in bridging the simulation-to-reality gap. Numerical simulations inherently contain discretization errors and modeling approximations that deviate from real-world measurements. In such scenarios, physics-informed fine-tuning can constrain predictions to remain within physically plausible ranges when adapting from simulation-trained models to real-world modeling, potentially offering substantial accuracy improvements for practical deployment.
Second, the current implementation performs physics-informed fine-tuning as a separate phase before inference, rather than integrating it directly into the inference process. This offline fine-tuning strategy, while effective, does not fully exploit the potential for online adaptation. Future work can explore online physics-informed fine-tuning, where the model dynamically updates its parameters upon receiving each new input during inference, using both the input state and governing equations to perform real-time adaptation. This capability would be particularly critical for control applications and feedback decision-making systems, where the model must continuously adapt to evolving system states and operating conditions. Additionally, the current method employs full-parameter fine-tuning, which incurs considerable computational costs and time. Inspired by recent advances in test-time training and low-rank adaptations, future research can investigate strategies that update only a small subset of model parameters during inference. Such approaches can dramatically reduce computational cost while maintaining the benefits of physics-informed adaptation, making the method more practical for large-scale models and real-time applications.
Moreover, the core of our work is to propose a novel methodology. We validated the method on two systems in Section 4 and cylinder wake flow in Appendix D. In the future, this method can be applied to and developed for more complex flow field modeling, such as turbulence and multi-phase flow modeling [49,50], to improve the accuracy of existing neural operator models and reduce computational time compared to classical numerical solvers.

Author Contributions

H.F.: conceptualization (equal); data curation; formal analysis (equal); investigation; methodology; project administration (equal); software; validation; visualization; and writing—original draft. Y.Z.: conceptualization (equal); funding acquisition (equal); formal analysis (equal); project administration (equal); and writing—review and editing (equal). D.F.: conceptualization (equal); funding acquisition (equal); formal analysis (equal); project administration (equal); and writing—review and editing (equal). All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Zhejiang Province Leading Goose Plan Project (Grant No. 2025C02017) and the Research Center for Industries of the Future at Westlake University under Grant No. WU2024C001.

Data Availability Statement

The data that support this work will be made available upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Pseudocode

The pseudocode of the proposed physics-informed fine-tuned neural operator is shown in Algorithm A1.
Algorithm A1 Physics-informed fine-tuned neural operator
Require: Training dataset D = { ( u t ( i ) , u t + 1 ( i ) ) } i = 1 N t r a i n
Require: Evaluation dataset B = { u t ( j ) } j = 1 N e v a l (only inputs)
Require: Neural operator G θ , learning rates η p r e , η f t
Require: Optimizers Opt p r e (for pretraining), Opt f t (for fine-tuning)
Require: Number of pretraining epochs E p r e , fine-tuning epochs E f t
  1:
Phase I: Data-Driven Pretraining
  2:
for  e p o c h = 1 to E p r e  do
  3:
    for each batch ( u t ( i ) , u t + 1 ( i ) ) D  do
  4:
         u ^ t + 1 ( i ) G θ ( u t ( i ) )                         ▹ Forward
  5:
         L d a t a 1 N t r a i n i u ^ t + 1 ( i ) u t + 1 ( i ) 2 u t + 1 ( i ) 2              ▹ Compute data loss
  6:
         θ Opt p r e . step ( θ L d a t a )                 ▹ Update parameters
  7:
    end for
  8:
end for
  9:
Phase II: Physics-Informed Fine-Tuning
10:
θ f t θ                     ▹ Initialize from pretrained model
11:
for  e p o c h = 1 to E f t  do
12:
    for each evaluation input batch u t ( j ) B  do
13:
         u ^ t + 1 ( j ) G θ f t ( u t ( j ) )                         ▹ Forward
14:
         L p h y s i c s u ^ t + 1 ( j ) t F ( u t ( j ) , u t ( j ) , 2 u t ( j ) ) 2         ▹ Compute physics loss
15:
         θ f t Opt f t . step ( θ f t L p h y s i c s )              ▹ Fine-tune parameters
16:
    end for
17:
end for

Appendix B. Model Parameters

The hyperparameters of PFT-CNO both for training/fine-tuning and the model architecture are listed in Table A1.
Table A1. Hyperparameters for PFT-CNO training (fine-tuning) and model architecture.
Table A1. Hyperparameters for PFT-CNO training (fine-tuning) and model architecture.
Training HyperparametersModel Hyperparameters
ParameterValueParameterValue
OptimizerAdam N b l o c k 3
Learning rate0.001 N r e s i d u a l _ l a y e r 3
LR schedulerStep N n e c k _ l a y e r 5
Scheduler step size100Channel multiplier32
Gamma0.5ActivationLeakyReLU
Train batch size16Kernel size3
Eval batch size64Cutoff denominator2.0001
Training updates1000Filter size6
Physics loss weight ( λ )1LReLU upsampling2
Half-width mult.0.8
Batch normTrue
Latent dim128

Appendix C. Visualizations of Results

We present more visualizations of prediction results here.
Figure A1. Visualizations of predictions on the shallow water equation.
Figure A1. Visualizations of predictions on the shallow water equation.
Jmse 14 00201 g0a1
Figure A2. Visualizations of predictions on the shallow water equation.
Figure A2. Visualizations of predictions on the shallow water equation.
Jmse 14 00201 g0a2
Figure A3. Visualizations of predictions on the advection–diffusion equation.
Figure A3. Visualizations of predictions on the advection–diffusion equation.
Jmse 14 00201 g0a3
Figure A4. Visualizations of predictions on the advection–diffusion equation.
Figure A4. Visualizations of predictions on the advection–diffusion equation.
Jmse 14 00201 g0a4

Appendix D. Experiment on Cylinder Wake Flow

In addition to the experiments on the two systems mentioned in the main text, we also conduct flow field modeling on more complex cylinder wake flow data to further demonstrate the advantages of the proposed method. Here, we use the BDIM solver [51] to calculate 10 numerical simulations of the u and v velocity fields and pressure p field for flow around a cylinder, with Reynolds numbers including [1781, 1875, 2062, 2156, 2250, 2343, 2531, 2718, 2812, 2906]. The first five configurations are used as the training dataset, and the last five configurations are used as the testing dataset. The method includes the model and the calculation method of physics loss, which are consistent with those described in Section 3.4.
Table A2. Performance comparison of different methods on cylinder wake flow prediction. Best in bold.
Table A2. Performance comparison of different methods on cylinder wake flow prediction. Best in bold.
MethodRel L 2 Error ↓RMSE ↓MAE ↓ R 2 Eval Time (s) ↓
DeepONet0.303490.054950.029180.620430.00192
CNO0.086640.015860.008170.968370.00178
PFT-CNO0.043480.007850.003870.992250.00171
The experimental results are shown in Table A2. We can see that PFT-CNO reduces the error by 49.8% compared to data-driven CNO. Therefore, the conclusion is consistent with the previous system modelings. Our proposed physics-informed fine-tuning strategy for flow field modeling can improve the accuracy of neural operators.

References

  1. Uchida, T.; Ohya, Y. Numerical simulation of atmospheric flow over complex terrain. J. Wind. Eng. Ind. Aerodyn. 1999, 81, 283–293. [Google Scholar] [CrossRef]
  2. Klein, R. Scale-dependent models for atmospheric flows. Annu. Rev. Fluid Mech. 2010, 42, 249–274. [Google Scholar] [CrossRef]
  3. Liu, X.; Zhang, Y.; Song, D.; Yang, H.; Li, X. A multifractal cascade model for energy evolution and dissipation in ocean turbulence. J. Mar. Sci. Eng. 2023, 11, 1768. [Google Scholar] [CrossRef]
  4. Xu, H.; Wang, M.; Zhou, Z.; Sun, T.; Zhang, G. Unsteady Flow and Loading Characteristics of Rotating Spheres During Underwater Ejection. J. Mar. Sci. Eng. 2025, 13, 2331. [Google Scholar] [CrossRef]
  5. Maximillian, J.; Brusseau, M.L.; Glenn, E.P.; Matthias, A.D. Pollution and environmental perturbations in the global system. In Environmental and Pollution Science, 3rd ed.; Academic Press: Cambridge, MA, USA, 2019; pp. 457–476. [Google Scholar]
  6. Van Gennip, S.J.; Popova, E.E.; Yool, A.; Pecl, G.T.; Hobday, A.J.; Sorte, C.J.B. Going with the flow: The role of ocean circulation in global marine ecosystems under a changing climate. Glob. Change Biol. 2017, 23, 2602–2617. [Google Scholar] [CrossRef]
  7. Bodnar, C.; Bruinsma, W.P.; Lucic, A.; Stanley, M.; Allen, A.; Brandstetter, J.; Garvan, P.; Riechert, M.; Weyn, J.A.; Dong, H.; et al. A foundation model for the Earth system. Nature 2025, 641, 1180–1187. [Google Scholar] [CrossRef]
  8. Liu, P.; Georgiadis, M.C.; Pistikopoulos, E.N. Advances in energy systems engineering. Ind. Eng. Chem. Res. 2011, 50, 4915–4926. [Google Scholar] [CrossRef]
  9. Brunton, S.L.; Noack, B.R.; Koumoutsakos, P. Machine learning for fluid mechanics. Annu. Rev. Fluid Mech. 2020, 52, 477–508. [Google Scholar] [CrossRef]
  10. Cai, Z. On the finite volume element method. Numer. Math. 1990, 58, 713–735. [Google Scholar] [CrossRef]
  11. Teixeira, F.L.; Sarris, C.; Zhang, Y.; Na, D.-Y.; Berenger, J.-P.; Su, Y.; Okoniewski, M.; Chew, W.C.; Backman, V.; Simpson, J.J. Finite-difference time-domain methods. Nat. Rev. Methods Prim. 2023, 3, 75. [Google Scholar] [CrossRef]
  12. Pirozzoli, S. Numerical methods for high-speed flows. Annu. Rev. Fluid Mech. 2011, 43, 163–194. [Google Scholar] [CrossRef]
  13. Cheng, M.; Fang, F.; Pain, C.C.; Navon, I.M. Data-driven modelling of nonlinear spatio-temporal fluid flows using a deep convolutional generative adversarial network. Comput. Methods Appl. Mech. Eng. 2020, 365, 113000. [Google Scholar] [CrossRef]
  14. Vinuesa, R.; Brunton, S.L.; McKeon, B.J. The transformative potential of machine learning for experiments in fluid mechanics. Nat. Rev. Phys. 2023, 5, 536–545. [Google Scholar] [CrossRef]
  15. Feng, H.; Wang, Y.; Fan, D. How to re-enable PDE loss for physical systems modeling under partial observation. Proc. Aaai Conf. Artif. Intell. 2025, 39, 182–190. [Google Scholar] [CrossRef]
  16. Gilpin, W. Generative learning for nonlinear dynamics. Nat. Rev. Phys. 2024, 6, 194–206. [Google Scholar] [CrossRef]
  17. Solomatine, D.P.; Ostfeld, A. Data-driven modelling: Some past experiences and new approaches. J. Hydroinform. 2008, 10, 3–22. [Google Scholar] [CrossRef]
  18. Marcus, E.; Teuwen, J. Artificial intelligence and explanation: How, why, and when to explain black boxes. Eur. J. Radiol. 2024, 173, 111393. [Google Scholar] [CrossRef]
  19. Li, Z.; Zheng, H.; Kovachki, N.; Jin, D.; Chen, H.; Liu, B.; Azizzadenesheli, K.; Anandkumar, A. Physics-informed neural operator for learning partial differential equations. ACM/IMS J. Data Sci. 2024, 1, 1–27. [Google Scholar] [CrossRef]
  20. Goswami, S.; Bora, A.; Yu, Y.; Karniadakis, G.E. Physics-informed deep neural operator networks. In Machine Learning in Modeling and Simulation: Methods and Applications; Springer: Berlin/Heidelberg, Germany, 2023; pp. 219–254. [Google Scholar]
  21. Rosofsky, S.G.; Al Majed, H.; Huerta, E.A. Applications of physics informed neural operators. Mach. Learn. Sci. Technol. 2023, 4, 025022. [Google Scholar] [CrossRef]
  22. Wang, S.; Teng, Y.; Perdikaris, P. Understanding and mitigating gradient flow pathologies in physics-informed neural networks. SIAM J. Sci. Comput. 2021, 43, A3055–A3081. [Google Scholar] [CrossRef]
  23. Wang, S.; Yu, X.; Perdikaris, P. When and why PINNs fail to train: A neural tangent kernel perspective. J. Comput. Phys. 2022, 449, 110768. [Google Scholar] [CrossRef]
  24. Sahoo, M.; Chakraverty, S. Dynamics of tsunami wave propagation in uncertain environment. Comput. Appl. Math. 2024, 43, 266. [Google Scholar] [CrossRef]
  25. Feng, H.; Hu, P.; Wang, Y.; Fan, D.; Wu, T.; Zhang, Y. Physics-informed super-resolution and forecasting method based on inaccurate partial differential equations and partial observation. Phys. Fluids 2025, 37, 066625. [Google Scholar] [CrossRef]
  26. Raonic, B.; Molinaro, R.; De Ryck, T.; Rohner, T.; Bartolucci, F.; Alaifari, R.; Mishra, S.; de Bézenac, E. Convolutional neural operators for robust and accurate learning of PDEs. Adv. Neural Inf. Process. Syst. 2023, 36, 77187–77200. [Google Scholar]
  27. Garcia-Fernandez, R.; Portal-Porras, K.; Irigaray, O.; Ansa, Z.; Fernandez-Gamiz, U. CNN-based flow field prediction for bus aerodynamics analysis. Sci. Rep. 2023, 13, 21213. [Google Scholar] [CrossRef]
  28. Deng, Z.; Chen, Y.; Liu, Y.; Kim, K.C. Time-resolved turbulent velocity field reconstruction using a long short-term memory (LSTM)-based artificial intelligence framework. Phys. Fluids 2019, 31, 075108. [Google Scholar] [CrossRef]
  29. Zhuang, D.L.; Wu, Y.; Wang, Q. A recurrent neural network for particle trajectory prediction in complex flow fields. Phys. Fluids 2025, 37, 105101. [Google Scholar] [CrossRef]
  30. Li, J.; Li, Y.; Liu, T.; Zhang, D.; Xie, Y. Multi-fidelity graph neural network for flow field data fusion of turbomachinery. Energy 2023, 285, 129405. [Google Scholar] [CrossRef]
  31. Wu, T.; Wang, Q.; Zhang, Y.; Ying, R.; Cao, K.; Sosic, R.; Jalali, R.; Hamam, H.; Maucec, M.; Leskovec, J. Learning large-scale subsurface simulations with a hybrid graph network simulator. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, 14–18 August 2022; pp. 4184–4194. [Google Scholar]
  32. Li, Z.; Huang, D.Z.; Liu, B.; Anandkumar, A. Fourier neural operator with learned deformations for pdes on general geometries. J. Mach. Learn. Res. 2023, 24, 1–26. [Google Scholar]
  33. Lu, L.; Jin, P.; Pang, G.; Zhang, Z.; Karniadakis, G.E. Learning nonlinear operators via DeepONet based on the universal approximation theorem of operators. Nat. Mach. Intell. 2021, 3, 218–229. [Google Scholar] [CrossRef]
  34. Wu, H.; Luo, H.; Wang, H.; Wang, J.; Long, M. Transolver: A fast transformer solver for PDEs on general geometries. In Proceedings of the 41st International Conference on Machine Learning, Vienna, Austria, 21–27 July 2024; pp. 53681–53705. [Google Scholar]
  35. Hu, P.; Wang, R.; Zheng, X.; Zhang, T.; Feng, H.; Feng, R.; Wei, L.; Wang, Y.; Ma, Z.-M.; Wu, T. Wavelet Diffusion Neural Operator. In Proceedings of the Thirteenth International Conference on Learning Representations, Singapore, 24–28 April 2025. [Google Scholar]
  36. Raissi, M.; Perdikaris, P.; Karniadakis, G.E. Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. J. Comput. Phys. 2019, 378, 686–707. [Google Scholar] [CrossRef]
  37. Cai, S.; Wang, Z.; Wang, S.; Perdikaris, P.; Karniadakis, G.E. Physics-informed neural networks for heat transfer problems. J. Heat Transf. 2021, 143, 060801. [Google Scholar] [CrossRef]
  38. Rasht-Behesht, M.; Huber, C.; Shukla, K.; Karniadakis, G.E. Physics-informed neural networks (PINNs) for wave propagation and full waveform inversions. J. Geophys. Res. Solid Earth 2022, 127, e2021JB023120. [Google Scholar] [CrossRef]
  39. Wang, S.; Wang, H.; Perdikaris, P. Learning the solution operator of parametric partial differential equations with physics-informed DeepONets. Sci. Adv. 2021, 7, eabi8605. [Google Scholar] [CrossRef] [PubMed]
  40. Ma, X.; Alkhalifah, T. An effective physics-informed neural operator framework for predicting wavefields. arXiv 2025, arXiv:2507.16431. [Google Scholar] [CrossRef]
  41. Yamazaki, Y.; Harandi, A.; Muramatsu, M.; Viardin, A.; Apel, M.; Brepols, T.; Reese, S.; Rezaei, S. A finite element-based physics-informed operator learning framework for spatiotemporal partial differential equations on arbitrary domains. Eng. Comput. 2025, 41, 1–29. [Google Scholar] [CrossRef]
  42. Zhang, Z.; Moya, C.; Lu, L.; Lin, G.; Schaeffer, H. Deeponet as a multi-operator extrapolation model: Distributed pretraining with physics-informed fine-tuning. J. Comput. Phys. 2025, 547, 114537. [Google Scholar] [CrossRef]
  43. Kang, S.; Park, G.; Kum, D. A Transfer Learning Physics-Informed Neural Network Framework for Solving Multi-Coupled Partial Differential Equations with Complex Geometries: Fuel Cell Case Study. SSRN 2025. [Google Scholar] [CrossRef]
  44. Wang, Y.; Bai, J.; Eshaghi, M.S.; Anitescu, C.; Zhuang, X.; Rabczuk, T.; Liu, Y. Transfer Learning in Physics-Informed Neurals Networks: Full Fine-Tuning, Lightweight Fine-Tuning, and Low-Rank Adaptation. Int. J. Mech. Syst. Dyn. 2025, 5, 212. [Google Scholar] [CrossRef]
  45. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  46. Ren, P.; Rao, C.; Liu, Y.; Ma, Z.; Wang, Q.; Wang, J.X.; Sun, H. PhySR: Physics-informed deep super-resolution for spatiotemporal data. J. Comput. Phys. 2023, 492, 112438. [Google Scholar] [CrossRef]
  47. Akbar, Z.; Murtaza, N.; Pasha, G.A.; Iqbal, S.; Ghumman, A.R.; Abbas, F.M. Predicting scour depth in a meandering channel with spur dike: A comparative analysis of machine learning techniques. Phys. Fluids 2025, 37, 045158. [Google Scholar] [CrossRef]
  48. Azizzadenesheli, K.; Kovachki, N.; Li, Z.; Liu-Schiaffini, M.; Kossaifi, J.; Anandkumar, A. Neural operators for accelerating scientific simulations and design. Nat. Rev. Phys. 2024, 6, 320–328. [Google Scholar] [CrossRef]
  49. Wang, Y.; Li, Z.; Yuan, Z.; Peng, W.; Liu, T.; Wang, J. Prediction of turbulent channel flow using Fourier neural operator-based machine-learning strategy. Phys. Rev. Fluids 2024, 9, 084604. [Google Scholar] [CrossRef]
  50. Zhang, K.; Zuo, Y.; Zhao, H.; Ma, X.; Gu, J.; Wang, J.; Yang, Y.; Yao, C.; Yao, J. Fourier neural operator for solving subsurface oil/water two-phase flow partial differential equation. SPE J. 2022, 27, 1815–1830. [Google Scholar] [CrossRef]
  51. Weymouth, G.D.; Yue, D.K. Boundary data immersion method for Cartesian-grid simulations of fluid-body interaction problems. J. Comput. Phys. 2011, 230, 6233–6247. [Google Scholar] [CrossRef]
Figure 1. Illustration of the training pipline of the physics-informed fine-tuned operator. The figure shows the calculation logic of data loss, physics-informed loss, and physics loss. PFT-CNO is first pretrained using data loss and then fine-tuned using physics loss.
Figure 1. Illustration of the training pipline of the physics-informed fine-tuned operator. The figure shows the calculation logic of data loss, physics-informed loss, and physics loss. PFT-CNO is first pretrained using data loss and then fine-tuned using physics loss.
Jmse 14 00201 g001
Figure 2. Visualizations of predictions on the shallow water equation. The left side shows the prediction and error of height h, while the right side shows the prediction and error of velocity field u. The first line shows the ground true values of both, and each line below is the prediction of the method. It can be seen that our method PFT-CNO has the smallest error in the last line.
Figure 2. Visualizations of predictions on the shallow water equation. The left side shows the prediction and error of height h, while the right side shows the prediction and error of velocity field u. The first line shows the ground true values of both, and each line below is the prediction of the method. It can be seen that our method PFT-CNO has the smallest error in the last line.
Jmse 14 00201 g002
Figure 3. Multi-step prediction performance. PFT-CNO maintains consistently lower error accumulation throughout the prediction horizon.
Figure 3. Multi-step prediction performance. PFT-CNO maintains consistently lower error accumulation throughout the prediction horizon.
Jmse 14 00201 g003
Figure 4. Visualizations of predictions on the advection–diffusion equation. The left side shows the prediction and error of concentration field c, while the right side shows the prediction and error of velocity field u. The first line shows the ground true values of both, and each line below is the prediction of the method. It can be seen that our method PFT-CNO has the smallest error in the last line.
Figure 4. Visualizations of predictions on the advection–diffusion equation. The left side shows the prediction and error of concentration field c, while the right side shows the prediction and error of velocity field u. The first line shows the ground true values of both, and each line below is the prediction of the method. It can be seen that our method PFT-CNO has the smallest error in the last line.
Jmse 14 00201 g004
Table 1. Performance comparison of different methods on SWE single-step prediction. Best in bold and second best underlined.
Table 1. Performance comparison of different methods on SWE single-step prediction. Best in bold and second best underlined.
MethodRel L 2 Error ↓RMSE ↓MAE ↓ R 2 Eval Time (s) ↓
DeepONet0.163310.097200.07124−0.054120.00727
PFT-DeepONet0.158460.094750.06924−0.001900.00737
Transolver0.162550.096820.07093−0.044650.00156
CNO0.012840.007900.003890.993040.00326
PD-CNO0.016930.010310.005380.988130.00334
PI-CNO0.012560.007740.003840.993310.00352
PFT-CNO0.011780.007370.003720.993940.00314
Table 2. Average relative error of CNO and PFT-CNO on multi-step prediction. Best in bold.
Table 2. Average relative error of CNO and PFT-CNO on multi-step prediction. Best in bold.
MethodSWE ↓ADE ↓
CNO0.092500.19034
PFT-CNO0.065510.15664
Table 3. Performance comparison of different methods on ADE single-step prediction. Best in bold and second best underlined.
Table 3. Performance comparison of different methods on ADE single-step prediction. Best in bold and second best underlined.
MethodRel L 2 Error ↓RMSE ↓MAE ↓ R 2 Eval Time (s) ↓
DeepONet0.258956.946671.516320.466840.00719
PFT-DeepONet0.197025.359811.167200.682600.00736
Transolver0.157843.344941.274910.876380.00224
CNO0.064771.199960.413100.984090.00365
PD-CNO0.069861.304610.447170.981200.00360
PI-CNO0.062971.165990.396200.984980.00344
PFT-CNO0.047670.905960.315880.990930.00358
Table 4. Comprehensive statistical comparison of different models.
Table 4. Comprehensive statistical comparison of different models.
Model ComparisonMean
Difference
ANOVA
(F-Statistic)
T-Test
(T-Statistic)
Cohen’s d
SWECNO vs. PFT-CNO0.02771234.20106.845.04
CNO vs. PI-CNO0.01013.4214.500.68
PFT-DeepONet vs. PFT-CNO0.06812147.9554.932.59
ADECNO vs. PFT-CNO1.6386266.5133.561.58
CNO vs. PI-CNO1.224081.4019.690.93
PFT-DeepONet vs. PFT-CNO2.83702329.9246.602.20
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Feng, H.; Zhang, Y.; Fan, D. Physics-Informed Fine-Tuned Neural Operator for Flow Field Modeling. J. Mar. Sci. Eng. 2026, 14, 201. https://doi.org/10.3390/jmse14020201

AMA Style

Feng H, Zhang Y, Fan D. Physics-Informed Fine-Tuned Neural Operator for Flow Field Modeling. Journal of Marine Science and Engineering. 2026; 14(2):201. https://doi.org/10.3390/jmse14020201

Chicago/Turabian Style

Feng, Haodong, Yuzhong Zhang, and Dixia Fan. 2026. "Physics-Informed Fine-Tuned Neural Operator for Flow Field Modeling" Journal of Marine Science and Engineering 14, no. 2: 201. https://doi.org/10.3390/jmse14020201

APA Style

Feng, H., Zhang, Y., & Fan, D. (2026). Physics-Informed Fine-Tuned Neural Operator for Flow Field Modeling. Journal of Marine Science and Engineering, 14(2), 201. https://doi.org/10.3390/jmse14020201

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop