A Novel Hybrid Deep Learning Method for Predicting the Flow Fields of Biomimetic Flapping Wings

The physics governing the fluid dynamics of bio-inspired flapping wings is effectively characterized by partial differential equations (PDEs). Nevertheless, the process of discretizing these equations at spatiotemporal scales is notably time consuming and resource intensive. Traditional PDE-based computations are constrained in their applicability, which is mainly due to the presence of numerous shape parameters and intricate flow patterns associated with bionic flapping wings. Consequently, there is a significant demand for a rapid and accurate solution to nonlinear PDEs, to facilitate the analysis of bionic flapping structures. Deep learning, especially physics-informed deep learning (PINN), offers an alternative due to its great nonlinear curve-fitting capability. In the present work, a hybrid coarse-data-driven physics-informed neural network model (HCDD-PINN) is proposed to improve the accuracy and reliability of predicting the time evolution of nonlinear PDEs solutions, by using an order-of-magnitude-coarser grid than traditional computational fluid dynamics (CFDs) require as internal training data. The architecture is devised to enforce the initial and boundary conditions, and incorporate the governing equations and the low-resolution spatiotemporal internal data into the loss function of the neural network, to drive the training. Compared to the original PINN with no internal data, the training and predicting dynamics of HCDD-PINN with different resolutions of coarse internal data are analyzed on the problem relevant to the two-dimensional unsteady flapping wing, which involves unsteady flow features and moving boundaries. Additionally, a hyper-parametrical study is conducted to obtain an optimal model for the problem under consideration, which is then utilized for investigating the effects of the snapshot and fraction of the coarse internal data on the HCDD-PINN’s performances. The results show that the proposed framework has a sufficient stability and accuracy for solving the considered biomimetic flapping-wing problem, and its great potential means that it can be considered as an alternative to accelerate or replace traditional CFD solvers in the future. The interested variables of the flow field at any instant can be rapidly obtained by the trained HCDD-PINN model, which is superior to the traditional CFD method that usually needs to be re-run. For the three-dimensional and optimization problems of flapping wings, the advantages of the proposed method are supposedly even more apparent.


Introduction
Deep learning (DL) has gained great attention over the last decade owing to its enormous breakthroughs in many fields, such as image processing [1], speech recognition, and disease diagnosis [2].Recently, because of its excellent abilities in handling strong nonlinearity and high dimensionality, DL has been widely tested in solving fluid dynamics and for envisaging, accelerating, or replacing computational fluid dynamic (CFD) simulations in the future without compromising accuracy.Ling et al. [3] achieved the first combination of a deep neural network (DNN) and fluid mechanics by constructing the deep-learning RANS turbulence model.Convolution neural networks (CNNs) can accurately learn the characteristics of flow around a cylinder [4,5], and can be combined with Multilayer Perceptron (MLP) to efficiently and accurately learn the incompressible laminar steady flow fields over airfoils [6], while combining them with long short-term memory (LSTM) allows them to learn the spatial-temporal features of turbulence dynamics [7].However, these DL experiments heavily depend on a large amount of training data to ensure learning accuracy, and they may fail to operate when the data become sparse.Hence, it can be considered as a black box (purely data-driven and lacking physical interpretation).
Differently to the purely data-driven deep learning method, the application of a neural network for simulating the fluid physics governed by the partial differential equations (PDEs) was first tested by Dissanayake and Phan-Thien [8], where they incorporated the PDEs and boundary conditions into the residual form of the neural network.Van Milligen et al. [9] applied a similar method in the magnetohydrodynamic plasma equilibrium problem.Raissi et al. [10] reinvigorated the original approach with modern and accessible tools, referring to it as physics-informed neural network (PINN) for both forward and inverse problems.This has sparked considerable interest in improving and expanding the PINN approach to a wide range of physical situations, such as the vortex-induced vibrations (VIV) problem [11], multiscale problems [12], and the supersonic flows problem [13].Automatic differentiation and the back-propagation algorithm are currently the dominant training approaches, choosing their derivatives with respect to the parameters (e.g., weights and biases) of the PINN model, which is suitable for dealing with derivatives at complex boundaries and in arbitrary domains.Rao et al. [14] constrained the loss function formulation of PINN through a mixed-variable scheme of Navier-Stokes equations to simulate the steady and transient laminar flows past a cylinder at low Reynolds numbers, which has been verified to improve the PINN's trainability and the accuracy of the solution.Choi et al. [15] used mini-batch training and a weighted loss function to handle the memory error and divergence problems which occur when using a PINN model for training a chemical-reactor-like multi-reference frame system.Wu et al. [16] proposed a generative adversarial network framework by embedding Navier-Stokes equations into the residual, to efficiently and precisely generate the flow filed data past a cylinder.Cheng and Zhang [17] tested the PINN with residual neural network blocks for Burger's equation and the Navier-Stokes (N-S) equations, which has been proven to exhibit a stronger predictive ability in fluid dynamic problems.
The proposal of PINN is an important breakthrough, since it transforms solving numerical problems to an unconstrained minimization problem.PINN exhibits numerous advantages, such as effectively training with small or no datasets, training in any region to satisfy the governing equation, and solving the inverse problem of finding unknown parameters that converge toward the true values.Ideally, the unique solution should be captured by PINN without any labeled data when the initial and boundary conditions are well imposed, which represents that the corresponding PDEs problem is well defined [18].However, PINNs still encounter great challenges in achieving stable training and producing accurate predictions, especially when the underlying PDEs solutions contain high-frequencies or multi-scale features [19,20].Besides being difficult to converge, the prediction of a PINN may hardly satisfy the ground truth even when the residual loss has been reduced to a relatively low value.The shortcomings and training deficiencies of the existing PINN model are extremely notable in the more unsteady and complex problems, which may result in larger errors for unsteady characteristics as time proceeds [11].Raissi et al. [21] developed hidden fluid mechanics (HFMs) by using several snapshots of extracted concentration fields to quantitatively obtain the flow fields for sev-eral physical and biomedical problems.Kochkov et al. [22] introduced an end-to-end deep learning method to accelerate prediction and improve approximations for two-dimensional turbulent flows by using an order-of-magnitude-coarser grid than is traditionally required.
The physics of the flow past a flapping wing can be well-simulated by solving the corresponding PDEs, which involves unsteady features and large body motions.Extending the PINN approach to other complex fluid dynamic problems, such as the simulation of a flapping wing, is our main interest in the current work.To the best of our knowledge, this problem has not been solved using a PINN before.Inspired by micro air vehicles' (MAVs) potential military and civilian applications [23], research on the aerodynamics of insect flapping flight have drawn significant attention, including flow visualization wind experiments of real insects in tethered and free-flying conditions [24,25] or the dynamically scaled-up models with prescribed kinematics [26,27], and the numerical simulations using the traditional CFD methods [28,29] or the immersed boundary method (IBM) [30,31].It is widely proven that the unsteady aerodynamic mechanisms, including delayed leadingedge vortex (LEV) [32], wake capture [33], rapid pitch-up [33] and clap-and-fling [34], collectively contribute to the high-lift generation and energy harvest of flapping wings.Therefore, to reveal the underlying high-performance mechanism of a flapping wing, accurate and detailed information about aero-forces and capture and flow structures are of significant importance [35].
In the current work, we propose a hybrid coarse-data-driven physics-informed neural network model (HCDD-PINN), aiming to improve the PINN model to solve the CFD problem by utilizing sparse data obtained from the corresponding experimental and numerical studies.The verification case considered in the current study involves large body motions and unsteady features, like the flow past flapping wings.The variables of interest, such as the velocity and pressure fields, are predicted by the trained HCDD-PINN model, which is obtained by minimizing the loss of the neural network with the physics-constrained enforcement and the supplementary coarse internal data.One challenge of using HCDD-PINN is dealing with the presence of the flapping wing, which is solved by adding boundary collocation points with prescribed velocities and filtering the inner collocation points out of the domain.The vicinity of the flapping wing has sufficient collocation points to ensure prediction accuracy.The remainder of this study is structured as follows.Section 2 presents the details of the methodology for the proposed HCDD-PINN model.In Section 3, the training and predicting dynamics of the HCDD-PINN with different coarse internal data sources are analyzed in detail, and we discuss the results of the simulations.Subsequently, the effects of the snapshot and fraction of the coarse internal data are investigated in Section 4, based on the HCDD-PINN model with the optimal hyper-parameters, in order to test the stability of the HCDD-PINN and obtain the minimum amount of coarse data required to achieve sufficient accuracy.Finally, the main conclusions are summarized in Section 5.

Methodology 2.1. Governing Equations and CFD Solver Setup
The unsteady two-dimensional flow past a flapping wing is governed by the following incompressible Navier-Stokes equations: where ∇, u, and p are the gradient operator, velocity vector, and pressure, respectively.An immersed boundary method (IBM), staggered Cartesian grid, Navier-Stokes solver, which has been validated with sufficient accuracy against several problems with large body motions [30], is adopted to obtain a solution for the flapping wing problem.The wing model is a symmetrical NACA0012 airfoil, and the wing's kinematics are prescribed by two motions: (i) translating along y axis (plunge) described by h, and (ii) rotating about the wing center (pitch) described by α, given by the following: where h m and α m represent the plunge amplitude and pitch amplitude, respectively.f is the flapping frequency.A schematic of the wing kinematics can be found in Figure 1.The computational domain and the boundary conditions are also illustrated in the figure, which are described below.
The governing Equations ( 1) and ( 5) are solved by the IBM solver with the fractional step method, in which the presence of the solid flapping wing is represented by a force term fc, as illustrated by the following: All spatial derivatives are discretized using the second-order central difference scheme in a staggered grid, and the Courant-Friedrichs-Lewy (CFL) number is used for determining the timestep.The domain size is 15c × 10c, and the grid near the wing is uniform.Here, c is the chord length of the wing.The center of the flapping wing is placed at (0.5c, 0).The inlet boundary (u = 1, v = 0, and dp/dx = 0) is set 5c from the leading edge of the wing, while the outlet boundary (  ⁄ +    ⁄ = 0, and p = 0) is set 9c from the trailing edge of the wing.Here, vc is space-averaged streamwise velocity at the domain exit.The distances of the top/bottom to the wing is set to 5c, and the boundary condition of top/bottom is du/dy = 0, v = 0, and dp/dy = 0.More details of the IBM solver can be found in [30,36].The governing Equations ( 1) and ( 5) are solved by the IBM solver with the fractional step method, in which the presence of the solid flapping wing is represented by a force term fc, as illustrated by the following: All spatial derivatives are discretized using the second-order central difference scheme in a staggered grid, and the Courant-Friedrichs-Lewy (CFL) number is used for determining the timestep.The domain size is 15c × 10c, and the grid near the wing is uniform.Here, c is the chord length of the wing.The center of the flapping wing is placed at (0.5c, 0).The inlet boundary (u = 1, v = 0, and dp/dx = 0) is set 5c from the leading edge of the wing, while the outlet boundary (∂u i /∂t + v c ∂u i /∂x = 0, and p = 0) is set 9c from the trailing edge of the wing.Here, v c is space-averaged streamwise velocity at the domain exit.The distances of the top/bottom to the wing is set to 5c, and the boundary condition of top/bottom is du/dy = 0, v = 0, and dp/dy = 0.More details of the IBM solver can be found in [30,36].

Hybrid Coarse Data-Driven with Physics-Informed Neural Network (HCDD-PINN)
The structure of the neural network consists of N l fully connected layers, which is shown in Figure 2. The input layer X n = (x, y, t) includes a set of independent space and time variables, while the output layer Y n = (u, v, p, σ) comprises velocity and pressure variables, and shear stress vector.The physical parameters of interest, for flow field comparisons later, consist of u, v, and p.All of them are evaluated because they are constrained through Navier-Stokes equations and boundary conditions simultaneously, which is different from the traditional CFD method.The hidden layers are used to solve the complex nonlinearity relations between the inputs and outputs, which construct a map 5 of 25

-Informed Neural Network (HCDD-PINN)
consists of Nl fully connected layers, which is , , ) includes a set of independent space and = (, , , ) comprises velocity and pressure sical parameters of interest, for flow field comhem are evaluated because they are constrained ndary conditions simultaneously, which is difhe hidden layers are used to solve the complex and outputs, which construct a map  to nonts  as: ts the nonlinear compositional function parahich can be expanded as below: to nonlinearly relate the inputs X n to the outputs Y n as:

Hybrid Coarse Data-Driven with Physics-Informed Neural Network (HCDD-PINN)
The structure of the neural network consists of Nl fully connected layers, which is n in Figure 2. The input layer  = (, , ) includes a set of independent space and variables, while the output layer  = (, , , ) comprises velocity and pressure ables, and shear stress vector.The physical parameters of interest, for flow field comsons later, consist of u, v, and p.All of them are evaluated because they are constrained ugh Navier-Stokes equations and boundary conditions simultaneously, which is difnt from the traditional CFD method.The hidden layers are used to solve the complex linearity relations between the inputs and outputs, which construct a map  to nonarly relate the inputs  to the outputs  as:

Hybrid Coarse Data-Driven with Physics-Informed Neural Network (H
The structure of the neural network consists of Nl fully connec shown in Figure 2. The input layer  = (, , ) includes a set of ind time variables, while the output layer  = (, , , ) comprises ve variables, and shear stress vector.The physical parameters of interest parisons later, consist of u, v, and p.All of them are evaluated because through Navier-Stokes equations and boundary conditions simultan ferent from the traditional CFD method.The hidden layers are used t nonlinearity relations between the inputs and outputs, which constru linearly relate the inputs  to the outputs  as: where the symbol ∘ is the composition operation.The output vector 2 … Nl) is calculated by the feed-forward algorithm, given by boundary conditions losses and the coarse internal data loss, respectively, in order to balance the weights of each part of the total loss. , * is the boundary condition function, in which * denotes DNN's predicted values.N(.) denotes the number of collocation points (subscripts g for governing equation, I for initial condition, bc for boundary condition, and int for coarse internal training data). (.) (.) and  (.) are the input and output variables for all the collocation points of the physics-informed neural network, respectively.The structure of the neural network consists of Nl fully connected layers, which is shown in Figure 2. The input layer  = (, , ) includes a set of independent space and time variables, while the output layer  = (, , , ) comprises velocity and pressure variables, and shear stress vector.The physical parameters of interest, for flow field comparisons later, consist of u, v, and p.All of them are evaluated because they are constrained through Navier-Stokes equations and boundary conditions simultaneously, which is different from the traditional CFD method.The hidden layers are used to solve the complex nonlinearity relations between the inputs and outputs, which construct a map  to nonlinearly relate the inputs  to the outputs  as: Here, (⋅, ) ∈ ℝ ⟶ ℝ represents the nonlinear compositional function parametrized by (, ) (weight W; bias b), which can be expanded as below: where the symbol ∘ is the composition operation.The output vector at the ith layer (i = 1, 2 … Nl) is calculated by the feed-forward algorithm, given by  (;  ) =  (  , ] ) =  (∑   +  ) where  (⋅;  ) ∈ ℝ ⟶ ℝ and  (⋅) ∈ ℝ ⟶ ℝ.Here, li is the size of the output at the ith layer and  ∈ ℝ * comprises the weights and biases corresponding to the layer i.The  ≡ ( ,⋅⋅⋅,  ,⋅⋅⋅  ) vectors are evaluated by training the neural network.The initial weights and biases are determined by the Xavier method to accelerate the convergence of the neural network. (⋅) denotes the activation function, and the tanh function is adopted here due to its infinitely differentiable capability.
The physics-informed neural network is constructed to calculate the fluid flow field through automatic differentiation (AD), which can be directly utilized in deep-learningframework Tensorflow.To reduce the order of the derivative in Equation ( 2), the Cauchy stress tensor  is introduced into Equation (2) as follows: which is added to the output vector of DNN.It is proved that such continuum-mechanicsbased formulation benefits improve the trainability of DNN [14].As shown in Figure 2, AD is adopted to calculate the identity operator I and the differential operators ∂ , ∂ ,  ∂ , to obtain the partial derivatives in Equations ( 1), (9), and (10) for embedding the physical regularities into the loss function [37].
The loss function  consists of two terms: (i) the physics-constrained term, which includes the governing partial differential equations' loss  (Equations (1), (9), and (10)) and the boundary condition loss  ; and (ii) the data correction term, representing the mean squared errors between the exact and predicted values, consists of the initial condition loss  and the coarse internal loss  , which are derived as follows: The structure of the neural network consists of Nl fully connected layers, which is shown in Figure 2. The input layer  = (, , ) includes a set of independent space and time variables, while the output layer  = (, , , ) comprises velocity and pressure variables, and shear stress vector.The physical parameters of interest, for flow field comparisons later, consist of u, v, and p.All of them are evaluated because they are constrained through Navier-Stokes equations and boundary conditions simultaneously, which is different from the traditional CFD method.The hidden layers are used to solve the complex nonlinearity relations between the inputs and outputs, which construct a map  to nonlinearly relate the inputs  to the outputs  as: Here, (⋅, ) ∈ ℝ ⟶ ℝ represents the nonlinear compositional function parametrized by (, ) (weight W; bias b), which can be expanded as below: where the symbol ∘ is the composition operation.The output vector at the ith layer (i = 1, 2 … Nl) is calculated by the feed-forward algorithm, given by  (;  ) =  (  , ] ) =  (∑   +  ) where  (⋅;  ) ∈ ℝ ⟶ ℝ and  (⋅) ∈ ℝ ⟶ ℝ.Here, li is the size of the output at the ith layer and  ∈ ℝ * comprises the weights and biases corresponding to the layer i.The  ≡ ( ,⋅⋅⋅,  ,⋅⋅⋅  ) vectors are evaluated by training the neural network.The initial weights and biases are determined by the Xavier method to accelerate the convergence of the neural network. (⋅) denotes the activation function, and the tanh function is adopted here due to its infinitely differentiable capability.
The physics-informed neural network is constructed to calculate the fluid flow field through automatic differentiation (AD), which can be directly utilized in deep-learningframework Tensorflow.To reduce the order of the derivative in Equation ( 2), the Cauchy stress tensor  is introduced into Equation (2) as follows: which is added to the output vector of DNN.It is proved that such continuum-mechanicsbased formulation benefits improve the trainability of DNN [14].As shown in Figure 2, AD is adopted to calculate the identity operator I and the differential operators ∂ , ∂ ,  ∂ , to obtain the partial derivatives in Equations ( 1), (9), and (10) for embedding the physical regularities into the loss function [37].
The loss function  consists of two terms: (i) the physics-constrained term, which includes the governing partial differential equations' loss  (Equations ( 1), (9), and ( 10)) and the boundary condition loss  ; and (ii) the data correction term, representing the mean squared errors between the exact and predicted values, consists of the initial condi- The structure of the neural network consists of Nl fully connected layers, which is shown in Figure 2. The input layer  = (, , ) includes a set of independent space and time variables, while the output layer  = (, , , ) comprises velocity and pressure variables, and shear stress vector.The physical parameters of interest, for flow field comparisons later, consist of u, v, and p.All of them are evaluated because they are constrained through Navier-Stokes equations and boundary conditions simultaneously, which is different from the traditional CFD method.The hidden layers are used to solve the complex nonlinearity relations between the inputs and outputs, which construct a map  to nonlinearly relate the inputs  to the outputs  as: Here, (⋅, ) ∈ ℝ ⟶ ℝ represents the nonlinear compositional function parametrized by (, ) (weight W; bias b), which can be expanded as below: where the symbol ∘ is the composition operation.The output vector at the ith layer (i = 1, 2 … Nl) is calculated by the feed-forward algorithm, given by  (;  ) =  (  , ] ) =  (∑   +  ) where  (⋅;  ) ∈ ℝ ⟶ ℝ and  (⋅) ∈ ℝ ⟶ ℝ.Here, li is the size of the output at the ith layer and  ∈ ℝ * comprises the weights and biases corresponding to the layer i.The  ≡ ( ,⋅⋅⋅,  ,⋅⋅⋅  ) vectors are evaluated by training the neural network.The initial weights and biases are determined by the Xavier method to accelerate the convergence of the neural network. (⋅) denotes the activation function, and the tanh function is adopted here due to its infinitely differentiable capability.
The physics-informed neural network is constructed to calculate the fluid flow field through automatic differentiation (AD), which can be directly utilized in deep-learningframework Tensorflow.To reduce the order of the derivative in Equation ( 2), the Cauchy stress tensor  is introduced into Equation (2) as follows: which is added to the output vector of DNN.It is proved that such continuum-mechanicsbased formulation benefits improve the trainability of DNN [14].As shown in Figure 2, AD is adopted to calculate the identity operator I and the differential operators ∂ , ∂ ,  ∂ , to obtain the partial derivatives in Equations ( 1), (9), and (10) for embedding the physical regularities into the loss function [37].
The loss function  consists of two terms: (i) the physics-constrained term, which includes the governing partial differential equations' loss  (Equations ( 1), (9), and ( 10)) and the boundary condition loss  ; and (ii) the data correction term, representing the mean squared errors between the exact and predicted values, consists of the initial condi- The i itial weights and biases are determined by the Xavier method to accelerate the conve gence of the neural network. (⋅) denotes the activation function, and the tanh functio is adopted here due to its infinitely differentiable capability.The physics-informed neural network is constructed to calculate the fluid flow fie through automatic differentiation (AD), which can be directly utilized in deep-learnin framework Tensorflow.To reduce the order of the derivative in Equation ( 2), the Cauch stress tensor  is introduced into Equation (2) as follows: which is added to the output vector of DNN.It is proved that such continuum-mechanic based formulation benefits improve the trainability of DNN [14].As shown in Figure AD is adopted to calculate the identity operator I and the differential oper tors ∂ , ∂ ,  ∂ , to obtain the partial derivatives in Equations ( 1), (9), and (10) for em bedding the physical regularities into the loss function [37].
The loss function  consists of two terms: (i) the physics-constrained term, whic includes the governing partial differential equations' loss  (Equations ( 1), (9), and (10 and the boundary condition loss  ; and (ii) the data correction term, representing th mean squared errors between the exact and predicted values, consists of the initial cond comprises the weights and biases c i.The  ≡ ( ,⋅⋅⋅,  ,⋅⋅⋅  ) vectors are evaluated by training th itial weights and biases are determined by the Xavier method gence of the neural network. (⋅) denotes the activation funct is adopted here due to its infinitely differentiable capability.
The physics-informed neural network is constructed to ca through automatic differentiation (AD), which can be directly framework Tensorflow.To reduce the order of the derivative in stress tensor  is introduced into Equation (2) as follows: which is added to the output vector of DNN.It is proved that su based formulation benefits improve the trainability of DNN [1 AD is adopted to calculate the identity operator I and tors ∂ , ∂ ,  ∂ , to obtain the partial derivatives in Equation bedding the physical regularities into the loss function [37].
The loss function  consists of two terms: (i) the physic includes the governing partial differential equations' loss  (E and the boundary condition loss  ; and (ii) the data correcti mean squared errors between the exact and predicted values, co where the symbol • is the composition operation.The output vector at the ith layer (i = 1, 2 . . .N l ) is calculated by the feed-forward algorithm, given by Biomimetics 2024, 9, x FOR PEER REVIEW 5 of 25

Hybrid Coarse Data-Driven with Physics-Informed Neural Network (HCDD-PINN)
The structure of the neural network consists of Nl fully connected layers, which is shown in Figure 2. The input layer  = (, , ) includes a set of independent space and time variables, while the output layer  = (, , , ) comprises velocity and pressure variables, and shear stress vector.The physical parameters of interest, for flow field comparisons later, consist of u, v, and p.All of them are evaluated because they are constrained through Navier-Stokes equations and boundary conditions simultaneously, which is different from the traditional CFD method.The hidden layers are used to solve the complex nonlinearity relations between the inputs and outputs, which construct a map  to nonlinearly relate the inputs  to the outputs  as: Here, (⋅, ) ∈ ℝ ⟶ ℝ represents the nonlinear compositional function parametrized by (, ) (weight W; bias b), which can be expanded as below: where the symbol ∘ is the composition operation.The output vector at the ith layer (i = 1, 2 … Nl) is calculated by the feed-forward algorithm, given by where  (⋅;  ) ∈ ℝ ⟶ ℝ and  (⋅) ∈ ℝ ⟶ ℝ.Here, li is the size of the output at the ith layer and  ∈ ℝ * comprises the weights and biases corresponding to the layer i.The  ≡ ( ,⋅⋅⋅,  ,⋅⋅⋅  ) vectors are evaluated by training the neural network.The initial weights and biases are determined by the Xavier method to accelerate the convergence of the neural network. (⋅) denotes the activation function, and the tanh function is adopted here due to its infinitely differentiable capability.
The physics-informed neural network is constructed to calculate the fluid flow field through automatic differentiation (AD), which can be directly utilized in deep-learningframework Tensorflow.To reduce the order of the derivative in Equation ( 2), the Cauchy stress tensor  is introduced into Equation (2) as follows: which is added to the output vector of DNN.It is proved that such continuum-mechanicsbased formulation benefits improve the trainability of DNN [14].As shown in Figure 2, AD is adopted to calculate the identity operator I and the differential operators ∂ , ∂ ,  ∂ , to obtain the partial derivatives in Equations ( 1), ( 9), and ( 10) for em- Biomimetics 2024, 9, x FOR PEER REVIEW 5 of 25

Hybrid Coarse Data-Driven with Physics-Informed Neural Network (HCDD-PINN)
The structure of the neural network consists of Nl fully connected layers, which is shown in Figure 2. The input layer  = (, , ) includes a set of independent space and time variables, while the output layer  = (, , , ) comprises velocity and pressure variables, and shear stress vector.The physical parameters of interest, for flow field comparisons later, consist of u, v, and p.All of them are evaluated because they are constrained through Navier-Stokes equations and boundary conditions simultaneously, which is different from the traditional CFD method.The hidden layers are used to solve the complex nonlinearity relations between the inputs and outputs, which construct a map  to nonlinearly relate the inputs  to the outputs  as: Here, (⋅, ) ∈ ℝ ⟶ ℝ represents the nonlinear compositional function parametrized by (, ) (weight W; bias b), which can be expanded as below: where the symbol ∘ is the composition operation.The output vector at the ith layer (i = 1, 2 … Nl) is calculated by the feed-forward algorithm, given by  (;  ) =  (  , ] ) =  (∑   +  ) where  (⋅;  ) ∈ ℝ ⟶ ℝ and  (⋅) ∈ ℝ ⟶ ℝ.Here, li is the size of the output at the ith layer and  ∈ ℝ * comprises the weights and biases corresponding to the layer i.The  ≡ ( ,⋅⋅⋅,  ,⋅⋅⋅  ) vectors are evaluated by training the neural network.The initial weights and biases are determined by the Xavier method to accelerate the convergence of the neural network. (⋅) denotes the activation function, and the tanh function is adopted here due to its infinitely differentiable capability.
The physics-informed neural network is constructed to calculate the fluid flow field through automatic differentiation (AD), which can be directly utilized in deep-learningframework Tensorflow.To reduce the order of the derivative in Equation ( 2), the Cauchy stress tensor  is introduced into Equation (2) as follows: which is added to the output vector of DNN.It is proved that such continuum-mechanicsbased formulation benefits improve the trainability of DNN [14].As shown in Figure 2, AD is adopted to calculate the identity operator I and the differential operators ∂ , ∂ ,  ∂ , to obtain the partial derivatives in Equations ( 1), (9), and (10) for embedding the physical regularities into the loss function [37].
The loss function  consists of two terms: (i) the physics-constrained term, which includes the governing partial differential equations' loss  (Equations ( 1), (9), and ( 10)) and the boundary condition loss  ; and (ii) the data correction term, representing the mean squared errors between the exact and predicted values, consists of the initial condi- where  (⋅;  ) ∈ ℝ ⟶ ℝ and  (⋅) ∈ ℝ ⟶ ℝ.Here, li is the size of the outpu ith layer and  ∈ ℝ * comprises the weights and biases corresponding to th i.The  ≡ ( ,⋅⋅⋅,  ,⋅⋅⋅  ) vectors are evaluated by training the neural network.itial weights and biases are determined by the Xavier method to accelerate the c gence of the neural network. (⋅) denotes the activation function, and the tanh fu is adopted here due to its infinitely differentiable capability.
The physics-informed neural network is constructed to calculate the fluid flo through automatic differentiation (AD), which can be directly utilized in deep-lea framework Tensorflow.To reduce the order of the derivative in Equation ( 2), the C stress tensor  is introduced into Equation (2) as follows: which is added to the output vector of DNN.It is proved that such continuum-mec based formulation benefits improve the trainability of DNN [14].As shown in Fi AD is adopted to calculate the identity operator I and the differential tors ∂ , ∂ ,  ∂ , to obtain the partial derivatives in Equations ( 1), (9), and (10) f bedding the physical regularities into the loss function [37].
The loss function  consists of two terms: (i) the physics-constrained term, includes the governing partial differential equations' loss  (Equations ( 1), ( 9), an and the boundary condition loss  ; and (ii) the data correction term, represent mean squared errors between the exact and predicted values, consists of the initial where 9, x FOR PEER REVIEW 5 of 25

Hybrid Coarse Data-Driven with Physics-Informed Neural Network (HCDD-PINN)
The structure of the neural network consists of Nl fully connected layers, which is shown in Figure 2. The input layer  = (, , ) includes a set of independent space and time variables, while the output layer  = (, , , ) comprises velocity and pressure variables, and shear stress vector.The physical parameters of interest, for flow field comparisons later, consist of u, v, and p.All of them are evaluated because they are constrained through Navier-Stokes equations and boundary conditions simultaneously, which is different from the traditional CFD method.The hidden layers are used to solve the complex nonlinearity relations between the inputs and outputs, which construct a map  to nonlinearly relate the inputs  to the outputs  as: Here, (⋅, ) ∈ ℝ ⟶ ℝ represents the nonlinear compositional function parametrized by (, ) (weight W; bias b), which can be expanded as below: (; ) =  (⋅;  ) ∘  (⋅;  ) ∘⋅⋅⋅∘  (;  ) where the symbol ∘ is the composition operation.The output vector at the ith layer (i = 1, 2 … Nl) is calculated by the feed-forward algorithm, given by where  (⋅;  ) ∈ ℝ ⟶ ℝ and  (⋅) ∈ ℝ ⟶ ℝ.Here, li is the size of the output at the ith layer and  ∈ ℝ * comprises the weights and biases corresponding to the layer i.The  ≡ ( ,⋅⋅⋅,  ,⋅⋅⋅  ) vectors are evaluated by training the neural network.The initial weights and biases are determined by the Xavier method to accelerate the convergence of the neural network. (⋅) denotes the activation function, and the tanh function is adopted here due to its infinitely differentiable capability.
The physics-informed neural network is constructed to calculate the fluid flow field through automatic differentiation (AD), which can be directly utilized in deep-learningframework Tensorflow.To reduce the order of the derivative in Equation ( 2), the Cauchy stress tensor  is introduced into Equation (2) as follows: which is added to the output vector of DNN.It is proved that such continuum-mechanicsbased formulation benefits improve the trainability of DNN [14].As shown in Figure 2, Biomimetics 2024, 9, x FOR PEER REVIEW 5 of 25

Hybrid Coarse Data-Driven with Physics-Informed Neural Network (HCDD-PINN)
The structure of the neural network consists of Nl fully connected layers, which is shown in Figure 2. The input layer  = (, , ) includes a set of independent space and time variables, while the output layer  = (, , , ) comprises velocity and pressure variables, and shear stress vector.The physical parameters of interest, for flow field comparisons later, consist of u, v, and p.All of them are evaluated because they are constrained through Navier-Stokes equations and boundary conditions simultaneously, which is different from the traditional CFD method.The hidden layers are used to solve the complex nonlinearity relations between the inputs and outputs, which construct a map  to nonlinearly relate the inputs  to the outputs  as: where the symbol ∘ is the composition operation.The output vector at the ith layer (i = 1, 2 … Nl) is calculated by the feed-forward algorithm, given by  (;  ) =  (  , ] ) =  (∑   +  ) where  (⋅;  ) ∈ ℝ ⟶ ℝ and  (⋅) ∈ ℝ ⟶ ℝ.Here, li is the size of the output at the ith layer and  ∈ ℝ * comprises the weights and biases corresponding to the layer i.The  ≡ ( ,⋅⋅⋅,  ,⋅⋅⋅  ) vectors are evaluated by training the neural network.The initial weights and biases are determined by the Xavier method to accelerate the convergence of the neural network. (⋅) denotes the activation function, and the tanh function is adopted here due to its infinitely differentiable capability.
The physics-informed neural network is constructed to calculate the fluid flow field through automatic differentiation (AD), which can be directly utilized in deep-learningframework Tensorflow.To reduce the order of the derivative in Equation ( 2), the Cauchy stress tensor  is introduced into Equation (2) as follows: which is added to the output vector of DNN.It is proved that such continuum-mechanicsbased formulation benefits improve the trainability of DNN [14].As shown in Figure 2, AD is adopted to calculate the identity operator I and the differential operators ∂ , ∂ ,  ∂ , to obtain the partial derivatives in Equations ( 1), ( 9), and (10) for embedding the physical regularities into the loss function [37].
The loss function  consists of two terms: (i) the physics-constrained term, which includes the governing partial differential equations' loss  (Equations ( 1), (9), and ( 10)) and the boundary condition loss  ; and (ii) the data correction term, representing the Here, l i is the size of the output at the ith layer and Θ i ∈ R l i−1 * l i +1 comprises the weights and biases corresponding to the layer i.The θ ≡ (Θ 1 , . . ., Θ i , . . ., Θ N l ) vectors are evaluated by training the neural network.The initial weights and biases are determined by the Xavier method to accelerate the convergence of the neural network.

Hybrid Coarse Data-Driven with Physics-Informed Neural Network (HCDD-PINN)
The structure of the neural network consists of Nl fully connected layers, which is shown in Figure 2. The input layer  = (, , ) includes a set of independent space and time variables, while the output layer  = (, , , ) comprises velocity and pressure variables, and shear stress vector.The physical parameters of interest, for flow field comparisons later, consist of u, v, and p.All of them are evaluated because they are constrained through Navier-Stokes equations and boundary conditions simultaneously, which is different from the traditional CFD method.The hidden layers are used to solve the complex nonlinearity relations between the inputs and outputs, which construct a map  to nonlinearly relate the inputs  to the outputs  as: Here, (⋅, ) ∈ ℝ ⟶ ℝ represents the nonlinear compositional function parametrized by (, ) (weight W; bias b), which can be expanded as below: (; ) =  (⋅;  ) ∘  (⋅;  ) ∘⋅⋅⋅∘  (;  ) where the symbol ∘ is the composition operation.The output vector at the ith layer (i = 1, 2 … Nl) is calculated by the feed-forward algorithm, given by  (;  ) =  (  , ] ) =  (∑   +  ) where  (⋅;  ) ∈ ℝ ⟶ ℝ and  (⋅) ∈ ℝ ⟶ ℝ.Here, li is the size of the output at the ith layer and  ∈ ℝ * comprises the weights and biases corresponding to the layer i.The  ≡ ( ,⋅⋅⋅,  ,⋅⋅⋅  ) vectors are evaluated by training the neural network.The initial weights and biases are determined by the Xavier method to accelerate the convergence of the neural network. (⋅) denotes the activation function, and the tanh function is adopted here due to its infinitely differentiable capability.
The physics-informed neural network is constructed to calculate the fluid flow field through automatic differentiation (AD), which can be directly utilized in deep-learningframework Tensorflow.To reduce the order of the derivative in Equation (2), the Cauchy stress tensor  is introduced into Equation (2) as follows: which is added to the output vector of DNN.It is proved that such continuum-mechanicsbased formulation benefits improve the trainability of DNN [14].As shown in Figure 2, AD is adopted to calculate the identity operator I and the differential operators ∂ , ∂ ,  ∂ , to obtain the partial derivatives in Equations ( 1), ( 9), and (10) for em-i (•) denotes the activation function, and the tanh function is adopted here due to its infinitely differentiable capability.
The physics-informed neural network is constructed to calculate the fluid flow field through automatic differentiation (AD), which can be directly utilized in deep-learningframework Tensorflow.To reduce the order of the derivative in Equation ( 2), the Cauchy stress tensor σ is introduced into Equation (2) as follows: which is added to the output vector of DNN.It is proved that such continuum-mechanicsbased formulation benefits improve the trainability of DNN [14].As shown in Figure 2, AD is adopted to calculate the identity operator I and the differential operators ∂ t , ∂ x , and ∂ y , to obtain the partial derivatives in Equations ( 1), (9), and (10) for embedding the physical regularities into the loss function [37].
The loss function ε loss consists of two terms: (i) the physics-constrained term, which includes the governing partial differential equations' loss ε g (Equations ( 1), (9), and ( 10)) and the boundary condition loss ε bc ; and (ii) the data correction term, representing the mean squared errors between the exact and predicted values, consists of the initial condition loss ε I and the coarse internal loss ε int , which are derived as follows: Here, λ bc/I and λ int are adopted as the weighting coefficients for the initial and boundary conditions losses and the coarse internal data loss, respectively, in order to balance the weights of each part of the total loss.y * bc,j is the boundary condition function, in which * denotes DNN's predicted values.N (.) denotes the number of collocation points (subscripts g for governing equation, I for initial condition, bc for boundary condition, and int for coarse internal training data).x i (.) i=1 and y (.) are the input and output variables for all the collocation points of the physics-informed neural network, respectively.
The total loss, as derived in Equation (11), is minimized by the ADAM and L-BFGS-B optimizers [38], which can constantly adjust the learning rates and possess good convergence speed.Theoretically, the solution for the N-S equations can be achieved, when the loss function is reduced to zero.
For the flapping wing problem considered in the current work, the domain and boundary setups utilized in the PINN are the same as in IBM simulation, as depicted in Figure 1.The presence of the flapping wing is achieved by some specific collocation points: (i) points at the moving boundary are added with prescribed velocities (v x i , v y i ), as described in Equations ( 16) and (17); and (ii) points located within the flapping wing are all filtered out of the fluid computational space.That is, where r i represents the length between the rotation center and the ith point at the moving boundary.(n x i , n y i ) is the outward normal on the flapping wing the ith point.The entire spatial-temporal space is filled with all the collocation points, and each point has certain spatial-temporal coordinates.These collocation points are generated and inputted into the DNN for training and to match the corresponding output variables.Similar to traditional CFD simulation, the training process requires a certain minimum number of collocation points to ensure the accuracy of PINN.The total collocation points N collo used for PINN can be calculated with Equation (18), which are shown in Figure 3 as a representation.That is, where N LHS and N Lins represent global and local collocation points, respectively.N LHS are obtained by taking a random sample using the Latin hypercube sampling (LHS) method.They are distributed in the entire spatial-temporal space, and are refined in the region of 7c × 5c (adjacent to the flapping wing) to better capture the details of the flow.N Lins are generated by the linspace method [14] within a cycle shape embracing the flapping wing.They only distributed in the local spatial space at a certain instant, to precisely illustrate the shape and the movements of the flapping wing.For the whole stroke cycle, they are sampled with uniform interval time, i.e., N Lins is generated for all the timestep.It is claimed that such a method for collocation-points generation can effectively improve the accuracy of PINN prediction for the current problem with a large body motion and unsteady features, which will be discussed later.
where  and  represent global and local collocation points, respectively. are obtained by taking a random sample using the Latin hypercube sampling (LHS) method.They are distributed in the entire spatial-temporal space, and are refined in the region of 7c × 5c (adjacent to the flapping wing) to better capture the details of the flow. are generated by the linspace method [14] within a cycle shape embracing the flapping wing.They only distributed in the local spatial space at a certain instant, to precisely illustrate the shape and the movements of the flapping wing.For the whole stroke cycle, they are sampled with uniform interval time, i.e.,  is generated for all the timestep.It is claimed that such a method for collocation-points generation can effectively improve the accuracy of PINN prediction for the current problem with a large body motion and unsteady features, which will be discussed later.

Problem Description
In the current study, the flapping kinematics, hm, αm, and f in Equations ( 3) and ( 4), are fixed at 0.2c, π/4, and 0.25, respectively.c is the chord length of the wing, and the Reynolds number Re is based on the chord of the flapping wing; a free-stream velocity

Problem Description
In the current study, the flapping kinematics, h m , α m , and f in Equations ( 3) and ( 4), are fixed at 0.2c, π/4, and 0.25, respectively.c is the chord length of the wing, and the Reynolds number Re is based on the chord of the flapping wing; a free-stream velocity and a fluid viscosity of 100 are adopted.
A grid convergence study is performed by decreasing the minimum grid interval ∆s by a factor of 2, to ensure the CFD solution accuracy and prepare the PINN training data to validate the corresponding prediction performance of the proposed HCDD-PINN model.The corresponding lift and drag forces' histories under different resolutions are compared in Figure 5.It can be seen that all the cases attain periodical state within three flapping cycles.Therefore, the CFD results from the fourth cycle are selected to be the data training and for comparison.For our PINN simulation, the fourth flapping cycle is trained and predicted.The reference field information at the beginning of the fourth cycle is used as the initial condition data (t = 0) to help with the PINN training, which introduces the initial condition loss ε I to the total loss.The high-resolution simulation (950 × 1024) is used as the ground truth and as a reference, and the three coarser-resolution simulations, i.e., 149 × 128, 269 × 256, and 502 × 512, are adopted as the coarse internal data, to improve the PINN's capability and prediction accuracy, which introduces the coarse internal loss ε int to the total loss.The coarse internal data are scattered randomly at various spatial-temporal locations, which can be easily utilized in the DNN framework.

Prediction Results
In this section, the effects of HCDD-PINN with different-resolution internal data are investigated, by comparing it with a PINN with no internal data.The three CFD results ((149 × 128), (269 × 256), and (502 × 512)) that are used as the coarse internal data are 51×, 14×, and 4× coarser resolution than the reference case (950 × 1024), and indicate ~259-fold, ~70-fold, and ~9-fold computational speedup, respectively.Three scalar state variables, i.e., velocity u, v, and pressure p, are saved in 25 field snapshots, and the number of snapshots used as internal data for PINN training is 25 for the three coarser cases.The fraction per snapshot is 0.0364, 0.01, and 0.00268, respectively, to ensure the same internal data are used for PINN.A case without internal data is also performed for comparison.To demonstrate the robustness of HCDD-PINN, the architecture and hyper-parameters are selected in a consistent fashion for all four of the cases.The architecture of the DNN used here is composed of eight hidden layers with 50 neurons per layer, to broadly balance the tradeoff between the expressivity and trainability of the neural network.The weighting coefficients  ⁄ and  are set as one for both the initial and boundary conditions loss and the coarse internal loss.The total collocation point N collo is set to 3.

Prediction Results
In this section, the effects of HCDD-PINN with different-resolution internal data are investigated, by comparing it with a PINN with no internal data.The three CFD results ((149 × 128), (269 × 256), and (502 × 512)) that are used as the coarse internal data are 51×, 14×, and 4× coarser resolution than the reference case (950 × 1024), and indicate ~259-fold, ~70-fold, and ~9-fold computational speedup, respectively.Three scalar state variables, i.e., velocity u, v, and pressure p, are saved in 25 field snapshots, and the number of snapshots used as internal data for PINN training is 25 for the three coarser cases.The fraction per snapshot is 0.0364, 0.01, and 0.00268, respectively, to ensure the same internal data are used for PINN.A case without internal data is also performed for comparison.To demonstrate the robustness of HCDD-PINN, the architecture and hyperparameters are selected in a consistent fashion for all four of the cases.The architecture of the DNN used here is composed of eight hidden layers with 50 neurons per layer, to broadly balance the trade-off between the expressivity and trainability of the neural network.The weighting coefficients λ bc/I and λ int are set as one for both the initial and boundary conditions loss and the coarse internal loss.The total collocation point N collo is set to 3.1 × 10 5 (N LHS = 1.6 × 10 5 , N Lins = 1.5 × 10 3 , and timestep = 100).The learning rate and max batch size of Adam optimizer are set as 5 × 10 −4 and 5 × 10 3 , respectively.The network construction, training, and prediction are all performed in Tensorflow.All of the training and predictions are performed on a high-performance computer cluster with 80 GB memory, twenty CPUs, and one NVIDIA Tesla V100 GPU.Note that a comprehensive parameter study and an analysis of the optimization of the DNN's architecture will be conducted in the next section.
The training and predicting performances on these four cases are summarized in Table 1, where the training loss is defined by Equation (11).The relative l 2 error rl 2 e _y of the output variables is defined as: where y is the physical quantity of interest, including u, v, and p; N is the total number of the reference point.The predicting error is obtained by calculating the cycle-averaged relative l 2 error rl 2 e _y ( rl 2 e _y = 1/T ∑ rl 2 e _y) to evaluate the prediction performance.It can be seen from Table 1 that the PINN model without internal data inaccurately predicted the flow field past the flapping wing, despite the fact that the training loss had been minimized to less than 0.003.With the help of the limited coarse internal data, nearly 700 points per snapshot, the predicting error is largely reduced for all three of the other cases, at the cost of the increment in training cost and training loss.This is mainly caused by the introduction of ε int .The reference CFD simulation takes 1967 s to complete four stroke cycles under the same CPU cores as the PINN.The PINN training time increases ~5.8-fold (no internal), ~15.6-fold (149 × 128), ~9.4-fold (269 × 256), and ~6.8-fold (502 × 512), respectively.However, once the training process of HCDD-PINN model is finished, the flow information at arbitrary spatial-temporal coordinates can be rapidly predicted with reasonable accuracy.This will be suitable for wing-shape optimization, stroke-trajectory kinematics optimization, and aero-forces optimization.The advantages of HCDD-PINN will be more apparent when the situation is extended to the corresponding three-dimensional problems, which needs an excellent mesh quality to ensure the accuracy of the traditional CFD solvers.The 3D CFD simulations usually need to solve billions of degrees of freedom in the relevant spatial-temporal flow fields, leading to unfavorable trade-offs between accuracy and computational cost.The weighted total residual for all cases is less than 0.0051.As expected, the coarserresolution the internal data are, the more time and iterations the HCDD-PINN model needs to train.It is worth mentioning that the HCDD-PINN with 269 × 256 resolution internal data shows the best predicting performance compared to the other internal data resolutions, which results from the longer training time of the relatively accurate data with relatively few points.It can reconstruct the velocity field and the pressure field with relative accuracy.
More details about the residual loss convergence histories, including the total loss ε loss , the governing equations loss ε g , the boundary condition loss ε bc , the initial condition loss ε I , and the coarse internal loss ε int , can be found in Figure 6.It requires nearly 60,000~90,000 iterations to reach a properly trained HCDD-PINN model that is deemed ready to be used for predictions.The sudden drop in the loss histories is caused by switching from the ADAM optimizer to the L-BFGS-B optimizer.After training is complete, the predictions of the velocity and pressure fields can be obtained almost instantaneously.Nearly one hundred thousand points scattered in space are used for prediction at a certain instant, and 51 time snapshots are saved for the entire flapping cycle, including the initial and final time instants.The corresponding prediction cost for the four cases can be found in Table 1.The relative l2 errors of velocity and pressure fields, between the predictions of the HCDD-PINN model with different coarse internal data resolutions and the exact values of the corresponding CFD reference, are calculated for one cycle, as illustrated in Figure 7.The results of the PINN model with no internal data are given for comparison.It is clear that the proposed framework is capable of constructing the entire velocity and pressure fields, and has significantly improved the predicting accuracy compared to the original PINN model.The HCDD-PINN with very coarse internal data (149 × 128) still has a good prediction, which validates that the proposed method possesses a high degree of robustness.For a more in-depth investigation, cloud pictures for the predicted velocity (u, v) and pressure (p) fields of the HCDD-PINN model with different resolutions of coarse internal data are drawn at t/T = 0.18 and t/T = 0.78035, as shown in Figure 8 and Figure 9, respectively.They are compared with cases of no internal data and the ground truth reference.After training is complete, the predictions of the velocity and pressure fields can be obtained almost instantaneously.Nearly one hundred thousand points scattered in space are used for prediction at a certain instant, and 51 time snapshots are saved for the entire flapping cycle, including the initial and final time instants.The corresponding prediction cost for the four cases can be found in Table 1.The relative l 2 errors of velocity and pressure fields, between the predictions of the HCDD-PINN model with different coarse internal data resolutions and the exact values of the corresponding CFD reference, are calculated for one cycle, as illustrated in Figure 7.The results of the PINN model with no internal data are given for comparison.It is clear that the proposed framework is capable of constructing the entire velocity and pressure fields, and has significantly improved the predicting accuracy compared to the original PINN model.The HCDD-PINN with very coarse internal data (149 × 128) still has a good prediction, which validates that the proposed method possesses a high degree of robustness.After training is complete, the predictions of the velocity and pressure fields can be obtained almost instantaneously.Nearly one hundred thousand points scattered in space are used for prediction at a certain instant, and 51 time snapshots are saved for the entire flapping cycle, including the initial and final time instants.The corresponding prediction cost for the four cases can be found in Table 1.The relative l2 errors of velocity and pressure fields, between the predictions of the HCDD-PINN model with different coarse internal data resolutions and the exact values of the corresponding CFD reference, are calculated for one cycle, as illustrated in Figure 7.The results of the PINN model with no internal data are given for comparison.It is clear that the proposed framework is capable of constructing the entire velocity and pressure fields, and has significantly improved the predicting accuracy compared to the original PINN model.The HCDD-PINN with very coarse internal data (149 × 128) still has a good prediction, which validates that the proposed method possesses a high degree of robustness.For a more in-depth investigation, cloud pictures for the predicted velocity (u, v) and pressure (p) fields of the HCDD-PINN model with different resolutions of coarse internal data are drawn at t/T = 0.18 and t/T = 0.78035, as shown in Figure 8 and Figure 9, respectively.They are compared with cases of no internal data and the ground truth reference.It can be seen that, compared to the PINN with no internal case, the flow fields predicted by the HCDD-PINN closely follow the governing partial differential equations laws and For a more in-depth investigation, cloud pictures for the predicted velocity (u, v) and pressure (p) fields of the HCDD-PINN model with different resolutions of coarse internal data are drawn at t/T = 0.18 and t/T = 0.78035, as shown in Figures 8 and 9, respectively.They are compared with cases of no internal data and the ground truth reference.It can be seen that, compared to the PINN with no internal case, the flow fields predicted by the HCDD-PINN closely follow the governing partial differential equations laws and the enforced initial and boundary conditions at any instant, and the flows in the areas adjacent to the flapping wing and the wake can be precisely captured by the proposed framework.To distinguish the differences of flow fields between the prediction and the ground truth, we define the absolute error as  =  −  .The corresponding absolute error cloud pictures for the u, v velocities and p pressure fields are drawn at t/T = 0.18 and t/T = 0.78035, as shown in Figure 10 and Figure 11, respectively.The positive absolute error, illustrated by the red color, means that the predicted values are larger than the extracted ones, while the blue negative error represents that the predicted values are smaller than the extracted ones.The corresponding relative l2 errors of u_diff, v_diff, and p_diff, defined by Equation (20), are given in the upper left corner of the respective picture for the whole domain.It can be seen that the HCDD-PINN models with different resolutions of coarse internal data are able to produce a good prediction performance compared to the original PINN model with no data to help.The relatively large absolute error occurs near the flapping wing and some specific areas, while the field information for the most parts of the domain is predicted with sufficient accuracy.The prediction of the velocity field is more accurate than that for the pressure field for all cases.The relatively large To distinguish the differences of flow fields between the prediction and the ground truth, we define the absolute error as y di f f = y PI NN − y CFD .The corresponding absolute error cloud pictures for the u, v velocities and p pressure fields are drawn at t/T = 0.18 and t/T = 0.78035, as shown in Figures 10 and 11, respectively.The positive absolute error, illustrated by the red color, means that the predicted values are larger than the extracted ones, while the blue negative error represents that the predicted values are smaller than the extracted ones.The corresponding relative l 2 errors of u_diff, v_diff, and p_diff, defined by Equation (20), are given in the upper left corner of the respective picture for the whole domain.It can be seen that the HCDD-PINN models with different resolutions of coarse internal data are able to produce a good prediction performance compared to the original PINN model with no data to help.The relatively large absolute error occurs near the flapping wing and some specific areas, while the field information for the most parts of the domain is predicted with sufficient accuracy.The prediction of the velocity field is more accurate than that for the pressure field for all cases.The relatively large difference in magnitude between the exact and the predicted pressure fields may be attributed to the very nature of incompressible Navier-Stokes equations, since the pressure field is only identifiable up to a constant.difference in magnitude between the exact and the predicted pressure fields may be attributed to the very nature of incompressible Navier-Stokes equations, since the pressure field is only identifiable up to a constant.Despite the relatively large difference between the predicted and exact pressure fields, the distribution of the pressure coefficient Cp for the flapping wing appears to be accurately predicted by the proposed HCDD-PINN with limited coarse scattered data.Despite the relatively large difference between the predicted and exact pressure fields, the distribution of the pressure coefficient Cp for the flapping wing appears to be accurately predicted by the proposed HCDD-PINN with limited coarse scattered data.

Discussions
Firstly, in this section, an optimal parameter-search study is conducted to find the best DNN parameters for the flapping wing problem.Secondly, the effects of coarse internal data on the training and prediction performances, with respect to different snapshots fractions of internal data, are investigated by adopting the best combination of these parameters.Based on the above analysis, the coarse results obtained with the 269 × 256 grid resolution are selected as the internal data source for the HCDD-PINN model in this section, as it gives the best predicting performance.

Optimal Parameters Search
The training and predicting performances of DNNs are significantly affected by the architecture, hyper-parameters, and the number of collocation points, which are optimized by a coarse grid search.A summary of the training and predicting performances for all the test experiments are presented in Table 2.The baseline parameters were selected to be snapshot = 25, fraction = 0.01, N LHS = 1.6 × 10 5 , N Lins = 1.5 × 10 3 , timestep = 100, layer = 8, neurons per layer = 50, λ bc/I = 1, λ int = 1, learning rate = 5 × 10 −4 , and iteration = 5 × 10 3 .All these test parameters are varied individually by fixing the other parameters.It can be seen from Table 2 that, as N LHS increases, the more training time is needed, but the training loss and the predicting errors are decreased.N LHS of 2.4 × 10 5 is selected by weighting the computational cost and the prediction accuracy.It can be seen that the introduction of N Lins can significantly improve the HCDD-PINN's capability of predicting the flow past the flapping wing.However, the improvement tends to slow down while using larger points of N Lins .N Lins of 500 and timestep of 50 are selected in conformity with the smallest-collocation-points principle, which contributes to the total N collo of 2.65 × 10 5 .empirical findings indicate that deeper and wider networks are usually more expressive (i.e., they can capture a larger class of functions, but are often costlier to train, which represents a feed-forward evaluation as the neural network takes more time and the optimizer requires more iterations to converge).A large number of layers prevents a model from being generalized, while a small number of layers is insufficient to represent the system.Therefore, an architecture of DNN with the minimum number of layers and neurons and the desired performance is preferable.Among all cases considered here, the best structure of the HCDD-PINN includes six hidden layers with 100 neurons per layer.
With the loss weighting coefficients increasing, the corresponding loss terms decrease.Considering that the information in the internal data is inaccurate, it is unnecessary to reduce the internal loss to a very low value.Therefore, the combination of a λ bc/I of three and a λ int of two suitable to balance the weights of different items in the total residual.Additionally, the learning rate and the max number of iterations in the ADAM optimizer are 5 × 10 −5 and 1 × 10 4 , respectively, resulting in the better prediction performance of our proposed framework.
In summary, throughout the coarse grid search study, with the snapshot and fraction of coarse internal data fixed as 25 and 0.01, the best combination of the other parameters for DNN are N LHS = 2.4 × 10 5 , N Lins = 500, timestep = 50, layer = 6, neurons per layer = 100, λ bc/I = 3, λ int = 2, learning rate = 5 × 10 −5 , and iteration = 5 × 10 4 .These sets of parameters will be adopted when investigating the effects of the fraction and snapshot of coarse internal data in the next section.

Effect of the Fraction of Coarse Internal Data
A train-and-test data split method is used for preparing the coarse internal data of the HCDD-PINN.The fraction parameter illustrates the percentage of the training data in the source CFD data, and the test data are ignored in the training process.The effect of the fraction of coarse internal data on the training and predicting performances of our proposed HCDD-PINN model is investigated by fixing the snapshot number of the coarse internal data to 25.The source of the coarse internal data is a 269 × 256 coarse grid CFD simulation, and five values of the fraction are considered for each snapshot, i.e., 0.1, 0.05, 0.01, 0.005, and 0.001.That is to say, the total internal points at each snapshot are nearly 6900, 3450, 690, 345, and 69, respectively.
The overall training and predicting results of HCDD-PINN, with different fractions of coarse internal data, are summarized in Table 3.Compared to the case with the old DNN parameters, the HCDD-PINN with optimal parameters has a better predicting performance with an increment in the cost of training time.The weighted total residuals throughout this section are minimized to be less than 0.0025.The overall effect of the fraction is relatively small.As the fraction increases, the predicting errors of pressure can be effectively decreased.Except for the pressure field, the training and predicting performances of the HCDD-PINN hardly benefit from the large fraction of the coarse internal data.On the contrary, the HCDD-PINN can predict the flow fields of the flapping wing with relative accuracy with a limited dataset, as long as the internal points at each snapshot are larger than a hundred.It can also be seen in Figure 13, which exhibits cloud pictures for the predicted and exact velocity and pressure at t/T = 0.18, with respect to different fractions, that the proposed HCDD-PINN model is robust for accurately reconstructing the velocity and pressure flow fields when the fraction of coarse internal data varies.
The differences in u, v velocities, and p pressure fields are compared among the predictions of the HCDD-PINN with different fractions of coarse internal data and the exact CFD reference, which are given as t/T = 0.18 in Figure 14.The corresponding relative l 2 errors are illustrated below each respective picture.It can be seen that the absolute errors of u and v rarely change, with respect to the different fractions of the coarse internal data.The absolute error of p increases to a certain extent as the fraction decreases from 0.1 to 0.001.Considering that the internal points are reduced by nearly 100 times, such adverse effects are still deemed to be acceptable by weighting the data-driven amount and predicting The predicting capability of the proposed HCDD-PINN model is stable enough only if an appropriate snapshot number of coarse internal data is selected, no matter how much the fraction changes.
tion is relatively small.As the fraction increases, the predicting errors of pressure can be effectively decreased.Except for the pressure field, the training and predicting performances of the HCDD-PINN hardly benefit from the large fraction of the coarse internal data.On the contrary, the HCDD-PINN can predict the flow fields of the flapping wing with relative accuracy with a limited dataset, as long as the internal points at each snapshot are larger than a hundred.It can also be seen in Figure 13, which exhibits cloud pictures for the predicted and exact velocity and pressure at t/T = 0.18, with respect to different fractions, that the proposed HCDD-PINN model is robust for accurately reconstructing the velocity and pressure flow fields when the fraction of coarse internal data varies.

3.
Training and predicting performance of HCDD-PINN with respect to different snapshots and fraction of the coarse internal data.The differences in u, v velocities, and p pressure fields are compared among the predictions of the HCDD-PINN with different fractions of coarse internal data and the exact CFD reference, which are given as t/T = 0.18 in Figure 14.The corresponding relative l2 errors are illustrated below each respective picture.It can be seen that the absolute errors of u and v rarely change, with respect to the different fractions of the coarse internal data.The absolute error of p increases to a certain extent as the fraction decreases from 0.1 to 0.001.Considering that the internal points are reduced by nearly 100 times, such adverse effects are still deemed to be acceptable by weighting the data-driven amount and predicting accuracy.The predicting capability of the proposed HCDD-PINN model is stable enough only if an appropriate snapshot number of coarse internal data is selected, no matter how much the fraction changes.

Effect of the Number of Coarse Internal Data
To investigate the effects of the snapshot number of coarse internal data, seven values are considered for the snapshots of coarse internal data, i.e., twenty-fie, thirteen, nine, seven, five, four, and three.In particular, the snapshot = 0 represents the original PINN model without internal data.The fraction of coarse internal data is adopted as 0.005 to ensure the internal data are in the order of hundreds.The source CFD data are saved in 25 field snapshots, and the timestep per cycle of the HCDD-PINN is 50.Therefore, nearly every two, four, six, eight, twelve, sixteen, and twenty-four timestep intervals, the corresponding coarse internal data at a certain instant are introduced into the HCDD-PINN for training, respectively.The summaries of the training and predicting results of HCDD-PINN with different snapshots of coarse internal data are given in Table 3.The training loss is observed to increase, while a decrease is detected for the predicting errors and the training cost, as the snapshot of the coarse internal data increases.When all 25 of the coarse internal data snapshots are used for the model training, it would take the shortest training time to minimize the loss function to the lowest value, and perform best in reconstructing the flow fields as expected.The flow fields predicted by the original PINN model without internal data include large errors compared to the ground truth, while the predicting errors of HCDD-PINN sharply decrease with the inclusion of coarse internal data.When gradually adding the snapshot of coarse internal data to more than four, the predicting errors are reduced to an acceptable number, and the predicting performances remain stable as the snapshot increases.
The cloud pictures for the velocity and pressure fields at t/T = 0.18, predicted by HCDD-PINN with respect to different snapshots of coarse internal data, are summarized in Figure 15.It can be seen that the proposed HCDD-PINN model can accurately reconstruct the velocity and pressure flow fields by utilizing the coarse internal data.The flow information can be captured with sufficient accuracy, especially near the leading edge of the flapping wing, as long as the snapshot of coarse internal data is increased up to five.It can be better distinguished by drawing the corresponding cloud pictures of the differences between the predicted and exact velocity and pressure fields at this moment, as shown in Figure 16.It can clearly be concluded that the absolute errors between the HCDD-PINN and CFD decrease significantly when the snapshot of coarse internal data changes from 0 to 25.
It can be distinguished by drawing the corresponding cloud pictures of the differences between the predicted and exact velocity and pressure fields at this moment, as shown in Figure 16.It can clearly be concluded that the absolute errors between the HCDD-PINN and CFD decrease significantly when the snapshot of coarse internal data changes from 0 to 25.

Conclusions
Traditional deep learning requires a large amount of training data to solve the flow problems, which is usually a very time-consuming process.According to the known partial differential equations, physics-constrained deep learning has the potential to create faster and more accurate solutions fluid flow over an arbitrary domain of interest.However, the existing PINN model may generate unreasonable or unrealistic predictions for specific problems due to the innate lack of understanding in the complexities of neural network.In the current study, taking the flow past the flapping wing as a specific example, a hybrid coarse-data-driven physics-informed neural network (HCDD-PINN) model is proposed for solving such highly unsteady problems with large body motions.The sources of internal data are coarser in magnitude than is required by the traditional high-resolution CFD.
The training and predicting performances of HCDD-PINN with different resolutions of internal data are analyzed, by comparing it to the original PINN model without internal data.Additionally, the effects of the snapshot and fraction of the coarse internal data on the HCDD-PINN performances are investigated, based on the best-parameters combination of DNN obtained from an optimal parameter-search study.
By introducing the coarse internal data, the proposed HCDD-PINN model has sufficient accuracy, and performs reasonable and realistic predictions of the flow past the flapping wing.The results are compared to the ground-truth reference, which is obtained by an immersed-boundary method-based solver.In general, the velocity and pressure fields can be precisely predicted by HCDD-PINN with different grid resolutions for the coarse internal data.The absolute errors of velocity are lower than that of pressure p.The errors in predicting pressure can be effectively decreased by appropriately increasing the snapshot and fraction of the coarse internal data.The proposed HCDD-PINN framework is considered to have sufficient stability and accuracy for solving specific problems with unsteady features and large body motions.This method offers an alternative way to solve arbitrarily complex fluid dynamic problems.Its advantages in accuracy, stability, and efficiency are likely to be even more apparent for the corresponding optimization and three dimensional researches, since they are notoriously time consuming to solve with the standard CFD methods; this issue awaits our future work.

Figure 1 .
Figure 1.Illustration of the computational domain, boundary conditions, and kinematics of the flapping wing.

Figure 1 .
Figure 1.Illustration of the computational domain, boundary conditions, and kinematics of the flapping wing.

Figure 2 .
Figure 2. Architecture of the physics-informed neural network.For illustration purposes only, the network depicted in the figure comprises two hidden layers and five neurons per hidden layer.

Figure 2 . 25 2. 2 .
Figure 2. Architecture of the physics-informed neural network.For illustration purposes only, the network depicted in the figure comprises two hidden layers and five neurons per hidden layer.Here,

25 2. 2 .
N b represents the nonlinear compositional function parametrized by θ(W, b) (weight W; bias b), which can be expanded as below: Biomimetics 2024, 9, x FOR PEER REVIEW 5 of Hybrid Coarse Data-Driven with Physics-Informed Neural Network (HCDD-PINN)

25 2. 2 .
(ξ; θ) = Biomimetics 2024, 9, x FOR PEER REVIEW 5 of Hybrid Coarse Data-Driven with Physics-Informed Neural Network (HCDD-PINN) N l •; Θ N l • Biomimetics 2024, 9, x FOR PEER REVIEW 5 of 2.2.Hybrid Coarse Data-Driven with Physics-Informed Neural Network (HCDD-PINN) The structure of the neural network consists of Nl fully connected layers, which shown in Figure 2. The input layer  = (, , ) includes a set of independent space an time variables, while the output layer  = (, , , ) comprises velocity and pressu variables, and shear stress vector.The physical parameters of interest, for flow field com parisons later, consist of u, v, and p.All of them are evaluated because they are constraine through Navier-Stokes equations and boundary conditions simultaneously, which is di ferent from the traditional CFD method.The hidden layers are used to solve the comple nonlinearity relations between the inputs and outputs, which construct a map  to no linearly relate the inputs  to the outputs  as:  = ( ) = ( , ). ( Here, (⋅, ) ∈ ℝ ⟶ ℝ represents the nonlinear compositional function par metrized by (, ) (weight W; bias b), which can be expanded as below: (; ) =  (⋅;  ) ∘  (⋅;  ) ∘⋅⋅⋅∘  (;  ) ( where the symbol ∘ is the composition operation.The output vector at the ith layer (i = 2 … Nl) is calculated by the feed-forward algorithm, given by  (;  ) =  (  , ] ) =  (∑   +  ) ( where  (⋅;  ) ∈ ℝ ⟶ ℝ and  (⋅) ∈ ℝ ⟶ ℝ.Here, li is the size of the output at th ith layer and  ∈ ℝ * comprises the weights and biases corresponding to the lay i.The  ≡ ( ,⋅⋅⋅,  ,⋅⋅⋅  ) vectors are evaluated by training the neural network.

2. 2 .
Hybrid Coarse Data-Driven with Physics-Informed Neural Netw The structure of the neural network consists of Nl fully c shown in Figure 2. The input layer  = (, , ) includes a set time variables, while the output layer  = (, , , ) compri variables, and shear stress vector.The physical parameters of in parisons later, consist of u, v, and p.All of them are evaluated bec through Navier-Stokes equations and boundary conditions sim ferent from the traditional CFD method.The hidden layers are nonlinearity relations between the inputs and outputs, which c linearly relate the inputs  to the outputs  as:  = ( ) = ( , ).Here, (⋅, ) ∈ ℝ ⟶ ℝ represents the nonlinear com metrized by (, ) (weight W; bias b), which can be expande (; ) =  (⋅;  ) ∘  (⋅;  ) ∘⋅⋅⋅∘  where the symbol ∘ is the composition operation.The output v 2 … Nl) is calculated by the feed-forward algorithm, given by  (;  ) =  (  , ] ) =  (∑   where  (⋅;  ) ∈ ℝ ⟶ ℝ and  (⋅) ∈ ℝ ⟶ ℝ.Here, li is th ith layer and  ∈ ℝ * Biomimetics 2024, 9, x FOR PEER REVIEW 2.2.Hybrid Coarse Data-Driven with Physics-Informed Neural Network (HCDD-PINN) The structure of the neural network consists of Nl fully connected layers, w shown in Figure 2. The input layer  = (, , ) includes a set of independent spa time variables, while the output layer  = (, , , ) comprises velocity and p variables, and shear stress vector.The physical parameters of interest, for flow fiel parisons later, consist of u, v, and p.All of them are evaluated because they are cons through Navier-Stokes equations and boundary conditions simultaneously, which ferent from the traditional CFD method.The hidden layers are used to solve the co nonlinearity relations between the inputs and outputs, which construct a map  linearly relate the inputs  to the outputs  as:  = ( ) = ( , ).Here, (⋅, ) ∈ ℝ ⟶ ℝ represents the nonlinear compositional function metrized by (, ) (weight W; bias b), which can be expanded as below: (; ) =  (⋅;  ) ∘  (⋅;  ) ∘⋅⋅⋅∘  (;  ) where the symbol ∘ is the composition operation.The output vector at the ith laye 2 … Nl) is calculated by the feed-forward algorithm, given by  (;  ) =  (  , ] ) =  (∑   +  ) Here, (⋅, ) ∈ ℝ ⟶ ℝ represents the nonlinear compositional function parametrized by (, ) (weight W; bias b), which can be expanded as below: (; ) =  (⋅;  ) ∘  (⋅;  ) ∘⋅⋅⋅∘  (;  )

Figure 3 .
Figure 3. Illustration of the collocation points distribution adopted by the HCDD-PINN model.NLHS represents the collocation points for the entire spatial-temporal space.NLins represents the instant collocation points around the flapping wing within a cycle shape.

Figure 3 . 25 A
Figure 3. Illustration of the collocation points distribution adopted by the HCDD-PINN model.N LHS represents the collocation points for the entire spatial-temporal space.N Lins represents the instant collocation points around the flapping wing within a cycle shape.A flow chat for the whole method can be viewed in Figure 4.The steps of the NN training process can be described as follows: Step1: establish the physical model of flapping wing, including the governing equations, and boundary conditions, and prepare the training data for the coarser CFD results, including the initial condition and coarse internal data; Step2: design the suitable deep neural network, including the NN structure, function, learning rate, iteration, and so on; and create the corresponding loss function according to the physical model; Step3: train the DNN using the coarser internal data, and establish the map relationship between the input and output layers; use Adam and L-BFGS-B optimizers to update the weight and bias of DNN and reduce the error of the loss function; end training when the converged conditions are satisfied; Step4: save the trained model, residual histories, and predicted flow field information.

Biomimetics 2024, 9 ,
x FOR PEER REVIEW 9 of 25 initial condition loss  to the total loss.The high-resolution simulation (950 × 1024) is used as the ground truth and as a reference, and the three coarser-resolution simulations, i.e., 149 × 128, 269 × 256, and 502 × 512, are adopted as the coarse internal data, to improve the PINN's capability and prediction accuracy, which introduces the coarse internal loss  to the total loss.The coarse internal data are scattered randomly at various spatialtemporal locations, which can be easily utilized in the DNN framework.

Figure 5 .
Figure 5. (a) The lift and (b) drag forces under different resolutions calculated by CFD solver.

Figure 5 .
Figure 5. (a) The lift and (b) drag forces under different resolutions calculated by CFD solver.

Biomimetics 2024, 9 , 25 Figure 6 .
Figure 6.Required iterations during optimization of the total loss and other loss components in the training process of HCDD-PINN with different coarse internal data.

Figure 7 .
Figure 7. Relative l2 errors between the HCDD-PINN model predictions and the CFD reference for velocities (a) u, (b) v, and pressure (c) p fields.The results of the PINN model with no internal data are given for comparison.

Figure 6 .
Figure 6.Required iterations during optimization of the total loss and other loss components in the training process of HCDD-PINN with different coarse internal data.

Biomimetics 2024, 9 , 25 Figure 6 .
Figure 6.Required iterations during optimization of the total loss and other loss components in the training process of HCDD-PINN with different coarse internal data.

Figure 7 .
Figure 7. Relative l2 errors between the HCDD-PINN model predictions and the CFD reference for velocities (a) u, (b) v, and pressure (c) p fields.The results of the PINN model with no internal data are given for comparison.

Figure 7 .
Figure 7. Relative l2 errors between the HCDD-PINN model predictions and the CFD reference for velocities (a) u, (b) v, and pressure (c) p fields.The results of the PINN model with no internal data are given for comparison.
and boundary conditions at any instant, and the flows in the areas adjacent to the flapping wing and the wake can be precisely captured by the proposed framework.

Figure 10 .
Figure 10.Absolute error cloud pictures for the velocity (u_diff -left column, v_diff -middle column) and pressure (p_diff -right column) flow fields between the CFD and HCDD-PINN, with respect to different coarse internal data at t/T = 0.18: (a-c) PINN with no internal; (d-f) HCDD-PINN with 149 × 128; (g-i) HCDD-PINN with 269 × 256; (j-l) HCDD-PINN with 502 × 512.Comparisons of the distribution pressure coefficient C p in the flapping wing, between the CFD reference values (black line) and PINN predicted values with no internal data (grey line), 149 × 128 (red line), 269 × 256 (green line), and 502 × 512 (blue line) coarse internal data, at t/T = 0.18 and t/T = 0.78035, are illustrated in Figure12aand 12b, respectively.The pressure coefficient C p is obtained by dividing 0.5ρU 2 re f c by the pressure force.Despite the relatively large difference between the predicted and exact pressure fields, the distribution of the pressure coefficient C p for the flapping wing appears to be accurately predicted by the proposed HCDD-PINN with limited coarse scattered data.

Figure 11 .
Figure 11.Absolute error cloud pictures for the velocity (u_diff-left column, v_diff-middle column) and pressure (p_diff-right column) flow fields between the CFD and HCDD-PINN, with respect to different coarse internal data at t/T = 0.78035: (a-c) PINN with no internal; (d-f) HCDD-PINN with 149 × 128; (g-i) HCDD-PINN with 269 × 256; (j-l) HCDD-PINN with 502 × 512.Comparisons of the distribution pressure coefficient Cp in the flapping wing, between the CFD reference values (black line) and PINN predicted values with no internal data (grey line), 149 × 128 (red line), 269 × 256 (green line), and 502 × 512 (blue line) coarse internal data, at t/T = 0.18 and t/T = 0.78035, are illustrated in Figure 12a and 12b, respectively.The pressure coefficient Cp is obtained by dividing 0.5ρU ref 2 c by the pressure force.

Figure 11 .
Figure 11.Absolute error cloud pictures for the velocity (u_diff-left column, v_diff-middle column) and pressure (p_diff-right column) flow fields between the CFD and HCDD-PINN, with respect to different coarse internal data at t/T = 0.78035: (a-c) PINN with no internal; (d-f) HCDD-PINN with 149 × 128; (g-i) HCDD-PINN with 269 × 256; (j-l) HCDD-PINN with 502 × 512.Comparisons of the distribution pressure coefficient Cp in the flapping wing, between the CFD reference values (black line) and PINN predicted values with no internal data (grey line), 149 × 128 (red line), 269 × 256 (green line), and 502 × 512 (blue line) coarse internal data, at t/T = 0.18 and t/T = 0.78035, are illustrated in Figure 12a and 12b, respectively.The pressure coefficient Cp is obtained by dividing 0.5ρU ref 2 c by the pressure force.

Figure 12 .
Figure 12.The distribution of the pressure coefficient C p for the flapping wing predicted by the HCDD-PINN with respect to different coarse internal data at (a) t/T = 0.18 and (b) t/T = 0.78035.The results of the model with no internal data are given for comparison.

Figure 14 .
Figure 14.Absolute error cloud pictures for the velocity (u_diff-left column, v_diff-middle column) and pressure (p_diff-right column) flow fields between the CFD and HCDD-PINN, with respect to different fraction of coarse internal data at t/T = 0.18: (a-c) fraction = 0.001; (e-g) fraction = 0.005; (ik) fraction = 0.01; (m-o) fraction = 0.05; (q-s) fraction = 0.1.The corresponding relative l2 errors are given below the picture.The internal data used for different fractions are illustrated by red points at the corresponding position of the last column (d,h,i,p,t).

Figure 14 .
Figure 14.Absolute error cloud pictures for the velocity (u_diff -left column, v_diff -middle column) and pressure (p_diff -right column) flow fields between the CFD and HCDD-PINN, with respect to different fraction of coarse internal data at t/T = 0.18: (a-c) fraction = 0.001; (e-g) fraction = 0.005; (i-k) fraction = 0.01; (m-o) fraction = 0.05; (q-s) fraction = 0.1.The corresponding relative l 2 errors are given below the picture.The internal data used for different fractions are illustrated by red points at the corresponding position of the last column (d,h,l,p,t).

Table 1 .
Training and predicting performance for HCDD-PINN with different coarse internal data.

Table 2 .
Training and predicting performance of HCDD-PINN for optimal parameters search.
mse_u cycle-averaged mean square of u mse_v cycle-averaged mean square error of v mse_p cycle-averaged mean square error of p