Geo-PhysNet: A Geometry-Aware and Physics-Constrained Graph Neural Network for Aerodynamic Pressure Prediction on Vehicle Fluid–Solid Surfaces
Abstract
1. Introduction
1.1. Research Background and Significance
1.2. The Bottlenecks and Challenges of Traditional Research Methods
1.3. The Rise and Limitations of Deep Learning Proxy Models
1.4. The Innovation and Contribution of This Article
- A Geo-PhysNet framework aimed at learning field differential structures is proposed, with its core lying in a dual-output head architecture customized for vehicle surface manifolds. Unlike previous methods that only predict scalar values [8,9,10,13], our network can simultaneously predict the pressure scalar field and its tangential gradient vector field on the surface manifold within a unified framework. The core motivation of this design lies in elevating the gradient from an implicit quantity that requires post-processing to an explicit and supervised learning objective, thereby compelling the network to learn and understand the local mechanical causes that lead to pressure changes, rather than merely memorizing and imitating discrete pressure values. This architecture deeply integrates multi-level geometric priors including curvature through a multi-scale graph attention network, providing a solid foundation for precise differential structure prediction.
- A hybrid loss function for enforcing mathematical consistency and physical rationality was constructed, providing a new paradigm for imposing field structure constraints in geometric deep learning. This is the key mechanism for achieving the proposed “differential representation learning”, which is significantly different from the traditional data-driven methods [14]. The core of this loss function is an innovative “pressure/gradient consistency loss”, which mathematically mandates that the pressure field output by the network must be the potential function of the gradient field they jointly predict. This ensures the intrinsic structural integrity of the predicted physical field and the irrotational property of the gradient field—two essential characteristics that the network is guided to learn. In addition, a physical regularization term based on the surface Laplacian operator was introduced to penalize non-smooth solutions that violate fluid physics priors, effectively suppressing high-frequency noise in the prediction results. This guides the model to generate a pressure field that is not only numerically precise but also structurally smooth, continuous, and physically realistic.
2. Methodology
2.1. Geometry-Aware Encoder
2.1.1. Problem Formulation: A Non-Dimensional Approach
2.1.2. Input Feature Representation
2.1.3. Multi-Scale Graph Attention Network (MS-GAT)
- Define the scale set: First, we define a set containing S different integers. Each integer will define an independent graph scale. For example, we can establish a set that contains three scales, and they respectively represent small-scale, mesoscale, and large-scale.
- Parallel construction of adjacency relationship: Next, for the same point cloud P and its features, we run the k-NN algorithm S times in parallel:
- For the small-scale graph (, ): We calculate the feature space distance between each point in the point cloud and all other points, and find the 16 points with the closest distances as its neighbor set . This defines the edge set of graph . In this graph, each node is only connected to its 16 closest neighbors, and the range it can “see” is very local, focusing on capturing fine geometric details, such as sharp chambers or minor curved surface changes.
- For the medium-scale graph (, ): We repeat the above-mentioned process, but this time, for each point , we find 32 nearest neighbors to form the neighbor set , and the edge set of graph . In this graph, the “field of view” of the nodes is expanded, and it can perceive the shape of the local region composed of more points, such as the overall contour of a car door handle or a rearview mirror.
- For the large-scale graph (, ): Similarly, we find 64 nearest neighbors for each point to form the neighbor set and the edge set of graph . At this scale, the model can aggregate information from a rather broad region, thereby understanding the macroscopic geometric trends over a larger range, such as the streamline shape of the car roof, the overall posture of the car body, and so on.
- Parallel Feature Learning: After constructing these S graphs (), their node sets are accurately the same, and only the edge sets are different. We feed the shared initial node features into S parallel Graph Attention (GAT) network branches with the same structure but non-shared weights. The GAT layer of the s-th branch will aggregate information strictly according to the adjacency relationship of graph . This means that in the small-scale branch, information transfer and update only occur among 16 neighbors; while in the large-scale branch, information will be exchanged among 64 neighbors.
2.2. Dual Output Head
2.2.1. Motivation and Rationale
- Richer physical characterization: The pressure gradient serves as a bridge connecting the pressure field and fluid dynamics. For example, the Adverse Pressure Gradient is the direct cause of the separation of the flow boundary layer and the generation of pressure difference resistance. By allowing the network to directly learn and predict gradients, we force it to understand the local mechanical mechanisms that lead to these key aerodynamic phenomena rather than merely learning the numerical mapping of pressure.
- Stronger learning constraints and generalization ability: Compared to an unstructured pressure scalar field, a gradient field provides stronger supervised signals and geometric constraints. To accurately predict the magnitude and direction of the gradient, the network must have a deeper understanding of the complex relationships among local geometric curvature, orientation, and flow direction. This more challenging and structured learning task itself is a powerful regularization which can effectively prevent the model from “overfitting” on the data, thereby enabling it to have stronger generalization ability when facing unfamiliar situations.
2.2.2. Differential Representation Learning
- Step 1: Define the local tangent plane .
- Step 2: Construct the orthogonal basis of the local tangent plane .
- First, we select a global, non-collinear reference vector, usually a unit vector of the coordinate axis. For example, . To avoid numerical singularity that may occur when is approximately parallel to , we will check if (where is a very small tolerance), then switch the reference vector to .
- The first tangent vector is obtained by performing a cross-product operation between the reference vector and the normal vector and then normalizing the result. This ensures that is orthogonal to , that is, it lies within the tangent plane.
- The second tangent vector is obtained by performing a cross-product of the normal vector and the first tangent vector . Since and are already unit orthogonal vectors, the result of their cross-product is naturally a unit vector and is orthogonal to both of them. Thus, the construction of the standard orthogonal basis is completed. also lies within the tangent plane.
- Step 3: Definition of Network Output.
- Pressure prediction head (): Outputs a scalar , which directly corresponds to the predicted pressure value.
- Gradient prediction head (): Outputs a two-dimensional vector . The two components are accurately the coordinates of the predicted tangential gradient vector in the local orthogonal basis we just constructed. Therefore, in the three-dimensional global coordinate system, the complete predicted gradient vector can be reconstructed as
2.3. Physics-Informed Hybrid Loss Function
2.3.1. Data-Driven Loss ()
- Pressure-fitting loss (): This is the most basic supervision signal. The mean squared error (MSE) is used to measure the deviation between the predicted pressure and the real CFD pressure .
- Gradient-fitting loss (): To provide direct supervision for the gradient prediction head, we first calculate the surface tangential gradient at each point from the real pressure field , and then obtain the label . Then, we use this “simulated value” to supervise the output of the gradient head.
2.3.2. Pressure/Gradient Consistency Loss ()
2.3.3. Surface Physical Regularization Loss ()
3. Verification Experiment
3.1. Experimental Setup
3.1.1. Dataset and Preprocessing
3.1.2. Benchmark Model
- PointNet++: As a landmark work in processing point cloud data, PointNet++ performs prominently in 3D understanding tasks through its multi-scale and multi-level feature extraction mechanism. We take it as a benchmark to test the ability of advanced point cloud networks in directly processing geometry for field prediction.
- DGCNN: This is the original model proposed in the DrivAerNet paper, which captures local geometric information through dynamic graph convolution. We take it as the benchmark because it represents a powerful geometric encoder that has been verified on the same dataset.
- GCN (Graph Convolutional Network): A basic graph convolutional network. This model uses a simple graph Laplacian operator for neighborhood information aggregation and does not incorporate attention or multi-scale mechanisms. It serves as a “minimalist” benchmark for evaluating the performance gains brought about by more complex graph structure designs.
- Transolver: This model efficiently captures physical correlations in complex geometric domains through a slicing and attention mechanism based on physical states, achieving leading PDE solution performance and superior scalability.
- Geo-PhysNet (w/o Physics): This is a key ablation version of our model. It has accurately the same geometry-aware encoder and dual-output head architecture as Geo-PhysNet. However, during training, only the data-driven loss () is used, and the two physical constraint loss terms and are completely excluded. This model is aimed at accurately quantifying the actual contribution of the physical constraint paradigm we proposed, compared to a purely data-driven model with the same architecture.
3.1.3. Evaluation Index
3.1.4. Implementation Details
3.2. Implementation Details
3.2.1. Computational Cost Analysis
- Traditional CFD (LES): The generation of ground truth data for a single vehicle geometry is computationally intensive. The process, including geometry cleanup, manual meshing, and solving, typically takes several days to complete. The Large Eddy Simulation (LES) solver step alone requires approximately 500 CPU-hours on a High-Performance Computing (HPC) cluster. This long turnaround time makes conventional CFD unsuitable for rapid, iterative design exploration.
Our Proposed Method
- Preprocessing: The initial data preprocessing, which involves sampling an 8192-point cloud from a surface mesh, is fully automated and highly efficient, taking only ∼15 s per geometry.
- Training: The model training is a one-time, offline process. Training our model for 1000 epochs on the full training set of 3000 geometries took approximately .
- Inference: Once the model is trained, its efficiency becomes evident. Predicting the complete surface pressure field for a new, unseen vehicle takes less than .
3.2.2. Experimental Comparison of Effectiveness of Pressure Gradient Consistency Measurement Methods
3.2.3. Ablation Study
3.2.4. Visual Comparison of Prediction Results
3.2.5. Visual Comparison of Prediction Errors
3.3. Computational Performance
- Training Phase: The training of the deep neural network is a one-time, offline process. The model was trained on our full dataset comprising 1000 unique vehicle geometries. The entire training process took approximately 48 h to converge on a single A6000 GPU. While this represents a significant computational investment, it is a one-off cost to generate a reusable, high-speed predictive tool.
- Inference Phase: The primary advantage of our data-driven approach is its exceptional inference speed. Once the model is trained, it can predict the surface pressure coefficient () field for a new, unseen vehicle geometry in a single forward pass. This process takes approximately 1.5 s.
- Comparison with Traditional CFD: The efficiency of our model becomes evident when compared to the traditional physics-based simulation used to generate our ground truth data. A single high-fidelity CFD simulation for one vehicle requires about 17 h of computation on a 128-core High-Performance Computing (HPC) cluster.
4. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Brand, C.; Anable, J.; Ketsopoulou, I.; Watson, J. Road to zero or road to nowhere? Disrupting transport and energy in a zero carbon world. Energy Policy 2020, 139, 111334. [Google Scholar] [CrossRef]
- Ahmed, S.R.; Ramm, G.; Faltin, G. Some salient features of the time-averaged ground vehicle wake. SAE Trans. 1984, 93, 473–503. [Google Scholar]
- Heft, A.I.; Indinger, T.; Adams, N.A. Experimental and numerical investigation of the DrivAer model. In Proceedings of the Fluids Engineering Division Summer Meeting, Rio Grande, PR, USA, 8–12 July 2012; American Society of Mechanical Engineers: New York, NY, USA, 2012; Volume 44755, pp. 41–51. [Google Scholar]
- Heft, A.I.; Indinger, T.; Adams, N.A. Introduction of a New Realistic Generic Car Model for Aerodynamic Investigations; Technical Report, SAE Technical Paper; SAE International: Warrendale, PA, USA, 2012. [Google Scholar]
- Aultman, M.; Wang, Z.; Auza-Gutierrez, R.; Duan, L. Evaluation of CFD methodologies for prediction of flows around simplified and complex automotive models. Comput. Fluids 2022, 236, 105297. [Google Scholar] [CrossRef]
- Thuerey, N.; Weißenow, K.; Prantl, L.; Hu, X. Deep learning methods for Reynolds-averaged Navier–Stokes simulations of airfoil flows. AIAA J. 2020, 58, 25–36. [Google Scholar] [CrossRef]
- Gunpinar, E.; Coskun, U.C.; Ozsipahi, M.; Gunpinar, S. A generative design and drag coefficient prediction system for sedan car side silhouettes based on computational fluid dynamics. Comput.-Aided Des. 2019, 111, 65–79. [Google Scholar] [CrossRef]
- Song, B.; Yuan, C.; Permenter, F.; Arechiga, N.; Ahmed, F. Surrogate modeling of car drag coefficient with depth and normal renderings. In Proceedings of the International Design Engineering Technical Conferences and Computers and Information in Engineering Conference, Boston, MA, USA, 20–23 August 2023; American Society of Mechanical Engineers: New York, NY, USA, 2023; Volume 87301, p. V03AT03A029. [Google Scholar]
- Remelli, E.; Lukoianov, A.; Richter, S.; Guillard, B.; Bagautdinov, T.; Baque, P.; Fua, P. Meshsdf: Differentiable iso-surface extraction. Adv. Neural Inf. Process. Syst. 2020, 33, 22468–22478. [Google Scholar]
- Jacob, S.J.; Mrosek, M.; Othmer, C.; Köstler, H. Deep learning for real-time aerodynamic evaluations of arbitrary vehicle shapes. arXiv 2021, arXiv:2108.05798. [Google Scholar] [CrossRef]
- Li, Z.; Kovachki, N.; Choy, C.; Li, B.; Kossaifi, J.; Otta, S.; Nabian, M.A.; Stadler, M.; Hundt, C.; Azizzadenesheli, K.; et al. Geometry-informed neural operator for large-scale 3d pdes. Adv. Neural Inf. Process. Syst. 2023, 36, 35836–35854. [Google Scholar]
- Pfaff, T.; Fortunato, M.; Sanchez-Gonzalez, A.; Battaglia, P. Learning mesh-based simulation with graph networks. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
- Wu, H.; Luo, H.; Wang, H.; Wang, J.; Long, M. Transolver: A fast transformer solver for pdes on general geometries. arXiv 2024, arXiv:2402.02366. [Google Scholar] [CrossRef]
- Kashefi, A.; Mukerji, T. Physics-informed PointNet: A deep learning solver for steady-state incompressible flows and thermal fields on multiple sets of irregular geometries. J. Comput. Phys. 2022, 468, 111510. [Google Scholar] [CrossRef]
- Deng, J.; Li, X.; Xiong, H.; Hu, X.; Ma, J. Geometry-guided conditional adaption for surrogate models of large-scale 3d PDEs on arbitrary geometries. In Proceedings of the ICLR 2024 Annual Conference, Vienna, Austria, 7–11 May 2024. [Google Scholar]
- Chang, A.X.; Funkhouser, T.; Guibas, L.; Hanrahan, P.; Huang, Q.; Li, Z.; Savarese, S.; Savva, M.; Song, S.; Su, H.; et al. Shapenet: An information-rich 3d model repository. arXiv 2015, arXiv:1512.03012. [Google Scholar]
- Elrefaie, M.; Dai, A.; Ahmed, F. DrivAerNet: A Parametric Car Dataset for Data-Driven Aerodynamic Design and Graph-Based Drag Prediction. In Proceedings of the ASME 2024 International Design Engineering Technical Conferences and Computers and Information in Engineering Conference, Washington, DC, USA, 25–28 August 2024. [Google Scholar]






| Model | MAE (↓) | RMSE (↓) | GME (↓) | GCS (↑) | Laplacian MAE (↓) |
|---|---|---|---|---|---|
| GCN | 0.0264 | 0.0381 | 0.1512 | 0.887 | 0.0286 |
| PointNet++ | 0.0217 | 0.0315 | 0.1305 | 0.911 | 0.0263 |
| DGCNN | 0.0165 | 0.0240 | 0.1189 | 0.946 | 0.0231 |
| Transolver | 0.0135 | 0.0200 | 0.0862 | 0.965 | 0.0225 |
| Geo-PhysNet (w/o Physics) | 0.0138 | 0.0201 | 0.0953 | 0.956 | 0.0230 |
| Geo-PhysNet (Ours) | 0.0112 | 0.0163 | 0.0616 | 0.981 | 0.0197 |
| Stage | Traditional CFD (LES) | Our Proposed Method |
|---|---|---|
| Setup/Preprocessing (per case) | Hours to days (manual meshing) | ∼15 s (automated point cloud sampling) |
| Solving/Inference (per case) | Days (∼500 CPU-hours on HPC) | <2 s (on a single GPU) |
| Training (one-time cost) | N/A | ∼24 h (on a single GPU) |
| Resulting Workflow | Batch processing, slow iteration | Interactive, near-real-time feedback |
| Consistency Measurement Method | GME (↓) | GCS (↑) | Convergence Stability |
|---|---|---|---|
| Simple Finite Difference | 0.055 (expected) | 0.968 (expected) | Oscillations may occur |
| Direct Curl-Free Constraint | 0.049 (expected) | 0.975 (expected) | Stable |
| Weighted Least Squares (Ours) | 0.0416 | 0.981 | Stable and efficient |
| Ablation Model | MAE (↓) | Laplacian-MAE (↓) | RMSE (↓) | GME (↓) | GCS (↑) |
|---|---|---|---|---|---|
| Geo-PhysNet(Full-Model) | 0.0112 | 0.0197 | 0.0163 | 0.0616 | 0.981 |
| w/o (Remove consistency loss) | 0.0129 | 0.0451 | 0.0188 | 0.0921 | 0.935 |
| w/o (Remove physical regularization loss) | 0.0115 | 0.0431 | 0.0169 | 0.0834 | 0.943 |
| w/o Geo-Features(Remove geometric features) | 0.0151 | 0.0487 | 0.0224 | 0.1097 | 0.921 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Liu, B.; Wang, H.; Xue, L.; Long, Y. Geo-PhysNet: A Geometry-Aware and Physics-Constrained Graph Neural Network for Aerodynamic Pressure Prediction on Vehicle Fluid–Solid Surfaces. Appl. Sci. 2025, 15, 11645. https://doi.org/10.3390/app152111645
Liu B, Wang H, Xue L, Long Y. Geo-PhysNet: A Geometry-Aware and Physics-Constrained Graph Neural Network for Aerodynamic Pressure Prediction on Vehicle Fluid–Solid Surfaces. Applied Sciences. 2025; 15(21):11645. https://doi.org/10.3390/app152111645
Chicago/Turabian StyleLiu, Bowen, Hao Wang, Liheng Xue, and Yin Long. 2025. "Geo-PhysNet: A Geometry-Aware and Physics-Constrained Graph Neural Network for Aerodynamic Pressure Prediction on Vehicle Fluid–Solid Surfaces" Applied Sciences 15, no. 21: 11645. https://doi.org/10.3390/app152111645
APA StyleLiu, B., Wang, H., Xue, L., & Long, Y. (2025). Geo-PhysNet: A Geometry-Aware and Physics-Constrained Graph Neural Network for Aerodynamic Pressure Prediction on Vehicle Fluid–Solid Surfaces. Applied Sciences, 15(21), 11645. https://doi.org/10.3390/app152111645

