Pedestrian Trajectory Prediction Based on Delaunay Triangulation and Density-Adaptive Higher-Order Graph Convolutional Network
Abstract
1. Introduction
1.1. Research Background
1.2. Literature Review
1.2.1. Physical Model-Based Methods
1.2.2. Deep Learning-Based Methods
1.2.3. Literature Summary
- Explicit modeling of social interaction relationships. By representing pedestrians as graph nodes and interaction relationships as graph edges, graph neural networks can explicitly model many-to-many social interactions with enhanced interpretability.
- Efficient parallel computation capability. Graph convolution operations possess inherent parallel computation characteristics, enabling simultaneous processing of feature aggregation for all pedestrians in the scene. This demonstrates superior computational efficiency in multi-pedestrian real-time prediction scenarios, meeting the real-time requirements of applications like autonomous driving.
- Balance issue in graph construction. Fully connected methods capture all potential interactions at the cost of substantial noise, while physics-constrained methods achieve computationally efficient but risk missing critical relationships. Therefore, maintaining important spatial proximity relationships while effectively avoiding redundant connections and constructing more reasonable graph structures is a core challenge for current GNN-based trajectory prediction methods.
- Limited scene adaptability. While higher-order graph convolution methods effectively model indirect influences, existing approaches employ fixed-order configurations, lacking adaptive adjustment mechanisms for different scenarios. Dynamically adjusting convolution orders based on scene characteristics, optimizing computational efficiency while ensuring prediction accuracy, is a key issue in enhancing the practical utility of higher-order graph convolution methods.
1.3. Main Contributions
- Delaunay triangulation-based sparse graph construction. We utilize the geometric properties of Delaunay triangulation to construct a spatial graph structure that maintains spatial proximity relationships while avoiding redundant connections. This provides a more reasonable topological foundation for subsequent graph convolution operations, thereby improving prediction accuracy.
- Density-adaptive higher-order graph convolution. We dynamically select the optimal graph convolution order based on local pedestrian density. The mechanism selects low-order convolution in high-density scenarios to avoid negative effects of visual occlusion, and convolution in low-density scenarios to adequately capture indirect interaction relationships. This adaptive mechanism balances prediction accuracy and computational efficiency across varying scene densities.
- Efficient computational optimization. We introduce a first-frame caching mechanism that computes density and optimal order only for the first frame of sequences, with subsequent frames reusing these values to avoid redundant computation. Additionally, a masked adaptive weight fusion module can dynamically fuse multi-order effective features, avoiding interference from invalid orders. This enhances feature representation capability, contributing to improved prediction accuracy, while ensuring the feature fusion process is efficient and redundancy-free.
1.4. Paper Structure
2. Method
2.1. Problem Definition
2.2. Overall Model
2.3. Delaunay Triangulation-Based Graph Construction
| Algorithm 1 Spatial Interaction Adjacency Matrix Construction Based on Delaunay Triangulation | |
| Input: —position set of pedestrians at a certain time, | |
| —spatial interaction adjacency matrix | |
| 1: | function BuildSpatialGraph (,) |
| 2: | # Initialize an zero matrix as adjacency matrix, where indicates initially no connection relationship between pedestrian i and pedestrian j |
| 3: | |
| 4: | # Handle special cases with insufficient pedestrian numbers, unable to form effective triangulation, using fully connected graph strategy |
| 5: | if then |
| 6: | |
| 7: | return |
| 8: | else |
| 9: | # Execute Delaunay triangulation algorithm on input pedestrian position point set , generating triangle set |
| 10: | |
| 11: | # Traverse all triangles generated by Delaunay triangulation, establishing connection relationships in the adjacency matrix for the three edges of each triangle |
| 12: | do |
| 13: | |
| 14: | |
| 15: | |
| 16: | end for |
| 17: | # Add self-loop connections for each node, self-loops represent that pedestrians’ own state information is preserved during graph convolution |
| 18: | to do |
| 19: | |
| 20: | end for |
| 21: | |
2.4. Density-Adaptive Higher-Order Graph Convolution
2.4.1. Density Calculation
2.4.2. Adaptive Maximum Order Selection
- In low-density scenarios, there are relatively few spatial constraints among individuals, pedestrians have relatively wide fields of view, enabling them to perceive and respond to social information from relatively distant locations. Third-order graph convolution can effectively capture distant indirect influences.
- In medium-density scenarios, individuals need to focus on both direct neighbors and indirect influences transmitted by two-hop neighbors through intermediate nodes. Second-order graph convolution balances the modeling of direct and indirect interactions.
- In high-density scenarios, visual occlusion significantly limits individuals’ attentional resources, focusing attention primarily on immediate neighbors’ behaviors. First-order graph convolution can meet requirements while avoiding redundant information.
| Algorithm 2 Density-Adaptive Order Selection Algorithm | |
| Input: —position set of | |
| Output: —spatial interaction adjacency matrix | |
| 1: | ) |
| 2: | # Check if it is the first frame of the sequence. If yes, calculate density; otherwise, return cached value |
| 3: | then |
| 4: | # Calculate local density for each pedestrian |
| 5: | to do |
| 6: | |
| 7: | to do |
| 8: | then |
| 9: | |
| 10: | end for |
| 11: | |
| 12: | end for |
| 13: | # Calculate scene average local density |
| 14: | |
| 15: | # Adaptively select maximum convolution order based on density thresholds |
| 16: | |
| 17: | |
| 18: | |
| 19: | # Cache first frame calculated order |
| 20: | |
| 21: | else |
| 22: | # Non-first frames directly read order from cache |
| 23: | |
| 24: | |
2.4.3. Higher-Order Graph Convolution
2.4.4. Adaptive Weight Fusion
2.5. Temporal Interaction Modeling
2.6. Temporal Convolutional Network and Trajectory Prediction
2.6.1. Spatio-Temporal Feature Fusion
2.6.2. Multimodal Trajectory Prediction
2.7. Loss Function
3. Experiments
3.1. Datasets
3.2. Evaluation Metrics
3.3. Experimental Settings
3.4. Quantitative Analysis
3.4.1. Comparison with Existing Methods
3.4.2. Model Parameters and Inference Time
3.5. Ablation Experiments
3.5.1. Component Ablation Experiments
3.5.2. Graph Construction Method Analysis
3.5.3. Graph Convolution Order Analysis
3.5.4. Density Threshold Sensitivity Analysis
3.5.5. First-Frame Caching Mechanism Analysis
3.6. Qualitative Analysis
4. Conclusions
4.1. Summary
- Delaunay triangulation-based sparse graph construction: We introduce Delaunay triangulation from computational geometry into pedestrian social interaction graph construction. By exploiting its geometrically optimal empty circle property, this approach maintains essential spatial proximity while eliminating redundant connections effectively, providing more reasonable topological foundations for subsequent graph convolution operations, thereby improving prediction accuracy. Compared to the fully connected graph method, this method reduces ADE and FDE by 10.5% and 13.4%, respectively. Compared to the physics-constrained method, it achieves reductions of 8.1% in ADE and 10.8% in FDE, demonstrating this method’s advantages in constructing reasonable graph structures.
- Density-adaptive higher-order graph convolution: We design a density-adaptive order selection mechanism that dynamically adjusts graph convolution order based on scene density. Low-density scenarios employ third-order convolution to capture long-range indirect interactions, medium-density scenarios use second-order convolution to balance direct and indirect influences, and high-density scenarios adopt first-order convolution to avoid visual occlusion interference. Ablation experiments show that compared to mainstream fixed first-order settings, this mechanism reduces ADE and FDE by 5.6% and 14.7%, respectively. Compared to the best-performing fixed third-order settings, it maintains comparable accuracy while reducing inference time from 0.0085 s to 0.0062 s, achieving a balance between accuracy and efficiency.
- Efficient computational optimization strategy: For characteristics of sequence prediction tasks, we design a first-frame caching mechanism to reduce algorithm time complexity. Simultaneously, we propose a masked adaptive weight fusion module achieving dynamic weighted combination of different order features, effectively addressing feature alignment issues under dynamic order configurations.
4.2. Future Work
- Fusion and Modeling of Multimodal Interaction Information: Current methods primarily focus on modeling pedestrian social interactions based on positional geometric relationships, with limited consideration of environmental constraint factors (such as static obstacles and road topology) and pedestrian motion attributes (such as velocity direction and target intent). Although spatial attention mechanisms can implicitly learn some pedestrian information from historical trajectories, when handling geometrically adjacent pedestrian pairs moving in opposite directions, explicitly incorporating velocity direction and target intent information may further enhance the model’s discriminative capability in more complex scenarios. Future work will explore integrating multimodal information such as semantic scene information, velocity vectors, and target intent into graph node features or edge weight calculations, further improving the model’s prediction accuracy and interpretability in complex heterogeneous scenarios.
- More Refined Scene-Adaptive Mechanisms: This paper employs an order selection strategy based on scene average density, but this method still relies on manually set hyperparameters that may require readjustment for specific application scenarios. Moreover, using a single global variable within the same frame may have limitations in scenarios with severe density heterogeneity. Future work will explore designing end-to-end learnable order selection networks, with dynamic decision modules based on deep reinforcement learning enabling the model to autonomously learn order decision strategies. For scenarios with severe density heterogeneity, we will investigate the feasibility of dynamically selecting independent orders for each pedestrian node, achieving more refined scene-adaptive modeling while ensuring computational efficiency, further enhancing the model’s adaptive capabilities across different scenarios.
- Engineering deployment and optimization: Current experiments are limited to offline evaluations. Future work will deploy the method to actual autonomous driving platforms, validating its performance in real environments. Additionally, for resource constraints of edge computing devices like mobile robots, we will research model compression and quantization acceleration techniques. These efforts aim to achieve further balance between real-time performance and accuracy, promoting algorithm transition from theoretical research to practical applications.
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Rudenko, A.; Palmieri, L.; Herman, M.; Kitani, K.M.; Gavrila, D.M.; Arras, K.O. Human Motion Trajectory Prediction: A Survey. Int. J. Robot. Res. 2020, 39, 895–935. [Google Scholar] [CrossRef]
- Jiang, J.; Yan, K.; Xia, X.; Yang, B. A Survey of Deep Learning-Based Pedestrian Trajectory Prediction: Challenges and Solutions. Sensors 2025, 25, 957. [Google Scholar] [CrossRef]
- Huang, Y.; Du, J.; Yang, Z.; Zhou, Z.; Zhang, L.; Chen, H. A Survey on Trajectory-Prediction Methods for Autonomous Driving. IEEE Trans. Intell. Veh. 2022, 7, 652–674. [Google Scholar] [CrossRef]
- Bharilya, V.; Kumar, N. Machine Learning for Autonomous Vehicle’s Trajectory Prediction: A Comprehensive Survey, Challenges, and Future Research Directions. Veh. Commun. 2024, 46, 100733. [Google Scholar] [CrossRef]
- Zhang, C.; Berger, C. Pedestrian Behavior Prediction Using Deep Learning Methods for Urban Scenarios: A Review. IEEE Trans. Intell. Transp. Syst. 2023, 24, 10279–10301. [Google Scholar] [CrossRef]
- Schöller, C.; Aravantinos, V.; Lay, F.; Knoll, A. What the Constant Velocity Model Can Teach Us About Pedestrian Motion Prediction. IEEE Robot. Autom. Lett. 2020, 5, 1696–1703. [Google Scholar] [CrossRef]
- Zhang, L.; Yuan, K.; Chu, H.; Huang, Y.; Ding, H.; Yuan, J.; Chen, H. Pedestrian Collision Risk Assessment Based on State Estimation and Motion Prediction. IEEE Trans. Veh. Technol. 2022, 71, 98–111. [Google Scholar] [CrossRef]
- Helbing, D.; Molnár, P. Social Force Model for Pedestrian Dynamics. Phys. Rev. E 1995, 51, 4282–4286. [Google Scholar] [CrossRef]
- Bi, H.; Fang, Z.; Mao, T.; Wang, Z.; Deng, Z. Joint Prediction for Kinematic Trajectories in Vehicle-Pedestrian-Mixed Scenes. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 10382–10391. [Google Scholar]
- Cheng, H.; Sester, M. Modeling Mixed Traffic in Shared Space Using LSTM with Probability Density Mapping. In Proceedings of the 2018 21st International Conference on Intelligent Transportation Systems (ITSC), Maui, HI, USA, 4–7 November 2018; pp. 3898–3904. [Google Scholar]
- Lee, N.; Choi, W.; Vernaza, P.; Choy, C.B.; Torr, P.H.S.; Chandraker, M. DESIRE: Distant Future Prediction in Dynamic Scenes with Interacting Agents. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2165–2174. [Google Scholar]
- Alahi, A.; Goel, K.; Ramanathan, V.; Robicquet, A.; Fei-Fei, L.; Savarese, S. Social LSTM: Human Trajectory Prediction in Crowded Spaces. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 961–971. [Google Scholar]
- Gupta, A.; Johnson, J.; Fei-Fei, L.; Savarese, S.; Alahi, A. Social GAN: Socially Acceptable Trajectories with Generative Adversarial Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 2255–2264. [Google Scholar]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 6000–6010. [Google Scholar]
- Giuliari, F.; Hasan, I.; Cristani, M.; Galasso, F. Transformer Networks for Trajectory Forecasting. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 10335–10342. [Google Scholar]
- Yuan, Y.; Weng, X.; Ou, Y.; Kitani, K. AgentFormer: Agent-Aware Transformers for Socio-Temporal Multi-Agent Forecasting. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 9793–9803. [Google Scholar]
- Mohamed, A.; Qian, K.; Elhoseiny, M.; Claudel, C. Social-STGCNN: A Social Spatio-Temporal Graph Convolutional Neural Network for Human Trajectory Prediction. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 14412–14420. [Google Scholar]
- Huang, Y.; Bi, H.; Li, Z.; Mao, T.; Wang, Z. STGAT: Modeling Spatial-Temporal Interactions for Human Trajectory Prediction. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6271–6280. [Google Scholar]
- Shi, L.; Wang, L.; Long, C.; Zhou, S.; Zhou, M.; Niu, Z.; Hua, G. SGCN:Sparse Graph Convolution Network for Pedestrian Trajectory Prediction. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 8990–8999. [Google Scholar]
- Sang, H.; Chen, W.; Wang, J.; Zhao, Z. RDGCN: Reasonably Dense Graph Convolution Network for Pedestrian Trajectory Prediction. Measurement 2023, 213, 112675. [Google Scholar] [CrossRef]
- Zhang, X.; Angeloudis, P.; Demiris, Y. Dual-Branch Spatio-Temporal Graph Neural Networks for Pedestrian Trajectory Prediction. Pattern Recognit. 2023, 142, 109633. [Google Scholar] [CrossRef]
- Chen, W.; Sang, H.; Wang, J.; Zhao, Z. IMGCN: Interpretable Masked Graph Convolution Network for Pedestrian Trajectory Prediction. Transp. B Transp. Dyn. 2024, 12, 2389896. [Google Scholar] [CrossRef]
- Abu-El-Haija, S.; Perozzi, B.; Kapoor, A.; Alipourfard, N.; Lerman, K.; Harutyunyan, H.; Steeg, G.V.; Galstyan, A. MixHop: Higher-Order Graph Convolutional Architectures via Sparsified Neighborhood Mixing. In Proceedings of the 36th International Conference on Machine Learning (ICML), Long Beach, CA, USA, 9–15 June 2019; pp. 21–29. [Google Scholar]
- Kim, S.; Chi, H.; Lim, H.; Ramani, K.; Kim, J.; Kim, S. Higher-Order Relational Reasoning for Pedestrian Trajectory Prediction. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024; pp. 15251–15260. [Google Scholar]
- Chen, W.; Sang, H.; Zhao, Z. PCHGCN: Physically Constrained Higher-Order Graph Convolutional Network for Pedestrian Trajectory Prediction. IEEE Internet Things J. 2025, 12, 25033–25045. [Google Scholar] [CrossRef]
- Wen, D.; Xu, H.; He, Z.; Wu, Z.; Tan, G.; Peng, P. Density-Adaptive Model Based on Motif Matrix for Multi-Agent Trajectory Prediction. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024; pp. 14822–14832. [Google Scholar] [CrossRef]
- Chen, W.; Sang, H.; Wang, J.; Zhao, Z. DSTIGCN: Deformable Spatial-Temporal Interaction Graph Convolution Network for Pedestrian Trajectory Prediction. IEEE Trans. Intell. Transp. Syst. 2025, 26, 6923–6935. [Google Scholar] [CrossRef]
- Fall, A.; Fortin, M.-J.; Manseau, M.; O’Brien, D. Spatial Graphs: Principles and Applications for Habitat Connectivity. Ecosystems 2007, 10, 448–461. [Google Scholar] [CrossRef]
- Feng, J.; Fu, J.; Shang, C.; Lin, Z.; Niu, X.; Li, B. Efficient Generation Strategy for Hierarchical Porous Scaffolds with Freeform External Geometries. Addit. Manuf. 2020, 31, 100943. [Google Scholar] [CrossRef]
- Yuan, W.; Zhu, J.; Wang, N.; Zhang, W.; Dai, B.; Jiang, Y.; Wang, Y. A Dynamic Large-Deformation Particle Finite Element Method for Geotechnical Applications Based on Abaqus. J. Rock Mech. Geotech. Eng. 2023, 15, 1859–1871. [Google Scholar] [CrossRef]
- Fradi, H.; Luvison, B.; Quoc, C.P. Crowd Behavior Analysis Using Local Mid-Level Visual Descriptors. IEEE Trans. Circuits Syst. Video Technol. 2017, 27, 589–602. [Google Scholar] [CrossRef]
- Hall, E.T. The Hidden Dimension; Anchor Books: New York, NY, USA, 1990. [Google Scholar]
- Pellegrini, S.; Ess, A.; Schindler, K.; Van Gool, L. You’ll Never Walk Alone: Modeling Social Behavior for Multi-Target Tracking. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 27 September–4 October 2009; pp. 261–268. [Google Scholar]
- Lerner, A.; Chrysanthou, Y.; Lischinski, D. Crowds by Example. Comput. Graph. Forum 2007, 26, 655–664. [Google Scholar] [CrossRef]
- Robicquet, A.; Sadeghian, A.; Alahi, A.; Savarese, S. Learning Social Etiquette: Human Trajectory Understanding in Crowded Scenes. In Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 October 2016; pp. 549–565. [Google Scholar]
- Monti, A.; Bertugli, A.; Calderara, S.; Cucchiara, R. DAG-Net: Double Attentive Graph Neural Network for Trajectory Forecasting. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 2551–2558. [Google Scholar]
- Mohamed, A.; Zhu, D.; Vu, W.; Elhoseiny, M.; Claudel, C. Social-Implicit: Rethinking Trajectory Prediction Evaluation and The Effectiveness of Implicit Maximum Likelihood Estimation. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; pp. 463–479. [Google Scholar]
- Chen, J.; Cao, J.; Lin, D.; Kitani, K.; Pang, J. MGF: Mixed Gaussian Flow for Diverse Trajectory Prediction. In Proceedings of the 38th Conference on Neural Information Processing Systems (NeurIPS 2024), Vancouver, BC, Canada, 9–15 December 2024; pp. 57539–57563. [Google Scholar]
- Zhang, P.; Ouyang, W.; Zhang, P.; Xue, J.; Zheng, N. SR-LSTM: State Refinement for LSTM towards Pedestrian Trajectory Prediction. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 12077–12086. [Google Scholar]
- Lian, J.; Ren, W.; Li, L.; Zhou, Y.; Zhou, B. PTP-STGCN: Pedestrian Trajectory Prediction Based on a Spatio-Temporal Graph Convolutional Neural Network. Appl. Intell. 2023, 53, 2862–2878. [Google Scholar] [CrossRef]
- Fang, Y.; Jin, Z.; Cui, Z.; Yang, Q.; Xie, T.; Hu, B. Modeling Human-Human Interaction with Attention-Based High-Order GCN for Trajectory Prediction. Vis. Comput. 2022, 38, 2257–2269. [Google Scholar] [CrossRef]
- Youssef, T.; Zemmouri, E.; Bouzid, A. STM-GCN: A Spatiotemporal Multi-Graph Convolutional Network for Pedestrian Trajectory Prediction. J. Supercomput. 2023, 79, 20923–20937. [Google Scholar] [CrossRef]
- Sun, C.; Wang, B.; Leng, J.; Zhang, X.; Wang, B. SDAGCN: Sparse Directed Attention Graph Convolutional Network for Spatial Interaction in Pedestrian Trajectory Prediction. IEEE Internet Things J. 2024, 11, 39225–39235. [Google Scholar] [CrossRef]
- Yu, C.; Ma, X.; Ren, J.; Zhao, H.; Yi, S. Spatio-Temporal Graph Transformer Networks for Pedestrian Trajectory Prediction. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; pp. 507–523. [Google Scholar]













| Method | Year | ETH | HOTEL | UNIV | ZARA1 | ZARA2 | AVG |
|---|---|---|---|---|---|---|---|
| Social-LSTM [12] | 2016 | 1.09/2.35 | 0.79/1.76 | 0.67/1.40 | 0.47/1.00 | 0.56/1.17 | 0.72/1.54 |
| Social-GAN [13] | 2018 | 0.81/1.52 | 0.72/1.61 | 0.60/1.26 | 0.34/0.69 | 0.42/0.84 | 0.58/1.18 |
| SR-LSTM [39] | 2019 | 0.63/1.25 | 0.37/0.74 | 0.51/1.10 | 0.41/0.90 | 0.32/0.70 | 0.45/0.94 |
| Social-STGCNN [17] | 2020 | 0.64/1.11 | 0.49/0.85 | 0.44/0.79 | 0.34/0.53 | 0.30/0.48 | 0.44/0.75 |
| SGCN [19] | 2021 | 0.63/1.03 | 0.32/0.55 | 0.37/0.70 | 0.29/0.53 | 0.25/0.45 | 0.37/0.65 |
| PTP-STGCN [40] | 2022 | 0.63/1.04 | 0.34/0.45 | 0.48/0.87 | 0.37/0.61 | 0.30/0.46 | 0.42/0.68 |
| High-order GCN [41] | 2022 | 0.54/1.09 | 0.24/0.44 | 0.53/1.14 | 0.41/0.89 | 0.32/0.70 | 0.41/0.85 |
| STMGCN [42] | 2023 | 0.73/1.13 | 0.31/0.42 | 0.45/0.85 | 0.33/0.53 | 0.29/0.46 | 0.42/0.67 |
| IMGCN [22] | 2024 | 0.61/0.82 | 0.31/0.45 | 0.37/0.67 | 0.29/0.51 | 0.24/0.42 | 0.36/0.57 |
| SDAGCN [43] | 2024 | 0.73/1.20 | 0.34/0.46 | 0.48/0.87 | 0.35/0.55 | 0.32/0.52 | 0.44/0.72 |
| HighGraph [24] | 2024 | 0.60/0.93 | 0.31/0.40 | 0.40/0.70 | 0.33/0.49 | 0.29/0.45 | 0.39/0.59 |
| DSTIGCN w/o LHS [27] | 2025 | 0.60/1.00 | 0.33/0.56 | 0.37/0.70 | 0.28/0.50 | 0.24/0.43 | 0.36/0.64 |
| Ours | - | 0.56/0.88 | 0.29/0.45 | 0.35/0.68 | 0.29/0.45 | 0.23/0.43 | 0.34/0.58 |
| STGAT [18] | DAG-Net [36] | SGCN [19] | Social-Implicit [37] | IMGCN [22] | Ours | |
|---|---|---|---|---|---|---|
| ADE | 0.58 | 0.53 | 0.46 | 0.47 | 0.46 | 0.43 |
| FDE | 1.11 | 1.04 | 0.75 | 0.89 | 0.74 | 0.72 |
| Method | Year | ETH | HOTEL | UNIV | ZARA1 | ZARA2 | AVG |
|---|---|---|---|---|---|---|---|
| Social-STGCNN [17] | 2020 | 0.40/0.63 | 0.59/0.92 | 0.33/0.50 | 0.49/0.76 | 0.42/0.66 | 0.45/0.69 |
| IMGCN [22] | 2024 | 0.48/0.87 | 0.49/0.90 | 0.35/0.62 | 0.34/0.59 | 0.28/0.48 | 0.39/0.69 |
| DSTIGCN w/o LHS [27] | 2025 | 0.52/1.08 | 0.39/0.67 | 0.33/0.56 | 030/0.56 | 0.25/0.46 | 0.36/0.67 |
| Ours | - | 0.59/1.08 | 0.46/0.83 | 0.33/0.57 | 0.34/0.59 | 0.28/0.49 | 0.40/0.71 |
| Method | Characteristic | Model Parameters | Inference Times |
|---|---|---|---|
| Social-LSTM [12] | RNN | 264 K | 0.2188 s |
| SR-LSTM [39] | RNN | 64.9 K | 0.0708 s |
| Social-GAN [13] | RNN | 46.3 K | 0.0551 s |
| TF [15] | Transformer | 33,082.8 K | 0.0532 s |
| STAR [44] | Transformer | 964.9 K | 0.0214 s |
| Social-STGCNN [17] | GNN | 7.6 K | 0.0020 s |
| SGCN [19] | GNN | 25 K | 0.0023 s |
| RDGCN [20] | GNN | 28 K | 0.0025 s |
| IMGCN [22] | GNN | 23.3 K | 0.0030 s |
| Ours | GNN | 23.6 K | 0.0062 s |
| DTM | HGM | ETH | HOTEL | UNIV | ZARA1 | ZARA2 | AVG |
|---|---|---|---|---|---|---|---|
| - | - | 0.64/1.13 | 0.34/0.56 | 0.38/0.69 | 0.29/0.54 | 0.25/0.42 | 0.38/0.66 |
| √ | - | 0.56/0.86 | 0.33/0.59 | 0.37/0.57 | 0.29/0.51 | 0.25/0.45 | 0.36/0.62 |
| - | √ | 0.62/1.01 | 0.34/0.59 | 0.36/0.66 | 0.29/0.52 | 0.25/0.43 | 0.37/0.64 |
| √ | √ | 0.56/0.88 | 0.29/0.45 | 0.35/0.68 | 0.29/0.45 | 0.23/0.43 | 0.34/0.58 |
| Graph Construction Method | ETH | HOTEL | UNIV | ZARA1 | ZARA2 | AVG |
|---|---|---|---|---|---|---|
| Fully connected | 0.62/1.01 | 0.34/0.59 | 0.36/0.66 | 0.29/0.52 | 0.25/0.43 | 0.37/0.64 |
| Physics-constrained | 0.59/1.03 | 0.32/0.53 | 0.37/0.67 | 0.30/0.50 | 0.24/0.42 | 0.36/0.63 |
| Delaunay triangulation (Ours) | 0.56/0.88 | 0.29/0.45 | 0.35/0.68 | 0.29/0.45 | 0.23/0.43 | 0.34/0.58 |
| θ2 | 0.26 | 0.30 | 0.34 |
|---|---|---|---|
| θ1 | |||
| 0.10 | 0.43/0.73 | 0.43/0.73 | 0.44/0.72 |
| 0.14 | 0.44/0.73 | 0.43/0.72 | 0.4/0.72 |
| 0.18 | 0.44/0.74 | 0.44/0.73 | 0.44/0.73 |
| Method | ETH | HOTEL | UNIV | ZARA1 | ZARA2 | AVG | Inference Times |
|---|---|---|---|---|---|---|---|
| Frame-by-frame computation | 0.56/0.86 | 0.28/0.45 | 0.35/0.69 | 0.29/0.44 | 0.23/0.42 | 0.34/0.57 | 0.0072 s |
| First-frame caching | 0.56/0.88 | 0.29/0.45 | 0.35/0.68 | 0.29/0.45 | 0.23/0.43 | 0.34/0.58 | 0.0062 s |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Published by MDPI on behalf of the International Society for Photogrammetry and Remote Sensing. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Chen, L.; Li, J.; Xiao, J.; Liu, R. Pedestrian Trajectory Prediction Based on Delaunay Triangulation and Density-Adaptive Higher-Order Graph Convolutional Network. ISPRS Int. J. Geo-Inf. 2026, 15, 42. https://doi.org/10.3390/ijgi15010042
Chen L, Li J, Xiao J, Liu R. Pedestrian Trajectory Prediction Based on Delaunay Triangulation and Density-Adaptive Higher-Order Graph Convolutional Network. ISPRS International Journal of Geo-Information. 2026; 15(1):42. https://doi.org/10.3390/ijgi15010042
Chicago/Turabian StyleChen, Lei, Jiajia Li, Jun Xiao, and Rui Liu. 2026. "Pedestrian Trajectory Prediction Based on Delaunay Triangulation and Density-Adaptive Higher-Order Graph Convolutional Network" ISPRS International Journal of Geo-Information 15, no. 1: 42. https://doi.org/10.3390/ijgi15010042
APA StyleChen, L., Li, J., Xiao, J., & Liu, R. (2026). Pedestrian Trajectory Prediction Based on Delaunay Triangulation and Density-Adaptive Higher-Order Graph Convolutional Network. ISPRS International Journal of Geo-Information, 15(1), 42. https://doi.org/10.3390/ijgi15010042

