Skip to Content
  • Article
  • Open Access

8 April 2025

An Inverted Transformer Framework for Aviation Trajectory Prediction with Multi-Flight Mode Fusion

,
,
,
,
,
and
1
The 10th Research Institute of China Electronics Technology Group, Chengdu 610036, China
2
College of Computer Science and Technology, Harbin Engineering University, Harbin 150001, China
3
Heilongjiang Dasanyuan Dairy Machinery Co., Ltd., Harbin 150069, China
4
Laboratory of Computer Security Problems, St. Petersburg Federal Research Center of the Russian Academy of Sciences (SPC RAS), Saint-Petersburg 199178, Russia
This article belongs to the Section Air Traffic and Transportation

Abstract

As globalization and rapid economic development drive a surge in air transportation demand, the need for enhanced efficiency and safety in flight operations has become increasingly critical. However, the exponential growth in flight numbers has exacerbated airspace congestion, creating a stark contrast with the limited availability of airspace resources. This imbalance poses significant challenges to flight punctuality and operational efficiency. To mitigate these issues, existing models often rely solely on individual flight data, which restricts the breadth and depth of feature learning. In this study, we propose an innovative Inverted Transformer framework for aviation trajectory prediction enhanced by multi-flight mode fusion. This framework leverages multi-flight inputs and inverted data processing to enrich feature representation and optimize the modeling of multi-variate time series. By treating the entire time series of each variable as an independent token, our model effectively captures global temporal dependencies and enhances correlation analysis among multiple variables. Extensive experiments on real-world aviation trajectory datasets demonstrate the superiority of our proposed framework. The results show significant improvements in prediction accuracy. Moreover, the integration of multi-flight data enables the model to learn more comprehensive flight patterns, leading to robust performance across varying flight conditions. This research provides a novel perspective and methodology for aviation trajectory prediction, contributing to the efficient and safe development of air transportation systems.

1. Introduction

The rapid progress of globalization and economic development have led to a significant increase in the demand for air transportation. This surge has imposed higher requirements for efficiency and safety in flight operations. The expanding aviation industry has not only intensified airport surface traffic but also led to a proliferation of taxiways and runways. As a result, the complexity of interconnections among these components has escalated. However, the limited availability of airspace resources is increasingly at odds with the exponential growth in flight numbers, exacerbating airspace congestion and negatively impacting flight punctuality and passenger comfort. Efficient ground operations become particularly challenging during extensive flight delays or unexpected changes in taxiway conditions [1]. Research into aviation trajectory prediction has a long history, dating back to the mid-20th century, and has evolved in tandem with advancements in the aviation industry and computer technology. Initially, trajectory predictions were predominantly based on mathematical models, leveraging Newton’s laws of motion and fundamental aerodynamic principles to formulate straightforward predictive models for aircraft trajectories. However, with the maturation of big data and machine learning technologies, data-driven approaches have come to the forefront. These techniques harness historical flight data, employing statistical analyses and machine learning algorithms to forecast future flight paths. Contemporary planning methods integrate multi-objective optimization, considering parameters such as flight duration and fuel efficiency, to devise optimal flight routes. Meanwhile, deep learning, through neural networks, processes vast spatiotemporal datasets, extracting intricate features to achieve high-precision trajectory predictions.
The recent advent of Transformer models has offered innovative solutions to a myriad of problems [2]. However, incorporating raw data into these models can result in the loss of correlations among multiple variables during the prediction phase. Furthermore, a significant portion of contemporary research predominantly utilizes trajectory data from a single flight for both training and prediction. This reliance on limited data from a single flight implies that the knowledge features acquired during the model training process are correspondingly restricted.
To address these limitations, we propose an improved Transformer model with inverted input. The inverted operation restructures the input data so that the complete temporal evolution of each flight variable (e.g., speed, altitude, etc.) is treated as an independent token before encoding. This fundamental architectural innovation provides three key benefits:
  • Global Feature Aggregation: Each token encapsulates an entire variable’s history, enabling the self-attention mechanism to directly model long-term intra-variable dependencies and cross-variable interactions at the system level.
  • Multi-Flight Knowledge Fusion: By processing data from multi-flights simultaneously, this enables model generalization to unseen flight phases through shared feature learning.
  • Physical Consistency Preservation: The inverted structure naturally aligns with aviation domain constraints, and temporal causality is preserved within each token.
The remainder of this paper is organized as follows: Section 2 provides an overview of the related work in trajectory prediction. Section 3 outlines the architecture of the proposed inverted Transformer model. Section 4 details the methodology and experimentation undertaken in this research. Section 5 discusses the performance outcomes derived from the experimental results, offering a comprehensive analysis. Finally, Section 6 concludes the paper.

3. Network Architecture

The conventional approach in time series forecasting using Transformer-based models involves processing data in a manner where each token represents a single time step across multiple variables, as illustrated in Figure 1. This approach influences the extraction of correlations among various variables, Specifically, the approach deals with the following challenges:
  • Heterogeneity of Simultaneous Measurements: Data points recorded at the same time step often represent distinct physical phenomena. Due to inconsistent recording practices across different variables, aggregating these points into a single token can obscure the inherent correlations between multiple variables. This aggregation can lead to a loss of critical information about the relationships between different physical processes, thereby hindering the model’s ability to capture multi-variate dependencies.
  • Complexity of Temporal Representation: The presence of a large number of local receptive fields, combined with the representation of temporally inconsistent events at the same time point, makes it challenging for tokens formed at a single time step to convey meaningful information. This complexity arises because the same time step may capture diverse and nonsynchronous events, which, when combined, can introduce noise and ambiguity into the token representation, thereby reducing the model’s effectiveness in processing temporal patterns.
  • Underutilization of Permutation-Invariant Attention: Although variations in sequences are significantly influenced by the order of the data, permutation-invariant attention mechanisms are not effectively utilized across the temporal dimension in traditional Transformer models. This limitation arises because the self-attention mechanism, while capable of capturing long-range dependencies, does not fully leverage the temporal structure of the data. As a result, the model may fail to effectively utilize the historical information and temporal context, leading to suboptimal performance in time series forecasting tasks.
Figure 1. The upper part depicts the flow of the original approach inputting into Transformer-based forecasters, while the lower part illustrates the flow of the inverted approach inputting into the Transformer-based forecasters.
However, this approach has inherent limitations when dealing with complex, multi-variate time series data, such as aviation trajectory data. In this study, we introduce an innovative inverted Transformer framework that significantly enhances the model’s ability to capture global temporal dependencies and multi-variate correlations. The network structure of the proposed model is illustrated in Figure 2.
Figure 2. (a) The overview of the inverted Transformer. (b) The framework of the multi-variate attention model. (c) The framework of the feedforward network model.

3.1. Inverted Input Embedding

In traditional Transformer models, the input data are organized such that each token represents a single time step across all variables. This can obscure the correlations between different variables, especially when the variables represent distinct physical phenomena. To address this, we employ an inverted input operation, where the entire time series of each variable is treated as an independent token. This approach allows the model to capture the global dependencies and trends across the entire time series while preserving the independence between variables.
Given the original input matrix X R T × N , where T is the number of time steps, and N is the number of features, the inverted matrix becomes X R N × T . Each feature’s entire time series is treated as a single token, enabling the model to better capture the dependencies and trends across the entire time series. The embedding operation is implemented using a Multi-Layer Perceptron (MLP).
X = X . t r a n s p o s e , X R N × T
h n 0 = E m b e d d i n g ( X : , n )
where H = { h 1 , h 2 , , h N } R N × T represents the set of N tokens with dimension D, and X : , n refers to the time series of the n-th feature.
Aviation trajectory variables exhibit well-defined physical couplings (such as nonlinear speed–altitude relationships during climb phases). Converntional Transformers bundle variables at the same timestep into a single token, forcing the model to learn these relationships within local temporal windows. In contrast, our inverted operation treats the entire time series of each variable as an independent token, enabling the self-attention mechanism to directly model the following:
  • Long-term intra-variable patterns;
  • Global inter-variable correlations.

3.2. Self-Attention Mechanism

The self-attention mechanism is a core component of the Transformer model (see Figure 3), enabling it to capture long-range dependencies in the data. In our inverted Transformer framework, since each token represents the entire time series of a single variable, the attention mechanism focuses on capturing the correlations between different variables.
Figure 3. The structure of self-attention.
For each token, the feature vector is multiplied by parameter matrices W Q , W K , and W V to obtain the query vector Q i , key vector K i , and value vector V i . The similarity between each variable and all other variables is computed using the dot product of the query and key vectors, followed by a softmax function to obtain the attention weights:
Q i = h i W Q , K i = h i W K , V i = h i W V , K j = h j W K
A i = s o f t m a x Q i K j T d k V i
A t t e n t i o n ( Q , K , V ) = A 1 A n
where d k is the embedding dimension of the attention head, which is the feature dimension of the query and key vectors. The scaling factor 1 d k normalizes each temporal token across its dimensions.
In the inverted Transformer, where tokens represent variable-level sequences, the attention score Q i K j T in Equation (4) fundamentally differs from conventional Transformer. The score computes the correlations between entire temporal evolutions of different variables.

3.3. Layer Normalization

To ensure stable training and prevent gradient vanishing, we apply layer normalization to the attention matrix A obtained in Section 3.2. This is done with the aim of averting gradient vanishing during the training phase, which is a phenomenon that can complicate the process of fitting the loss function. The detailed implementation procedure is as follows:
A = A + H
where H represents the original data block, and A = { h 1 , h 2 , , h n } , with n = 1 , 2 , , N .
Layer normalization standardizes the data, mitigating the risk of gradient explosion during the model training phase:
L a y e r N o r m ( A ) = h n M e a n ( h n ) V a r ( h n ) n = 1 , 2 , , N
In traditional Transformer models, normalization is applied to the representations of multiple variables at an identical timestamp, which can introduce interaction noise in noncausal or delayed processes. In contrast, our inverted input Transformer model applies normalization to the time series representations of individual variables, effectively mitigating discrepancies caused by inconsistent measurements.

3.4. Feedforward Network

The feedforward network (FFN) applies more sophisticated nonlinear transformations to the output of the attention mechanism. The FFN architecture comprises two Conv1d layers with an activation function:
F F N ( A ) = m a x ( 0 , w 1 A + b 1 ) w 2 + b 2
where w 1 and b 1 represent the weight matrix and bias vector for the first linear transformation, respectively, while w 2 and b 2 represent the weight matrix and bias vector for the second linear transformation.
Traditional Transformer models typically treat variables as independent entities, often neglecting their temporal dynamics within the broader sequence. This approach may result in insufficient detailed information for accurate prediction. In contrast, the inverted Transformer, by tokenizing the entire sequence of variables, excels at handling complex time series data. This method provides a more comprehensive representation, capturing the intricate temporal relationships and dependencies among variables, thereby enhancing the model’s predictive capabilities.

3.5. Projection

After the encoding block processing, which involves stacking self-attention, layer normalization, and feedforward networks, the resulting outputs are fed into a Multi-Layer Perceptron (MLP) for the projection operation. The MLP layers consist of an input layer, hidden layers, and an output layer. Through both linear and nonlinear combinations, the MLP promotes comprehensive interactions among various dimensions of the feature vectors, allowing the model to capture richer information on nonlinear and combined features. The projection can be succinctly expressed as
Y ^ : , n = P r o j e c t i o n ( h n L )
where h n L represents the output of the embedding vector with a lookback window size of n at the L-th layer, and Y ^ : , n denotes the predicted result for the flight trajectory.

4. Experiments

4.1. Dataset

The ADS-B data used in this study were obtained through a collaborative research program between our laboratory and the Tenth Research Institute of China Electronics Technology Group Corporation (CETC-10). The dataset comprises actual operational records in May 2021. The ADS-B dataset encompasses a comprehensive range of attributes, including flight number, call sign, timestamp, longitude, latitude, speed, and altitude, as detailed in Table 1. For the purposes of this study, the dataset was organized by flight numbers to extract the trajectory data for each specific flight. To ensure the robustness of the training inputs, a length check was applied to each flight’s data, filtering out sequences that might be too short and thus potentially less informative.
Table 1. Description of dataset feature.
As an automatic surveillance technology, ADS-B operates using a broadcast mechanism. Aircraft transmit state vectors as position, velocity, and altitude via an Extended Squitter at 0.5–1 Hz intervals. While this provides richer data streams than traditional radar systems, the technology inherently suffers from several limitations that become particularly pronounced in multi-flight integration scenarios. Space weather disturbances, such as those documented during the 2003 Halloween solar storms, can induce ionospheric scintillation that manifests as intermittent signal loss. Furthermore, the dependence on GNSS synchronization makes the system vulnerable to clock drift in older transponders, while terrain occlusion in mountainous regions creates spatial inconsistencies in sampling rates. Furthermore, during the data collection process, challenges such as inconsistent sampling intervals, missing values, and outliers may emerge. Compared to conventional single-flight studies, multi-flight integration faces the critical challenge of temporal misalignment caused by heterogeneous sampling rates. Our temporal alignment method employs adaptive interpolation to synchronize multi-flight data. In this context, it is imperative to preprocess the gathered ADS-B data prior to their utilization in training. The subsequent steps delineate the preprocessing procedure:
  • Elimination of Redundant Features: The flight number and call sign were identified as noncontributory to the prediction objective and were consequently excluded as superfluous features.
  • Temporal Feature Integration We amalgamated the data and temporal features to diminish the quantity of features, thereby streamlining operations.
  • Uniform Sampling: The method of uniform sampling was employed to diminish the quantity of data points while concurrently maintaining their representativeness.
  • Treatment of Missing Values: The missing data were imputed using the mean value method.
  • Dataset Partitioning: The dataset was divided into training, validation, and test sets following a 70%, 20%, and 10% ratio, respectively.
The primary objective of this experiment is to investigate the impact of inputting different scales of datasets during model training. To achieve this goal, the experiment simulates different scales of datasets by limiting the number of flight records. Specifically, as the limitation imposed increases, the number of data points decreases. Detailed experimental results are presented in Section 5.1.

4.2. Evaluation Metrics

In assessing the model’s performance, this paper employs three crucial indicators to comprehensively measure prediction accuracy: Mean Squared Error (MSE), Root Mean Squared Error (RMSE), and Mean Absolute Error (MAE).
The Mean Squared Error (MSE) is the average of the squares of the prediction errors, which is calculated as follows:
M S E = i = 1 N ( y i y ^ i ) 2 N
The Root Mean Squared Error (RMSE) is the square root of MSE, which is formulated as follows:
R M S E = i = 1 N ( y i y ^ i ) 2 N
The Mean Absolute Error (MAE) represents the average of the absolute values of the prediction errors, which is formulated as follows:
M A E = i = 1 N | y i y ^ i | N
where y i denotes the predicted value, and y ^ i represents the true value.

5. Experiment Results

5.1. Experiment Performance

In this study, we aimed to examine the influence of varying data scales on model training. To achieve this, we utilized different data length constraints to create our training sets. We then proceeded to make predictions for a flight with the ID CBJ5270-1619861400-schedule-0516:0.
By manipulating the data length constraint, we were able to input data at varying scales to examine the impact of these scales on the model’s performance. Additionally, we modified the number of encoder and decoder layers in pursuit of optimal results. The findings are detailed in Table 2:
Table 2. Detailed results for scaling up the dataset scale and the parameters of inverted Transformer. The best results are in bold.
The results indicate that the largest scale inputs during the training of the model yielded the best prediction evaluation. Consequently, we can infer that the scales of input data significantly influence the performance of the model.

5.2. Inverted Transformer Performance

In this study, we assessed the efficacy of the inverted Transformer by separately inputting inverted and original data into various Transformer-based models. Our aim was to determine if the invert operation could augment the capacity of Transformer models to process time series data more effectively. We compared the performance of the inverted data input with that of the unprocessed data input on the Reformer [31], Informer [32], Flowformer [28], and Flashattention [33]. The experimental results are presented in Table 3 and Figure 4. These results suggest that the invert operation enhances the ability of Transformer models to process and comprehend time series data. Specifically, it boosts the efficiency of Transformer predictors, bolsters their generalization capability to unseen variables, and promotes a more effective use of historical observation data.
Table 3. Result on various models based on Transformer using original and inverted inputting.
Figure 4. The effectiveness of inverted input vs original input on various forecasters; the light blue represents the results of original data inputting, while the dark blue represents the results of inverted data inputting.
From Figure 4, we can see that the inverted input outperformed the original input on all Transformer-based forecasters. Table 3 shows the improvement of the invert operation on each forecaster.
In conclusion, when data processed through the invert operation were fed into various Transformer-based forecasters, the experimental outcomes indicate a range of improvements relative to the direct input of raw data. This underscores the efficacy of the invert operation in enhancing time series data for Transformer-based forecasting models.

5.3. Result Visualization

In order to better evaluate the prediction effect, this paper visualizes the predicted results in the geographical coordinate system. Using the folium library, we mapped the flight trajectory of flight ID CBJ5270-1619861400-schedule-0516:0. To distinguish the predicted trajectory from the real trajectory, we used the color blue to represent the predicted trajectory and the color red to represent the real trajectory. The visualization is shown in Figure 5. It can be seen that the predicted trajectory of the model is consistent with the real trajectory.
Figure 5. Trajectory visualization in red line.

6. Conclusions

This study presents a novel approach to aviation trajectory prediction by introducing an inverted Transformer framework enhanced with multi-flight mode fusion. The proposed method addresses the limitations of traditional Transformer models, which often treat variables as independent entities and neglect their temporal dynamics within the broader sequence. By tokenizing the entire sequence of variables, the inverted Transformer effectively captures global temporal dependencies and enhances the correlation analysis among multiple variables. The experimental results demonstrate the superior performance of the proposed framework. The use of multi-flight data significantly enriches the feature representation, enabling the model to learn more comprehensive flight patterns and characteristics. This, in turn, leads to robust performance across varying flight conditions. The study also highlights the importance of data scale in model training, as larger input data scales yielded better prediction accuracy.
The findings of this research contribute to the field of aviation trajectory prediction by providing a new perspective and methodology. The inverted Transformer framework not only improves prediction accuracy but also enhances the model’s ability to generalize to unseen variables. This approach has the potential to be applied in various aviation related applications, such as air traffic management, flight scheduling, and aviation safety systems. Specifically, the variable-level temporal modeling could extend beyond positional prediction to energy-aware trajectory optimization—a task previously requiring a separate fuel model [34]. Future work may focus on further optimizing the model architecture and exploring additional data sources to further improve prediction accuracy. Additionally, this framework could empower Free Route Airspace operations through adaptive constraint learning while remaining applicable to maritime/vehicle domains.

Author Contributions

Conceptualization, G.L. and W.L.; Data curation, Y.O.; Formal analysis, G.L.; Funding acquisition, W.L.; Investigation, W.L. and Z.Z.; Methodology, W.L.; Project administration, G.L. and Y.O.; Resources, Z.Z.; Software, Y.O., X.Z. and D.H.; Validation, X.Z., Z.Z., D.H. and I.K.; Visualization, X.Z. and D.H.; Writing—original draft, X.Z. and D.H.; Writing—review and editing, G.L., W.L. and I.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was sponsored by the Fundamental Research Funds for the Central Universities, 3072024XX0604 and KYWZ120240606, and the Natural Science Foundation of Heilongjiang Province, LH2023F020.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

We gratefully acknowledge the valuable comments and suggestions provided by the editors and reviewers.

Conflicts of Interest

Authors Gaoyong Lu and Yang Ou were employed by the company The 10th Research Institute of China Electronics Technology Group. Author Dongcheng Huang was employed by the company Heilongjiang Dasanyuan Dairy Machinery Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Liu, S.; Zhou, S.; Miao, J.; Shang, H.; Cui, Y.; Lu, Y. Autonomous Trajectory Planning Method for Stratospheric Airship Regional Station-Keeping Based on Deep Reinforcement Learning. Aerospace 2024, 11, 753. [Google Scholar] [CrossRef]
  2. Dong, X.; Tian, Y.; Dai, L.; Li, J.; Wan, L. A New Accurate Aircraft Trajectory Prediction in Terminal Airspace Based on Spatio-Temporal Attention Mechanism. Aerospace 2024, 11, 718. [Google Scholar] [CrossRef]
  3. Xue, D.; Du, S.; Wang, B.; Shang, W.L.; Avogadro, N.; Ochieng, W.Y. Low-carbon benefits of aircraft adopting continuous descent operations. Appl. Energy 2025, 383, 125390. [Google Scholar] [CrossRef]
  4. Ammoun, S.; Nashashibi, F. Real time trajectory prediction for collision risk estimation between vehicles. In Proceedings of the 2009 IEEE 5Th International Conference on Intelligent Computer Communication and Processing, Cluj-Napoca, Romania, 27–29 August 2009; pp. 417–422. [Google Scholar]
  5. Schubert, R.; Richter, E.; Wanielik, G. Comparison and evaluation of advanced motion models for vehicle tracking. In Proceedings of the 2008 11th International Conference on Information Fusion, Cologne, Germany, 30 June–3 July 2008; pp. 1–6. [Google Scholar]
  6. Lytrivis, P.; Thomaidis, G.; Amditis, A. Cooperative path prediction in vehicular environments. In Proceedings of the 2008 11th International IEEE Conference on Intelligent Transportation Systems, Beijing, China, 12–15 October 2008; pp. 803–808. [Google Scholar]
  7. Barth, A.; Franke, U. Where will the oncoming vehicle be the next second? In Proceedings of the 2008 IEEE Intelligent Vehicles Symposium, Eindhoven, The Netherlands, 4–6 June 2008; pp. 1068–1073. [Google Scholar]
  8. Batz, T.; Watson, K.; Beyerer, J. Recognition of dangerous situations within a cooperative group of vehicles. In Proceedings of the 2009 IEEE Intelligent Vehicles Symposium, Xi’an, China, 3–5 June 2009; pp. 907–912. [Google Scholar]
  9. Kaempchen, N.; Weiss, K.; Schaefer, M.; Dietmayer, K.C. IMM object tracking for high dynamic driving maneuvers. In Proceedings of the IEEE Intelligent Vehicles Symposium, Parma, Italy, 14–17 June 2004; pp. 825–830. [Google Scholar]
  10. Jin, B.; Jiu, B.; Su, T.; Liu, H.; Liu, G. Switched Kalman filter-interacting multiple model algorithm based on optimal autoregressive model for manoeuvring target tracking. IET Radar Sonar Navig. 2015, 9, 199–209. [Google Scholar] [CrossRef]
  11. Okamoto, K.; Berntorp, K.; Di Cairano, S. Driver intention-based vehicle threat assessment using random forests and particle filtering. IFAC-PapersOnLine 2017, 50, 13860–13865. [Google Scholar] [CrossRef]
  12. Wang, Y.; Liu, Z.; Zuo, Z.; Li, Z.; Wang, L.; Luo, X. Trajectory planning and safety assessment of autonomous vehicles based on motion prediction and model predictive control. IEEE Trans. Veh. Technol. 2019, 68, 8546–8556. [Google Scholar] [CrossRef]
  13. Virjonen, P.; Nevalainen, P.; Pahikkala, T.; Heikkonen, J. Ship movement prediction using k-NN method. In Proceedings of the 2018 Baltic Geodetic Congress (BGC Geomatics), Olsztyn, Poland, 21–23 June 2018; pp. 304–309. [Google Scholar]
  14. Liu, J.; Shi, G.; Zhu, K. Vessel trajectory prediction model based on AIS sensor data and adaptive chaos differential evolution support vector regression (ACDE-SVR). Appl. Sci. 2019, 9, 2983. [Google Scholar] [CrossRef]
  15. Rong, H.; Teixeira, A.; Soares, C.G. Ship trajectory uncertainty prediction based on a Gaussian Process model. Ocean. Eng. 2019, 182, 499–511. [Google Scholar] [CrossRef]
  16. Silveira, P.; Teixeira, A.; Guedes-Soares, C. AIS based shipping routes using the Dijkstra algorithm. Transnav Int. J. Mar. Navig. Saf. Sea Transp. 2019, 13. [Google Scholar] [CrossRef]
  17. Martinsen, A.B.; Lekkas, A.M.; Gros, S. Optimal model-based trajectory planning with static polygonal constraints. IEEE Trans. Control. Syst. Technol. 2021, 30, 1159–1170. [Google Scholar] [CrossRef]
  18. Karbowska-Chilinska, J.; Koszelew, J.; Ostrowski, K.; Kuczynski, P.; Kulbiej, E.; Wolejsza, P. Beam search Algorithm for ship anti-collision trajectory planning. Sensors 2019, 19, 5338. [Google Scholar] [CrossRef] [PubMed]
  19. Lazarowska, A. Comparison of discrete artificial potential field algorithm and wave-front algorithm for autonomous ship trajectory planning. IEEE Access 2020, 8, 221013–221026. [Google Scholar] [CrossRef]
  20. Maw, A.A.; Tyan, M.; Nguyen, T.A.; Lee, J.W. iADA*-RL: Anytime graph-based path planning with deep reinforcement learning for an autonomous UAV. Appl. Sci. 2021, 11, 3948. [Google Scholar] [CrossRef]
  21. Wang, S.; Yan, X.; Ma, F.; Wu, P.; Liu, Y. A novel path following approach for autonomous ships based on fast marching method and deep reinforcement learning. Ocean. Eng. 2022, 257, 111495. [Google Scholar] [CrossRef]
  22. Akter, R.; Doan, V.S.; Lee, J.M.; Kim, D.S. CNN-SSDI: Convolution neural network inspired surveillance system for UAVs detection and identification. Comput. Netw. 2021, 201, 108519. [Google Scholar] [CrossRef]
  23. Jung, O.; Seong, J.; Jung, Y.; Bang, H. Recurrent neural network model to predict re-entry trajectories of uncontrolled space objects. Adv. Space Res. 2021, 68, 2515–2529. [Google Scholar] [CrossRef]
  24. Zhang, Y.; Zhang, Q.; Zhang, Y.; Zhu, Z. VGAE-AMF: A novel topology reconstruction algorithm for invulnerability of ocean wireless sensor networks based on graph neural network. J. Mar. Sci. Eng. 2023, 11, 843. [Google Scholar] [CrossRef]
  25. Lin, Z.; Zhang, X.; He, F. A GNN-LSTM-Based Fleet Formation Recognition Algorithm. In Proceedings of the International Conference on Guidance, Navigation and Control, Tianjin, China, 5–7 August 2022; pp. 7272–7281. [Google Scholar]
  26. Vaswani, A. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar]
  27. Wu, H.; Xu, J.; Wang, J.; Long, M. Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. Adv. Neural Inf. Process. Syst. 2021, 34, 22419–22430. [Google Scholar]
  28. Wu, H.; Wu, J.; Xu, J.; Wang, J.; Long, M. Flowformer: Linearizing transformers with conservation flows. arXiv 2022, arXiv:2202.06258. [Google Scholar]
  29. Wu, H.; Hu, T.; Liu, Y.; Zhou, H.; Wang, J.; Long, M. Timesnet: Temporal 2d-variation modeling for general time series analysis. arXiv 2022, arXiv:2210.02186. [Google Scholar]
  30. Nie, Y.; Nguyen, N.H.; Sinthong, P.; Kalagnanam, J. A time series is worth 64 words: Long-term forecasting with transformers. arXiv 2022, arXiv:2211.14730. [Google Scholar]
  31. Kitaev, N.; Kaiser, L.; Levskaya, A. Reformer: The Efficient Transformer. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  32. Zhou, H.; Zhang, S.; Peng, J.; Zhang, S.; Li, J.; Xiong, H.; Zhang, W. Informer: Beyond efficient transformer for long sequence time-series forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; Volume 35, pp. 11106–11115. [Google Scholar]
  33. Dao, T.; Fu, D.; Ermon, S.; Rudra, A.; Ré, C. Flashattention: Fast and memory-efficient exact attention with io-awareness. Adv. Neural Inf. Process. Syst. 2022, 35, 16344–16359. [Google Scholar]
  34. Khan, W.A.; Ma, H.L.; Ouyang, X.; Mo, D.Y. Prediction of aircraft trajectory and the associated fuel consumption using covariance bidirectional extreme learning machines. Transp. Res. Part E Logist. Transp. Rev. 2021, 145, 102189. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.