Next Article in Journal
Driving for More Moore on Computing Devices with Advanced Non-Volatile Memory Technology
Previous Article in Journal
Reactive Load Balancing for Sentient Spaces in Absence of Cloud and Fog
Previous Article in Special Issue
Research on the Enhancement of Provincial AC/DC Ultra-High Voltage Power Grid Security Based on WGAN-GP
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Radar Track Fusion Method Based on Parallel Track Fusion Model

1
Nanjing Research Institute of Electronic Technology, Nanjing 610500, China
2
School of Electronic and Information Engineering, Beihang University, Beijing 100191, China
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(17), 3461; https://doi.org/10.3390/electronics14173461
Submission received: 1 August 2025 / Revised: 25 August 2025 / Accepted: 28 August 2025 / Published: 29 August 2025
(This article belongs to the Special Issue Applications of Computational Intelligence, 3rd Edition)

Abstract

With the development of multi-sensor collaborative detection technology, radar track fusion has become a key means to improve target tracking accuracy. Traditional fusion methods based on Kalman filtering and weighted averaging have the problem of insufficient adaptability in complex environments. This paper proposes an end-to-end deep learning track fusion method, which achieves high-precision track reconstruction through residual extraction and parallel network fusion, providing a new end-to-end method for track fusion. The method combines the attention mechanism and the long short-term memory network in parallel and optimizes the computational complexity. Through the uncertainty weighting mechanism, the fusion weight is dynamically adjusted according to the reliability of the track features. Experimental results show that the mean absolute error of fusion accuracy of this method is 79% lower than the Kalman filter algorithm and about 87% lower than the mainstream deep learning model, providing an effective way for multi-radar track fusion in complex scenarios.

1. Introduction

Modern radar systems face challenges from complex electromagnetic environments and multi-target interference. Track estimation from a single sensor often suffers from insufficient accuracy and poor robustness. Track fusion can effectively improve the accuracy and reliability of target state estimation by integrating multi-sensor data. It is a core technology in military reconnaissance, intelligent transportation, and other fields.
As the core area of multi-sensor information fusion, track fusion technology aims to achieve high-precision joint estimation of the state of moving targets by collaboratively processing distributed sensor data. Linear weighted fusion methods, such as variance weighting and convex combination fusion, only use single-time position information and have low accuracy. Kalman filter fusion achieves dynamic updates through state estimation, but it has a strong dependence on motion models and noise parameters, and its performance degrades significantly when the model mismatches. Deep learning technology can establish a direct mapping from raw sensor data to fused track data, breaking through the limitations of traditional methods that rely on staged processing. Optimal fusion of multisource information can be achieved through end-to-end global optimization.
The progression of multi-radar track fusion methodologies has been shaped by persistent challenges in complex operational environments. Early linear fusion techniques, such as variance weighting and convex combination, relied exclusively on instantaneous position data, leading to significant accuracy degradation during target maneuvers due to their neglect of temporal dynamics [1]. This limitation spurred the development of Kalman filter-based approaches, where Kumar et al. interacting multiple model (IMM) filter enabled asynchronous fusion of radar and identification of friend or foe (IFF) data, substantially reducing maneuvering target position errors [2]. However, these model-driven methods maintained stringent dependencies on predefined motion patterns and noise parameters, causing performance collapse under model mismatch [3].
Subsequent innovations focused on sequential processing frameworks to enhance adaptability. Bu et al. combined IMM with unscented Kalman filters (UKF) to compensate for spatiotemporal biases under nearly coordinated turn (NCT) and nearly constant acceleration (NCA) models [3], while Zhang et al. avoided covariance matrix inversions through information filtering for equivalent centralized fusion [4]. Though the improved robustness, scalability issues persisted in distributed networks. Gayraud et al. addressed this through track graph construction using space-time discretization, processing asynchronous covariance via random maximum consensus [5]. Crucially, Dunham et al. revealed the fundamental trade-off between system-level fusion and point track fusion, that is, higher position accuracy and better velocity estimation [6], highlighting the need for context-aware architectures.
The shift toward complex systems yielded specialized solutions. Xia et al. unified active/passive radar data through measurement dimension expansion [7], while Li et al. Newton difference field fusion eliminated non-uniform errors in joint array domains [8]. For dynamic weighting, Qiao et al. comprehensive adaptive weighted data fusion (CAWDF) incorporated historical relationships to suppress outliers [9], and Zhang et al. asynchronous fusion based on track quality with multiple model (AFTQMM) allocated weights via multi-model track quality [10]. By fusing multi-period local tracks with time correlation, the mean square error of the fused track is lower than that of the local track [11]. Yet these methods struggled with real-time reliability shifts during interference, as fixed heuristics could not adapt to sudden sensor degradation.
Breakthroughs in deep learning have enabled new approaches. Yun et al. fully convolutional network avoided parameter explosion through differential loss weighting [12], while Li et al. cloud-edge collaborative architecture combined offline training with online updates [13]. Chen et al. deep learning track fusion algorithm bypassed traditional correlation assumptions for over-the-horizon radar (OTHR) fusion [14], and He et al. hybrid consensus strategy improved multi-target tracking through decentralized node estimation [15]. However, these approaches revealed three unresolved deficiencies: sequential spatiotemporal processing neglected feature synergy [12], static fusion weights ignored real-time confidence fluctuations [16], and multitask optimization imbalances persisted [13]. For radar networks, Wu et al. accelerated correlation via multithreading optimization [17], and Gao et al. optimized power combinations to extend detection ranges [18]. Despite these advances, the field’s core limitations remain: inadequate parallel spatiotemporal integration, inability to dynamically weight features based on live reliability, and unresolved conflicts in multitasking optimization deficiencies that require fundamentally new architectural approaches.
In order to solve the problem of insufficient spatiotemporal fusion, that is, the sequential processing of spatial and temporal features ignores their interdependence, and the static fusion strategy cannot be dynamically adjusted according to the changes in physical features, this paper proposes an end-to-end track fusion method with parallel fusion of time and space. This method enhances track features through a fully connected neural network and extracts features using a parallel attention mechanism and a long short-term memory network. Through selection of adaptive feature weights, it reconstructs high-precision fused tracks. The main contributions are as follows:
  • This study presents a series of data normalization methods, processing data based on different input physical features separately. Track data are segmented, rectified, and normalized using a z-score transformation. Sensor-specific physical properties are preserved by integrating a feature group enhancement layer and a temporal embedding mechanism.
  • The parallel track fusion model (PTFM) is based on a spatiotemporal-aware parallel fusion architecture, combines an attention mechanism for adaptive feature calibration, and uses a gated long short-term memory network for long-term temporal dependency modeling, effectively improving the fusion accuracy of target positions.
  • The model features an adaptive multitask loss mechanism that dynamically balances optimization conflicts among estimation tasks such as position, velocity, and heading. Uncertainty weights focus on high-confidence features, significantly improving fusion robustness and reducing error propagation in complex scenarios.
Verification on the multisource track segment association dataset shows that the proposed method outperforms existing traditional method and deep learning models in core metrics. The core metrics include position root mean square error (RMSE), mean absolute error (MAE), variance, etc. This method provides a technical method for intelligent collaborative sensing technology in distributed radar networks.

2. Multi-Radar Track Fusion Model

2.1. End-to-End Track Fusion Problem Description

Multi-sensor track fusion involves fusing track information from multiple sensors on the same target to obtain a more accurate, continuous, and robust target track estimate. This problem can be formalized as a state estimation problem: given a sequence of observations from a heterogeneous sensor system at discrete time steps, estimate the target’s true track. Within the framework of deep learning, we consider an end-to-end fusion approach. This method directly learns the mapping function from the original multi-sensor time series data to the optimal fusion track, realizes global optimization learning from input to output, and outputs the final target track.
Suppose we have M sensors, each of which outputs an observation vector at time step t . For the sensor m , its time series is
  X m = x m ( 1 ) , x m ( 2 ) , , x m ( T ) T R T × D m ,
where T is the time step length and D m is the dimension of the sensor m ’s observation vector, usually including position, velocity, heading, etc. For example, a typical observation vector can be x m ( t ) = [   x ,   y ,   v ,   θ   ] T . The fused target track sequence is
  Y = y ( 1 ) , y ( 2 ) , , y ( T ) T R T × D y ,
where D y is the output dimension, usually corresponding to the position, speed, and heading in the input, such as four dimensions: longitude, latitude, speed, and heading angle.
Our goal is to learn a mapping function F that maps a multi-sensor input sequence to a target track sequence:
  Y = F X 1 , X 2 , , X M ; Θ ,
where Θ is the parameter of the mapping function and the weights of the deep learning model. Therefore, we propose the parallel track fusion model (PTFM), which achieves the optimal fusion of multisource information through end-to-end global optimization.

2.2. Overall Framework Design

The core challenge of multi-radar track fusion is to solve the spatiotemporal asynchrony, feature heterogeneity, and robustness problems of heterogeneous sensor data in complex scenarios. The parallel track fusion model builds on time-step alignment and track correlation, employing a core model parallel architecture to achieve high-precision multi-radar track fusion through feature enhancement, spatiotemporal collaboration, dynamic optimization, and multitask output processing. The innovation of this architecture is reflected in three aspects. First, it preserves the physical characteristics of different radars through physical perception. Second, it integrates an attention mechanism with a gated long short-term memory (LSTM) network to address the problem of extracting spatiotemporal features from asynchronous data. Third, it designs an adaptive multitask loss function to dynamically balance the optimization conflicts among position, velocity, and heading estimation.
The input of PTFM is the time series observation tensor X R N × T × 8 of the dual radar sensors, where N is the number of samples, T is the fixed time step, and the feature dimension includes the longitude and latitude, speed, and heading of the two radar sources. Each radar contributes four-dimensional features, totaling eight dimensions. The output layer generates a fused track Y R N × T × 4 , which contains the fused longitude and latitude, speed, and heading. The overall framework is shown in Figure 1.
First, the sensor data is segmented, and the input tensor is decoupled into independent data subsets of sensor 1 and sensor 2 according to the radar source to avoid mutual interference of heterogeneous features. The bidirectional data is nonlinearly mapped through an independent fully connected network to achieve bidirectional feature data enhancement and retain the feature distribution patterns of different radar sensors. Triangular periodic coding is used to generate time embedding vectors so the model can separate and process the spatial position changes and time evolution laws caused by target movement. The importance of spatial features is dynamically adjusted through the attention weighting mechanism, and the gated LSTM is combined to capture the timing dependency to achieve the coordinated fusion of spatiotemporal features. Finally, the adaptive weight learning is combined to balance the multitask loss to generate position, speed, and heading fusion.

2.3. Data Preprocessing Module

Data preprocessing is fundamental to track fusion. Its core goal is to eliminate format differences in heterogeneous sensor input data through three-dimensional tensor reshaping and standardization. To eliminate time-step misalignment between the input sensor data and the true track data, the track segment clipping method utilizes a multidimensional track data preprocessing algorithm [19].
Different sensors have varying data collection rates and latencies, leading to inconsistent track timestamps and observation counts. To overcome this, a preprocessing pipeline ensures temporal alignment and data uniformity. First, the shared time coverage between sensors is determined by finding their earliest common start and latest common end timestamp. Next, for a data point from sensor A at time t , a time window is defined around t . Sensor B’s data falling within this window is averaged, aligning its value to t . The algorithm then segments the aligned tracks into sequential, standardized time windows defined by window size. At the same time, the real track data is aligned to the same timestamp using interpolation method. This process mitigates sensor timing differences, creates standardized temporal windows, computes inter-sensor disparities within them, and normalizes the data for downstream analysis or model training.
PTFM preprocessing includes dimensionality correction, sensor data separation, and linear normalization, providing high-quality input for subsequent feature extraction. Sensor stream separation is used to extract multidimensional observations from different sensors as input streams, and the true value of real data is used as the target stream to form a physically isolated supervision topology. In the process of mathematically modeling the standardization process, given a three-dimensional data tensor X R N × T × 4 , where N represents the number of samples and T represents the time step. Each feature dimension f { 1,2 , 3,4 } is independently normalized. Reshape the 3D feature slices into a 2D matrix M f :
  M f = x 1,1 , f x 1 , T , f x N , 1 , f x N , T , f R N × T .
Calculate the mean and standard deviation of the feature matrix:
  μ f = 1 N · T i = 1 N j = 1 T x i , j , f ,
σ f = 1 N · T i = 1 N j = 1 T x i , j , f μ f 2 .
x i , j , f represents the value of the feature of the f th dimension in the j th time step in the i th sample. μ f and σ f are the mean and standard deviation of the feature matrix, respectively. The preprocessing uses a standardized z-score transformation to eliminate the dimensional effect. The transformation formula is Z : x ~ i , j , f = ( x i , j , f μ f ) / σ f . At the same time, the parameters are saved to achieve a strictly reversible mapping Z 1 : x i , j , f = σ f · x ~ i , j , f + μ f , guarantees the stability of the reverse mapping, and ensures that the prediction results can be reverse-mapped to the geographic coordinate system to avoid information distortion.

2.4. Feature Enhancement Layer

The goal of the feature enhancement layer is to improve the representation ability of the original radar data through nonlinear transformation while retaining the physical property differences of the sensors. The data of each sensor passes through an independent two-layer fully connected network. The first layer uses the R e L U activation function to map the original 4-dimensional features, namely latitude, longitude, speed, and heading, to a 32-dimensional space. The second layer uses the ReLU activation function again to map the 32-dimensional feature vector to a 64-dimensional space. The purpose is to increase the expressiveness of each sensor data so subsequent fusion can be based on richer features. The whole process realizes a hierarchical transformation from original low-dimensional features to high-dimensional abstract representations. Suppose the input data of sensor i is s ( i ) R T × 4 , where T is the sequence length and h ( i ) represents the feature vector of sensor i after being mapped by the activation function:
  h 1 ( i ) = ReLU s ( i ) W 1 ( i ) + b 1 ( i ) ,                   W 1 ( i ) R 4 × 32 , b 1 ( i ) R 32 ,
h 2 ( i ) = ReLU h 1 ( i ) W 2 ( i ) + b 2 ( i ) ,                   W 2 ( i ) R 32 × 64 , b 2 i R 64 .
The two enhanced features are concatenated in the last dimension:
  h combined   = h 2 ( 1 ) , h 2 ( 2 ) R T × 128 .

2.5. Time Embedding Mechanism

Asynchrony is one of the core problems of multi-radar fusion. Therefore, the track data in the model is a uniform time-sampling sequence after preprocessing, and the original delay of the sensor has been corrected in the preprocessing stage. At the same time, LSTM itself has the ability to remember the time sequence, but it is difficult to accurately identify the absolute position information. Therefore, the time embedding part uses all-zero delay as a position encoding technique so each row in the embedding matrix corresponds to a different position rather than using real delay embedding. The encoding method is as follows:
  PE ( t , 2 i ) = sin t 1000 0 2 i / d PE ( t , 2 i + 1 ) = cos t 1000 0 ( 2 i + 1 ) / d ,
t is the time step index, i is the dimension index, and d = 16 is the embedding dimension. This design makes the embedding vectors of different time steps distinguishable through the frequency attenuation mechanism while retaining a certain periodicity so the model can separate and process the spatial position changes and time evolution laws caused by the movement of the target. Its core value is to inject absolute time-step position information into the model while improving the spatiotemporal modeling capability with minimal computational cost.
The enhanced 64-dimensional sensor features are spliced with the 16-dimensional time embedding vector in the channel dimension to form a fusion tensor with a dimension of T × 144. This operation deeply binds the time information with the spatial features so the model can use the information of the time dimension and the space dimension for track inference at the same time. In a variety of motion scenarios, the spliced features can more accurately capture the correlation between time delay and position change and improve the accuracy of fusion.

2.6. Parallel Fusion Core

The parallel fusion core is a key module of PTFM. Through the synergy of the attention weighting mechanism and gated LSTM, dynamic calibration of spatial features and long-term modeling of time dependence are achieved. This module contains four subunits, attention feature calibration, gated LSTM time series modeling, parallel fusion, and residual connection.

2.6.1. Attention Weighting Mechanism

The reliability of different radars in complex scenarios varies dynamically. The role of this attention mechanism is to dynamically select features, give greater weights to important features, and suppress unimportant features. Since the output dimension is the same as the input dimension, it is feature-by-feature attention, that is, a weight is given to each feature at each time step. Let the input feature tensor be X R B × T × D , where B is the batch size, T is the time series length, and D is the fusion features’ dimension. Its mathematical transformation follows a dual-mapping relationship, and the first layer is a high-dimensional nonlinear projection:
  H = ReLU X W 1 + b 1 ,
where W 1 R D × 128 , b 1 R 128 , the nonlinear dimensionality increase transformation of the input features, is realized through the 128-dimensional latent space activation, and the cross-sensor interaction features are extracted. The second level is dimensional adaptive gating:
  A = σ H W 2 + b 2 ,
where W 2 R 128 × D , b 2 R D , the Sigmoid activation function σ ( z ) = 1 + e z 1 , is used to generate the D -dimensional space-gating vector A of [0, 1], satisfying X attended   = X A , where ⊙ represents the Hadamard product and X attended represents the output feature tensor.

2.6.2. LSTM Time Series Modeling

The long short-term memory network (LSTM) solves the gradient vanishing problem of traditional RNN through the gating mechanism and is suitable for capturing the long-range dependency of the track. The LSTM time series model framework is shown in Figure 2.
The LSTM design of PTFM adopts an orthogonal initialization strategy and contains 64 memory units, and the cell state update equation is
  C t = f t C t 1 + i t tanh W c · h t 1 , x t ,
where the forget gate is f t ; the input gate is i t , which respectively control the fusion ratio of the historical state and the current input; W c is the weight matrix; h t 1 is the hidden state of the previous time step; and x t is the current input feature. L2 regularization is introduced, and λ = 1 0 4 is taken to constrain the norm of the weight matrix to reduce the risk of gradient vanishing. In the 50-step ship track prediction, the long-range dependency modeling error of LSTM is further reduced.

2.6.3. Gated Parallel Fusion and Residual Connection

To achieve the coordinated optimization of spatiotemporal features, PTFM innovatively adopts a dual-branch parallel fusion architecture. The temporal feature branch uses LSTM to output 64-dimensional temporal features H LSTM R T × 64 to capture the temporal evolution of target motion, while the spatial feature branch uses the 144-dimensional spatial features X attended   R T × 144 after attention weighting to retain the spatial correlation of the sensor.
The dual-path features are spliced into a 208-dimensional tensor along the channel dimension, and feature selection is achieved through the gating function:
  G = σ W g · H LSTM , X attended ,
where σ is the Sigmoid function, W g is the gating weight matrix, and G [ 0 ,   1 ] T × 208 is the gating coefficient. The fusion result is
  O = G H LSTM , X attended .
In order to avoid information loss in the deep network, the residual connection is introduced: O final   = O + W r · X attended , where W r is the linear projection matrix. The ablation experiment confirms that this design improves the fusion accuracy of dense target scenes and verifies the effectiveness of spatiotemporal feature complementarity.

2.7. Adaptive Weights and Multi-Task Output Layer

One of the core innovations of this study is the proposal of an adaptive multitask loss mechanism based on uncertainty modeling. The adaptive weight branch converts the input into a three-dimensional normalized weight coefficient through a fully connected layer and a Softmax activation function. This coefficient is used to dynamically adjust the loss contribution weight of each subtask during multitask learning, thereby achieving an adaptive balance between tasks and preventing low-quality features from contaminating the core feature prediction through a shared layer. The core breakthrough of this module is that there is no need to preset feature validity priors. The output dimension is expanded to seven dimensions, of which the newly added three dimensions are weight parameters. The output of the model at time step t includes three predicted values of position, velocity, and heading and the corresponding weight vector w t = w p t , w v t , w θ t R 3 , which is generated by the terminal Softmax layer:
  w t = softmax z t = e z p ( t ) e z t , e z v ( t ) e z t , e z θ ( t ) e z t ,
where z t is the linear projection output of the 128-dimensional fusion feature. The model dynamically balances the position, speed, and heading losses through the coefficients learned by the adaptive weight layer:
L = 1 exp σ p L pos   + 1 exp σ v L vel   + 1 exp σ c L course   + σ p + σ v + σ c ,
L pos   = 1 T t = 1 T p t true   p t pred 2 2 ,
L vel   = 1 T t = 1 T v t true   v t pred 2 ,
L course   = 1 T t = 1 T m i n θ t true   θ t pred , 360 θ t true   θ t pred ,
among them, L pos is the Euclidean distance loss of position, L vel is the mean square error loss of speed, and L course is the periodic loss of heading.
The dynamic weight parameters σ p , σ v , σ θ are calculated using sequential averaging:
  σ p = 1 T t = 1 T l n 1 w p t + ϵ .
This design results in a negative correlation between σ p and the weight w p t when the confidence level of the position prediction is high, w p t increases, and σ p decreases accordingly, thereby increasing the contribution of the position loss term L pos to the total loss. During backpropagation, the gradient of the loss function with respect to each weight parameter is
  L σ p = L pos exp σ p + 1 ,
L w p ( t ) = 1 T · L σ p · 1 w p t + ϵ .
In the process of training based on experimental data, the position weight is further increased, and the speed and heading weight are relatively low, which meets the actual needs of position accuracy priority in target tracking.
The goal of the multitask output layer is to simultaneously optimize the estimation accuracy of position, speed, and heading and solve the conflict of multi-objective optimization and the problem of low data quality of some input features. This layer adopts a decoupled output structure and an adaptive loss mechanism. The output layer of the model is designed as a multi-output architecture. The position estimation branch maps the feature into a 2D latitude and longitude coordinate output representing the geographic location through a fully connected layer and a linear activation function. The speed and heading estimation branch compresses the input into scalar outputs representing 1D instantaneous speed and heading, respectively, through a fully connected layer and a linear activation function.

3. Experiment and Result Analysis

3.1. Experimental Settings

3.1.1. Experimental Data Design

This study uses the multisource track association dataset (MTAD) for model validation [20]. The dataset covers multiple high-confidence simulation scenarios. The experimental data contains more than 18,000 sets of associated track segments, with a training set to validation set ratio of 3:2. Each set of track segment contains one track of data measured by two radar sensors and one AIS real track data. The fusion task involves challenging conditions, such as target maneuvering behavior, sensor noise interference, and complex system errors. The example diagram of sensor tracks and actual tracks is shown in Figure 3. Subfigures a, b, and c represent three examples of track scenarios. As can be seen from the example tracks, the initial radar detection tracks have large errors and sensor noise.
In the calculation process of the indicator evaluation part, the model is trained using standardized data with a mean of 0 and a standard deviation of 1. Therefore, when calculating results such as geographic distance error, it is calculated on the original data after denormalization, and the Haversine formula is used to calculate the actual geographic distance. Only after denormalization do these values have clear physical meanings, while the mean absolute error value in the training graph is the mean absolute error of the standardized data, which is independent of the original feature scale. Therefore, the values in the training graph are mainly used to present the training process, while the values in the evaluation table are used to objectively compare the performance of models and methods.

3.1.2. Training Parameter Configuration

To enhance experimental reproducibility, this section systematically describes the hyperparameter configuration and optimization strategy for model training. The model is implemented in TensorFlow 2.15 using the Adam optimizer. The initial learning rate is set to 0.001, and weight parameters are dynamically adjusted. The learning rate is adaptively decayed using a callback function: if the validation loss does not improve for 10 consecutive training epochs, the learning rate is decayed to 0.5 times the current value, with a lower threshold of 10 6 . The training batch size is fixed at 32, the maximum number of epochs is set to 150, and an early stopping mechanism is implemented to prevent overfitting. This mechanism intelligently terminates training by continuously monitoring the validation set loss trend. If the validation loss does not show a downward trend for 20 consecutive training epochs, the training process is automatically terminated, and the model weights corresponding to the lowest validation loss are automatically restored. Regarding the regularization strategy, L2 weight regularization is introduced in the LSTM layer with a coefficient of λ = 1 0 4 , and weight initialization uses an orthogonal initialization method. During training, the validation set position MAE, velocity MAE, and heading MAE metrics are monitored in real time, and the trends of key metrics are visualized using convergence curves.

3.2. Comparative Experiments

The comparative methods cover classic signal processing algorithms and cutting-edge deep learning models, including four traditional methods: weighted fusion (WF), Kalman filter (KF), convex combination fusion (CCF), covariance intersection fusion (CIF), and long short-term memory network (LSTM), Transformer, and fully connected models (Dense) based on deep learning. In terms of the evaluation system, the track association performance is quantitatively analyzed through three core indicators: position root mean square error (RMSE), mean absolute error (MAE), and variance.
The experimental results clearly show that the parallel fusion model proposed in this study shows comprehensive advantages in the track prediction task. As shown in Table 1, this method is significantly better than various comparative methods in the three core indicators of root mean square error, mean absolute error, and variance. This breakthrough performance is attributed to the innovative design of the model architecture. The attention mechanism and the long short-term memory network parallel processing structure are constructed; key feature focus and time-dependent modeling are taken into account. A gated fusion mechanism and residual connection are used to optimize the information integration process.
The PTFM method is significantly better than the traditional method and the single deep learning method in terms of position estimation accuracy, especially in complicated scenarios. The RMSE of the PTFM method is 32.7% lower than that of KF, indicating that it has a stronger generalization ability for complex motion patterns. The change in the mean absolute error of the position during training is shown in Figure 4a, and the model training loss is shown in Figure 4b. The vertical axis of the figure is a dimensionless indicator value used to measure model error or performance.
It is worth paying special attention to the outstanding performance of the mean absolute error, which is 72.4% lower than the optimal traditional method and 85.7% lower than the 2.4953 of the mainstream deep learning model. The parallel structure effectively leverages the complementary capabilities of the two mechanisms. The attention layer dynamically focuses on key track nodes, while the LSTM branch strengthens the modeling of long-term dependencies across the continuous track dynamics. This design overcomes the inherent limitations of single models. The LSTM model alone, due to its lack of feature focus, results in an MAE as high as 2.4953 km, a 604% degradation compared to PTFM. The Transformer model, on the other hand, suffers from a MAE of 3.1070 km due to the dilution of local details by global attention, a 778% degradation compared to PTFM. At the same time, the low level of prediction variance of 9.2169 verifies the stable performance of the gated fusion structure, which is 91.8% lower than the traditional method and 82.7% lower than the densely connected model.
In the data processing stage, standardized preprocessing ensures the scale consistency of multisource sensor inputs, and serialization processing completely retains the dynamic characteristics of the track. In the experiment, the performance of traditional filtering methods is limited by the linear assumption, such as the root mean square error of Kalman filtering is 10.667; although the mainstream time series model is better than the traditional method in root mean square error, it still has a significant gap with this method in key indicators due to the lack of a dedicated module for multisensor fusion. The comparison between the predicted track and the actual track and the change of the distance range error with the track point are shown in Figure 5.
The PTFM model achieves a root mean square error of 33.3% higher than the closest comparison method through multitask collaborative prediction of the position decoding layer, speed decoding layer, and heading decoding layer. Experimental data proves that in complex track prediction scenarios, the parallel fusion architecture has an advantage over the serial processing mode. Its attention weighting mechanism can effectively deal with sensor noise interference, and the gated fusion design significantly improves feature utilization.

3.3. Feature Weight Comparison Experiment

This experiment deeply analyzes the performance difference of the adaptive weight module in multitask fusion. The experimental data reveal the significant optimization effect of this mechanism on position estimation and the complex trade-off relationship between other tasks. The results of the feature weight experiment are shown in Table 2.
The adaptive weight mechanism shows excellent performance in the position prediction task. The fixed weight model has a position MAE of 1.1136 km and an RMSE of 4.0714 km, while the adaptive weight model significantly reduces the MAE to 0.3542 km, a decrease of 68.2%, and the RMSE to 2.6180 km, a decrease of 35.7%. This breakthrough improvement stems from the synergistic mechanism of multitask feature fusion and loss adaptation in the model architecture: in the feature fusion stage, the position, speed, and heading features interact in multidimensional space; the adaptive weight layer dynamically adjusts the task loss contribution through the learned probability distribution so the position feature obtains a higher gradient priority in the back propagation. Experiments have shown that this mechanism can effectively strengthen the spatiotemporal correlation modeling of high-credibility input features in the dataset.
While improving the position accuracy, the adaptive weight module causes the performance of the speed and heading tasks to decline. The speed MAE increased from 1.1998 to 2.6786, and the RMSE increased from 9.0650 to 12.931. The heading MAE increased from 21.873° to 86.130°, and the RMSE increased from 44.969° to 97.654°. The results show that there is a coupling between the optimization of the position task and the speed and heading estimation. When the weight of the position task increases, the speed and heading weights decay relatively, making it difficult for their feature extraction layers to obtain sufficient gradient updates. At the same time, it reflects the model’s ability to select high-precision features. It can effectively select features with high accuracy and strong stability in multidimensional input features, thereby enhancing the utilization of effective data.
Experimental data further demonstrates that, even in scenarios where input feature validity is unknown, this module can dynamically filter key features, achieving global interference rejection by actively sacrificing secondary features. In complex multi-radar collaborative detection scenarios, the reliability and noise level of sensor data vary dynamically, leading to inherent optimization conflicts between position, velocity, and heading estimation tasks. Traditional fusion methods assign fixed weights to various features and are unable to adapt to environmental variations such as electromagnetic interference and target maneuvers. While the PTFM model significantly improves position accuracy, it also increases velocity and heading errors. The root cause lies in gradient competition in multitask optimization: when the model prioritizes high-confidence position features through the attention mechanism, gradient updates for velocity and heading tasks are relatively insufficient, automatically recognizing the signal-to-noise ratio advantage of position information.
In engineering practice, position data is typically more stable than velocity and heading data. In applications such as military reconnaissance, position accuracy often holds greater tactical value, and velocity and heading data can be directly calculated from position data. Weight decay for velocity and heading tasks essentially cuts off noise propagation paths, preventing low-quality features from contaminating the core position prediction through shared layers. The core breakthrough of this module lies in the absence of a predefined feature validity prior. In fixed-weight models, sensor confidence weights must be preassigned, but this static assignment struggles to cope with dynamic interference. The adaptive model addresses this dilemma through end-to-end training. When the training and validation datasets have advantages in other feature dimensions, the model automatically prioritizes optimization based on the input data’s signal-to-noise ratio, selecting the most advantageous features for fusion judgment while suppressing the propagation of low-quality features.

3.4. Parallel Architecture Comparison Experiment

To verify the superiority of PTFM’s parallel architecture over its sequential counterpart, we conducted comparative experiments using a rigorous controlled experimental methodology, constructing four different deep learning fusion architectures for performance comparison. These experiments compared the PTFM’s parallel architecture, a sequential LSTM–attention architecture, a sequential attention–LSTM architecture, and an ungated hybrid parallel architecture. This experimental design ensured that performance differences were solely due to the architecture itself rather than external factors. The detailed results of the comparative experiments are presented in Table 3 and Figure 6. The subfigures a, b and c represent the experimental results of RMSE, MAE and Variance respectively.
To address the dimensionality mismatch inherent in the sequential architecture, a dimensional adapter layer was introduced. This layer, through a parameter matrix and mathematical transformations, enables bidirectional conversion between 144-dimensional and 64-dimensional models. The adapter layer was activated only in the corresponding architecture, minimizing the impact of parameter differences on experimental results.
The experimental results showed that the different fusion architectures exhibited significant performance differences in the track prediction task. Based on error metrics, the attention-first sequential architecture achieved the best RMSE (RMSE) of 2.5926 km and a variance of 6.5211, demonstrating its superiority in overall prediction accuracy and stability. Notably, while the parallel fusion without gating architecture achieved the lowest MAE of 0.3522 km, its variance of 13.6332 was significantly higher than the other architectures, indicating significant prediction volatility. This demonstrates that PTFM effectively controls extreme errors while maintaining high prediction accuracy. The PTFM architecture achieved optimal performance, with an RMSE of 2.6180 km and a MAE of 0.3542 km, while maintaining a low variance of 6.7284. This demonstrates that its parallel fusion mechanism and gating strategy effectively coordinate the spatiotemporal feature extraction process, maintaining both prediction accuracy and output stability. The core advantage of the gating mechanism lies in its ability to dynamically adjust the contributions of different feature streams, enabling adaptive optimization of feature selection through learnable weight parameters. This design avoids the inconsistency associated with simple feature concatenation in parallel fusion without gating architectures and overcomes the error accumulation inherent in serial architectures. This performance advantage may stem from the model’s decoupling of sensor spatiotemporal characteristics and the optimized integration of heterogeneous data from multiple sources through the adaptive weighting mechanism.

3.5. Different Number of Track Points

This section discusses the performance under different numbers of track points for a systematic analysis. The experimental results reveal the sensitivity of the model to track length and the efficiency boundary of key components, as shown in Table 4. The model performance changes with the data of the track points, as shown in Figure 7.
When the number of track points is between 10 and 25, the model shows relatively stable fusion accuracy. The RMSE fluctuates between 1.8014 and 2.1320, and the MAE is stable in the range of 0.2910 to 0.3200, indicating that the model has a strong ability to extract spatiotemporal features of short sequences. This advantage is mainly attributed to the synergy of the dual-branch parallel architecture. The sensor’s feature enhancement layer extracts sensor-specific features separately through the fully connected layer, while the LSTM time series modeling layer effectively captures local time series dependencies. At the same time, the attention mechanism branch optimizes the feature fusion process through dynamic weight allocation, and its output and the splicing of LSTM features significantly improve the ability to represent key information of short sequences.
When the number of track points increases to 30, there is a significant performance degradation: RMSE jumps to 3.6236, an increase of 94.4% compared with 25 points; MAE reaches 0.5633, an increase of 76.0%; and the variance surges to 12.813. This phenomenon exposes the limitations of the model in dealing with medium time series dependencies. Although the LSTM layer is designed to retain complete sequence information, the fixed-length time series modeling capability may face the risk of gradient diffusion under long sequences. It is worth noting that the time embedding layer uses static sine and cosine encoding and lacks the ability to adaptively adjust the sequence length, which may lead to insufficient representation of time information in medium sequences. In addition, the feature interaction of the gated fusion mechanism in high-dimensional space may introduce noise accumulation effects.
When the number of track points continues to increase to 40 or 50, the model shows a certain degree of self-regulation ability. RMSE drops to 2.6180, MAE drops to 0.3542, and variance converges to 6.7284. The local recovery of performance benefits from the synergy of the adaptive loss weight mechanism and the residual connection structure. The former alleviates the conflict problem of multi-objective optimization by dynamically balancing the loss weights of various input features; the latter suppresses the degradation of deep networks through cross-layer connections. However, the variance index is still significantly higher than that of short sequences, indicating that uncertainty propagation under long sequences still has influence.

3.6. Ablation Experiment

In order to verify the effectiveness of each module, an ablation experiment is conducted. The ablation experiment results systematically reveal the functional contribution of each module of the model. The complete model performs best in terms of RMSE, MAE, and variance, verifying the effectiveness of the overall architecture design. The comparison of different ablation experiment results is shown in Table 5, where w/o denotes the ablation of specified component from the complete PTFM architecture. The model performance of different ablation experiments is shown in Figure 8. The subfigures a, b and c represent the experimental results of RMSE, MAE and Variance respectively.
When the LSTM module is removed, the RMSE increases to 3.0469, an increase of 16.4%; the MAE increases to 0.6288, an increase of 77.6%; and the variance expands to 8.8882. This phenomenon stems from the key modeling ability of the LSTM layer for time series dependence. Its orthogonally initialized 64-unit network retains the complete sequence information through configuration. The model after ablation loses the ability to characterize the continuous state of the track, resulting in a significant degradation in the performance of time dimension feature extraction.
The removal of the attention module caused the most severe performance degradation, with RMSE rising to 5.2418 and variance increasing to 27.162, both of which were significantly higher than the base model. This result confirms the dual value of the attention branch, the dynamic weights generated by its 128-dimensional attention layer not only achieve differentiated fusion of sensor features, but, more importantly, fix the dimensional alignment problem after splicing the temporal embedding and sensor features. After ablation, the model cannot adaptively focus on key features, resulting in a significant increase in noise sensitivity, which is particularly evident in the abnormal increase in the variance index.
The ablation of residual connections caused a decrease in model stability, with the variance index rising to 24.016. Its cross-module connection effectively alleviates the vanishing gradient, and the residual path allows deep features to be directly transmitted back to the operation node, ensuring the efficiency of backpropagation. The lack of this mechanism makes it difficult for the optimization process to converge, and the RMSE deteriorates to 4.9235.
The influence of the temporal embedding module is reflected in the temporal alignment dimension. After removal, the RMSE increased to 3.8576, an increase of 47.3%, verifying the role of sine and cosine position encoding. Although zero-delay embedding is used in the experiment, its fixed encoding mode still provides sequence position priors and enhances the perception of the temporal dimension. The ablation of the sensor enhancement layer increases the MAE to 0.4069, reflecting the necessity of the dual-path Dense layer to upgrade the feature dimension of the original sensor data. Through dual-sensor branch processing, the model improves the representation compatibility of heterogeneous sensor data.
The above conclusions show that LSTM and temporal attention constitute the core feature extraction framework, the residual mechanism ensures the optimization stability, and temporal embedding and sensor enhancement, respectively, enhance the position sensitivity and data compatibility. The modules work together to achieve a balance between accuracy and robustness in the track fusion task.

3.7. Computational Efficiency and Real-Time Experiments

To comprehensively evaluate the computational efficiency and real-time performance of the PTFM framework, we conducted a series of benchmark tests. These experiments compared PTFM with traditional Kalman filtering and LSTM models across multiple performance metrics, providing a deep understanding of the PTFM framework’s computational characteristics. The experimental platform is based on a laptop computer with Windows system, 24 physical CPU cores, 32 logical CPU cores, 31.63 GB total memory, and NVIDIA GeForce RTX 4070 Laptop GPU. The indicators compared in the experiment are cold start time, average delay, maximum delay, CPU usage, and memory usage. The performance analysis results, shown in Table 6, demonstrate competitive overall performance in terms of computational efficiency and real-time performance.
PTFM achieved the best cold start performance, with a cold start time of 35.34 ms, the shortest among the three models. This demonstrates PTFM’s high efficiency in model initialization and initial inference preparation, which is crucial for real-time applications requiring fast response times. However, in terms of average latency during the continuous inference phase, PTFM’s 44.35 ms was slightly higher than the Kalman filter’s 31.68 ms, but better than the LSTM’s 39.73 ms. This result reflects PTFM’s balanced approach between computational complexity and accuracy. However, from the perspective of latency stability, PTFM has some shortcomings, and its maximum latency needs further improvement. PTFM’s peak latency of 117.66 ms is significantly higher than the other two methods, suggesting potential performance fluctuations under extreme conditions.
In terms of resource usage, in the experimental configuration of 24 cores and 32 threads, the PTFM framework’s CPU utilization of 158.38% demonstrates relatively low CPU consumption and computational pressure. This indicates that the PTFM framework only utilized the computing resources of approximately 1.58 logical cores during operation, or only 4.95% of the total CPU capacity of the system’s 32 logical cores. This reflects the computational efficiency of the PTFM framework’s algorithmic design and demonstrates that it can complete track fusion tasks with relatively few computational resources. Compared to the Kalman and LSTM methods, its CPU utilization, while relatively high, is still adequate. Memory usage shows that all three methods have peak memory usage of around 600 MB, with no significant difference, indicating that memory requirements primarily stem from the underlying framework and data processing pipeline rather than the model itself.
Combining these results, the PTFM framework achieves reasonable computational efficiency through high CPU utilization while maintaining a relatively reasonable memory footprint. Its advantage in cold start makes it suitable for application scenarios that require fast initialization, while its lack of latency stability suggests that further optimization of the computational graph structure or the introduction of a dynamic scheduling mechanism may be necessary. Compared with traditional Kalman filtering and LSTM, PTFM has a certain gap in average latency, but this gap may be exchanged for performance advantages when dealing with complex nonlinear problems.

4. Conclusions

This paper proposes a radar track fusion method based on parallel fusion model of deep learning. Through model-driven residual extraction, the attention mechanism and LSTM parallel processing simultaneously capture key spatiotemporal features and long-term dependencies, solving the single-dimensional defects of the LSTM model and the local information loss of the Transformer. Through the coordinated optimization of various modules, high-precision track fusion in complex scenarios is achieved. Experimental results show that this method is significantly better than traditional algorithms in fusion accuracy and motion pattern generalization ability, providing a new solution for radar track fusion technology. Future research will further explore the application of graph neural networks in multi-sensor spatiotemporal correlation modeling to improve the real time and robustness of the system.

Author Contributions

J.Q.: Conceptualization, methodology, software, formal analysis, writing—original draft. X.L.: Conceptualization, formal analysis, validation, writing—review and editing. J.S.: Conceptualization, resources, writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, Grant No. 62131001 and 62171029.

Data Availability Statement

The MTAD used in this paper is publicly available at https://www.scidb.cn/en/detail?dataSetId=c7d8dc56fe854ec2b084d075feb887fd (accessed on 31 December 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
IMMInteracting multiple model
IFFIdentification of friend or foe
CCFConvex combination fusion
CIFCovariance intersection fusion
UKFUnscented Kalman filter
NCTNearly coordinated turn
NCANearly constant acceleration
CAWDFComprehensive adaptive weighted data fusion
OTHROver-the-horizon radar
AFTQMMAsynchronous fusion based on track quality with multiple model
PTFMParallel track fusion model
MTADMultisource track association dataset
LSTMLong short-term memory network
RMSERoot mean square error
MAEMean absolute error

References

  1. Chen, S.; Dou, H.; Liu, Z.; Mao, Z. Multi-sensor track fusion algorithm research and simulation analysis. In Proceedings of the 2022 Global Conference on Robotics, Artificial Intelligence and Information Technology (GCRAIT), Chicago, IL, USA, 15–17 July 2022; pp. 131–135. [Google Scholar] [CrossRef]
  2. Kumar, A.; Muthukumar, A.; Rajesh, R. Centralized Multi Sensor Data Fusion Scheme for Airborne Radar and IFF. In Proceedings of the 2024 IEEE Space, Aerospace and Defence Conference (SPACE), Bangalore, India, 6–8 March 2024; pp. 24–27. [Google Scholar] [CrossRef]
  3. Bu, S.; Zhou, G. Sequential Spatiotemporal Bias Compensation and Data Fusion for Maneuvering Target Tracking. IEEE Trans. Aerosp. Electron. Syst. 2023, 59, 241–257. [Google Scholar] [CrossRef]
  4. Zhang, D.; Duan, Z. Recursive LMMSE Sequential Fusion with Multi-Radar Measurements for Target Tracking. In Proceedings of the 2021 IEEE 24th International Conference on Information Fusion (FUSION), Sun City, South Africa, 1–4 November 2021; pp. 1–8. [Google Scholar] [CrossRef]
  5. Lionel, G.; Frédéric, D. Association of labelled tracks with low reliability covariance information: A track graph partitioning approach. In Proceedings of the 2023 26th International Conference on Information Fusion (FUSION), Charleston, SC, USA, 27–30 June 2023; pp. 1–8. [Google Scholar] [CrossRef]
  6. Dunham, D.T.; Ogle, T.L.; Miceli, P.A. Measurement and Track Fusion at the System Level. In Proceedings of the 2023 26th International Conference on Information Fusion (FUSION), Charleston, SC, USA, 27–30 June 2023; pp. 1–5. [Google Scholar] [CrossRef]
  7. Xia, K.-Q.; Xing, M.; Du, C.-Y.; Fu, N. Multi-mode Composite Track Estimation Method Based on Distributed Measurement Fusion. In Proceedings of the 2021 IEEE 4th Advanced Information Management, Communicates, Electronic and Automation Control Conference (IMCEC), Chongqing, China, 18–20 June 2021; pp. 1418–1421. [Google Scholar] [CrossRef]
  8. Li, B.; Li, J.; Zhu, B.; Zhang, X. Multiple Unknown Emitters Direct Tracking With Distributed Sensor Arrays: Non-Homogeneous Data Fusion and Fast Position Update. IEEE Sens. J. 2022, 22, 10965–10973. [Google Scholar] [CrossRef]
  9. Qiao, S.; Fan, Y.; Zhang, H. An Improved Multi-Radar Track Weighted Data Fusion Algorithm. In Proceedings of the 2023 42nd Chinese Control Conference (CCC), Tianjin, China, 24–26 July 2023; pp. 3152–3156. [Google Scholar] [CrossRef]
  10. Zhang, K.; Wang, Z.; Guo, L.; Peng, Y.; Zheng, Z. An Asynchronous Data Fusion Algorithm for Target Detection Based on Multi-Sensor Networks. IEEE Access 2020, 8, 59511–59523. [Google Scholar] [CrossRef]
  11. Xu, J.H.; Liu, Y.; Che, J.; Zhang, L.J.; Wang, J.; Zhang, Y. Multi-target track fusion algorithm based on CNN. Mode. Radar 2019, 41, 45–48. [Google Scholar]
  12. Yun, T.; Pan, Q.; Yang, J.L. Track Fusion Algorithm Based on Fully Convolutional Network with Multi-order Difference Loss. Mod. Radar 2024, 46, 2. [Google Scholar] [CrossRef]
  13. Li, Y.G.; Wang, H.M.; Yang, H.M. Multi-sensor Online Track Fusion Algorithm Based on Multi-output. Mod. Radar 2024, 46, 8. [Google Scholar] [CrossRef]
  14. Chen, A.; Chen, W.; Liu, R.; Liu, Q.; Sun, B.; Yang, Y. Track fusion algorithm for OTHR network based on deep learning. In Proceedings of the 2023 2nd International Conference on Artificial Intelligence and Computer Information Technology (AICIT), Yichang, China, 22–24 September 2023; pp. 1–4. [Google Scholar] [CrossRef]
  15. He, S.; Shin, H.-S.; Tsourdos, A. Distributed Joint Probabilistic Data Association Filter With Hybrid Fusion Strategy. IEEE Trans. Instrum. Meas. 2020, 69, 286–300. [Google Scholar] [CrossRef]
  16. Zhang, Y.; Ran, J.H. Dynamic weighted track fusion algorithm based on track comparability degree. In Proceedings of the 2010 IEEE International Conference on Information Theory and Information Security, Beijing, China, 17–19 December 2010; pp. 710–713. [Google Scholar] [CrossRef]
  17. Wu, J.; Xie, X.X.; Ai, X.F. Parallel optimization of track fusion algorithm based on OpenMP. J. Terahertz Sci. Electron. Inf. 2024, 22, 1021–1028. [Google Scholar] [CrossRef]
  18. Gao, X.Z. Analysis of Detection Performance and Fusion Algorithm for Distributed Surface Wave OTHR. Ph.D. Thesis, Harbin Institute of Technology, Harbin, China, 2015. [Google Scholar] [CrossRef]
  19. Qi, J.D.; Lu, X.K.; Sun, J.P. An Intelligent Track Segment Association Method Based on Characteristic-Aware Attention LSTM Network. Sensors 2025, 25, 3465. [Google Scholar] [CrossRef]
  20. Cui, Y.Q.; Xu, P.L.; Gong, C.; Yu, Z.C.; Zhang, J.T.; Yu, H.B.; Dong, K. Multisource Track Association Dataset Based on the Global AIS. J. Electron. Inf. Technol. 2023, 45, 746–756. [Google Scholar] [CrossRef]
Figure 1. The overall model framework.
Figure 1. The overall model framework.
Electronics 14 03461 g001
Figure 2. The LSTM time series model framework.
Figure 2. The LSTM time series model framework.
Electronics 14 03461 g002
Figure 3. Example diagram of sensor tracks and actual tracks.
Figure 3. Example diagram of sensor tracks and actual tracks.
Electronics 14 03461 g003
Figure 4. Position mean absolute error and model training loss during training.
Figure 4. Position mean absolute error and model training loss during training.
Electronics 14 03461 g004
Figure 5. Tracks and error plots of predicted results.
Figure 5. Tracks and error plots of predicted results.
Electronics 14 03461 g005
Figure 6. Performance comparison of different architectures.
Figure 6. Performance comparison of different architectures.
Electronics 14 03461 g006
Figure 7. Performance comparison of different track points.
Figure 7. Performance comparison of different track points.
Electronics 14 03461 g007
Figure 8. The model performance of different ablation experiments.
Figure 8. The model performance of different ablation experiments.
Electronics 14 03461 g008
Table 1. Performance comparison of different methods.
Table 1. Performance comparison of different methods.
PTFMWAFKFCCFCIFLSTMTransformerDense
RMSE (km)2.618010.57610.66710.59710.5614.68114.57977.6242
MAE (km)0.35421.42441.66721.57861.29182.49533.10702.2228
Variance (km2)6.7284109.83111.01109.81109.8615.68711.32053.187
Table 2. Comparison of different feature output performance.
Table 2. Comparison of different feature output performance.
ModelPosition (km)Velocity (m/s)Course (°)
MAEBase1.11361.199821.873
Adaptive0.35422.678686.130
RMSEBase4.07149.065044.969
Adaptive2.618012.93197.654
Table 3. Performance comparison of different architectures.
Table 3. Performance comparison of different architectures.
PTFMSerial LSTM FirstSerial Attention FirstHybrid
RMSE (km)2.61802.63542.59263.7091
MAE (km)0.35420.43330.44760.3522
Variance (km2)6.72846.75776.521113.633
Table 4. Performance comparison of different track points.
Table 4. Performance comparison of different track points.
10152025304050
RMSE (km)2.13201.80141.99991.86363.62362.79972.6180
MAE (km)0.33130.29100.31520.32000.56330.44530.3542
Variance (km2)4.43573.16043.90043.370612.8137.64006.7284
Table 5. Comparison of different ablation experiment results.
Table 5. Comparison of different ablation experiment results.
PTFMw/o LSTMw/o Attentionw/o Residualw/o Sensor Enhancew/o Time Embedding
RMSE (km)2.61803.04695.24184.92353.16983.8576
MAE (km)0.35420.62880.56070.47390.40690.3738
Variance (km2)6.72848.888227.16224.0169.882414.741
Table 6. Comparison of computational efficiency and real-time performance of different methods.
Table 6. Comparison of computational efficiency and real-time performance of different methods.
Cold Start (ms)Avg Latency (ms)Max Latency (ms)CPU Usage (%)Mem Usage (MB)
PTFM35.3444.35117.66158.38601.89
Kalman42.4231.6842.91105.76602.28
LSTM41.9339.7343.82127.90603.61
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qi, J.; Lu, X.; Sun, J. Multi-Radar Track Fusion Method Based on Parallel Track Fusion Model. Electronics 2025, 14, 3461. https://doi.org/10.3390/electronics14173461

AMA Style

Qi J, Lu X, Sun J. Multi-Radar Track Fusion Method Based on Parallel Track Fusion Model. Electronics. 2025; 14(17):3461. https://doi.org/10.3390/electronics14173461

Chicago/Turabian Style

Qi, Jiadi, Xiaoke Lu, and Jinping Sun. 2025. "Multi-Radar Track Fusion Method Based on Parallel Track Fusion Model" Electronics 14, no. 17: 3461. https://doi.org/10.3390/electronics14173461

APA Style

Qi, J., Lu, X., & Sun, J. (2025). Multi-Radar Track Fusion Method Based on Parallel Track Fusion Model. Electronics, 14(17), 3461. https://doi.org/10.3390/electronics14173461

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop