Next Article in Journal
Recognition of Landslide Triggering Mechanisms and Dynamics Using GNSS, UAV Photogrammetry and In Situ Monitoring Data
Previous Article in Journal
A Sensor Bias Correction Method for Reducing the Uncertainty in the Spatiotemporal Fusion of Remote Sensing Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep-Learning-Based Multiple Model Tracking Method for Targets with Complex Maneuvering Motion

1
Radar Research Lab, School of Information and Electronics, Beijing Institute of Technology, Beijing 100081, China
2
Advanced Technology Research Institute, Beijing Institute of Technology, Jinan 250300, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(14), 3276; https://doi.org/10.3390/rs14143276
Submission received: 16 May 2022 / Revised: 25 June 2022 / Accepted: 5 July 2022 / Published: 7 July 2022

Abstract

:
The effective detection of unmanned aerial vehicle (UAV) targets is of great significance to guarantee national military security and social stability. In recent years, with the development of communication and control technology, the movement of UAVs has become increasingly flexible and complex, presenting diverse trajectory forms and different motion models in different phases. The Gaussian mixture probability hypothesis density filter incorporating the linear Gaussian jump Markov system approach (LGJMS-GMPHD) provides an efficient method for tracking multiple maneuvering targets, as applied to the switching of motions between a set of models in a Markovian chain. However, in practice, the motion model parameters of targets are generally unknown and the model switching is uncertain. When the preset filtering model parameters are mismatched, the tracking performance is dramatically degraded. In this paper, within the framework of the LGJMS-GMPHD filter, a deep-learning-based multiple model tracking method is proposed. First, an adaptive turn rate estimation network is designed to solve the filtering model mismatch caused by unknown turn rate parameters in coordinate turn models. Second, a filter state modification network is designed to solve the large tracking errors in the maneuvering phase caused by uncertain motion model switching. Finally, based on simulations of multiple maneuvering targets in cluttered environments and experimental field data verification, it can be concluded that the proposed method has strong adaptability to multiple maneuvering forms and can effectively improve the tracking performance of targets with complex maneuvering motion.

Graphical Abstract

1. Introduction

Unmanned aerial vehicles (UAVs), with their advantages of flexible usage and easy deployment, have been widely used in many fields, including military applications such as battlefield surveillance, air reconnaissance, and tracking, as well as in civilian applications such as disaster monitoring, geophysical exploration, and agricultural plant protection [1]. At the same time, their malicious use also poses a severe threat to aviation security and the maintenance of social stability [2]. Radar has proven to be an essential technical method of UAV detection [3,4]. However, the strong maneuverability of UAVs, including their unknown motion model parameters and uncertain model switching [5], brings about great challenges in the stable tracking of targets.
Random finite set (RFS)-based filters provide a theoretically optimal approach to multi-target tracking and are widely used in many fields [6,7,8]. In the RFS formulation, the collection of target states at any given time is regarded as a set-valued multi-target state, and the collection of measurements is regarded as a set-valued multi-target observation [9,10]. Then, the problem of dynamically estimating multiple targets can be formulated in a Bayesian filtering framework by propagating the posterior distribution of the multi-target state in time. However, multi-target filtering solutions in their standard form are generally implemented by a fixed motion model. Thus, to characterize the a rapidly maneuvering target with multiple models, the jump Markov system (JMS) approach has been incorporated into the existing RFS-based filters. Mahler derived the probability hypothesis density (PHD) filter and cardinalized probability hypothesis density (CPHD) filter approximations of the jump Markov multitarget Bayes filter [11]. On this basis, a closed-form solution to the PHD recursion for the linear Gaussian jump Markov multi-target model was derived, and this is referred as the LGJMS-GMPHD filter [12]. In addition, the JMS approach has also been incorporated into CPHD [13], multi-Bernoulli [14], and labeled multi-Bernoulli filters [7]. However, in the above multiple model filters, the constant velocity (CV) and coordinate turn (CT) models are generally combined as model sets to describe the maneuvering motion. When these are applied to practical UAV tracking scenarios, there are two key problems: (i) the turn rate of the target is generally unknown, resulting in a mismatch between the target’s motion and the CT filtering model; (ii) for complex maneuvering trajectories, the uncertainty of motion model switching leads to the degradation of tracking performance.
There are two common approaches to estimating the turn rate. The first class estimates the turn rate online by using the estimated acceleration magnitude over the estimated speed [15]. However, the estimated acceleration is generally not accurate enough, leading to inaccurate turn rate estimates. For the second class, the turn rate parameter is augmented into the state vector and the turn rate is estimated as part of the state vector recursively, which creates a difficult nonlinear problem [16]. Furthermore, target Doppler measurements usually have higher precision compared with position estimates, and thus they have been incorporated to improve the accuracy of the turn rate estimate. In [17], four possible turn rates were estimated based on the Doppler measurement. Due to the lack of prior information on target motion, the minimum turn rate and its opposite value were chosen to be the possible turn rates, and an interacting multiple model (IMM) algorithm consisting of one CV model and two CT models was adopted, increasing the computational burden.
The uncertainty of motion model switching is also a challenging problem in multiple model (MM) filters. In this case, the filtering state estimate has a delay in its response to target maneuvers, so the tracking precision degrades rapidly in the maneuvering phase. A smoothing process can produce delayed estimates, i.e., obtaining the target state estimates at time k given the measurements up to time k + d ( d > 0 ) , which contributes to improving the tracking performance of maneuvering targets and has been incorporated into the MM algorithms [18,19,20]. A sequential Monte Carlo (SMC) implementation for MMPHD smoothing was derived in [21] to improve the capability of PHD-based tracking algorithms. However, the backward smoothing procedure is still model-based and cannot extract temporal correlations, so longer lags do not yield better estimates of the current state. Moreover, closed-form recursion with a smoothing process needs to be derived for different filters respectively and the computation is generally intractable, leading to limitations in its applications. Deep neural networks have strong capability of fitting if there are sufficient training data [22], which is conducive to solving the problems of model mismatch and estimation delays in existing MM filtering algorithms for targets with complex maneuvering motions. In [23], a deep learning maneuvering target tracking algorithm was proposed. However, the input of the network was the filtering states estimated by a single-model unscented Kalman filter. For strong maneuvering trajectories, large tracking errors increase the training burden of the network and reduce the convergence rate of training losses. Additionally, an LSTM-based deep recurrent neural network was presented in [24] to overcome the limits of traditional model-based methods, but only CV and constant acceleration (CA) motions were verified, without considering CT maneuvers. Furthermore, the above methods are not suitable for scenarios with multiple maneuvering targets.
In view of the above problems, within the framework of the LGJMS-GMPHD filter, we propose a deep-learning-based multiple model tracking method for multiple maneuvering targets with complex maneuvering motions. The main contributions of this study are summarized as follows:
  • An adaptive turn rate estimation network (ATN) is designed. The feature matrix, including multi-frame and multi-dimensional kinematic states, is constructed and used as the input of the network. The temporal correlation information at each previous time step and the weight of different variables are extracted to improve the estimation accuracy of the turn rate. Then, the parameter is fed back to the CT model to enhance the consistency of the target motion and filtering model.
  • A filter state modification network (FMN) is designed to smooth the state estimates of the filter. The relationship information between all time steps of the state vector, including position estimates, is extracted to improve the adaptability of the filter to complex maneuvering motions, thereby reducing the tracking errors caused by uncertain model switching.
The remainder of this paper is structured as follows. Section 2 reviews the PHD recursion for the linear Gaussian jump Markov multi-target model. Section 3 presents the principle of the proposed deep-learning-based multiple model tracking method. Section 4 designs the network parameters and verifies the effectiveness of the method in improving tracking performance. The simulation results and experimental data verification are analyzed in Section 5. The discussion and future work are presented in Section 6, and our conclusions are presented in Section 7.

2. Principle of LGJMS-GMPHD

The PHD filter [25] recursively propagates the first order moment or the intensity function associated with the multi-target posterior density, which has the advantage of low computational complexity and has been widely used. Furthermore, a closed-form solution to the PHD recursion for LGJMS derived in [12] provides an efficient method for tracking multiple maneuvering targets. In this section, a short review of RFS and the PHD recursion for the linear Gaussian jump Markov multi-target model is presented.

2.1. Random Finite Sets in Multi-Target Tracking

(1)
Dynamic model
For a given multi-target state X k 1 at time k 1 , each x k 1 X k 1 either continues to exist at time k with probability p S , k ( x k 1 ) , or dies with probability 1 p S , k ( x k 1 ) . Consequently, at the next time step, the behavior of the given state x k 1 X k 1 is modelled as the RFS
S k | k 1 ( x k 1 )
which can take on either { x k } when the target survives, or when the target dies. For the sake of simplicity, spawned targets are not considered here. The multi-target state X k at time k is given by the union of the surviving targets and the spontaneous births:
X k = { x k 1 X k 1 S k | k 1 ( x k 1 ) } Γ k ,
where Γ k is the RFS of spontaneous birth at time k . Each target state x k = [ x k x . k y k y . k ] T denotes the position and velocity of the x -axis and y -axis.
For the linear Gaussian multi-target model [12], the multi-target transition density is represented as:
f k | k 1 ( x k | x k 1 ) = N ( x k ; F k 1 x k 1 , Q k 1 ) ,
where N ( · ; m , P ) denotes a Gaussian density with mean m and covariance P ; F k 1 is the state transition matrix; Q k 1 is the process noise covariance.
In general, motion along a fixed heading at constant speed can be described by a CV model, and a level turn can be described by a CT model. At a turn rate of 0 ° s 1 , the CT model reduces to the CV model. Therefore, the transition matrix can be uniformly expressed as:
F k 1 = [ 1 sin ω T ω 0 1 cos ω T ω 0 cos ω T 0 sin ω T 0 1 cos ω T ω 1 sin ω T ω 0 sin ω T 0 cos ω T ] ,
where T is the sample interval and ω is the turn rate, which is defined as the rate of change of the (velocity) heading angle in the horizontal plane and estimated as the magnitude of the acceleration divided by the speed of the target [15]. In this paper, we refer to this method as the physical definition method (PDM).
(2)
Measurement model
The RFS measurement model, which accounts for detection uncertainty and clutter, is described as follows. A given target x k X k is either detected with probability p D , k ( x k ) or missed with probability 1 p D , k ( x k ) . At time k , each state x k X k generates an RFS
Θ k ( x k )
that can take on either { z k } when the target is detected, or when the target is not detected. Given a multi-target state X k at time k , the multi-target measurement Z k received at the sensor is formed by the union of target generated measurements and clutter, i.e.,
Z k = K k [ x X k Θ k ( x ) ] .
For the linear Gaussian multi-target model, the multitarget likelihood is represented as:
g k ( z k | x k ) = N ( z k ; H k x k , R k ) ,
where H k is the observation matrix and R k is the measurement noise covariance.
Target Doppler measurements may provide additional information about the target’s kinematic state and therefore can be incorporated into the tracker to enhance the tracking performance [26]. After the Doppler is introduced, each measurement at time k is denoted as z k = [ x k m y k m r . k m ] , where x k m and y k m are the positions of x -axis and y -axis, and r . k m is the Doppler. The modified measurement equation is expressed as follows:
z k = h k ( x k ) + n k = [ x k m y k m r . k m ] = [ x k y k r . k ] + [ n k x n k y n k d ] ,
r . k = h d , k ( x k ) = x k x . k + y k y . k x k 2 + y k 2 ,  
where n k is Gaussian measurement noise with mean zero and covariance R k = d i a g ( σ x 2 , σ y 2 , σ d 2 ) , σ x 2 and σ y 2 are the position noise variance, and σ d 2 is the Doppler noise variance. Obviously, the Doppler measurement function h d , k ( · ) is nonlinear.
In this paper, to simplify the operation, we take the Taylor series expansion of (9) at x k to linearize the measurement equation. Then, the observation matrix is modified as:
H k = [ 1   0   0   0 0   1   0   0     r . k x k     r . k y k     r . k x . k     r . k y . k ] ,  
where
r . k x k = ( x . k y k 2 x k y k · y . k ) / ( x k 2 + y k 2 ) 3 / 2 r . k y k = ( y . k x k 2 x k y k · x . k ) / ( x k 2 + y k 2 ) 3 / 2 r . k x . k = x k / ( x k 2 + y k 2 ) 1 / 2 r . k y . k = y k / ( x k 2 + y k 2 ) 1 / 2

2.2. PHD Recursion for LGJMS Multi-Target Model

JMS provides a natural means to model a maneuvering target for which the behavior cannot always be characterized by a single model. It is described by a set of parameterized state space models, the underlying parameters of which evolve with time according to a finite state Markov chain [12]. Suppose that μ k M is the label of the model in effect at time k , where M denotes the (discrete) set of all model labels. The models follow a discrete Markov chain with transition probability t k | k 1 ( μ k | μ k 1 ) . The state vector is augmented as x k = [ ξ k T , μ k ] T χ = n × M and the transition is governed by
f k | k 1 ( x k | x k 1 ) = f ˜ k | k 1 ( ξ k | ξ k 1 , μ k ) t k | k 1 ( μ k | μ k 1 ) .
The corresponding measurement likelihood is
g k ( z k | x k ) = g k ( z k | ξ k , μ k ) .
For a JMS with linear Gaussian models, the state transition density and measurement likelihood conditioned on mode μ k are given by
f ˜ k | k 1 ( ξ k | ξ k 1 , μ k ) = N ( ξ k ; F k 1 ( μ k ) ξ k 1 , Q k 1 ( μ k ) ) ,
g k ( z k | ξ k , μ k ) = N ( z k ; H k ( μ k ) ξ k , R k ( μ k ) ) ,
where F k 1 ( μ k ) and H k ( μ k ) denote the transition and the observation matrices of model μ k , and Q k 1 ( μ k ) and R k ( μ k ) denote covariance matrices of the process noise and measurement noise.
The purpose of maneuvering target tracking is to estimate the augmented state x k at time k from the sequence of measurement sets z 1 : k = ( z 1 , , z k ) . The closed-form PHD recursion for the LGJMS multi-target model [12] is presented as follows:
  • Prediction step
For an LGJMS multi-target model, if the posterior intensity v k 1 at time k 1 has the form
v k 1 ( ξ k 1 , μ k 1 ) = i = 1 J k 1 ( μ k 1 ) w k 1 ( i ) ( μ k 1 ) N ( ξ k 1 ; m k 1 ( i ) ( μ k 1 ) , P k 1 ( i ) ( μ k 1 ) ) ,
then the predicted intensity v k | k 1 is given by
v k | k 1 ( ξ k , μ k ) = v S , k | k 1 ( ξ k , μ k ) + v γ , k ( ξ k , μ k ) ,
where v γ , k ( ξ k , μ k ) is the intensity of birth targets, and the intensity of survival targets is represented as:
v S , k | k 1 ( ξ k , μ k ) = μ k 1 j = 1 J k 1 ( μ k 1 ) w S , k | k 1 ( j ) ( μ k , μ k 1 ) N ( ξ k ; m S , k | k 1 ( j ) ( μ k , μ k 1 ) , P S , k | k 1 ( j ) ( μ k , μ k 1 ) ) .
The weight, mean, and covariance of the survival Gaussian density are represented as:
w S , k | k 1 ( j ) ( μ k , μ k 1 ) = τ k | k 1 ( μ k | μ k 1 ) p S , k ( μ k 1 ) w k 1 ( j ) ( μ k 1 ) ,
m S , k | k 1 ( j ) ( μ k , μ k 1 ) = F S , k 1 ( μ k ) m k 1 ( j ) ( μ k 1 ) ,  
P S , k | k 1 ( j ) ( μ k , μ k 1 ) = F S , k 1 ( μ k ) P k 1 ( j ) ( μ k 1 ) F S , k 1 T ( μ k ) + Q S , k 1 ( μ k ) .
  • Update step
For an LGJMS multi-target model, if the predicted intensity v k | k 1 has the form
v k | k 1 ( ξ k , μ k ) = i = 1 J k | k 1 ( μ k ) w k | k 1 ( i ) ( μ k ) N ( ξ k ; m k | k 1 ( i ) ( μ k ) , P k | k 1 ( i ) ( μ k ) ) ,
then the posterior intensity v k at time k is given by
v k ( ξ k , μ k ) = ( 1 p D , k ( μ k ) ) v k | k 1 ( ξ k , μ k ) + z Z k v D , k ( ξ k , μ k ; z ) ,
where ( 1 p D , k ( μ k ) ) v k | k 1 ( ξ k , μ k ) is the mis-detection term, and v D , k ( · ; z ) refers to the detection terms for each measurement z Z k . The calculation expression is as follows:
v D , k ( ξ k , μ k ; z ) = j = 1 J k | k 1 ( μ k ) w k ( j ) ( μ k ; z ) N ( ξ k ; m k | k 1 ( j ) ( μ k ; z ) , P k | k ( j ) ( μ k ) ) ,
w k ( j ) ( μ k ; z ) = p D , k ( μ k ) w k | k 1 ( j ) ( μ k ) q k ( j ) ( μ k ; z ) κ k ( z ) + μ k p D , k ( μ k ) i = 1 J k | k 1 ( μ k ) w k | k 1 ( i ) ( μ k ) q k ( i ) ( μ k ; z ) ,
q k ( j ) ( μ k ; z ) = N ( z ; η k | k 1 ( j ) ( μ k ) , S k | k 1 ( j ) ( μ k ) ) ,  
m k | k 1 ( j ) ( μ k ; z ) = m k | k 1 ( j ) ( μ k ) + G k ( j ) ( μ k ) ( z H k ( μ k ) m k | k 1 ( j ) ( μ k ) ) ,
P k | k ( j ) ( μ k ) = ( I G k ( j ) ( μ k ) H k ( μ k ) ) P k | k 1 ( j ) ( μ k ) ,  
G k ( j ) ( μ k ) = P k | k 1 ( j ) ( μ k ) H k T ( μ k ) ( H k ( μ k ) P k | k 1 ( j ) ( μ k ) H k T ( μ k ) + R k ( μ k ) ) 1 .
Thus, the posterior intensity v k at time k is given by
v k ( ξ k , μ k ) = i = 1 J k ( μ k ) w k ( i ) ( μ k ) N ( ξ k ; m k ( i ) ( μ k ) , P k ( i ) ( μ k ) ) .
Given the above, the intensities v k | k 1 and v k are analytically propagated in time under the LGJMS multi-target model and the number of Gaussian components of the predicted and posterior intensity increases with time. Therefore, some simple pruning procedures need to be carried out. Finally, the estimate of the multi-target state is the set of N ^ k ordered pairs of means and modes ( m k ( i ) ( μ k ) , μ k ) with the largest weights w k ( i ) ( μ k ) , μ k M , i = 1 , , J k ( μ k ) , where N ^ k is the estimate of the number of targets and equals to i = 1 J k ( μ k ) w k ( i ) ( μ k ) rounded to the nearest integer.

3. Deep-Learning-Based Multiple Model Tracking Method

The LGJMS-GMPHD filter described in Section 2 can realize the tracking of multiple maneuvering targets in a cluttered environment. However, there are two key problems in practical applications in relation to targets with complex maneuvering characteristics:
(i)
The turn rate parameters of targets are generally unknown and changing, causing a mismatch between the target motion and the filtering model;
(ii)
The motion models are varied and switch with uncertainty, and the filter state estimate always lags the current target maneuver, causing the degradation of tracking precision in the maneuvering phase.
Therefore, a deep-learning-based multiple model tracking method is proposed to improve the adaptability of the LGJMS-GMPHD filter to complex maneuvering motions. The relevant flow diagram is shown in Figure 1. Firstly, the multi-target state estimate is obtained by the LGJMS-GMPHD filter. At the same time, track management [27] is performed to obtain the track labels of individual targets. Then, an adaptive turn rate estimation network is designed to realize the real-time estimation of the turn rate, and it is fed back to the filtering process to update the parameters of the CT model, thereby improving the matching degree of the filtering model. On this basis, a filter state modification network is designed to smooth the state estimates. Finally, a trajectory reconstruction is performed to implement the reconstruction of the track segments output by the network, and then the entire target track can be obtained.

3.1. Adaptive Turn Rate Estimation Network

In JMS, the determination of the turn rate parameter is the key point in the successful application of the CT model. However, in existing methods [15], the turn rate is estimated simply based on the filtering state estimates of the last frame. Additionally, the information dimension is less and the relationship between different variables is not fully utilized, so it is difficult to obtain an accurate estimate of the turn rate.
Deep learning provides an effective means to overcome the limitations of the traditional estimation method [28,29]. A proper network module design can mine more dimensional features, and a diverse set of trajectories can be constructed for network training, which contributes to obtaining more accurate parameter estimates for target trajectories with unknown maneuvering motions. Given all this, we designed an adaptive turn rate estimation network and integrated it into the LGJMS-GMPHD filtering process, thereby improving the tracking precision for CT maneuvering targets. The structure of the network is shown in Figure 2.
Based on the previous track information from time K w a + 1 to K obtained by the filtering process, the turn rate estimate at time K can be obtained by the adaptive turn rate estimation network, where w a is the length of the sequence. The length of the sliding window is s a . First, the feature matrix F v , including multi-dimensional kinematic information, can be obtained and used as the input of the network. It is represented as follows:
F v K w a + 1 : K { F v K w a + 1 , F v K w a + 2 , , F v K } .
The feature vector at time k is represented as follows:
F v k = [ x ^ k , y ^ k , x . ^ k , y . ^ k , r . k , x k m , y k m ] , k [ K w a + 1 , K ] ,
where x ^ k and y ^ k are the position estimates, and x . ^ k and y . ^ k are the velocity estimates. Meanwhile, based on the track labels, we can also obtain the measurements associated with the track, and then obtain the corresponding Doppler r . k m , and positions x k m and y k m .
The measurement precision of variables in the feature matrix F v is different. For example, the target Doppler usually has a higher precision than the position state estimates. Additionally, for each variable, the variation in the temporal dimension reflects the maneuvering characteristics of the target. Therefore, multi-perspective feature extraction is performed to mine more abundant features and improve the accuracy of the turn rate estimates. The temporal dimension feature of each variable is extracted by means of bidirectional long short-term memory (Bi-LSTM). Then, temporal pattern attention (TPA) [30] is introduced to determine the weights of different variables. The concrete implementation steps are as follows:
  • Bi-LSTM
The Bi-LSTM structure is used to capture the temporal correlation information of the feature matrix F v at each previous time step. The structure of Bi-LSTM is shown in Figure 3. The output vector h i k corresponding to the k -th time point of the i -th Bi-LSTM is the element-wise sum of the forward and backward LSTM outputs h i k and h i k at the k -th time point, and is calculated as follows:
h i k = [ h i k h i k ] .
In the proposed adaptive turn rate estimate network, two Bi-LSTM layers are integrated, and the output hidden state matrix is represented as H A = { h K w a + 1 a , h K w a + 2 a , , h K a } , where H A m × w a , m is the number of hidden units, and h k a is the hidden state at time step k .
  • Temporal pattern attention
For different features in F v , the temporal patterns across multiple time steps and the weights are extracted by the temporal pattern attention module. First, the f convolutional neural network (CNN) filters C i 1 × w a are applied on the row vector of H A to extract temporal pattern features. The convolutional operations yield H C m × f , where H i , j C represents the convolutional value of the i -th row vector and the j -th filter. Formally, this operation is expressed as
H i , j C = k = 1 w a H i , k × C j , k .
Then, an attention mechanism is carried out. v K is calculated as a weighted sum of the row vectors of H C to capture the relational information between different variables. Defined below is the scoring function: g : k × m to evaluate relevance:
g ( H i C , h K ) = ( H i C ) Τ W a h K ,
where H i C is the i -th row of H C , and W a f × m . The attention weight α i is obtained as
α i = s i g m o i d ( g ( H i C , h K ) ) .
The row vectors of H C are weighted by α i to obtain the context vector v K f
v K = i = 1 n α i H i C .
Finally, integrating v K and h K to yield the turn rate estimate:
h K = W h h K + W v v K ,
ω ˜ K = W h h K ,
where h K , h K m , W h m × m , W v m × f , W h n × m , and ω ˜ K n is the turn rate estimate.
In the training stage, the root mean square error (RMSE) is used as the loss function:
L = k = K w a + 1 K ( ω ˜ k ω k ) 2 ,
where ω k is the true turn rate. The model is trained by minimizing (40) and is optimized through the adaptive moment estimation (Adam) algorithm [31] over the training datasets.

3.2. Filter State Modification Network

For strong maneuvering targets, the motion switches between multiple models with uncertainty. The multiple model filtering state estimates have a delay in their response to target maneuvering, resulting in the degradation of the tracking performance. Therefore, a filter state modification network is designed here to smooth the state estimates and advance the tracking precision. The diagram is shown in Figure 4.
The state estimate of each target obtained by the LGJMS-GMPHD filter is first cropped into a state vector of uniform length w r , and the length of the sliding window is s r . It is used as the input of the network and is represented as follows:
x ^ K w r + 1 : K c { x ^ K w r + 1 c , x ^ K w r + 2 c , , x ^ K c } ,
where x ^ k c = [   x ^ k   y ^ k ] T , k [ K w r + 1 , K ] is the state vector including position estimates.
The variation of the state estimate vector along the temporal dimension can reflect the maneuvering characteristics of the target. Therefore, in the proposed filter state modification network, Bi-LSTM is also applied to extract the temporal correlation information in the forward and reverse directions. In addition, temporal attention (TA) [32] is integrated to weight the hidden state vectors at each time step.
The specific implementation of each module is described as follows.
  • Bi-LSTM
Two Bi-LSTM layers are integrated to capture the temporal correlation information of state vector x ^ K w r + 1 : K c at each previous time step. The output hidden state matrix is represented as H R = { h K w r + 1 R , h K w r + 2 R , , h K R } , where H R m × w r , m is the number of hidden units, and h k R is the hidden state at time step k .
  • Temporal attention
The hidden state matrix after the temporal attention module is represented as H t a = { h K w r + 1 t a , h K w r + 2 t a , , h K t a } , H t a m × w r , which is obtained by weighting H R at each time step. The weight at time k is represented as α k t a and is calculated as follows:
M k = tanh ( h k R ) ,
α k t a = exp ( ε k T M k ) j = 1 w r exp ( ε j T M j ) ,
h k t a = tanh ( α k t a h k R ) ,
where εk is a trained parameter vector and the dimension is m . Then H t a passes through the full connection layer to obtain the modified filter state output:
  x ~ K w r + 1 : K c = W f c H t a + b f c ,
where Wfc and b f c represent the weight and bias matrix of the full connection layer, respectively.
In the training stage, the loss function is:
L = k = K w r + 1 K (   x ~ k c x k c ) 2 ,
where   x ~ k c is the modified filter state estimate and x k c is the ground-truth of the trajectory. The model is trained by minimizing (46) and is optimized through the Adam algorithm over the training datasets.

3.3. Trajectory Reconstruction

In the proposed deep-learning-based multiple model tracking method, the historical track information of the target is clipped into segments of appropriate length and used as the input of the deep learning network. Therefore, to obtain the state estimate of the entire trajectory, the modified trajectory segments x ~ K w r + 1 : K c output from the filter state modification network need to be connected through a reconstruction step [23].
When the length of a trajectory segment is w r and the sliding window is sl, the overlap region of the adjacent segments is from K w r + s l + 1 to K . The value of the target states in the overlap region is the average of the two adjacent segments, and the state parameters remain constant for non-overlapping regions. The above process is repeated to complete the reconstruction of trajectory segments at each time, and the entire target trajectory can be obtained.

4. Network Parameter Design and Performance Analysis

To guarantee the flexibility of the designed deep learning network in real scenarios, a training dataset, including trajectories with different positions, speeds, turn rates, and various maneuvering modes, was constructed. In addition, process noise and measurement noise were added to simulate the target trajectories more realistically. Meanwhile, based on the estimation accuracy of the test dataset, the suitable network structure, network parameters, and input sequence length were designed. Finally, we verified the effectiveness of the two network modules in improving the tracking precision in relation to complex maneuver targets.
For the sake of convenience, in the comparisons of results presented below, LGJMS-GMPHD is abbreviated as GMPHD. After introducing adaptive turn rate estimation, it is denoted as GMPHD-ATN, and after introducing filtering state modification, it is denoted as GMPHD-ATN-FMN.

4.1. Dataset Preparation

4.1.1. Parameter Design

  • Target trajectory parameters
We designed five groups of maneuvering trajectories, as shown in Table 1, including different turn rate parameters. A turn rate equal to 0 ° / s represents the CV stage and a non- 0 ° / s rate represents the CT stage. Meanwhile, two groups of turn rate intervals were designed: ω 1 [ 8 ° / s , 8 ° / s ] represents a slow turn and ω 2 [ 15 ° / s , 8 ° / s ] [ 8 ° / s , 15 ° / s ] represents a fast turn. In addition, the durations of different motion stages are different. Specifically, CV is 20–25 s, ω 1 is 10–50 s, and ω 2 is 8–20 s. For each group of trajectories in the training datasets, the kinematic parameters were uniformly extracted according to Table 1.
  • Filter parameters
For the LGJMS-GMPHD filter process, μ = 1 is the CV model and μ = 2 is the CT model. The switching between the motion models is given by the Markov transition probability matrix as
[ τ k | k 1 ( μ k | μ k 1 ) ] = [ 0.9 0.1 0.1 0.9 ] .
For simplicity, we assume that the probability of target survival and detection is constant. p S , k | k 1 = 0.99 and p D , k = 0.98 are assumed for models μ = 1 , 2 . Additionally, the interval between the samples is T = 1   s . The standard deviation of process noise σ v = 1   m / s 2 and the standard deviation of measurement noise σ R = 10   m .

4.1.2. Dataset Construction

  • Training and validation datasets
The construction of the datasets for the two networks is summarized in the flowchart shown in Figure 5. The details are described as follows:
(a)
The kinematic parameters from Table 1 are randomly selected to generate real target trajectories.
(b)
Measurement noise is added and target measurements are obtained.
(c)
The LGJMS-GMPHD filter process is performed to obtain the state estimate, where the turn rate of CT model is calculated via a physical definition method to obtain the datasets of the adaptive turn rate estimate network, and the true value is used to obtain the datasets of the filter state modification network.
(d)
The trajectories are clipped to segments of uniform length. For the adaptive turn rate estimation network, the feature matrix F v defined in (31) is stored. For the filter state modification network, the position state estimate x ^ c defined in (41) is stored.
(e)
Data normalization is performed based on min-max scaling.
(f)
The order is disrupted and the training and validation datasets are divided.
In this paper, the number of trajectory segments in the training and validation datasets for the two networks was 300,000 and 60,000, respectively. In the training procedure, the learning rate was set to 10−4, the batch size was 100 samples, and the training epoch was 300.
  • Test datasets
Based on the trajectory parameters shown in Table 1, 100 maneuvering trajectories were constructed for each group to verify the effectiveness of each module in the proposed method in improving tracking performance.

4.2. Network Module Parameter Design

4.2.1. Adaptive Turn Rate Estimation Network

The network structures, parameters, and input sequence length were validated based on the five groups’ trajectories with different maneuvering forms in the test datasets. The evaluation criterion is the RMSE of turn rate estimate and defined as follows:
σ ω ( k ) = 1 N i = 1 N ( ω ^ ( k ) ω ( k ) ) 2
where ω ^ ( k ) is the output of the adaptive turn rate estimation network, and ω ( k ) is the true value. N is the number of simulations.
  • Network structure
The number of Bi-LSTM layers and the module effectiveness were verified based on the RMSE of the turn rate estimate, as shown in Table 2. Here, we set the number of hidden units of the Bi-LSTM to 64, the sequence length to 10, and the sliding window to 1. Compared to one Bi-LSTM layer, two Bi-LSTM layers can produce a higher accuracy, but the improvement is not significant as the number of layers continues to increase. Therefore, we set the number of Bi-LSTM layers as two. After the TPA module was introduced, it can be seen from the last column that the estimation error was greatly reduced. Especially for the second group, the RMSE was reduced by 0.71°, which verifies the effectiveness of the TPA module.
  • Network parameters
After the network structure was determined, we analyzed the influence of the number of hidden units of the Bi-LSTM layers on the performance. The RMSE of the turn rate estimates with different numbers of hidden units for the Bi-LSTM layers is shown in Table 3. Here, we set the sequence length to 10 and the sliding window to one. As the number of hidden units increased, the RMSE decreased, but the reduction was not significant for 64, 96, and 128. Therefore, considering operational efficiency and estimation precision comprehensively, we set the number of hidden units to 64.
  • Sequence length
Long sequences generally contain more abundant feature information, but this also degrades the real-time performance and increases the difficulty of network training. Therefore, we analyzed the influence of input sequence length on the estimate precision of the network. Here, we set the number of hidden units to 64 and the sliding window to one. The RMSE comparisons for different sequence length are shown in Table 4. When the sequence length increased from 6 to 12, the RMSE of the turn rate estimate decreased gradually. However, when the length was 14, the RMSE increased slightly compared with 12. Therefore, we set the input sequence length of the turn rate estimation network as 12.

4.2.2. Filter State Modification Network

The network structure, parameters, and input sequence length were validated based on the testing datasets. The evaluation criterion was the RMSE and was defined as follows:
σ M ( k ) = 1 N i = 1 N ( ( x ^ k x k ) 2 + ( y ^ k y k ) 2 )
where x ^ k and y ^ k represent the output of the network, and x k and y k are the real values; N is the number of simulations.
  • Network structure
Comparisons of the RMSE of different network structures are shown in Table 5. Here, we set the number of hidden units to 64, the length of input sequence to 15, and the sliding window to five. First, when the number of Bi-LSTM layers increased from one to two, the output precision of the network was obviously improved. However, when the number of layers increased to three and four, the RMSE of groups 1, 3, and 4 increased instead; that is, the network was overfitted. Therefore, we set the number of Bi-LSTM layers to two. After the temporal attention module was introduced, as shown in the last column in Table 5, the RSME was further reduced. Especially for groups 1, 3, and 4, the reduction was remarkable, which verifies the effectiveness of the TA module.
  • Network parameters
The RMSE of test datasets for different numbers of hidden units is shown in Table 6. Here, we set the length of the input sequence to 15 and the sliding window to five. First, with the increase in the number of hidden units, the network performance was improved, but the increase was not obvious. The RMSE for 64 hidden units was reduced by about 0.2 m compared with that for 32 hidden units. Compared with 64, the performance improvement for 96 and 128 hidden units was about 0.1 m. Meanwhile, for groups 1–4, overfitting also occurred when the parameter was 128. Hence, we set the number of hidden units to 64.
  • Sequence length
The influence of the input sequence length on the precision of the filter state modification network was analyzed. Comparisons of the RMSE of the testing datasets are shown in Table 7. Here, we set the number of hidden units to 64 and the sliding window to five. Horizontally, for each group of test trajectories, the RMSE decreased with the increase in the sequence length. However, when the sequence length was 15, the error was at its minimum, and when the sequence length increased to 16, the performance slightly decreased. Therefore, we set the input sequence length to 15.

4.3. Network Module Performance Validation

In this section, five groups of target trajectories with different maneuvering modes in the testing datasets were extracted to verify the validity of the two network modules in the proposed deep-learning-based multiple model tracking method. To show the result comparisons more intuitively, we extracted a target trajectory from each group of maneuvering modes, and the comparison of single simulation results is shown in Figure 6. The first column represents the comparison of the tracking results, the second column represents the comparison of the turn rate estimation results, and the third column represents the comparison of the tracking errors.
  • Effects of the adaptive turn rate estimation network
For the traditional physical definition method, only the filtering state estimates in the last frame were used. In the maneuvering phase, the estimation error of the turn rate was large and fluctuated greatly, which led to a decrease in the tracking precision. Especially for the strong maneuvering trajectory with a large turn rate, for example, the second target performed the turning motion with the parameter ω = 13 ° / s between frames 20 and 38, and the tracking RMSE in the corresponding period increased rapidly. For the proposed adaptive turn rate estimation network, multi-frame kinematic features can be utilized and more complex logical relationships between different variables can be extracted, which contributes to obtaining more accurate estimates of turn rates, as shown by the red curve in Figure 6b. Although the estimation error in the maneuvering phase was slightly larger than that in the non-maneuvering phase, the performance gradually became stable after several frames. The corresponding tracking results are indicated by the red curve in Figure 6c. Compared with the GMPHD filter, our method can effectively reduce the tracking error in the turn maneuvering phase.
  • Effects of the filter state modification network
As can be seen from the tracking results in Figure 6a and the tracking error curve in Figure 6c, an accurate estimate of the turn rate can improve the tracking performance to a certain extent. However, there is uncertainty in the switching of target motion models, resulting in large tracking errors. The filter state modification network can smooth the target state estimates that are produced by the GMPHD filter. As shown in Figure 6a, the tracking result of the purple curve is smoother than those of the red and green curves. Additionally, it can also be seen in Figure 6c that the RMSE corresponding to the blue curve greatly decreased during the turning maneuver phase.
Furthermore, from the tracking result comparisons of all trajectories in the testing datasets shown in Figure 7, it can be concluded that the proposed method has strong adaptability to multiple maneuvering modes and can effectively enhance the tracking precision of complex maneuvering targets.

5. Simulations and Experimental Results

In this section, the adaptability of the proposed method to the multiple maneuvering target tracking in a cluttered environment is demonstrated.

5.1. Simulations

5.1.1. Simulation Scenario and Parameters

Multiple maneuvering targets in a cluttered environment were constructed to verify the tracking performance of the proposed method. The filter parameter design was the same as that presented in Section 4.1.1. In addition, the detected measurements were immersed in clutter, which can be modelled as a Poisson RFS K k with intensity
κ k ( z ) = λ c V U ( z )
where U ( · ) is the uniform density over the surveillance region, V = 1.6 × 10 7 m 2 is the ‘volume’ of the surveillance region, and λ c = 3.2 × 10 6 m 2 is the average number of clutter returns per unit volume.
We designed five target trajectories, representing the five maneuvering forms shown in Table 1 respectively, and the trajectory parameters of each target are shown in Table 8. The simulation scenario is shown in Figure 8.

5.1.2. Result Comparisons

Based on the simulation scenarios shown in Figure 8, the tracking performance of the proposed method was verified, and the optimal subpattern assignment (OSPA) metric [33] was adopted to evaluate the performance of multi-target tracking. After 100 Monte Carlo simulation runs, the OSPA distance comparison is shown in Figure 9, and the turn rate estimate error is shown in Figure 10.
First, from 25–32 s and 71–82 s, targets 2, 3, and 5 all contained fast turning maneuvers with the parameter ω 2 , and the OSPA distance of the GMPHD filter increased dramatically. For GMPHD-ATN, it did not show a large fluctuation, and compared with the GMPHD filter, the maximum reduction was about 12 m, which verifies the importance of turn rate estimate accuracy for the stable tracking of strong maneuvering targets. Additionally, for the GMPHD-ATN-FMN, the introduction of the filter state modification network module was able to smooth the state estimate. Therefore, the tracking precision was further significantly improved in the entire tracking stage, and the performance was improved significantly in the strong maneuvering phase. Specifically, in the weak maneuvering stage, such as 35–70 s, the maximum reduction of OSPA distance was about 4 m, whereas in the strong maneuvering stage, such as 25–32 s, the maximum reduction was about 7.5 m.
Figure 11 shows comparisons of the tracking results obtained in a single simulation. The estimate results of the turn rate for the five targets are shown in Figure 12. For target 5, the large estimate error of the turn rate in GMPHD led to a mismatch in the filtering parameters, and the track was broken in area 3. After introducing the adaptive turn rate estimation process, as shown in Figure 12, more accurate estimates of the turn rate can be obtained. Although the estimate error increased slightly when the motion model switched, after several frames the error gradually decreased. As indicated by the green curve in Figure 11, the matched filtering parameters were able to effectively improve the tracking precision and avoid track breakages in the maneuvering stage. Additionally, as shown by the purple curve in Figure 11, the filter state modification process can smooth the trajectory and output higher quality maneuvering target tracks.

5.2. Experimental Data Validation

5.2.1. Experimental Scenarios and Parameters

In this section, the tracking performance of the proposed method is further verified based on experimental field data. The experimental equipment is a phased array radar working in the Ku band. The sample interval is 6 s. In the experiment, the flight height of a small UAV (DJI Matrice 600) is approximately 200 m with CT maneuvers. The experimental scenario is shown in Figure 13.

5.2.2. Result Comparisons

After the echo data collected by the radar were processed by constant false alarm rate detection, measurements of the area where the UAV was located could be obtained. To verify the tracking performance in the multiple targets scenario, we combined two sets of measurements of the single UAV target into one set.
Figure 14 shows the tracking result comparisons. Firstly, for the GMPHD with unknown turn rate parameters, the mismatch in the model parameter resulted in track breakages at the turning position. For the GMPHD-ATN, the track quality was significantly improved. Additionally, as shown in Figure 14c, after introducing filter state modification processing, the tracking result was smoothed, and the precision of the state estimate was effectively improved. Therefore, the tracking results of real UAV observation data verify the rationality of the parameters of the constructed training datasets, and the proposed method demonstrated its adaptability to practical scenarios with detection and measurement noise uncertainties.

5.2.3. Computational Complexity

For a fair comparison, the computational efficiency of all algorithms was tested on an Intel Xeon E5-2680 CPU at 2.4 GHz. For the adaptive turn rate estimation network, the testing runtime per iteration was 2 ms, and for the filter state modification network, the testing runtime per iteration was 4 ms. Although the training of the deep neural network was time consuming, its implementation in practice is highly efficient because the calculations are mainly matrix multiplications and element-wise operations. Therefore, the proposed method can ensure the real-time performance of the tracking process.

6. Discussion

The RFS-based filter incorporating the jump Markov system approach is an effective method for multiple maneuvering target tracking in cluttered environments. However, UAV targets have strong maneuverability and their trajectory forms are diverse, presenting unknown motion model parameters and uncertain model switching. The preset parameters of filter models are difficult to match to the time-varying target motion, leading to a serious decline in tracking performance. Therefore, within the framework of the LGJMS-GMPHD filter, we proposed a deep-learning-based multiple model tracking method to improve maneuvering adaptability.
The simulation results indicated that the traditional LGJMS-GMPHD filter has a low tracking precision and the track is prone to breaking in scenarios involving complex maneuvering targets. As shown in Figure 12, for the PDM, the estimation error of the turn rate was large and fluctuated greatly in the maneuvering phase as only the filtering state estimate in the last frame was considered, whereas the proposed adaptive turn rate estimation network made use of multiple dimensions of kinematic features and extracted the relationship between different features. As shown in the red curve in Figure 12, a more accurate estimation of turn rate can be obtained, which helps to improve the matching degree of the CT model in the filter. Therefore, as shown in Figure 9, compared with the GMPHD filter, the OSPA distance of GMPHD-ATN was greatly reduced, and the tracking results shown in Figure 11 indicate that track continuity was optimized.
Another reason for poor tracking performance is the uncertainty of motion model switching. Smoothing results in better estimates for target states by means of a time delay. However, as described in [21], longer lags do not yield better estimates of the current state because the temporal correlations cannot be extracted. Moreover, backward smoothing is still model-based, which cannot address the substantive problem of tracking performance degradations due to model switching. Deep neural networks have a strong capability of fitting any mapping, providing an effective way to handle target motion uncertainty. However, the LSTM-based deep recurrent neural network in [24] only considered CV and CA motions, and is not suitable for the scenario of multiple maneuvering targets in the CT model. However, in the proposed filter state modification network, the state estimates output by the LGJMS-GMPHD filter were used as the input, and the temporal correlation information of the state vector was captured, with the ability to handle complex maneuvering motions. Hence, as shown in Figure 9, the tracking precision was improved in the entire tracking stage, especially in the strong maneuvering phase. The tracking results, shown in Figure 11, demonstrated that the track was smoothed and the track quality was improved. Moreover, based on the tracking error comparisons of test datasets, shown in Figure 7, it can be concluded that the proposed method has strong adaptability to different maneuvering forms. The experimental data processing results indicate that the parameters of the training dataset designed in this paper can adapt to practical scenarios, and the method can effectively enhance the track quality of real UAV targets.
However, in this paper, we assumed that the clutter rate was known and remained constant. In future research, more complex detection environments, such as a time-varying clutter rate and detection uncertainty, will be considered. Furthermore, we will also consider extending the proposed method to group target tracking.

7. Conclusions

In this paper, we have proposed a deep-learning-based multiple model tracking method. The adaptive turn rate estimation network employs multi-frame and multi-dimensional kinematic information to improve the accuracy of the turn rate estimates for a maneuvering target with CT motion, thereby enhancing the consistency with the filtering model. Additionally, the filter state modification network uses the temporal features of the multi-frame state estimate to achieve the smoothing of target state estimates, thereby circumventing the large tracking errors caused by uncertain motion model switching. To ensure the applicability of the algorithm, in the training datasets, we designed five switching modes of the movement model. Therefore, the proposed method is suitable for practical maneuvering movements: (i) from CV to CT, (ii) from CT with ω 1 to CT with ω 2 , and (iii) from CT to CV. In the end, based on the simulation results of multiple maneuvering target tracking in cluttered environments, it can be concluded that the proposed method can obtain accurate turn rate estimates and output high-quality tracks, and the performance improvement is especially significant in the maneuvering phase. Moreover, for the real UAV observation scenario with measurement noise and detection uncertainties, the proposed method can output more stable and smooth trajectories, which verifies its applicability in real scenarios.

Author Contributions

Conceptualization, L.F. and W.T.; methodology, L.F. and W.T.; software, L.F.; validation, L.F. and R.W.; formal analysis, R.W. and H.L.; investigation, R.W.; resources, W.L. (Weigang Luo), N.N. and H.L.; data curation, H.L. and W.L. (Weidong Li); writing—original draft preparation, L.F. and W.T.; writing—review and editing, R.W. and L.F.; visualization, N.N. and W.L. (Weidong Li); supervision, C.H.; project administration, C.H.; funding acquisition, C.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grant No. U2133217), the National Natural Science Foundation of China (Grant No. 31727901) and the National Natural Science Foundation of China (Grant No. 62001021).

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gu, J.; Su, T.; Wang, Q.; Du, X.; Guizani, M. Multiple moving targets surveillance based on a cooperative network for multi-UAV. IEEE Commun. Mag. 2018, 56, 82–89. [Google Scholar] [CrossRef]
  2. Zheng, J.; Chen, R.; Yang, T.; Liu, X.; Liu, H.; Su, T.; Wan, L. An efficient strategy for accurate detection and localization of UAV Swarms. IEEE Internet Things J. 2021, 8, 15372–15381. [Google Scholar] [CrossRef]
  3. Chen, W.; Liu, J.; Li, J. Classification of UAV and bird target in low-altitude airspace with surveillance radar data. Aeronaut. J. 2019, 123, 191–211. [Google Scholar] [CrossRef]
  4. Li, W.; Hu, C.; Wang, R.; Kong, S.; Zhang, F. Comprehensive analysis of polarimetric radar cross-section parameters for insect body width and length estimation. Sci. China Inf. Sci. 2021, 64, 122302. [Google Scholar] [CrossRef]
  5. Chen, X.; Chen, W.; Rao, Y.; Huang, Y.; Guan, J.; Dong, Y. Progress and prospects of radar target detection and recognition technology for flying birds and unmanned aerial vehicles. J. Radars 2020, 9, 803–827. [Google Scholar] [CrossRef]
  6. Vo, B.N.; Ma, W.K. The Gaussian mixture probability hypothesis density filter. IEEE Trans. Signal Process. 2006, 54, 4091–4104. [Google Scholar] [CrossRef]
  7. Yi, W.; Jiang, M.; Hoseinnezhad, R. The multiple model Vo-Vo filter. IEEE Trans. Aerosp. Electron. Syst. 2017, 53, 1045–1054. [Google Scholar] [CrossRef]
  8. Jeong, T.T. Particle PHD filter multiple target tracking in sonar image. IEEE Trans. Aerosp. Electron. Syst. 2007, 43, 409–416. [Google Scholar] [CrossRef]
  9. Vo, B.N.; Vo, B.N.; Cantoni, A. Analytic implementations of the cardinalized probability hypothesis density filter. IEEE Trans. Signal Process. 2007, 55, 3553–3567. [Google Scholar] [CrossRef]
  10. Zhu, S.; Yang, B.; Wu, S. Measurement-driven multi-target tracking filter under the framework of labeled random finite set. Digit Signal Process. 2021, 112, 103000. [Google Scholar] [CrossRef]
  11. Mahler, R. On multitarget jump-Markov filters. In Proceedings of the International Conference on Information Fusion, Singapore, 9–12 July 2012; pp. 149–156. [Google Scholar]
  12. Pasha, S.A.; Vo, B.N.; Tuan, H.D.; Ma, W.K. A gaussian mixture PHD Filter for jump Markov system models. IEEE Trans. Aerosp. Electron. Syst. 2009, 45, 919–936. [Google Scholar] [CrossRef] [Green Version]
  13. Dong, P.; Jing, Z.; Gong, D.; Tang, B. Maneuvering multi-target tracking based on variable structure multiple model GMCPHD filter. Signal Process. 2017, 141, 158–167. [Google Scholar] [CrossRef]
  14. Dunne, D.; Kirubarajan, T. Multiple model multi-bernoulli filters for manoeuvering targets. IEEE Trans Aerosp Electron. Syst. 2013, 49, 2679–2692. [Google Scholar] [CrossRef]
  15. Li, X.; Jilkov, V.P. Survey of maneuvering target tracking. Part I. Dynamic models. IEEE Trans. Aerosp. Electron. Syst. 2003, 39, 1333–1364. [Google Scholar] [CrossRef]
  16. Punithakumar, K.; Kirubarajan, T.; Sinha, A. Multiple-model probability hypothesis density filter for tracking maneuvering targets. IEEE Trans. Aerosp. Electron. Syst. 2008, 44, 87–98. [Google Scholar] [CrossRef]
  17. Yuan, X.; Han, C.; Duan, Z.; Lei, M. Adaptive turn rate estimation using range rate measurements. IEEE Trans. Aerosp. Electron. Syst. 2006, 42, 1532–1541. [Google Scholar] [CrossRef]
  18. Chen, B.; Tugnait, J.K. Multisensor tracking of a maneuvering target in clutter using IMMPDA fixed-lag smoothing. IEEE Trans. Aerosp. Electron. Syst. 2000, 36, 983–991. [Google Scholar] [CrossRef]
  19. Koch, W. Fixed-interval retrodiction approach to Bayesian IMM-MHT for maneuvering multiple targets. IEEE Trans. Aerosp. Electron. Syst. 2000, 36, 2–14. [Google Scholar] [CrossRef]
  20. Lian, F.; Han, C.; Liu, W.; Yuan, X. Multiple-model probability hypothesis density smoother. Acta Autom. Sin. 2010, 36, 939–950. [Google Scholar] [CrossRef]
  21. Nadarajah, N.; Kirubarajan, T.; Lang, T.; Mcdonald, M.; Punithakumar, K. Multitarget tracking using probability hypothesis density smoothing. IEEE Trans. Aerosp. Electron. Syst. 2011, 47, 2344–2360. [Google Scholar] [CrossRef]
  22. Zheng, T.; Yao, Y.; He, F.; Ji, D.; Zhang, X. Active switching multiple model method for tracking a noncooperative gliding flight vehicle. Sci. China Inf. Sci. 2020, 63, 192202. [Google Scholar] [CrossRef]
  23. Liu, J.; Wang, Z.; Xu, M. DeepMTT: A deep learning maneuvering target-tracking algorithm based on bidirectional LSTM network. Inf. Fusion. 2020, 53, 289–304. [Google Scholar] [CrossRef]
  24. Gao, C.; Yan, J.; Zhou, S.; Varshney, P.K.; Liu, H. Long short-term memory-based deep recurrent neural networks for target tracking. Inf. Sci. 2019, 502, 279–296. [Google Scholar] [CrossRef]
  25. Mahler, R.P.S. Multitarget Bayes filtering via first-order multitarget moments. IEEE Trans. Aerosp. Electron. Syst. 2003, 39, 1152–1178. [Google Scholar] [CrossRef]
  26. Wang, X.; Musicki, D.; Ellem, R.; Fletcher, F. Efficient and enhanced multi-target tracking with doppler measurements. IEEE Trans. Aerosp. Electron. Syst. 2009, 45, 1400–1417. [Google Scholar] [CrossRef]
  27. Panta, K.; Clark, D.E.; Vo, B.N. Data association and track management for the gaussian mixture probability hypothesis density filter. IEEE Trans. Aerosp. Electron. Syst. 2009, 45, 1003–1016. [Google Scholar] [CrossRef] [Green Version]
  28. Milan, A.; Rezatofighi, S.H.; Dick, A.; Reid, I.; Schindler, K. Online multi-target tracking using recurrent neural networks. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–10 February 2017; pp. 4225–4232. [Google Scholar]
  29. Ondruska, P.; Posner, I. Deep tracking: Seeing beyond seeing using recurrent neural networks. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; pp. 3361–3367. [Google Scholar]
  30. Shih, S.Y.; Sun, F.K.; Lee, H.Y. Temporal pattern attention for multivariate time series forecasting. Mach. Learn. 2019, 108, 1421–1441. [Google Scholar] [CrossRef] [Green Version]
  31. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. In Proceedings of the 3rd International Conference for Learning Representations, San Diego, CA, USA, 7–9 May 2015; pp. 1–14. [Google Scholar] [CrossRef]
  32. Yang, Z.; Yang, D.; Dyer, C.; He, X.; Smola, A.; Hovy, E. Hierarchical attention networks for document classification. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, San Diego, CA, USA, 12–17 June 2016; pp. 1480–1489. [Google Scholar] [CrossRef] [Green Version]
  33. Schuhmacher, D.; Vo, B.T.; Vo, B.N. A consistent metric for performance evaluation of multi-object filters. IEEE Trans. Aerosp. Electron. Syst. 2008, 56, 3447–3457. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Flow diagram of the proposed deep-learning-based multiple model tracking method.
Figure 1. Flow diagram of the proposed deep-learning-based multiple model tracking method.
Remotesensing 14 03276 g001
Figure 2. Structure of the adaptive turn rate estimation network.
Figure 2. Structure of the adaptive turn rate estimation network.
Remotesensing 14 03276 g002
Figure 3. Bi-LSTM structure diagram.
Figure 3. Bi-LSTM structure diagram.
Remotesensing 14 03276 g003
Figure 4. Structure of the filter state modification network.
Figure 4. Structure of the filter state modification network.
Remotesensing 14 03276 g004
Figure 5. Flowchart of dataset construction.
Figure 5. Flowchart of dataset construction.
Remotesensing 14 03276 g005
Figure 6. Single simulation results for different maneuvering modes: (a) tracking results, (b) turn rate estimates, and (c) tracking errors.
Figure 6. Single simulation results for different maneuvering modes: (a) tracking results, (b) turn rate estimates, and (c) tracking errors.
Remotesensing 14 03276 g006aRemotesensing 14 03276 g006b
Figure 7. Tracking error comparisons of test datasets: (a) group 1, (b) group 2, (c) group 3, (d) group 4, and (e) group 5.
Figure 7. Tracking error comparisons of test datasets: (a) group 1, (b) group 2, (c) group 3, (d) group 4, and (e) group 5.
Remotesensing 14 03276 g007aRemotesensing 14 03276 g007b
Figure 8. Simulation scenario. (△—start position, ◯—stop position).
Figure 8. Simulation scenario. (△—start position, ◯—stop position).
Remotesensing 14 03276 g008
Figure 9. OSPA distance comparison.
Figure 9. OSPA distance comparison.
Remotesensing 14 03276 g009
Figure 10. RMSE of turn rate estimates.
Figure 10. RMSE of turn rate estimates.
Remotesensing 14 03276 g010
Figure 11. Tracking results of a single simulation.
Figure 11. Tracking results of a single simulation.
Remotesensing 14 03276 g011
Figure 12. Result comparisons of turn rate estimate: (a) target 1, (b) target 2, (c) target 3, (d) target 4, and (e) target 5.
Figure 12. Result comparisons of turn rate estimate: (a) target 1, (b) target 2, (c) target 3, (d) target 4, and (e) target 5.
Remotesensing 14 03276 g012aRemotesensing 14 03276 g012b
Figure 13. Experimental scenario.
Figure 13. Experimental scenario.
Remotesensing 14 03276 g013
Figure 14. Comparisons of real UAV tracking results: (a) GMPHD, (b) GMPHD-ATN, and (c) GMPHD-ATN-FMN.
Figure 14. Comparisons of real UAV tracking results: (a) GMPHD, (b) GMPHD-ATN, and (c) GMPHD-ATN-FMN.
Remotesensing 14 03276 g014
Table 1. The kinematic parameters of the target trajectory.
Table 1. The kinematic parameters of the target trajectory.
Group IndexInitial Velocity (m/s)Turn Rate
Stage 1Stage 2Stage 3Stage 4
1[−60, 60]0 ω 1 0/
2[−40, 40]0 ω 2 0/
3[−60, 60]0 ω 1 ω 2 0
4[−60, 60]0 ω 1 ω 1 0
5[−40, 40]0 ω 2 ω 1 0
Table 2. RMSE of different network structures (units: degree).
Table 2. RMSE of different network structures (units: degree).
Group
Index
Network Structure
One
Bi-LSTM
Layer
Two
Bi-LSTM
Layers
Three
Bi-LSTM
Layers
Four
Bi-LSTM
Layers
Two
Bi-LSTM
Layers + TPA
10.900.760.760.770.55
22.061.631.561.450.92
31.671.371.351.230.74
41.291.231.241.230.91
51.781.481.481.421.02
Table 3. RMSE of different network parameters (units: degree).
Table 3. RMSE of different network parameters (units: degree).
Group IndexNumber of Hidden Units
326496128
10.620.550.550.56
21.060.920.860.88
30.820.740.700.68
40.970.910.900.88
51.091.021.041.04
Table 4. RMSE of different input sequence lengths (units: degree).
Table 4. RMSE of different input sequence lengths (units: degree).
Group
Index
Sequence Length
68101214
10.640.600.550.540.58
20.980.870.920.840.86
30.800.760.740.690.72
41.040.980.910.880.92
51.131.071.021.001.01
Table 5. RMSE of different network structures (units: m).
Table 5. RMSE of different network structures (units: m).
Group
Index
Network Structure
One
Bi-LSTM
Layer
Two
Bi-LSTM
Layers
Three
Bi-LSTM
Layers
Four
Bi-LSTM
Layers
Two
Bi-LSTM
Layers + TA
111.9611.6212.5213.1311.28
210.239.939.9910.029.88
311.6111.2012.6813.0710.66
411.7111.2112.9113.4310.84
511.4110.8610.9810.9910.82
Table 6. RMSE of different network parameters (units: m).
Table 6. RMSE of different network parameters (units: m).
Group IndexNumber of Hidden Units
326496128
111.4611.2811.1911.32
29.919.889.779.79
310.8110.6610.7310.98
411.0110.8410.7910.95
510.8510.8210.7210.71
Table 7. RMSE of different input sequence lengths (units: m).
Table 7. RMSE of different input sequence lengths (units: m).
Group
Index
Sequence Length
81012141516
112.2912.0811.8111.3011.2811.71
210.3710.5210.3610.029.889.97
311.6711.7111.3411.1010.6610.89
412.1611.7011.5511.2210.8411.21
511.3611.3211.2411.1610.8211.06
Table 8. Target trajectory parameters.
Table 8. Target trajectory parameters.
IndexInitial State Turn   Rate   ( ° / s )
1[500 m; 20 m/s; 523 m; −25 m/s]1~24 s25~57 s58~84 s
0−70
2[243 m; −35 m/s; 266 m; −20.4 m/s]1~21 s22~33 s34~55 s
0−10.40
3[941 m; −27.4 m/s; 737 m; 35.5 m/s]1~21 s22~70 s71~82 s83~110 s
05.2−10.60
4[1000 m; 15 m/s; 1300 m; −20 m/s]1~25 s26~49 s50~94 s95~116 s
02.9−6.90
5[109 m; −30 m/s; 620 m; −30 m/s]1~23 s24~33 s34~54 s55~74 s
013.7−7.80
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tian, W.; Fang, L.; Li, W.; Ni, N.; Wang, R.; Hu, C.; Liu, H.; Luo, W. Deep-Learning-Based Multiple Model Tracking Method for Targets with Complex Maneuvering Motion. Remote Sens. 2022, 14, 3276. https://doi.org/10.3390/rs14143276

AMA Style

Tian W, Fang L, Li W, Ni N, Wang R, Hu C, Liu H, Luo W. Deep-Learning-Based Multiple Model Tracking Method for Targets with Complex Maneuvering Motion. Remote Sensing. 2022; 14(14):3276. https://doi.org/10.3390/rs14143276

Chicago/Turabian Style

Tian, Weiming, Linlin Fang, Weidong Li, Na Ni, Rui Wang, Cheng Hu, Hanzhe Liu, and Weigang Luo. 2022. "Deep-Learning-Based Multiple Model Tracking Method for Targets with Complex Maneuvering Motion" Remote Sensing 14, no. 14: 3276. https://doi.org/10.3390/rs14143276

APA Style

Tian, W., Fang, L., Li, W., Ni, N., Wang, R., Hu, C., Liu, H., & Luo, W. (2022). Deep-Learning-Based Multiple Model Tracking Method for Targets with Complex Maneuvering Motion. Remote Sensing, 14(14), 3276. https://doi.org/10.3390/rs14143276

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop