Next Article in Journal
Performance Analysis of Supervised Machine Learning Algorithms for Automatized Radiographical Classification of Maxillary Third Molar Impaction
Previous Article in Journal
Situation-Aware Survivable Network Design for Tactical Environments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Target Localization and Sensor Movement Trajectory Planning with Bearing-Only Measurements in Three Dimensional Space

1
School of Automation, Central South University, Changsha 410083, China
2
School of Electrical and Information Engineering, Changsha University of Science and Technology, Changsha 410114, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(13), 6739; https://doi.org/10.3390/app12136739
Submission received: 1 April 2022 / Revised: 10 June 2022 / Accepted: 29 June 2022 / Published: 3 July 2022

Abstract

:
In order to improve the accuracy of bearing-only localization in three dimensional (3D) space, this paper proposes a novel bias compensation method and a new single-sensor maneuvering trajectory algorithm, respectively. Compared with traditional methods, the bias compensation method estimates the unknown variance of bearing noise consistently, which is utilized in pseudo-linear target localization to achieve higher precision. The sensor maneuvering algorithm designs the next moment sensor location in consideration of all the past sensor locations, unlike other approaches that only consider finite past locations. Research shows that the trajectories generated by our algorithm have greater Fisher information matrix (FIM) determinants and better localization accuracy.

1. Introduction

Target localization has applications in both civilian and military domains including wireless communications, environmental monitoring [1], network security [2,3], medical diagnosis, etc. The commonly used methods for target localization can be characterized into four categories according to the type of sensors: bearing-only (also called angle of arrival (AOA)), range-only, time-of-arrival (time-difference-of-arrival) and received signal strength [4]. The problem of using only angle measurements to locate a target is referred to as bearing-only target localization which is the focus of this paper.
The bearing-only target positioning problem can usually be solved in the following two categories. The first category is based on Kalman filtering and its various evolutionary versions [5,6], Kalman filter, extended Kalman filter and unscented Kalman filter have been applied to the bearing-only target localization problem in [7,8]. Due to nonlinear-to-linear transformation of the angle measurement equations, there exists a serious deviation problem. Dogançay et al. [9] propose a new two-step pseudo-linear Kalman filter for bearing-only target tracking in 3D space. In the second category, many batch form methods based on the ordinary least squares are used to deal with target localization under small measurement noise [10,11,12]. To deal with the serious deviation of the pseudo-linear estimator (PLE) in two dimensional (2D) target localization problem, Dogançay [13] proposes a closed-form reduced-bias PLE to realize the asymptotic unbiased estimation of target motion parameters by using the instrumental variable (IV) method. In the case of small noise, Dogançay also proposes pseudo-linear estimator for AOA target motion analysis in 3D space [14]. Adib et al. [15] proposes a two-stage weighted Stansfield geolocation algorithm based on maximum likelihood estimation in 3D space, and it has the advantages of consistency and statistical efficiency. In addition, the position of the sensor has a great influence on positioning accuracy [4,16]. In order to minimize the uncertainty of the target estimation, the position of the sensor at the next moment needs to be reasonably planned. The optimal localization of the sensor is generally determined by maximizing the determinant of FIM or minimizing the trace of error covariance matrix of the estimator [2,17,18,19]. The optimal trajectory of a single mobile sensor is determined by maximizing the determinant of FIM under the condition of adding state constraints on the observer trajectory [20]. When the absolute elevation of the sensor is constant, Sheng et al. [21] proposes a new simple sensor optimal deployment criterion based on minimizing the trace of error covariance matrix. Based on the optimization criterion of minimizing the mean square error (MSE), an own-ship path optimization algorithm by using the gradient-descent method in x y plane and the grid search optimization method in z axis is presented [22]. A distributed path optimization algorithm is applied when the communication distance constraints and no-fly zone are considered [23].
Most estimation methods generally assume that the variance of measurement noise is known a priori for the purpose of bias compensation. In fact, the noise variance is generally unknown and needs to be approximated in advance. The least-square solution for the target dynamics is calculated and plugged into bearing measurement equations to compute the noise in the first step. Then the “true” noise is used to construct the pseudolinear vector and form the final bias-compensation PLE result which is still biased [24]. As the first contribution of this paper, we analyze the properties of Gaussian noise along with the angle measurement equations and newly discover the potential relationship between the determinant of the extended coefficient matrix of the equation and the noise variance. A BC method for target localization on x y plane and the target position in z-axis direction based on the pseudo-linear estimation method is proposed thereafter.
Optimizing the increment of FIM determinant over a finite time horizon rather than the whole time span makes trajectory planning less effective [25]. As the second contribution of this paper, a sensor trajectory planning algorithm based on greedy strategies maximizes the increment between two consecutive FIM determinants to realize the next location of a single sensor. It is proven that the increment equals the aggregation of FIM determinants constituted by the sensor location at next time step and the combination of all past locations. Considering the spherical constraints of target safety area and sensor movement area, the optimal solution always sits on the area surface. Based on this fact, an analytical solution is suggested to solve the maximization problem.
The rest of this paper is organized as follows. Section 2 formulates the problem of target localization in 3D space, including the notations and localization methods. Section 3 proposes a BC estimator and bias compensation weighted instrumental variable (BC-WIV) estimator. In Section 4, the sensor maneuvering algorithm in the composition of two trajectories is designed in detail with the analysis of the FIM increment. Numerical examples are studies to validate the method studies in Section 5. Section 6 concludes the whole paper and points out future research directions.

2. Bearing-Only Localization in 3D Space

As shown in Figure 1, the location of target T = x t y t z t is unknown in 3D space. T needs to be estimated based on the azimuth ϕ ^ i and elevation angle θ ^ i measurements collected from a mobile sensor S i = x s i y s i z s i ( i = 1 , , n ) at n locations. t and s i are the coordinates of the target and the sensor on x y plane, respectively. The true bearing information ϕ i and θ i received by the sensor from the target are:
ϕ i = tan 1 y t y s i x t x s i ,
θ i = tan 1 z t z s i ( x t x s i ) 2 + ( y t y s i ) 2 .
The bearing information ϕ ^ i and θ ^ i are the sensor measurements corrupted by the noise e i and δ i at n locations, i.e.,
ϕ ^ i = ϕ i + e i ,
θ ^ i = θ i + δ i ,
where e i and δ i are independent identically distributed white Gaussian noise with zero mean and unknown variance of σ 2 and ξ 2 , respectively.
The problem of target localization in 3D space is solved in two steps. Firstly, the bearing line is projected onto the x y plane. The corresponding 2D estimator is used to solve the resulting 2D positioning problem. Then the estimated position of T along the z axis is obtained from the estimated x and y coordinates. The 2D pseudolinear estimator (PLE) algorithm is used to estimate t by using n azimuth measurements ϕ ^ i . To turn a nonlinear problem into a linear one, small noise assumption also appears in the development and analysis of the Stansfield algorithm [26] and the total-least-squares solution [27] for localization. The pseudolinear equation from (1) and (3) between t and the given information is described as:
sin ϕ ^ i x t cos ϕ ^ i y t = sin ϕ ^ i x s i cos ϕ ^ i y s i + | | t s i | | sin e i ,
which can be expressed in matrix form as:
A n t = B n + n n
where
A n = sin ϕ ^ 1 cos ϕ ^ 1 sin ϕ ^ n cos ϕ ^ n , B n = sin ϕ ^ 1 x s 1 cos ϕ ^ 1 y s 1 sin ϕ ^ n x s n cos ϕ ^ n y s n , n n = | | t s 1 | | 0 0 0 0 0 0 | | t s n | | sin e 1 sin e n .
The 2D PLE estimator of (6) is:
t ^ p l e = ( A n A n ) 1 A n B n ,
where A n represents the transpose of A n .
The target position of the projection in 2D space has been estimated. According to the geometric relationship, the target position in the z axis direction can be estimated according to the elevation angle measurements and other information as:
z 0 = 1 n i = 1 n z i + x 0 x s i 2 + y 0 y s i 2 tan θ ^ i ,
where x 0 and y 0 are the target position estimated by different methods in 2D space.

3. Bias Compensation Estimator for Localization in 3D Space

3.1. Bias Compensation Localization in 2D Space

In [28], a bias compensation method for bearing-only localization is proposed, which can effectively solve the problem of nonlinear localization. Dividing cos ϕ ^ i on both sides of (5) gives:
tan ϕ ^ i x t y t = tan ϕ ^ i x s i y s i + | | t s i | | sin e i cos ϕ ^ i .
Rewrite (10) into the matrix form as:
Z n = W ¯ n t + C n N n ,
where
Z n = tan ϕ ^ 1 x s 1 y s 1 tan ϕ ^ n x s n y s n , W ¯ n = tan ϕ ^ 1 1 tan ϕ ^ n 1 ,
C n = | | t s 1 | | cos ϕ ^ 1 0 0 0 0 0 0 | | t s n | | cos ϕ ^ n , N n = sin e 1 sin e n .
Based on
σ ^ 2 = min λ , s . t . det 1 n W ¯ n Z n W ¯ n Z n λ 1 0 1 n i = 1 n x s i 0 0 0 1 n i = 1 n x s i 0 1 n i = 1 n x s i 2 = 0
of σ 2 , the Bai estimator is:
t ^ σ ^ 2 = 1 n W ¯ n W ¯ n σ ^ 2 1 0 0 0 1 1 n W ¯ n Z n + σ ^ 2 1 n i = 1 n x s i 0 .
Inspired by [28], a corresponding bias compensation estimator is designed for (5). Firstly, the matrix A n and B n are processed as:
1 n A n B n A n B n = 1 n i = 1 n a 11 i a 12 i b 1 i a 12 i a 22 i b 2 i b 1 i b 2 i c i ,
where
a 11 i = sin 2 ϕ i cos 2 e i + 2 sin ϕ i cos ϕ i cos e i sin e i + cos 2 ϕ i sin 2 e i , a 12 i = sin ϕ i cos ϕ i cos 2 e i cos 2 ϕ i cos e i sin e i
+ sin 2 ϕ i cos e i sin e i + cos ϕ i sin ϕ i sin 2 e i ,
a 22 i = cos 2 ϕ i cos 2 e i 2 cos ϕ i sin ϕ i cos e i sin e i + sin 2 ϕ i sin 2 e i ,
b 1 i = a 11 i x s i + a 12 i y s i ,
b 2 i = a 12 i x s i + a 22 i y s i ,
c i = a 11 i x s i 2 + 2 a 12 i x s i y s i + a 22 i y s i 2 .
To derive the bias compensation estimator in detail, two lemmas are proposed.
Lemma 1.
E ( cos e i sin e i ) = 0 , E ( sin 2 e i ) = ( 1 e 2 σ 2 ) / 2 , E ( cos 2 e i ) = ( 1 + e 2 σ 2 ) / 2 where E ( · ) represents the mathematical expectation.
Proof of Lemma 1.
Since cos e i sin e i = sin 2 e i / 2 is an odd function, we have:
1 2 2 π σ sin 2 e i · e e i 2 2 σ 2 d e i = 0 .
From
E ( cos 2 e i sin 2 e i ) = E ( cos 2 e i ) = i = 0 ( 1 ) i 2 2 i ( 2 i ) ! · 1 2 π σ e i 2 i e e i 2 2 σ 2 d e i = i = 0 ( 1 ) i 2 2 i ( 2 i ) ! · σ 2 i ( 2 i ) ! 2 i ( i ) ! = i = 0 ( 1 ) i ( 2 σ 2 ) i ( i ) ! = e 2 σ 2
and E ( cos 2 e i + sin 2 e i ) = 1 , E ( sin 2 e i ) = ( 1 e 2 σ 2 ) / 2 , E ( cos 2 e i ) = ( 1 + e 2 σ 2 ) / 2 . □
Define
P ( γ ) = 1 n A n B n A n B n γ n n 0 i = 1 n x s i 0 n i = 1 n y s i i = 1 n x s i i = 1 n y s i i = 1 n x s i 2 + i = 1 n y s i 2
as the extended coefficient matrix.
Lemma 2.
γ min = E ( sin 2 e i ) is the smallest real value s . t . det ( P ( γ ) ) = 0 in probability as n .
Proof of Lemma 2.
With Lemma 1, we have:
1 n A n B n A n B n P ( γ min ) = γ min n n 0 i = 1 n x s i 0 n i = 1 n y s i i = 1 n x s i i = 1 n y s i i = 1 n x s i 2 + i = 1 n y s i 2
in probability where:
P ( γ min ) = E ( cos ( 2 e i ) ) n i = 1 n sin ϕ i cos ϕ i sin ϕ i x s i cos ϕ i y s i sin ϕ i cos ϕ i sin ϕ i x s i cos ϕ i y s i .
Through observation, P ( γ min ) is positive semi-definite. Its null space is one dimensional in the form of: β x t y t 1 , where β R . Let vector v = v 1 v 2 v 3 be arbitrary in R 3 × 1 . So we have
v P ( γ ) v = v P ( γ min ) v γ γ min n v n 0 i = 1 n x s i 0 n i = 1 n y s i i = 1 n x s i i = 1 n y s i i = 1 n x s i 2 + i = 1 n y s i 2 v .
The second term
γ γ min n v n 0 i = 1 n x s i 0 n i = 1 n y s i i = 1 n x s i i = 1 n y s i . i = 1 n x s i 2 + i = 1 n y s i 2 v = γ γ min n i = 1 n ( v 1 2 + 2 v 1 v 3 x s i + v 2 2 + 2 v 2 v 3 y s i + v 3 2 x s i 2 + v 3 2 y s i 2 ) = ( γ γ min ) ( ( v 1 + v 3 n i = 1 n x s i ) 2 + ( v 2 + v 3 n i = 1 n y s i ) 2 + v 3 2 2 n 2 i = 1 n j = 1 n ( x s i x s j ) 2 + v 3 2 2 n 2 i = 1 n j = 1 n ( y s i y s j ) 2 ) > 0 ,
when γ < γ min , the first term v P ( γ min ) v in (26) is no less than 0. This makes v P ( γ ) v always greater than 0. Hence, the minimum γ that leads to v P ( γ ) v = 0 or det ( P ( γ ) ) = 0 satisfies γ = γ min when v = β x t y t 1 . □
It is worthwhile to point out that γ min can be found by getting the minimum root of the third-order polynomial function det ( P ( γ ) ) = 0 . To sum up, the 2D positioning results with bias compensation can be obtained, as
Theorem 1.
t ^ b c = 1 n A n A n γ min 0 0 γ min 1 1 n A n B n γ min n i = 1 n x s i y s i
converges totin probability as n when E ( c o s 2 e i ) 0 .
Proof of Theorem 1.
As n ,
1 n A n A n γ min 0 0 γ min = E ( cos ( 2 e i ) ) n i = 1 n sin ϕ i cos ϕ i sin ϕ i cos ϕ i ,
1 n A n B n γ min n i = 1 n x s i y s i = E ( cos ( 2 e i ) ) n i = 1 n sin ϕ i cos ϕ i sin ϕ i cos ϕ i x t y t .
From (29) and (30), Theorem 1 stands when E ( cos 2 e i ) 0 . □

3.2. Bias Compensation Method in Z-Axis

The geometric relationship in 3D space is:
z t cos θ ^ i = r ^ i sin θ ^ i + z i cos θ ^ i ,
where r ^ i = t ^ b c ( 1 ) x s i 2 + t ^ b c ( 2 ) y s i 2 . Rewrite (31) into the matrix form
Y n z t = F n ,
where
Y n = cos θ ^ 1 cos θ ^ n , F n = z s 1 cos θ ^ 1 + r ^ 1 sin θ ^ 1 z s n cos θ ^ n + r ^ n sin θ ^ n .
Similarly, we handle the matrix Y n and F n ,
1 n Y n F n Y n F n = 1 n i = 1 n d 1 i d 2 i d 2 i d 3 i ,
where
d 1 i = cos 2 θ i cos 2 δ i 2 cos θ i sin θ i cos δ i sin δ i + sin 2 θ i sin 2 δ i , d 2 i = z s i d 1 i + r ^ i ( sin θ i cos θ i cos 2 δ i sin 2 θ i cos δ i sin δ i
+ cos 2 θ i cos δ i sin δ i sin θ i cos θ i sin 2 δ i ) , d 3 i = z s i 2 d 1 i + 2 z s i r ^ i ( sin θ i cos θ i cos 2 δ i sin 2 θ i cos δ i sin δ i + cos 2 θ i cos δ i sin δ i
sin θ i cos θ i sin 2 δ i ) + r ^ i 2 ( sin 2 θ i cos 2 δ i + 2 sin θ i cos θ i sin δ i cos δ i + cos 2 θ i sin 2 δ i ) .
We also define
Q ( μ ) = 1 n Y n F n Y n F n μ n n i = 1 n z s i i = 1 n z s i i = 1 n z s i 2 + i = 1 n r ^ i 2 .
According to Lemma 2, the same reasoning process can be concluded. μ min = E ( sin 2 δ i ) is the smallest real value s . t . det ( Q ( μ ) ) = 0 in probability as n . The difference is μ min can be found by getting the minimum root of a second-order equation det ( Q ( μ ) ) = 0 .
Theorem 2.
z ^ b c = 1 n Y n Y n μ min 1 1 n Y n F n μ min n i = 1 n z s i
converges to z t in probability as n when E ( c o s 2 δ i ) 0 .
Proof of Theorem 2. 
As n ,
1 n Y n Y n μ min = E ( cos ( 2 δ i ) ) n i = 1 n cos θ i cos θ i ,
1 n Y n F n μ min n i = 1 n z s i = E ( cos ( 2 δ i ) ) n i = 1 n cos θ i cos θ i z t .
From (40) and (41), Theorem 2 stands when E ( cos 2 δ i ) 0 . □
From (28) and (39), the 3D bias compensation estimator result is T ^ b c = t ^ b c ; z ^ b c .

3.3. BC-WIV Estimator in 3D Space

To improve the accuracy, the researchers often apply the weighted instrumental variable algorithm by constructing instrumental variable matrix G w i v and weighted matrix W w i v ,
G w i v = sin ϕ ¯ 1 cos ϕ ¯ 1 sin ϕ ¯ n cos ϕ ¯ n , W w i v = | | t s 1 | | 2 0 0 0 0 0 0 | | t s n | | 2 ,
where ϕ ¯ i is calculated from 2D PLE estimator via (8) and
ϕ ¯ i = tan 1 t ^ p l e ( 2 ) y s i t ^ p l e ( 1 ) x s i .
Then the 2D weighted instrumental variable estimator can be expressed as:
t ^ w i v = ( G w i v W w i v 1 A n ) 1 G w i v W w i v 1 B n .
Similarly, the accuracy of bias compensation method in 2D space can be improved by constructing matrix,
G b c = sin ϕ 1 cos ϕ 1 sin ϕ n cos ϕ n , W b c = | | t ^ b c s 1 | | 2 0 0 0 0 0 0 | | t ^ b c s n | | 2 ,
where
ϕ i = tan 1 t ^ b c ( 2 ) y s i t ^ b c ( 1 ) x s i .
The 2D bias compensation weighted instrumental variable estimator is:
t ^ b c w i v = ( G b c W b c 1 A n ) 1 G b c W b c 1 B n .
Here
1 n G b c W b c 1 A n E ( cos e i ) n i = 1 n sin 2 ϕ i | | t ^ b c s i | | 2 sin ϕ i cos ϕ i | | t ^ b c s i | | 2 sin ϕ i cos ϕ i | | t ^ b c s i | | 2 cos 2 ϕ i | | t ^ b c s i | | 2 ,
1 n G b c W b c 1 B n E ( cos e i ) n i = 1 n sin 2 ϕ i x s i sin ϕ i cos ϕ i y s i | | t ^ b c s i | | 2 sin ϕ i cos ϕ i x s i + cos 2 ϕ i y s i | | t ^ b c s i | | 2
in probability as n . Hence t ^ b c w i v is an asymptotically unbiased estimator. Its covariance
E ( ( t ^ b c w i v t ) ( t ^ b c w i v t ) ) 1 n G b c W b c 1 G b c 1 1 n 2 G b c W b c 1 E ( ( B n A n t ) ( B n A n t ) ) W b c 1 G b c 1 n G b c W b c 1 G b c 1 E ( sin 2 e i ) n 1 n G b c W b c 1 G b c 1 C R L B
when E ( sin 2 e i ) is small as n . In z-axis direction, by constructing matrix
G z b c = cos θ ¯ 1 cos θ ¯ n , W z b c = | | T ^ b c S 1 | | 2 0 0 0 0 0 0 | | T ^ b c S n | | 2 ,
where
θ ¯ i = tan 1 z ^ b c z s i r ^ b c w i v ,
r ^ b c w i v = t ^ b c w i v ( 1 ) x s i 2 + t ^ b c w i v ( 2 ) y s i 2 .
The bias compensation weighted instrumental variable estimator in z axis is:
z ^ b c w i v = ( G z b c W z b c 1 Y n ) 1 G z b c W z b c 1 F n .

4. Sensor Trajectory Design

In this section we propose the one-step-ahead sensor maneuvering strategies for S k + 1 ( n k 3 ) based on S 1 , , S k and discuss the properties of such strategies. It is worth pointing out beforehand that S 1 , , S 3 are prefixed for motion model initialization and do not need to be designed.

4.1. Constraints

As shown in Figure 2, sensor trajectory planning is implemented under the following constraints [17,29]:
Target safety area: In order to ensure no collision between the sensor and the target, the distance between S k + 1 and T ^ k should be greater than or equal to the safe distance R.
Sensor motion area: Since the motion of the sensor is limited by the maximum speed, the sensor motion area is set to be spherical centering at S k with the radius of r (including the boundary of the sphere). Their boundaries are governed by polynomial equations f t s b ( x s k + 1 , y s k + 1 , z s k + 1 ; T ^ k ) = 0 and f s m b ( x s k + 1 , y s k + 1 , z s k + 1 ; S k ) = 0 , where
f t s b ( x s k + 1 , y s k + 1 , z s k + 1 ; T ^ k ) = | | S k + 1 T ^ k | | 2 R 2 ,
f s m b ( x s k + 1 , y s k + 1 , z s k + 1 ; S k ) = | | S k + 1 S k | | 2 r 2 .

4.2. Sensor Trajectory Planning Strategies

The position of sensor is deployed to improve the subsequent target positioning accuracy. According to [19], define:
M 1 = 1 σ 2 i = 1 k + 1 1 | | t s i | | 2 sin 2 ϕ i 1 2 sin 2 ϕ i 0 1 2 sin 2 ϕ i cos 2 ϕ i 0 0 0 0 ,
M 2 = 1 ξ 2 i = 1 k + 1 1 | | T S i | | 2 sin 2 θ i cos 2 ϕ i 1 2 sin 2 θ i sin 2 ϕ i 1 2 sin 2 θ i cos ϕ i 1 2 sin 2 θ i sin 2 ϕ i sin 2 θ i sin 2 ϕ i 1 2 sin 2 θ i sin ϕ i 1 2 sin 2 θ i cos ϕ i 1 2 sin 2 θ i sin ϕ i cos 2 θ i ,
where the matrix M 1 represents the azimuth FIM over k + 1 measurements, as it corresponds to the 2D projection of the localization problem onto x y plane, and the matrix M 2 represents the elevation FIM over k + 1 measurements. Therefore, the FIM over k + 1 measurements in 3D space is:
F I M k + 1 = M 1 + M 2 .
Lemma 3.
det ( F I M k + 1 ) = { p 1 , p 2 , p 3 } 2 ( k + 1 ) 3 det ( V p 1 + V p 2 + V p 3 ) ,
where the set { p 1 , p 2 , p 3 } is the combination selected from { 1 , , 2 ( k + 1 ) } .
V 2 i 1 and V 2 i ( i = 1 , 2 , 3 , . . . , k + 1 ) are the azimuth FIM and the elevation FIM, respectively.
Proof of Lemma 3. 
By the definition of matrix determinant, we have:
det ( F I M k + 1 ) = { p 1 , p 2 , p 3 } 2 ( k + 1 ) 3 det ( V p 1 + V p 2 + V p 3 ) + Ψ 1 { p 4 , p 5 } 2 ( k + 1 ) 2 det ( V p 4 + V p 5 ) + Ψ 2 { p 6 } 2 ( k + 1 ) 1 det ( V p 6 ) ,
where the set { p 4 , p 5 } and { p 6 } are the combinations selected from { 1 , , 2 ( k + 1 ) } . Ψ 1 and Ψ 2 are constant factors decided by the combination numbers. Under r a n k ( V p 4 + V p 5 ) 2 and rank ( V p 6 ) = 1 , det ( V p 4 + V p 5 ) = det ( V p 6 ) = 0 . Then (61) turns into (60). □
Theorem 3.
det ( F I M k + 1 ) = det ( F I M k ) + { p 7 , p 8 } 2 k 2 det ( V 2 k + 1 + V p 7 + V p 8 )
+ { p 9 , p 10 } 2 k + 1 2 det ( V 2 ( k + 1 ) + V p 9 + V p 10 ) ,
where the set { p 7 , p 8 } and { p 9 , p 10 } are the combination selected from { 1 , , 2 k } and { 1 , , 2 ( k + 1 ) } .
Proof of Theorem 3. 
From Lemma 3,
det ( F I M k ) = { p 11 , p 12 , p 13 } 2 k 3 det ( V p 11 + V p 12 + V p 13 ) ,
where the set { p 11 , p 12 , p 13 } is the combination selected from { 1 , 2 k } .
det ( F I M k + 1 ) det ( F I M k ) = { p 7 , p 8 } 2 k 2 det ( V 2 k + 1 + V p 7 + V p 8 )
+ { p 9 , p 10 } 2 k + 1 2 det ( V 2 ( k + 1 ) + V p 9 + V p 10 )
naturally follows. □
Let the location S k + 1 on the trajectory is generated pointwisely through the maximization of
max S 4 , S n det ( F I M n ) k = 4 n max S k + 1 ( det ( F I M k + 1 ) det ( F I M k ) ) + det ( F I M k ) ,
which shows the trajectory is suboptimal. According to Theorem 3, it is still a reasonable choice because the design of S k + 1 considers information given by S k + 1 and all the past sensor locations jointly. Defining
f ( S k + 1 ) = det ( F I M k + 1 ) det ( F I M k )
as the objective function of solving S k + 1 . Formally defining
Trajectory 1.
the location S k + 1 on the trajectory is generated pointwisely through the maximization of
max S k + 1 f ( S k + 1 )
s . t . | | S k + 1 S k | | r ,
| | S k + 1 T k | | R
with respect to S k + 1 given S 1 , , S k ( k 3 ) .
When optimizing f ( S k + 1 ) , unknown T, ϕ i , θ i , σ 2 and ξ 2 ( i = 1 , , k ) are replaced by the estimated T ^ k , ϕ ^ i ( T ^ k ) and θ ^ i ( T ^ k ) , σ ^ 2 ( T ^ k ) and ξ ^ 2 ( T ^ k ) . This method updates target location and re-estimates all the historical data in each step. To circumvent this weakness, we present
Trajectory 2.
the location S k + 1 on the trajectory is generated pointwisely through the maximization of
max S k + 1 f ^ ( S k + 1 )
s . t . | | S k + 1 S k | | r ,
| | S k + 1 T k | | R
with respect to S k + 1 given S 1 , , S k ( k 3 ) .
When optimizing f ^ ( S k + 1 ) , σ 2 and ξ 2 are replaced by ( i = 1 k σ ^ i 2 ( T ^ i 1 ) + σ ^ k + 1 2 ( T ^ k ) ) / ( k + 1 ) and ( i = 1 k ξ ^ i 2 ( T ^ i 1 ) + ξ ^ k + 1 2 ( T ^ k ) ) / ( k + 1 ) , respectively. Distinct with trajectory 1, ϕ ^ i and θ ^ i ( i = 1 , , k ) are not updated by T ^ k .
In Trajectory 1 and 2, the computation time to generate the sensor position at the ( k + 1 ) th moment includes the time to obtain the objective function and the time to optimize the objective function. Trajectory 1 needs ( k + 1 ) (49 additions/multiplications, three square roots, nine trigonometric function calculations) and 18 additions/multiplications for the calculation of det ( F I M k + 1 ) to obtain the objective function, while trajectory 2 requires only 49 additions/multiplications, three square roots, nine trigonometric functions and 18 additions/multiplications for the calculation of det ( F I M k + 1 ) . The generation of trajectory 2 reduces k ( 49 additions/multiplications, three square root, nine trigonometric function calculations) calculations compared to trajectory 1. In addition, the objective function structure is similar, the optimization time of the objective function is almost equal depending on the optimization method.

4.3. Optimal Solution Region

Through the analysis of (62), we have:
Theorem 4.
det ( F I M k + 1 ) det ( F I M k ) = c i { p 7 , p 8 , p 9 } c i { p 13 , p 14 } j i { 1 , 3 } i = 1 3 s g n ( j ) L i , j i ( c i ) ,
where the set { p 7 , p 8 , p 9 } is the combination selected from { 1 , 2 ( k + 1 ) } with p j ( 7 j 9 ) { 2 k + 1 , 2 ( k + 1 ) } . The set { p 13 , p 14 } is the combination selected from { p 7 , p 8 , p 9 } . j represents the permutation of { j 1 , j 2 , j 3 } . The sign function s g n ( j ) = 1 when the number of exchanges that { j 1 , j 2 , j 3 } takes to be the standard order { 1 , 2 , 3 } is odd. Otherwise, s g n ( j ) = 1 . L i , j i ( c i ) is the ith row and j i th column element in V c i .
Proof of Theorem 4. 
From Theorem 3 and the Leibniz formula,
det ( F I M k + 1 ) det ( F I M k ) = c i { p 7 , p 8 , p 9 } j i { 1 , 3 } i = 1 3 s g n ( j ) L i , j i ( c i ) .
In addition,
{ p 13 , p 14 } det ( V p 13 + V p 14 ) = c i { p 13 , p 14 } j i { 1 , 3 } i = 1 3 s g n ( j ) L i , j i ( c i ) = 0 .
Deducting (77) from (76) leads to (75). □
From Theorem 4, (62) can be rewritten uniformly into:
f ( S k + 1 ) = h = 0 4 q = 0 4 l = 0 2 α h q l x s k + 1 h y s k + 1 q z s k + 1 l h + = 0 10 q + = 0 10 l + = 0 6 β h + q + l + x s k + 1 h + y s k + 1 q + z s k + 1 l + ,
where the coefficient α p q l and β p + q + l + can be zero. According to Theorem 4, when the terms 2 k + 1 and 2 ( k + 1 ) in (75) are selected in set { p 7 , p 8 , p 9 } , the numerator and denominator of f ( S k + 1 ) have the highest order with respect to x, y and z. Since V 2 k + 1 and V 2 ( k + 1 ) contain x, y and z, when only the term 2 ( k + 1 ) in (75) is selected in set { p 7 , p 8 , p 9 } , neither of the numerator and denominator in (13) has the highest order. To make V 2 k + 1 and V 2 ( k + 1 ) in matrix form with x s k + 1 , y s k + 1 , z s k + 1 , we get the expression:
V 2 k + 1 = 1 σ k + 1 2 1 | | t ^ k s k + 1 | | 2 ( y s k + 1 T ^ k ( 2 ) ) 2 | | t ^ k s k + 1 | | 2 ( x s k + 1 T ^ k ( 1 ) ) ( y s k + 1 T ^ k ( 2 ) ) | | t ^ k s k + 1 | | 2 0 ( x s k + 1 T ^ k ( 1 ) ) ( y s k + 1 T ^ k ( 2 ) ) | | t ^ k s k + 1 | | 2 ( x s k + 1 T ^ k ( 1 ) ) 2 | | t ^ k s k + 1 | | 2 0 0 0 0 ,
V 2 ( k + 1 ) = 1 ξ k + 1 2 1 | | T ^ k S k + 1 | | 2 ( z s k + 1 T ^ k ( 3 ) ) 2 ( x s k + 1 T ^ k ( 1 ) ) 2 | | T ^ k S k + 1 | | 2 | | t ^ k s k + 1 | | 2 ( z s k + 1 T ^ k ( 3 ) ) 2 ( x s k + 1 T ^ k ( 1 ) ) ( y s k + 1 T ^ k ( 2 ) ) | | T ^ k S k + 1 | | 2 | | t ^ k s k + 1 | | 2 ( z s k + 1 T ^ k ( 3 ) ) ( x s k + 1 T ^ k ( 1 ) ) | | T ^ k S k + 1 | | 2 ( z s k + 1 T ^ k ( 3 ) ) 2 ( x s k + 1 T ^ k ( 1 ) ) ( y s k + 1 T ^ k ( 2 ) ) | | T ^ k S k + 1 | | 2 | | t ^ k s k + 1 | | 2 ( z s k + 1 T ^ k ( 3 ) ) 2 ( y s k + 1 T ^ k ( 2 ) ) 2 | | T ^ k S k + 1 | | 2 | | t ^ k s k + 1 | | 2 ( z s k + 1 T ^ k ( 3 ) ) ( y s k + 1 T ^ k ( 2 ) ) | | T ^ k S k + 1 | | 2 ( z s k + 1 T ^ k ( 3 ) ) ( x s k + 1 T ^ k ( 1 ) ) | | T ^ k S k + 1 | | 2 ( z s k + 1 T ^ k ( 3 ) ) ( y s k + 1 T ^ k ( 2 ) ) | | T ^ k S k + 1 | | 2 | | t ^ k s k + 1 | | 2 | | T ^ k S k + 1 | | 2 . .
Take the highest order of x s k + 1 as an example. Only when the item ( x s k + 1 T ^ k ( 1 ) ) 2 / | | t ^ k
s k + 1 | | 2 in V 2 k + 1 and the item ( z s k + 1 T ^ k ( 3 ) ) 2 ( x s k + 1 T ^ k ( 1 ) ) 2 / | | T ^ k S k + 1 | | 2 | | t ^ k s k + 1 | | 2 in V 2 ( k + 1 ) are selected, the numerator and denominator of f ( S k + 1 ) have the highest order with respect to x, which are the 4th order and the 10th order, respectively.
Ω ¯ = Ω Ω is a compact and connected set in 3D space which Ω stands for all internal points of the feasible area, Ω represents the boundary of the feasible area. When the target point T ^ k is outside the sensor motion area, make tangent lines from point T ^ k to the sensor motion area and all tangent points can form a circle. Denote the cone formed by vertex T ^ k and the circle as T A B in Figure 3 (the tangent points A and B are symmetrically distributed).
In order to solve the proposed optimization problem quickly, according to the relative position of the sensor motion area and the target safety area, the feasible region of the sensor trajectory optimization problem can be analyzed in following cases.
  • When the sensor motion area and the target safety area are separated, define the boundary part of the sensor motion area inside T A B as surface 1 in Figure 3a. The feasible region is the sensor motion area;
  • When the sensor motion area intersects with the target safety area, the feasible region is the sensor motion area where the part inside the target safety area is excluded, and the surface is where part of the target safety boundary is inside the sensor motion area. According to the position of T ^ k , it can be divided into three sub-cases: (i) when T ^ k is outside the sensor motion area and T A B is not contained by the target safety area, define the part of the sensor motion boundary inside T A B with the section inside the target safety area replaced by the section of the target safety boundary inside the sensor movement area, such as surface 2 in Figure 3b. E and F (E and F are symmetrically distributed) are the points where the two spheres intersect; (ii) when the T ^ k is outside the sensor motion area and if T A B is contained by the target safety area, define the part of the target safety boundary inside the sensor motion area as surface 3 in Figure 3c; (iii) when the T ^ k is inside the sensor motion area, define the part of the target safety boundary inside the sensor motion area as surface 4 in Figure 3d;
  • when the sensor motion area contains the target safety area, the feasible region is the sensor motion area where the interior section of the target safety area is excluded. The target safety boundary is defined as surface 5 in Figure 3e.
Proposition 1.
In case 1, the maximizer of the f ( S k + 1 ) exists on the spherical crown (surface 1). In case 2(i), the maximizer of the f ( S k + 1 ) exists on surface 2 where the two areas intersect in the sensor motion area, and the side of the circular truncated cone (lower bottom circle with A B as the diameter and upper bottom circle with E F as the diameter). In cases 2(ii) and 2(iii), the maximizer of the f ( S k + 1 ) exists on surfaces 3, 4 where the two areas intersect. In case 3, the maximizer of f ( S k + 1 ) is on the boundary of the target safety area (surface 5).
Proof of Proposition 1. 
Regardless of the relative location of the two areas, a straight blue dotted line is drawn between T ^ k and any point in the feasible region; all the points on this line have the same ϕ ^ k + 1 ( T ^ k ) and θ ^ k + 1 ( T ^ k ) as shown. Therefore the maximizer of f ( S k + 1 ) on the line segment is the closest one D * to T ^ k in the Euclidean sense. Gathering up these maximizers gives different surfaces such as the red part in Figure 3. □

4.4. Analytical Derivation for the Global Maximizer

According to (13), the partial derivative of f ( S k + 1 ) over x s k + 1 , y s k + 1 and z s k + 1 can be written into:
f ( S k + 1 ) x s k + 1 = h = 0 13 q = 0 14 l = 0 8 α h q l 1 ( y s k + 1 q , z s k + 1 l ) x s k + 1 h h + = 0 20 q + = 0 20 l + = 0 12 β h + q + l + + x s k + 1 h + y s k + 1 q + z s k + 1 l + ,
f ( S k + 1 ) y s k + 1 = h = 0 14 q = 0 13 l = 0 8 α h q l 2 ( y s k + 1 q , z s k + 1 l ) x s k + 1 h h + = 0 20 q + = 0 20 l + = 0 12 β h + q + l + + x s k + 1 h + y s k + 1 q + z s k + 1 l + ,
f ( S k + 1 ) z s k + 1 = h = 0 14 q = 0 14 l = 0 7 α h q l 3 ( y s k + 1 q , z s k + 1 l ) x s k + 1 h h + = 0 20 q + = 0 20 l + = 0 12 β h + q + l + + x s k + 1 h + y s k + 1 q + z s k + 1 l + ,
where the coefficient α h q l 1 , α h q l 2 , α h q l 3 and β h + q + l + + can be zero. From Proposition 1, the feasible region always lies on the area boundary. Take the case when the two areas are separated as an example. f ( S k + 1 ) / x s k + 1 f ( S k + 1 ) / y s k + 1 f ( S k + 1 ) / z s k + 1 should be colinear with the normal vector x s k + 1 x s k y s k + 1 y s k z s k + 1 z s k at the optimal, i.e.,
f ( S k + 1 ) x s k + 1 ( y s k + 1 y s k ) f ( S k + 1 ) y s k + 1 ( x s k + 1 x s k ) = 0 ,
f ( S k + 1 ) x s k + 1 ( z s k + 1 z s k ) f ( S k + 1 ) z s k + 1 ( x s k + 1 x s k ) = 0 .
Otherwise, any infinitesimal move along the tangent vector that constructs an obtuse angle with f ( S k + 1 ) / x s k + 1 f ( S k + 1 ) / y S k + 1 f ( S k + 1 ) / z S k + 1 increases the value of f ( S k + 1 ) , (84) is equivalent to:
h = 0 15 g h ( y s k + 1 , z s k + 1 ) x s k + 1 h = 0 ,
where g h ( y s k + 1 , z s k + 1 ) is the polynomial of y s k + 1 , z s k + 1 . Define η ( y s k + 1 , z s k + 1 ) = ( y s k + 1 y s k ) 2 + ( z s k + 1 z s k ) 2 + x s k 2 r 2 . The movement boundary condition is rewritten into:
η ( y s k + 1 , z s k + 1 ) 2 x s k x s k + 1 + x s k + 1 2 = 0 .
The joint solution of (84) and (87) can be developed as follows. The ( 15 + 2 ) × ( 15 + 2 ) Sylvester matrix of (84) and (87) with respect to y s k + 1 , z s k + 1 is:
S y l 1 ( ( 84 ) , ( 87 ) ; y s k + 1 , z s k + 1 ) = g 15 ( y s k + 1 , z s k + 1 ) 0 1 0 . . . 0 g 14 ( y s k + 1 , z s k + 1 ) g 15 ( y s k + 1 , z s k + 1 ) 2 x s k 1 . . . 0 g 13 ( y s k + 1 , z s k + 1 ) g 14 ( y s k + 1 , z s k + 1 ) η ( y s k + 1 , z s k + 1 ) 2 x s k . . . 0 g 1 ( y s k + 1 , z s k + 1 ) g 2 ( y s k + 1 , z s k + 1 ) 0 0 . . . 1 g 0 ( y s k + 1 , z s k + 1 ) g 1 ( y s k + 1 , z s k + 1 ) 0 0 . . . 2 x s k 0 g 0 ( y s k + 1 , z s k + 1 ) 0 0 . . . η ( y s k + 1 , z s k + 1 ) .
The Sylvester resultant det ( S y l 1 ( ( 84 ) , ( 87 ) ; y s k + 1 , z s k + 1 ) ) is a bivariate polynomial with an upper bound of the 58th order for y s k + 1 and the 46th order for z s k + 1 .
Similarly, the joint solution of (85) and (87) can be developed as the ( 15 + 2 ) × ( 15 + 2 ) Sylvester matrix of (85) and (87) is similar to (88). det ( S y l 2 ( ( 85 ) , ( 87 ) ; y s k + 1 , z s k + 1 ) ) is also a bivariate polynomial with an upper bound of the 58th order for y s k + 1 and the 46th order for z s k + 1 . The real solution can be obtained by solving the equations given by the determinants of the two Sylvester matrices.
Comparing only the real solutions lying on surface 1 in terms of f ( S k + 1 ) gives the global maximizer. For surface 2, the real solutions of the surface equation and (86) are compared with the two intersection points to generate the global maximizer. For surfaces 3 and 4, the real solutions of the surface equation and (86) on the surface are compared with the two intersection points to provide S k + 1 . When the movement area contains the safety area, the real solutions of the safety constraint and the (86) that has the biggest f ( S k + 1 ) are the global maximizer.
Remark 1.
The position of a sensor should not be collinear with the target position in the z-axis direction so that the target is observable. It is necessary to set the cylindrical as a non-flying area in Figure 4. It means to insert a cylinder centering at the target position with the radius of ρ in the target safety area, and the z-axis direction extends to infinity. The constraint is formulated as:
( x k + 1 T ^ k ( 1 ) ) 2 + ( y k + 1 T ^ k ( 2 ) ) 2 ρ 2 .
If the maximum solution of f ( S k + 1 ) falls into the section where the surface intersects the cylinder, we choose the other one which leads to the second largest f ( S k + 1 ), and so forth. If all the maximum solutions of f ( S k + 1 ) fall within the intersection section, a point is randomly selected from the feasible region as the solution of f ( S k + 1 ).

4.5. Single Sensor Trajectory Design Procedure

We summarize the sensor trajectory design procedure in the following algorithm.
Trajectory design based localization algorithm:
Initialization: Use the initial position of the sensor to generate the initial estimator parameter by localization algorithm, e.g., Kalman-related methods.
Repeat until n is reached:
  • Generate S k + 1 through the maximization of (69) or (72). The maximizer of (69) or (72) can be produced in an analytical way described in Section 4.4;
  • Gather the measurement information ϕ ^ k + 1 , θ ^ k + 1 , σ ^ 2 and ξ ^ 2 from the sensor at S k + 1 ;
  • Use localization technique to generate T ^ k + 1 ;
  • k = k + 1 .
Output the target location.
In many applications, various location-based services need to determine accurate positioning information, such as for public safety services, field rescue and other fields. It is necessary to track the target under motion constraints. Unlike static sensors, which have a fixed density and perception range, mobile sensors can cover a larger area over time without increasing their number. In addition, their spatial distribution can change dynamically to accommodate the movement of the target, thus providing measurements of information about its position. Choosing the best sensing location is particularly important, especially given time-constrained applications such as tracking hostile targets. For example, autonomous underwater vehicles (AUV) are underwater unmanned autonomous vehicle platforms without a cable connection that can be applied for underwater environment monitoring, offshore oil engineering operations, underwater search and mapping after loading appropriate sensors. In the main-slave UUVS (Unmanned Underwater Vehicles) system, the slave-UUV only carries the azimuth information of the main-UUV under a cluttered environment and realizes the convergence with the main-UUV as soon as possible through autonomous optimization of its movement trajectory. The localization method and trajectory planning method proposed in this paper can be applied to the above scenarios.

5. Simulation

In this section, the bias compensation method and sensor trajectory planning algorithm are compared with other benchmarks through MATLAB R2020a simulation. The location unit used for the following experiments is a metre.

5.1. Example 1

We now explore the positioning performance of the proposed method and estimate the deviation through simulation. The simulation is performed in 3D space, where the target is assumed to be at position T = 30 40 50 m. The sensor collects N = 100 azimuth/elevation angle measurements, which ensure the target is observable. The initial sensor is located at S 1 = 0 0 0 m, the motion velocity components of the sensor in x, y and z directions are 0.5 m/s, 0.8 m/s, 1 m/s, respectively. The bearing angle measurement noise standard deviations σ = ξ are set in between 0.01 and 0.08 radian with the variance difference of 0.01 radian 2 . One thousand independent Monte Carlo experiments are run. We use the mean square error (MSE) and bias norm as evaluation indicators. The MSE and bias norm are defined as:
M S E = i = 1 1000 j = 1 3 ( T ^ i ( j ) T ( j ) ) 2 1000 ,
b i a s = i = 1 1000 j = 1 3 | T ^ i ( j ) T ( j ) | 1000 ,
respectively. The first simulation is about the MSE and bias norm comparison between the proposed BC method and the original 3D localization method (PLE) in [15]. As shown in Figure 5 and Figure 6, the BC method has lower MSE and bias norm compared to the PLE method. In addition, the MSE and bias norm increase as measurement noise level increases. The positioning accuracy of the two methods is significantly improved by using the weighted instrumental variable method. The PLE-WIV estimator, as expected, suffers from more serious bias than the BC-WIV estimator.
The second simulation based on positioning methods is carried out with the increase of sensor number. The bearing angle measurement noise standard deviation is assumed to be 0.001 radian. The number of known sensors is changed from 60 to 260 with a difference of 40. The position of sensors is set as x s i ( 0 50 ) m, y s i ( 0 50 ) m, z s i ( 0 50 ) m ( i = 1 , , n ) .
Under the same conditions of each experiment, as shown in Figure 7 and Figure 8, the BC and the BC-WIV methods still show the best positioning performance. The MSE and bias norm decrease when the number of sensors increases. However, there can be low positioning accuracy occasionally when the number of sensors is large. This is because in the process of the random generation of sensors, the position of some sensors makes the target unobservable, resulting in low positioning accuracy.

5.2. Example 2

The third simulation is about sensor trajectory planning. For fairness and completeness, we use the pseudolinear Kalman filter method in [30] for target localization as the platform to compare the performance of trajectories. One fitness function,
L ( T ^ ) = 100 ( 1 | | T T ^ | | 2 | | T | | 2 ) ,
in [31] is used to demonstrate the performance of each estimator over n runs. It could even be much less than zero when the estimators deviate far from the real values. To downplay the effect of the extremum points, L ( T ^ ) is modified into:
L ´ ( T ^ ) = max ( 0 , 100 ( 1 | | T T ^ | | 2 | | T | | 2 ) ) .
Trajectory 3.
The location S k + 1 on the trajectory is generated through the maximization of det( FIM k + 1 )-det( FIM k 2 ) with respect to S k + 1 recursively. The design of S k + 1 is given by the ( k + 1 ) th measurement jointly with its two temporal nearest predecessors.
Trajectory 4.
The location S k + 1 on the trajectory is generated through the minimization of the state estimate covariance.
The four trajectories planning methods are simulated and compared. The positions of the target and sensors are located at T = 20 35 55 m, S 1 = 10 25 20 m, S 2 = 30 20 40 m, S 3 = 70 50 15 m, S 4 = 12 80 70 m, S 5 = 59 22 18 m, S 6 = 120 150 90 m, respectively. For each test, the four trajectories share the same six initial sensor locations. The noise of the angle measurements are Gaussian independently distributed with zero mean and the standard deviation is 0.1 radian. The anti-collision distance and the maximum distance of sensor movement are set to R = 5 m , r = 10 m , respectively. The radius of the cylindrical non-flying area is set to ρ = 0.5 R . One hundred independent Monte Carlo experiments are run.
Performance of the four trajectories is measured and displayed at two observation points when k = 10 and k = 15 . Table 1 shows that Trajectory 1 and 2 contribute to greater det ( F I M k + 1 ) in over 90 % of tests where trajectory 4 is used as a benchmark. This is consistent with the numerical results given in Table 2 where Trajectory 1 has the highest L ´ ( T ^ ) at k = 10 and k = 15 . Compared with Trajectory 3 and 4, the target estimations are more accurate under Trajectory 1 and 2 as the corresponding L ´ ( T ^ ) values are closer to 100. The accuracy of target localization is related to det ( F I M k ) . The larger the det ( F I M k ) , the higher the positioning accuracy.
From Theorem 3, det ( F I M k + 1 ) det ( F I M k ) for trajectory 1 and det ( F I M k + 1 ) det ( F I M k 2 ) for Trajectory 3 both are parts of det ( F I M k + 1 ) . Since Trajectory 3 only utilizes the sensor location of the past two moments, det ( F I M k + 1 ) det ( F I M k 2 ) weighs less in det ( F I M k + 1 ) than det ( F I M k + 1 ) det ( F I M k ) . The maximizer of det ( F I M k + 1 ) det ( F I M k 2 ) is less likely to be close to the global optimizer of det ( F I M k + 1 ) . Under such circumstances, Trajectory 3 contributes a smaller information increment than Trajectories 1 and 2. Trajectory 4 is achieved by minimizing the trace of the state estimation covariance matrix. Due to the inverse relation between the covariance and FIM, the generation of Trajectory 4 also utilizes more past information than Trajectory 3. However, compared with Trajectories 1 and 2, not all the information is used.
Comparison of the mean of det ( F I M k ) for four trajectories can be seen intuitively in Figure 9. When the sampling step increases, det ( F I M k ) of the proposed method increases more. In addition, more det ( F I M k ) can be obtained when the angles are updated in Trajectory 1 at the cost of increasing the calculation amount. Less computation time is consumed by generating Trajectory 2. When the measurement noise is small, the difference of the determinant obtained by Trajectories 1 and 2 is subtle as seen in the inner plot of Figure 9.
The trajectories simulation results are shown in Figure 10. In the process of trajectory planning, the predicted sensor position keeps getting closer to the target. When the sensor motion area is separated from the target safety area, S k + 1 sits on the movement boundary sphere centering at S k .
As shown in Figure 10, Trajectories 1 and 2 approach the target much faster because the trajectory planning method proposed in this paper takes into account sensor position information at all past moments and have higher FIM determinant. Since Trajectories 3 and 4 correspond to lower and mild FIM determinants, their curves reach the target in a slow and mild speed, respectively, amongst four trajectories; although it seems that Trajectory 3 distances itself from the target.
For each scenario, the whole computation complexity includes the following parts: initialization, Kalman-related localization, objective function construction and optimization. The average runtime for each trajectory is given in Table 3 to describe the complexity.
In Table 3, it can be seen that the time consumed for Trajectory 2 is the shortest. The generation of Trajectory 2 can effectively reduce the computational complexity in the optimization of (62) in return for no update of the bearing angles. The generation of Trajectory 3 takes a longer time since the objective function is different from Trajectories 1 and 2. For Trajectory 4, inverse operation of Fisher information matrix results in a longer calculation time.

6. Conclusions and Future Works

This paper studies the bearing-only localization from the improvement of target localization method and the reasonable planning of a sensor trajectory strategy. Under the condition of Gaussian noise, a potential relationship exists between the determinant of the extended coefficient matrix and the noise variance in the angle measurement equation. When the determinant of the extended coefficient matrix is zero, the variance of zero-mean Gaussian noise can be derived. Based on this estimated variance, the bias compensation localization of the target on the 3D space is realized. The estimator can be further refined by the BC-WIV method. In addition, a trajectory planning algorithm encapsulating two trajectory planning strategies are proposed to improve localization accuracy under several constraints. The next moment sensor position is given by optimizing quantified FIM determinant increment as the objective function that evaluates all the historical measurements. The optimal solution always lies on the constraint boundary. Simulation results show that Trajectory 1 leads to a higher localization accuracy at the cost of more computation time while Trajectory 2 has a comparable result with a mild computation burden.
In the future, we plan to extend our current approach to multiple sensors localization. The research work in this paper is only carried out at the theoretical level, the next research will test the method proposed in the practical communication environment. A team of three Pioneer robots is expected to be deployed in a rectangular area, with one Pioneer robot as the target and the other two Pioneers as tracking sensors to locate the target by the obtained bearing information.

7. Patents

Yiqun Zou, Bilu Gao, Xiafei Tang. 3D trajectory planning method, system, equipment and medium based on bearing-only measurements: 202111562286.9[P].2021.12.20.

Author Contributions

Conceptualization, Y.Z.; methodology, B.G.; software, B.G.; validation, Y.Z., B.G. and X.T.; formal analysis, Y.Z.; investigation, B.G.; resources, Y.Z.; data curation, Y.Z.; writing—original draft preparation, B.G.; writing—review and editing, B.G. and Y.Z.; visualization, X.T.; supervision, L.Y.; project administration, X.T.; funding acquisition, Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by National Natural Science Foundation of China (NSFC) [grant 61403427] and Hunan Provincial Natural Science Foundation of China [project 2020JJ5585 and project 2020JJ5777].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors are also grateful to Professor Erwei Bai in the university of Iowa for shedding light on some initial research in this paper.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
FIMFisher information matrix
AOAAngle of arrival
2DTwo dimensional
3DThree dimensional
MSEMean square error
IVInstrumental variable
CRLBCramér-Rao lower bound
UAVUnmanned aerial vehicle
BCBias compensation
BC-WIVBias compensation weighted instrumental variable
PLEPseudo-linear estimator

References

  1. Jatoi, M.A.; Kamel, N.; Malik, A.S.; Faye, I.; Begum, T. A survey of methods used for source localization using EEG signals-ScienceDirect. Biomed. Signal Process. Control 2014, 11, 42–52. [Google Scholar] [CrossRef]
  2. Dogançay, K. UAV Path Planning for Passive Emitter Localization. IEEE Trans. Aerosp. Electron. Syst. 2012, 48, 1150–1166. [Google Scholar] [CrossRef]
  3. Anup, P.; Takuro, S. Localization in Wireless Sensor Networks: A Survey on Algorithms, Measurement Techniques, Applications and Challenges. J. Sens. Actuator Netw. 2017, 6, 24. [Google Scholar]
  4. Bishop, A.N.; Fidan, B.; Anderson, B.D.; Doğançay, K.; Pathirana, P.N. Optimality analysis of sensor-target localization geometries. Automatica 2010, 46, 479–492. [Google Scholar] [CrossRef]
  5. Kim, Y.; Bang, H. Introduction to Kalman Filter and Its Applications. Introd. Implement. Kalman Filter 2018, 1, 1–16. [Google Scholar]
  6. Aidala, V.J. Kalman Filter Behavior in Bearings-Only Tracking Applications. IEEE Trans. Aerosp. Electron. Syst. 2007, 15, 29–39. [Google Scholar] [CrossRef]
  7. Aidala, V.; Hammel, S. Utilization of modified polar coordinates for bearings-only tracking. IEEE Trans. Autom. Control 1983, 28, 283–294. [Google Scholar] [CrossRef] [Green Version]
  8. Zhan, R.; Wan, J. Iterated Unscented Kalman Filter for Passive Target Tracking. IEEE Trans. Aerosp. Electron. Syst. 2007, 43, 1155–1163. [Google Scholar] [CrossRef]
  9. Nguyen, N.H.; Dogançay, K. 3D AOA Target Tracking with Two-step Instrumental-variable Kalman Filtering. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) IEEE, Brighton, UK, 12–17 May 2019; pp. 4390–4394. [Google Scholar]
  10. Dogançay, K.; Ibal, G. Instrumental Variable Estimator for 3D Bearings-Only Emitter Localization. In Proceedings of the 2005 International Conference on Intelligent Sensors, Sensor Networks and Information Processing, Melbourne, Australia, 5–8 December 2005; pp. 63–68. [Google Scholar]
  11. Nardone, S.C.; Lindgren, A.G.; Gong, K.F. Fundamental properties and performance of conventional bearings-only target motion analysis. IEEE Trans. Autom. Control 1984, 29, 775–787. [Google Scholar] [CrossRef]
  12. Badriasl, L.; Dogançay, K. Three-dimensional target motion analysis using azimuth/elevation angles. IEEE Trans. Aerosp. Electron. Syst. 2014, 50, 3178–3194. [Google Scholar] [CrossRef]
  13. Dogançay, K. Bias compensation for the bearings-only pseudolinear target track estimator. IEEE Trans. Signal Process. 2005, 54, 59–68. [Google Scholar] [CrossRef]
  14. Dogançay, K. 3D Pseudolinear Target Motion Analysis From Angle Measurements. IEEE Trans. Signal Process. 2015, 63, 1. [Google Scholar] [CrossRef]
  15. Adib, N.; Douglas, S.C. Extending the Stansfield Algorithm to Three Dimensions: Algorithms and Implementations. IEEE Trans. Signal Process. 2018, 66, 1106–1117. [Google Scholar] [CrossRef]
  16. Zhao, S.; Chen, B.M.; Lee, T.H. Optimal sensor placement for target localization and tracking in 2D and 3D. Int. J. Control 2015, 86, 1687–1704. [Google Scholar] [CrossRef]
  17. He, S.; Shin, H.S.; Tsourdos, A. Trajectory Optimization for Target Localization With Bearing-Only Measurement. IEEE Trans. Robot. 2019, 35, 653–668. [Google Scholar] [CrossRef] [Green Version]
  18. Dogançay, K.; Hmam, H. Optimal angular sensor separation for AOA localization. Signal Process. 2008, 88, 1248–1260. [Google Scholar] [CrossRef]
  19. Xu, S.; Dogançay, K. Optimal Sensor Placement for 3-D Angle-of-Arrival Target Localization. IEEE Trans. Aerosp. Electron. Syst. 2017, 53, 1196–1211. [Google Scholar] [CrossRef]
  20. Oshman, Y.; Davidson, P. Optimization of observer trajectories for bearings-only target localization. IEEE Trans. Aerosp. Electron. Syst. 1999, 35, 892–902. [Google Scholar] [CrossRef]
  21. Sheng, X.; Dogançay, K. Optimal sensor deployment for 3D AOA target localization. In Proceedings of the 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), South Brisbane, Australia, 19–24 April 2015; pp. 2544–2548. [Google Scholar]
  22. Sheng, X.; Dogançay, K.; Hmam, H. 3D pseudolinear Kalman filter with own-ship path optimization for AOA target tracking. In Proceedings of the 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Shanghai, China, 20–25 March 2016; pp. 3136–3140. [Google Scholar]
  23. Xu, S.; Dogançay, K.; Hmam, H. Distributed pseudolinear estimation and UAV path optimization for 3D AOA target tracking. Signal Process. 2017, 133, 64–78. [Google Scholar] [CrossRef]
  24. Nguyen, N.H.; Dogançay, K. Single-platform passive emitter localization with bearing and Doppler-shift measurements using pseudolinear estimation techniques. Signal Process. 2016, 125, 336–348. [Google Scholar] [CrossRef]
  25. Logothetis, A.; Isaksson, A.; Evans, R.J. Comparison of suboptimal strategies for optimal own-ship maneuvers in bearings-only tracking. In Proceedings of the 1998 American Control Conference, Philadelphia, PA, USA, 26 June 1998; Volume 6, pp. 3334–3338. [Google Scholar]
  26. Gavish, M.; Weiss, A.J. Performance analysis of bearing-only target location algorithms. IEEE Trans. Aerosp. Electron. Syst. 1992, 28, 817–828. [Google Scholar] [CrossRef]
  27. Dogançay, K. Bearings-only target localization using total least squares. Signal Process. 2005, 85, 1695–1710. [Google Scholar] [CrossRef]
  28. Bai, E. Parameter convergence in source localization. In Proceedings of the 2016 IEEE Conference on Control Applications (CCA), Buenos Aires, Argentina, 19–22 September 2016; pp. 929–933. [Google Scholar]
  29. Zhou, K. Roumeliotis S. Multirobot Active Target Tracking With Combinations of Relative Observations. IEEE Trans. Robot. 2011, 27, 678–695. [Google Scholar] [CrossRef]
  30. Bu, S.; Meng, A.; Zhou, G. A New Pseudolinear Filter for Bearings-Only Tracking without Requirement of Bias Compensation. Sensors 2021, 21, 5444. [Google Scholar] [CrossRef]
  31. Pillonetto, G.; Dinuzzo, F.; Chen, T.; De Nicolao, G.; Ljung, L. Kernel methods in system identification, machine learning and function estimation: A survey. Automatica 2014, 50, 657–682. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Target and sensor in R 3 .
Figure 1. Target and sensor in R 3 .
Applsci 12 06739 g001
Figure 2. Sensor motion area and target safety area.
Figure 2. Sensor motion area and target safety area.
Applsci 12 06739 g002
Figure 3. Surface 1 (red part in (a)) when the two areas are separated, surface 2 (red part in (b)), surface 3 (red part in (c)) and surface 4 (red part in (d)) when the two areas intersect, the target safety boundary (red part in (e)) when the target safety area is in the sensor movement area.
Figure 3. Surface 1 (red part in (a)) when the two areas are separated, surface 2 (red part in (b)), surface 3 (red part in (c)) and surface 4 (red part in (d)) when the two areas intersect, the target safety boundary (red part in (e)) when the target safety area is in the sensor movement area.
Applsci 12 06739 g003
Figure 4. Sensor motion area, target safety area and cylindrical non-flying area.
Figure 4. Sensor motion area, target safety area and cylindrical non-flying area.
Applsci 12 06739 g004
Figure 5. MSE with increasing noise standard deviation with a variance difference of 0.01 radian 2 for each method.
Figure 5. MSE with increasing noise standard deviation with a variance difference of 0.01 radian 2 for each method.
Applsci 12 06739 g005
Figure 6. Bias norm MSE with increasing noise standard deviation with the variance difference of 0.01 radian 2 for each method.
Figure 6. Bias norm MSE with increasing noise standard deviation with the variance difference of 0.01 radian 2 for each method.
Applsci 12 06739 g006
Figure 7. MSE with increasing sensor number at the difference of 40 for each method.
Figure 7. MSE with increasing sensor number at the difference of 40 for each method.
Applsci 12 06739 g007
Figure 8. Bias norm with increasing sensor number at the difference of 40 for each method.
Figure 8. Bias norm with increasing sensor number at the difference of 40 for each method.
Applsci 12 06739 g008
Figure 9. Comparison of the mean of det ( F I M k ) for four trajectories over 100 Monte Carlo runs.
Figure 9. Comparison of the mean of det ( F I M k ) for four trajectories over 100 Monte Carlo runs.
Applsci 12 06739 g009
Figure 10. Sensor trajectories in typical runs. The trajectories in other runs are similar to the ones in this run. Trajectories 1, 2, 3 and 4 are represented by circle, cross, upper triangle and square, respectively. The black sphere is the boundary of the target safety area. The target position represented by the five pointed star.
Figure 10. Sensor trajectories in typical runs. The trajectories in other runs are similar to the ones in this run. Trajectories 1, 2, 3 and 4 are represented by circle, cross, upper triangle and square, respectively. The black sphere is the boundary of the target safety area. The target position represented by the five pointed star.
Applsci 12 06739 g010
Table 1. Statistics that sum up the number of tests when det ( F I M k i ) det ( F I M k 4 ) > 0 ( i = 1 , 2 , 3 ) for trajectory i at two observations.
Table 1. Statistics that sum up the number of tests when det ( F I M k i ) det ( F I M k 4 ) > 0 ( i = 1 , 2 , 3 ) for trajectory i at two observations.
Observation Point det ( FIM k 1 ) det ( FIM k 4 ) > 0 det ( FIM k 2 ) det ( FIM k 4 ) > 0 det ( FIM k 3 ) det ( FIM k 4 ) > 0
k = 1098/10096/1003/100
k = 1599/10096/1002/100
Table 2. Mean values of L ´ ( T ^ ) for four trajectories at two observation points.
Table 2. Mean values of L ´ ( T ^ ) for four trajectories at two observation points.
Trajectory 1Trajectory 2Trajectory 3Trajectory 4
mean of L ´ ( T ^ ) ( k = 10 ) 91.1690.5889.6290.49
mean of L ´ ( T ^ ) ( k = 15 ) 93.4292.8688.8590.99
Table 3. Averaged runtimes for four trajectories.
Table 3. Averaged runtimes for four trajectories.
Trajectory 1Trajectory 2Trajectory 3Trajectory 4
runtime(s)0.55580.52400.69960.5899
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zou, Y.; Gao, B.; Tang, X.; Yu, L. Target Localization and Sensor Movement Trajectory Planning with Bearing-Only Measurements in Three Dimensional Space. Appl. Sci. 2022, 12, 6739. https://doi.org/10.3390/app12136739

AMA Style

Zou Y, Gao B, Tang X, Yu L. Target Localization and Sensor Movement Trajectory Planning with Bearing-Only Measurements in Three Dimensional Space. Applied Sciences. 2022; 12(13):6739. https://doi.org/10.3390/app12136739

Chicago/Turabian Style

Zou, Yiqun, Bilu Gao, Xiafei Tang, and Lingli Yu. 2022. "Target Localization and Sensor Movement Trajectory Planning with Bearing-Only Measurements in Three Dimensional Space" Applied Sciences 12, no. 13: 6739. https://doi.org/10.3390/app12136739

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop