Next Article in Journal
Altitude Measurement-Based Optimization of the Landing Process of UAVs
Next Article in Special Issue
Forecasting Nonlinear Systems with LSTM: Analysis and Comparison with EKF
Previous Article in Journal
An Optimum Deployment Algorithm of Camera Networks for Open-Pit Mine Slope Monitoring
Previous Article in Special Issue
Using Convolutional Neural Networks with Multiple Thermal Sensors for Unobtrusive Pose Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Sensor Fusion for Underwater Vehicle Localization by Augmentation of RBF Neural Network and Error-State Kalman Filter

1
Oceanic Engineering Research Institute, University of Malaga, 29010 Malaga, Spain
2
Department of Electrical and Computer Engineering, King Abdulaziz University, Jeddah 21589, Saudi Arabia
3
Center of Excellence in Intelligent Engineering Systems, King Abdulaziz University, Jeddah 21589, Saudi Arabia
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(4), 1149; https://doi.org/10.3390/s21041149
Submission received: 1 January 2021 / Revised: 29 January 2021 / Accepted: 1 February 2021 / Published: 6 February 2021
(This article belongs to the Special Issue Information Fusion and Machine Learning for Sensors)

Abstract

:
The Kalman filter variants extended Kalman filter (EKF) and error-state Kalman filter (ESKF) are widely used in underwater multi-sensor fusion applications for localization and navigation. Since these filters are designed by employing first-order Taylor series approximation in the error covariance matrix, they result in a decrease in estimation accuracy under high nonlinearity. In order to address this problem, we proposed a novel multi-sensor fusion algorithm for underwater vehicle localization that improves state estimation by augmentation of the radial basis function (RBF) neural network with ESKF. In the proposed algorithm, the RBF neural network is utilized to compensate the lack of ESKF performance by improving the innovation error term. The weights and centers of the RBF neural network are designed by minimizing the estimation mean square error (MSE) using the steepest descent optimization approach. To test the performance, the proposed RBF-augmented ESKF multi-sensor fusion was compared with the conventional ESKF under three different realistic scenarios using Monte Carlo simulations. We found that our proposed method provides better navigation and localization results despite high nonlinearity, modeling uncertainty, and external disturbances.

1. Introduction

The ocean floor has billions of dollars of natural resources in the form of precious elements and medicinal herbs. To take advantage of ocean resources, seabed mapping is the ultimate tool that depends on precise sensors and robust navigation fusion algorithms for autonomous underwater vehicles (AUVs) and remotely operated underwater vehicles (ROVs). Navigational accuracy is a key requirement for complex seabed mapping tasks [1]. However, primary sensors, three-axis gyros, and accelerometers have biases and drifts, which vary with time and are affected by noise. In contrast, most commonly used fusion algorithms based on extended Kalman filter (EKF) and its variant error-state Kalman (ESKF) suffer from divergence and degraded mean square error (MSE) performance in the nonlinear underwater condition because of linear approximation [2]. Thus, the higher the nonlinearity present in the system, the greater the error of the EKF state prediction, and it can also induce filter divergence.
Backpropagation multi-layer neural networks (BPNN) and radial basis neural networks (RBFNN) have excellent learning abilities and are well-known for their nonlinear system identification [3,4,5]. In comparison, the RBFNN has much greater accuracy of prediction and versatility in their choice of base functions [6]. Furthermore, they have fast convergence and less computation load compared to BPNN. These advantages of RBFNN lead to a major research question: Can we incorporate the strengths of the RBFNN to improve the underwater vehicle localization performance of ESKF?

1.1. State-of-the-Art Review

The basic form of an on-board navigation system on any underwater vehicle comprises an inertial measurement unit (IMU) that can determine positions by integrating three-axis acceleration and angular velocities [7,8]. This basic form of navigation suffers from drift, typically 1.8 km per day to 1.5 km per hour based on the grade of IMU [7,9,10], which makes them practically impossible to use for long missions. On top of that, most common off-board positioning by global positing system (GPS) satellites does not work underwater because of radio frequency attenuation [11]. Alternate communication means based on acoustic positioning are widely used underwater, which suffers from communication uncertainly and delays [12,13]. On-board aiding sensors, Doppler velocity log, pressure sensor, and magnetometers can also help to reduce the effect of IMU drift, but all these sensors are affected by noise. To improve the navigation accuracy and to minimize disturbance because of noise, EKF-based algorithms are the most commonly used in underwater navigation and localization [14,15,16,17,18,19].
On the other hand, the EKF algorithm has its shortcoming in that the accuracy of estimation is reduced under high nonlinear system dynamics. However, many variants of EKF have been proposed in academic research to cater to this problem [20,21]. Most of the EKF and neural network estimation algorithms are designed for land-based vehicles. In these approaches, when GPS data is present and valid, the neural network is trained and, when GPS information is not available, the neural network output improves the EKF prediction. For instance, a detailed study [22] proposed a hybrid offline trained RBFNN with time series prediction for measurement update during GPS outage. Another study [23] combined extreme learning machine neural network (ELM) and EKF to bridge the GPS outage. They claimed to have a better real-time performance by improving the computation load. In addition, some authors [24,25] also suggested using machine learning techniques to improve localization via intelligent communication networks, but their research is limited to land-based applications. Recently, an intelligent methodology was proposed that uses deep learning neural networks with EKF [26]. Moreover, they claimed that, by using recurrent neural networks (RNNs), state estimation can be improved and their model can also work well with low-cost sensors. Nevertheless, they did not incorporate oceanic parameters for state prediction besides the fact that RNNs have a high computational cost.
Another researcher group, in [27], used underwater model-aided dead reckoning to improve EKF response. They calculated aided velocity using an identified surge dynamic model. This work design reached a position accuracy of 92% during external position fix outage. However, using a model in this design makes the system difficult to tune under different sensors and working conditions because models depend on various factors such as size, the weight of the AUV, and the physical characteristics of the sensors. The authors of [23] proposed to embed underwater vehicle dynamic equations in the EKF and estimated error in the navigation. This work claimed to have less computation load and better accuracy compared to a full model integration. However, it is also dependent on the physical parameters of the vehicle and has more implementation complexity.
Likewise, a study [28] compared underwater EKF and its statistical linearization variant, also known as unscented Kalman filter (UKF) [29] in their project. They found that the statistical form provides better accuracy in highly nonlinear conditions. Similar results were found in [30]. Nonetheless, the UKF drawbacks include implementation complexity, high computational time and cost, and round-off error [31,32].
Traditionally, the Kalman filter is used to train RBFNN [33,34] or in offline training of the radial basis function (RBF) [35]. These methods underperform in uncertain conditions with unmodeled dynamics. The authors of [36] introduced a forgetting factor, which is based on RFBNN, to improve the performance of central difference Kalman filter (CDKF) for attitude-of-the-satellite estimation. They proposed the range of forgetting at 0.2 to 2 as a multiplier to the Kalman gain, but limited description of the selection of the forgetting factor was provided. Recently, the RBFNN-aided Kalman Filter was proposed to improve the state estimation accuracy for spacecraft navigation [37]. Moreover, they did not use multi-sensor fusion of high-rate and low-rate sensors.

1.2. Contributions of the Paper

The proposed work fills the gap by proposing a novel multi-sensor fusion architecture based on the strengths of the RBF neural network and error-state Kalman filter for underwater navigation, which has not been proposed to date to the authors’ knowledge. We named this algorithm RBF-ESKF. The augmentation of both algorithms improves the navigation of underwater vehicles in GPS-less environments. Our major contribution is the derivation of a multi-sensor fusion algorithm that improves the accuracy of underwater localization by taking advantage of a radial basis function (RBF) neural network that has the capability of nonlinear universal approximation via recursive learning [38]. Moreover, a simple structure of the RBF network can be trained online with less computation cost compared to the backpropagation neural network (BPNN) [39].
The structure of this paper is organized as follows. Section 2 discusses kinematic mathematical modeling with an ellipsoid Earth model. Section 3 discusses external and internal sensors for underwater navigation. In addition, the mathematical models of the sensors are described with their operating noise characteristics. The error dynamic model is presented in Section 4. In Section 5, a multi-sensor fusion algorithm is derived for accurate underwater integrated navigation. Section 6 shows the test results of the proposed multi-sensor fusion filter in three different conditions. In addition, acoustic communication lost and on-board sensor malfunctioning are tested and analyzed. Moreover, the performance of the RBF-ESKF is compared with ESKF under different scenarios. A comparative analysis is performed, showing that the proposed algorithm has promising results.

2. Mathematical Modeling

This section discusses the mathematical modeling of the underwater vehicle taken from [40,41,42,43,44]. The first subsection explains the notations used for modeling. The second subsection describes the frame of references used for the mathematical formulation of IMU. The third subsection formulates the kinematics equation of motion for six degrees of freedom. The last subsection provides brief information on the sensors used for underwater navigation.

2.1. Frame of References

The importance of frame of reference transformation for underwater navigation arises from the fact that sensors are mounted on the vehicle’s body. The origin of the body is defined as the center of body frame ( b ) External position fixes are in a rotating, Earth-centered, Earth-fixed (ECEF) frame ( e ) . Moreover, Newton’s laws are applicable to the Earth-centered inertial (ECI) frame ( i ) . However, the north-east-down (NED) frame or navigation frame ( n ) is the tangent plane to the Earth’s surface at the location of the underwater vehicle and its x-axis points toward the true north of the Earth. Figure 1 shows a frame of references used to develop navigation equations for underwater vehicles, with slight modification from [40].

2.2. Mathematical Notation

The mathematical notations used in this work are standard notations used to model underwater vehicle position, velocity, and attitude [40]. The position and velocity are three dimensional (3D) vectors in Euclidean space. Rotations are represented by quaternion. Subscripts and superscripts are used to represent the relationship between the frames. For example, ω i e e shows an angular velocity of frame ( e ) with respect to ( i ) represented in ( e ) frame. The following Table 1 shows a list of symbols used in the development of underwater vehicle navigation equations.

2.3. Navigation Equations

The underwater vehicles are most commonly equipped with a strapdown inertial navigation system [41]. In this configuration, measurements are obtained directly in the body frame and sensors experience full rotation during maneuvers. Moreover, this configuration requires an initial condition for position, velocity, and attitude. The kinematics equations for the underwater vehicles used in this work are described in [42,43,44]. The rate of change in position p e ˙ and velocity v e of vehicle in ( e ) frame is related by the following differential equation:
p ˙ e = v e
The rate of change in velocity v e ˙ of the underwater vehicle in ( e ) frame is dependent on accelerometer output f b , angular velocity ω i e e , and gravity vector g e in ECEF and can be expressed by the following differential equation:
v e ˙ = R b e f b 2 Ω i e e v e + g e
where Ω i e e is a skew symmetric matrix of angular velocity ω i e e and R b e is the rotation matrix that transforms a specific force vector from the ( b ) frame to the ( e ) frame.
The rate of change of rotation matrix R ˙ b e represented in frame ( e ) is dependent on angular velocity Ω i b b of the body with respect to frame ( i ) , and angular velocity ω i e b of Earth with respect to frame ( i ) is expressed by the following equation:
R ˙ b e = R b e ( Ω i b b Ω i e b )
The relationship between the change in latitude of the vehicle ϕ l a t ˙ and velocity is represented by the following differential equation:
ϕ l a t ˙ = v N / R M + h d
where R M = a 1 e 2 sin ϕ
The change in longitude of the vehicle λ ˙ l o n g in the form of the east velocity is represented by following mathematical relationship:
λ ˙ l o n g = v E / ( R N + h d ) cos ϕ l a t
where R N = R M 1 e 2 1 e 2 sin ϕ l a t
The change in height of the vehicle h d ˙ is expressed in the form of the down velocity as
h d ˙ = v D
The latitude ϕ l a t , longitude λ l o n g , and depth h d are given as
ϕ l a t = tan 1 z / x 2 + y 2 1 e 2 R M / R M + h d
λ l o n g = tan 1 [ y / x ]
h d = x 2 + y 2 cos ϕ l a t R M
The rate of change in velocity of the vehicle in the ( n ) frame v ˙ e b n is expressed by the following differential equation:
v ˙ e b n = R b n f i n b ( Ω e n n + 2 Ω i e n ) v e b n + g e b n
where Ω e n n v n is the centripetal acceleration related to the motion of the ( n ) frame with respect to the ( e ) frame and 2 Ω i e n v n is the Coriolis acceleration.
The local gravity vector g e b n = 0 0 g T depending on the latitude, longitude, radius of curvature of the meridian, and radius of curvature of the prime vertical is given by
g e b n = g 0 1 + h d R N R M 2
where g 0 = 9.780318 × 1 + 5.3024 × 10 3 sin 2 ϕ 5.9 × 10 6 sin 2 2 ϕ [44]. The rate of change in the rotation matrix R ˙ b n represented in frame ( n ) is dependent on the angular velocity of the body with respect to frame ( i ) and on the angular velocity of frame ( n ) with respect to frame ( i ) :
R ˙ b n = R b n ( Ω i b b Ω i n b )
The attitude of the vehicle is represented by quaternion. It has only one constraint compared to the direction cosine matrix (DCM), and this method is also singularity free in comparison to the Euler angle representation, which has a singularity problem [45].
The attitude of the vehicle represented in quaternion q ˙ b n is given as
q ˙ b n = 1 2 q b n o ω i b b 1 2 0 ω i n n q n b
where q b n has two parts: η is the scalar part; ϵ i is the vector part; and i = 1,2,3. The ⊗ sign represents a quaternion product.
This section developed the motion equations of an underwater vehicle in starpdown configuration. The measurements from accelerometers were obtained in ( b ) frame as a specific force vector and were not the true acceleration of the vehicle. To obtain the true acceleration or rate of change in velocity of the vehicle in e frame, as shown by Equation (2), the ( b ) frame specific forces are transformed into the e frame by a rotational matrix and by compensating for the gravitation and rotation effects of the Earth. Equation (1) shows that the rate of change in the position and acceleration has an integral relationship. Since the position from an acoustic fix is obtain in the geoid ECEF frame, for compatibility, Equations (4)–(9) convert the position of the vehicle in latitude, longitude, and depth. Equation (13) shows a quaternion representation of the roll, pitch, and yaw angles of the vehicle. Interested readers can find more details for mathematical modeling of an underwater vehicle in starpdown configuration in [40,41,42,43,44].

3. Sensors on the Vehicle

The following subsections give a brief overview of the sensors used in the AUV with a mathematical formulation of errors. For detail about the mathematical model of the sensors used for navigation, readers can refer to [40,46,47].

3.1. Inertial Measurement Unit

The inertial measurement unit (IMU) provides a three-axis accelerometer and three-axis gyro outputs [48]. These measurements are affected by variable factors such as temperature, manufacturing process, scale factor, noise, and drift.
The accelerometer measurement output vector f a c c b in (b) frame is modeled as
f a c c b = f i b b + b a c c + ϱ a c c
where ϱ a c c is white noise and b a c c is acceleration bias, which is modeled as a 1st-order Markov process.
The accelerometer bias b a c c is expressed by the random walk and the random constant shown as
b ˙ a c c = τ a c c 1 b a c c + ρ a c c
where τ a c c is the correlation time given by the manufacturer. Its value depends on the accelerometer sensors used in the IMU. The value ρ a c c depends on standard deviation, and it is known as a process driving noise.
The accelerometer in IMU does not directly measure the true kinematic acceleration of vehicles due to the presence of Earth’s gravitation. For that reason, the measurement from the accelerometer is known as the relative acceleration or specific force f i b b and it is related to the kinematics acceleration of the vehicle as follows:
f i b b = R ˙ i b p ¨ i g b
where p ¨ i is acceleration represented in an inertial frame and g b is gravitation sensed by the accelerometer in the body frame.
The actual output of the gyro ω g b is influenced by noise and bias that is given by
ω g b = ω i b b + b g + ϱ g
where ϱ g is the white noise and b g is the gyro bias, which is modeled as a 1st-order Markov process.
The gyro bias b g is represented by a random walk and a random constant given as follows:
b ˙ g = τ g 1 b g + ρ g
where τ g is the gyro correlation time obtained from manufacturer documentation. Its value depends on the quality of the accelerometer sensors used in the IMU. The value ρ g depends on thte standard deviation, and it is called process driving noise. The IMU used for simulation is tactical grade Emcore SDI-1500. It uses high-precision micro-electro-mechanical systems (MEMS) quartz sensor technology with 1 /h gyro bias and 1 mg accelerometer bias stability. The IMU offers the best cost to performance ratio compared to other technologies. It consists of three orthogonal accelerometer sensors that provide the measurement of specific forces. Three gyros provide the angular rates of the body with respect to the inertial frame of reference.

3.2. Underwater Acoustic Positing System

The underwater acoustic positing system measures the distance and direction of the vehicle from the reference positions. For this work, HiPaP 502 is used for simulation. This system provides a typical range detection accuracy of 0.2 m, with an operating range of 1 to 5000 m. It has an acoustic operating area of 200 /200 with the capability of narrow beamforming of 10 , which improves the signal-to-noise ratio. It can be interfaced by GPS to provide Earth-related coordinates. However, acoustic position estimate is effected by GPS accuracy, system installation, ship attitude, sound velocity profile, ray bending, and measurement noise.
Assuming the system is precisely calibrated, installation and ship attitude have negligible effects. The mathematical model of the system actual output p ^ is given as
p ^ h = p h + b h + ϱ h
where p h is true output position, whereas b h is the time-varying bias modeled as a 1st-order Markov process and depends on the sound velocity profile and ray bending effect. ϱ h represents measurement white noise.

3.3. Doppler Velocity Log

The Doppler velocity log (DVL) measures the change in acoustic frequency for determining the speed of the vehicle with reference to the seabed. In deep water, when the seabed is not available, the DVL measures the speed with respect to water. The DVL sends a known frequency signal to the seabed and receives the signal that bounces back to the vehicle. The speed of the underwater vehicle is a dependent doppler effect [49]. The work used Nortek DVL-500, which has a range of 0.3 to 200 m and long-term accuracy of ±0.1% ±0.1 cm/s. The DVL can be operated in a 4-beam Janus configuration complete performance testing of DVL; readers may refer to [50].
Considering that the attitude and installation error are negligible, the actual output of DVL v ^ d v l can be modeled for the true velocity measurement vector, noise, and bias given as [51].
v ^ d v l = v d v l + b d v l + ϱ d v l
where v d v l is the true output and random velocity error, b d v l is modeled as the 1st-order Markov process, and ϱ d v l is white noise.

3.4. Depth Sensor

Depth and underwater pressure have a direct relationship [52]. As the vehicle goes deep into water, the pressure reading increases linearly. The depth sensor modeled in this work is from Paroscientific, Inc. part number 8CDP700-I, which provides an accuracy of 0.01% and high stability under tough conditions.
The depth sensor actual output h ^ d is modeled by adding true depth h d with noise:
h ^ d = h d + ϱ d
where ϱ d is measurement noises modeled as white noise.

3.5. Magnetometer

The magnetometer or compass measures the magnitude and direction of Earth’s magnetic field [53]. The magnetometer used in the work is jewel instrument ECS-AC-RS232 e-compass, which has a 3-axis magnetometer and a 2-axis tilt sensor. The tilt sensor is used for the initialization of roll and pitch [40]. It offers an accuracy of ±0.5 root mean square RMS, a repeatability of ±0.3 , and a response time of 36 milliseconds. The pitch and roll are ±42 whereas the dip angle range ±80 . The major error sources of the magnetometer include the declination angle, which is the difference between true north and sensor north; the hard and soft magnetic distortion due to motors and ferromagnetic materials [54]; and sensor imperfection, misalignment, and noise. However, with proper calibration, most of the errors in measurement can be removed.
The actual output q m of the magnetometer is a combination of noise and true output q m written as
q ^ m = q m ϱ m
where ϱ m is the sensor noise modeled as white noise.

4. Error Dynamic Model

The position p , velocity v , attitude q , gyro bias b g , accelerometer bias b a c c , and acoustic fix bias b h in full state vector x form are given as
x = p v q b g b a c c b h s v T
where s v is the sound velocity model underwater used to calculation underwater acoustic transmission.
The estimated state vector x ^ is written as
x ^ = p ^ v ^ q ^ b ^ g b ^ a c c b ^ h s v ^ T
The design used in this worked estimate an error-state vector, which offers the main advantages of flexible sampling rate, robustness, and low computation burden [55,56]. The error-state vector δ x ˙ is the difference between the true state and estimated state x ^ ˙ of the model and is given as
δ x = x x ^
The error-state vector can be represented as
δ x = δ p δ v δ q δ b g δ b a c c δ b h δ s v T
where the position, velocity, and attitude error equations are provided as under. For complete derivation of this error model, readers can refer to [42,57].
Assuming that the underwater vehicle travel at low speed and that the depth of operation is much less than the Earth’s radius, the rate of change in errors in longitude δ λ l o n g ˙ , latitude δ ϕ l a t ˙ , and depth δ h d ˙ are given by
δ λ ˙ l o n g δ ϕ ˙ l a t δ h d ˙ = 0 0 v n R N + h d 2 v e sin ( λ l o n g ) R N + h d cos 2 ( λ l o n g ) 0 v e R N + h d 2 cos ( λ l o n g ) 0 0 0 δ λ l o n g δ ϕ l a t δ h d + 1 R M + h d 0 0 0 1 R N + h cos ( λ l o n g ) 0 0 0 1 δ v n δ v e δ v d
The velocity error δ v ˙ e b n is given by the following differential equation:
δ v ˙ e b n = δ g e b n + δ R b n f a c c b + R b n δ f i b b 2 δ Ω i e n + δ Ω e n n v ^ e b n 2 Ω i e n + Ω e n n δ v e b n
If the vehicle operates at low speed underwater, angler velocity ω i e n and δ ω i e n can be neglected. Gravity error is also neglected because of the small operating area and accurate estimation [44,58].
Thus, the velocity error δ v ˙ e b n differential equation can be re- written as
δ v ˙ e b n = δ R b n f a c c b + R b n δ f i b b
Assuming that the angular velocity of rotation of Earth with respect to the inertial frame ω i e is accurately known, using the attitude error model in the form of the quaternion, it is given as
δ q ˙ = 1 2 δ ω i b b + 1 2 δ Ω i b b δ q
The error of gyro bias δ b ˙ g is given as
δ b ˙ g = b ˙ g b ^ ˙ g
The error of accelerometer bias δ b a c c is given as
δ b a c c = b ˙ a c c b ^ ˙ a c c
The error of hydro-acoustic system bias δ b a c c is given as
δ b h = b ˙ h b ^ ˙ h
Since the ESKF employed in this work used an error-state or indirect-form-state vector z , it is obtained by subtracting the outputs of the inertial measurement unit (INS) and aiding sensors measurement [59].
The position error δ z p between the INS position p I N S and acoustic position system measurement p h is written as
δ z p = p I N S p h
The velocity error δ z v between the INS velocity v I N S and DVL measurement v d is given by the following equation:
δ z v = v I N S v d
The attitude error δ z v between the INS attitude q I N S and magnetometer measurement q m is not a vector quantity; it cannot be subtracted as position and velocity. Quaternion multiplication ⊗ is used to find the error term as follows:
δ z q = q m 1 q I N S

5. RBF-ESKF Mulit-Sensor Fusion

The proposed modifications improve ESKF performance and make use of the advantages of the RBF neural network. The RBF neural network can approximate any nonlinear function, and they are also known as universal function approximators [60]. The RBF center, its width, and the linear weights for each output neuron are altered at every iteration of a learning algorithm. When each RBF center is as close to the input vector as possible and the network output error is within the target limit, the training phase is completed. Therefore, it is possible to express the approximation of any functional dependency between variables as a linear combination of a best possible number of RBF neurons with appropriate weight and center. The top level block diagram of our proposed fusion algorithm is depicted in Figure 2.
As shown by Figure 2, the algorithm takes the error of the aiding sensors and INS as input. The RBF-ESKF fusion algorithm after processing gives an output to INS for error correction and reset. The RBF neural network used for processing has a basic three layer structure; input, hidden, and output layer. Furthermore, compared to BPNN, the RBF variants have less computational load and fast online learning. The first layer is the input layer, which provides an interface between the data and neural network. The data from the input layer to the second hidden layer is transferred in such a manner that the output value of every hidden neuron is inversely related to the Euclidean distance from that neuron’s input vector to the RBF neuron’s center. The third layer is the output layer, which takes into account the cumulative weights and biases of all RBF neuron outputs.
Several variants of RBF exist in literature that depend on the application [61]. However, in this work, we used the Gaussian-type RBF function. The weights of neurons in the hidden layer for the Gaussian RBF function indicate the center of the symmetrical Gaussian distribution curve. The novelty of the RBF-ESKF algorithm is that it includes system information within the weight and center learning update rules.

5.1. RBF-ESKF Mathematical Formulation

The underwater vehicle navigation system has nonlinear dynamics and measurement characteristics, which can be represented in state space form by Equation (37) and Equation (38) [40]
x ˙ ( t ) = f ( x ( t ) , u ( t ) , t ) + α ( t )
where the state of the system is represented by x ˙ ( t ) . The  u ( t ) is a known system input and α ( t ) is Gaussian white noise.
The measurement output z ( t ) in state space can be written as
z ( t ) = h ( x ( t ) , t ) + β ( t )
where f and h are nonlinear functions. The measurement is corrupted by Gaussian white noise β ( t ) .
Dropping time and noise in the above equations for simplification, the linear form of error state can represented by Equation (39) and Equation (40): [40]
δ x ˙ = F ( x ^ , u , t ) δ x
where F ( x ^ , u , t ) = f x x = x ^
However because the deterministic component F ( x ^ , u , t ) , is always incomplete, which means the model does not incorporate, for instance, AUV underwater movements due to waves, the stochastic component α ( t ) takes these effects into account.
The linear residual measurement output δ z can represented by
δ z = H ( x ^ , t ) δ x
where H ( x ^ , t ) = h x x = x ^ , x ^ is defined as the trajectory obtained by using vehicle kinematics equations.
The Kalman filter [62], in its basic form, is based on linear system and measurement models, which in reality might not be the case. In our underwater vehicle, navigation equations are used to construct the mathematical model and are not linear with respect to the state variables. This is done by linearization of any predicted trajectory, leading to an error-state model as discussed in Section 4. This linear approximation was proven to be incomplete and requires special consideration in underwater navigation, which leads to derivation of the RBF-ESKF algorithm.
For mathematical formulation of RBF-ESKF, we start with ESKF implementation, which utilizes definitions of measurement error and state dynamics error, as described by the Equations (39) and (40) [63]. The ESKF algorithm has three major components. The first component is initialization, in which state, covariance, process, and measurement covariance are initialized [64].
  • x 0 = Initialization of the state variables.
  • P 0 = Initialization covariance matrix.
  • Q 0 = Initialization process noise covariance.
  • R 0 = Initialization of measurement noise covariance.
where superscript minus denotes the a priori state that occurs before innovation is updated. Superscript plus + denotes the posteriori state after innovation calculation.
The second component is time update, in which the error state and error-state covariance are updated, give by
δ x k + 1 = Φ k δ x k
where δ x k + 1 is the predicted error state and Φ k is the state transition matrix in discrete form.
P k + 1 = Φ k P k + Φ k + Q k
where P k + 1 is the predicted error covariance and Q k is the process noise covariance v k .
The third component is measurement update, in which the residual of measurement is updated. The residual of measurement δ z k is given by the difference between the actual measurement z k and the prediction of measurement h x ^ k
δ z k = z k h x ^ k
The kalman K k gain is given as
K k = p k H k H k p k H k + R k 1
The posteriori error estimate δ x k + is given as
δ x k + = δ x k + K k δ z k H k δ x k
The expression δ z k H k δ x k is referred to as innovation. It is the difference between the error of observation and its expected error, represented by s k as
s k = δ z k H k δ x k
The posteriori error-state covariance P k + is given as
P k + = I K k H k P k
The complete corrected navigation state x ^ k + can be written as the sum of the error estimate from Equation (45) and prior full state estimate x ^ k as
x ^ k + = x ^ k + δ x k +
The major assumption for obtaining proper results from ESKF is that the time interval should be short for error calculation and nonlinearity should not be dominant in the calculation of the innovation term. To compensate the effect of nonlinearity in the innovation Equation (46), we propose to modify it by incorporating an RBF neural network. The modified innovation term s ˜ k is given as
s ˜ k = s k W k y k
where the term W k y k is the output of the RBF neural network. The term y k is the output of the hidden layer of the RBF neural network and W k is the weight matrix that provides the link between output and hidden layers of RBF neural network. These weights can be designed by minimizing the the mean square error (MSE) cost function J , defined by
J = s k W k y k 2
To improve the estimation of multi-sensor fusion under nonlinear conditions, the Gaussian RBF function utilize an a priori error-state estimate, which is given as
y k ( i ) = exp δ x k c i k 2 2 σ i 2 , i = 1 , 2 , , N c
where c i k is the neuron center and σ i is the width of the Gaussian RBF function. The weights of the neurons in matrix form w k are represented as
W k = w 1 k ( 1 ) w 1 k ( 2 ) w 1 k ( N c ) w 2 k ( 1 ) w 2 k ( 2 ) w 2 k ( N c ) w M k ( 1 ) w M k ( 2 ) w M k ( N c )
where M donates the size of the state vector and N c represents the number of neuron centers. The size of W k is M × N c . The size of W k y k is M × 1 , and the size of y k is N c × 1 .

5.2. Derivation of Weight Update of RBF-ESKF

For the weight matrix W k , we only consider weight associated with the mth-specific output. Thus, the weight vector of the mth output of RBF can be denoted by w m k . Therefore, the weight update rule of the specific mth output is given as
w m k + 1 = w m k η w 1 2 J w m k
where η w is the learning rate, and its value is selected by experimentation. Too small a value of η w can cause unsuitability, and too large a value can make the response sluggish. A gradient descent method, namely the steepest decent, is used to minimize the cost function J relative to the RBF neural network weight. By using Equation (50), the gradient of cost function can be written as
J w m k = s k W k y k 2 w m k
The result of the derivative is given as
J w m k = 2 ( δ z k ( m ) H k ( m ) δ x k ( m ) w m k y k ) y k T
Thus, putting Equation (54) in Equation (52), we get the complete weight update equation:
w m k + 1 = w m k + η w ( δ z k ( m ) H k ( m ) δ x k ( m ) w m k y k ) y k T

5.3. Derivation of Center Update of RBF-ESKF

The center update rule for the jth element of the mth neuron center is given as
c i k + 1 ( j ) = c i k ( j ) η c 1 2 J c i k ( j )
where η c is the learning rate of center update. It is determined experimentally. By applying the steepest descent method, we minimized cost function J with respect to weight. The result of derivative is shown in the following equation:
J c i k ( j ) = 2 i = 1 M ( s k ( m ) + w m k y k ) · ( w m k ( i ) y k ( i ) ( δ x k ( j ) c i k ( j ) ) σ i 2 )
Plugging Equation (57) into Equation (56), the new center update equation becomes
c i k + 1 ( j ) = c i k ( j ) + η c i = 1 M ( s k ( m ) + w m k y k ) · ( w m k ( i ) y k ( i ) ( δ x k ( j ) c i k ( j ) ) σ i 2 )
It is evident from the weight update Equation (55) and center update Equation (58) that our algorithm uses system information to train the RBF neural network to overcome the drawbacks of ESKF.
The complementary form block diagram representation of the RBF-ESKF fusion algorithm is shown in Figure 3. The vehicle kinematics combined with IMU high-rate sensors of IMU provide the estimated output. This estimated output is then subtracted from low-rate sensors and fed into RBF-ESKF. RBF and ESKF work together to find the best error estimate, which is then added into the underwater vehicle kinematics in feed-forward fashion to obtain the total state.
Algorithm 1 shows iterative steps of the proposed method. The term s ˜ k is RBF modified innovation term used by RBF-ESKF to improve accuracy of filter in highly non-linear conditions.
Algorithm 1 RBF-ESKF multi-sensor fusion for underwater navigation
   
Initialization:
1:
Initialize ESKF variables x 0 , P 0 , Q 0 , R 0
2:
Initialize RBF variables w 0 , c 0 , σ 0 , η w , η c
   
Kalman gain update:
3:
calculate Kalman gain
K k = P k H k H k P k H k + R k 1
   
RBF Gaussian function update:
4:
Learning non-linearity of error state vector
y k ( i ) = exp δ x k c i k 2 2 σ i 2 , i = 1 , 2 , , N c
   
Innovation update:
5:
Non-linearity influence is minimized by using output of RBF neural network in innovation term
s ˜ k = δ z k ( H k δ x k + W k y k )
   
Measurement update:
6:
Estimate error state by using innovation term and Kalman gain
δ x k + = δ x k + K k s ˜ k
7:
Error state covariance update.
P k + = I K k H k P k
   
Full State correction
8:
Full state is corrected by error adding error estimate.
x = x ^ + δ x k +
   
RBF Neural Network Weight and Center update:
9:
RBF weight update
w m k + 1 = w m k + η w ( δ z k ( m ) H k ( m ) δ x k ( m ) w m k y k ) y k T
10:
RBF center update
c m k + 1 ( j ) = c i k ( j ) + η c ( w i k ( j ) y k ( j ) c i k ( j ) σ i 2 ) . i = 1 M ( δ z k ( i ) ( H k ( i ) δ x k ( i ) + w i k y k ) )
  
Time propagation:
11:
Time propagation of error state and covariance
δ x k + 1 = Φ k δ x k +
P k + 1 = Φ k p k + Φ k + Q k
12:
Next iteration (posterior becomes prior)

5.4. Complexity of RBF-ESKF

The proposed algorithm uses RBFNN to enhance the performance of underwater vehicle localization with a slight increased in time and space complexity compared to ESKF due to matrix multiplications. Compared to BPNN and deep learning neural networks, RBFNN has less complexity because of its simple three-layer structure because the time and memory space complexity of the neural networks is directly related to the structure and number of layers. However, faster matrix multiplication algorithms such as the Strassen algorithm [65] can be used to decrease execution time. Table 2 shows the structure of the RBFNN used in this work.

6. Results and Discussion

In order to compare performances, the proposed algorithm and ESKF were simulated in three different realistic scenarios. As the ESKF structure was modified by RBF in our fusion algorithm, low-level functions were written for simulation. The noise specifications of the sensors used in this work are comparable to their datasheets. The main purpose of the simulation was to compare the maximum error (max) and root means square error (RMSE) of position, velocity, and attitude. Practical failure mode tests cases were developed and simulated with DVL and acoustic positioning loss of measurements for a short duration. The simulation results were compared with conventional ESKF for performance evaluation. Furthermore, for simulation, the assumption was made that underwater vehicles can move in any direction and with any roll, pitch, and yaw angle. A reference trajectory of the vehicle was generated by angular velocities and acceleration.
To consider the effects of random variations in the accuracy of the fusion algorithms, a Monte Carlo simulation was used. The test consisted of 100 runs. Two filters processed the same data during the test to ensure a fair contrast. For all three cases, the same RBF neural network structure was used, as listed in Table 2. For the simulation, the RBF weights, centers, and sigma were initialized randomly.

6.1. Test Case 1: Normal Working Condition

In the first case, the vehicle was considered to be working in a normal operating mode without any on-board and off-board sensor failure. To simulate a real situation, the noise and drift characteristics of the sensors used for simulation were almost the same as listed in the manufacturer’s documentation stated under Section 3. Both ESKF and RBF-ESKF were tested on the same operating conditions. The performances of ESKF and RBF-ESKF are compared side by side in Table 3.
From the results of the north position prediction displayed in Table 3, it can be observed that ESKF has an almost three times higher maximum error than that of the RBF-ESKF. Furthermore, for the east position, the ESKF maximum error was twice as high as the RBF-ESKF. In the case of maximum error in the prediction of depth, the performance of the two filters are approximately identical. The RBF-ESKF RMSE for position estimation was even better, almost twice as high. Significant improvement was seen in the north, where it were three times higher than the one achieved by the ESKF. Overall, the RMSE sum for the RBF-ESKF position estimate was approximately two times better than that of the ESKF.
The estimation of velocity showed considerable improvement compared to the ESKF. The overall northern velocity error was almost double than that of the ESKF compared to the RBF-ESKF. Small but noticeable improvements were seen in the maximum error for east and depth velocities. The RMSE of the RBF-ESKF was roughly two times better than the ESKF for the north, east, and down velocities. Overall, with our proposed fusion algorithm, the sum of all RMSE states was almost three times better than ESKF.
Overall, in terms of attitude, relative to ESKF, the sum of all state RMSEs improved significantly by about double as much.
Figure 4 shows the 2D and 3D trajectories. It can be seen that the trajectory is not linear. The position, velocity, and attitude errors are estimated on this trajectory. The star symbol in the 2D trajectory shows an acoustic fix from the external source. The performances of these ESKF and ESKF-RBF were evaluated by how close the estimated path was to the actual value.
Figure 5 shows comparison of ESKF and RBF-ESKF estimated the position, velocity and attitude.

6.2. Test Case 2: Acoustic Fix Not Available

In this case, the robustness of the multi-sensor fusion algorithm was tested by simulating the loss of the underwater acoustic fix for a short period. The unavailability of the position information from acoustic fix mostly influenced the position estimate of both filters, predominantly ESKF. In this case, the position estimate was only available from the integration of the DVL velocity, which compensated the acoustic fix loss effect and reduced the drift error. The detailed comparison is shown in Table 4.
It can be noted from Table 4 that the maximum error of RBF-ESKF was almost five times better for the north and east directions as compared to ESKF. The down position estimate was also two times better than that of ESKF. The overall RMSE in all three directions is considerably improved by almost one and a half times that of ESKF. Overall, compared to normal working conditions, the position accuracy had a detrimental effect, but RBF-ESKF has proven to be more robust.
In contrast to standard operating conditions, the velocity estimation was marginally influenced by acoustic fix unavailability. The maximum ESKF error was worse than that of RBF-ESKF for velocity in all directions. For RBF-ESKF, the RMSE of velocity was significantly better by almost two times that for ESKF. Overall, with our proposed method of fusion, the sum of all RMSE velocity states was almost two times better than ESKF.
For RBF-ESKF, the RMSE of the roll, pitch, and yaw angles were substantially better by about one and a half times that of ESKF. The maximum roll, pitch, and yaw angle errors were two times better for RBF-ESKF. Overall, the sum of all state for RBF-ESKF was better than that for ESKF.
Figure 6 below illustrates a comparison of the position and velocity errors when acoustic fix was not available for a short period from 1500 to 1700 s. The performances of ESKF and ESKF-RBF were evaluated on similar trajectory. Moreover, the RBF-EKF is consistent in the estimation and is close to the actual value.

6.3. Test Case 3: DVL Unavailable

In this case, the robustness of both filters was tested when DVL was not available for a short duration. The velocity estimate had a larger influence than the position estimate because, when DVL was not available, it was calculated by taking the derivative of the position measurement, which suffers from noise amplification from the differentiation process. Moreover, acoustic fix measurement bias was negatively influenced by DVL measurements that contribute to increasing the position error. Table 5 compares the performance robustness multi-sensor fusion algorithm with DVL failure.
It can be observed from the above data that position estimate in this case was better than case 2 but slightly less accurate than normal working conditions. However, in the north direction, ESKF maximum error was almost two times worse and almost one and a half times worse for east and depth. Furthermore, the RBF-ESKF RMSE for the position was notably better than that for ESKF by around one and a half times in all directions. Thus, the position of estimation for the overall mission was improved in all directions by the RBF-ESKF algorithm.
The velocity estimation without DVL was less accurate compared to case 1 and case 2. However, the overall results of our method are much better than those of ESKF, which were almost doubly improved with respect to the maximal error and RMSE. In addition, the sum of all estimated velocity states from RBF-ESKF in all directions was approximately two times better than that of ESKF. Hence, the RBF-ESKF velocity estimation was more robust than that of ESKF. The RBF-ESKF attitude estimation of the roll, pitch, and yaw angles had a lower maximum estimation error. Moreover, the RMSE for RBF-ESKF showed considerable improvement. Overall, the sum of all estimated attitude states was one and half times better than that of ESKF.
Figure 7 below shows comparison velocity errors and velocity when the DVL measurement was not available for a short duration from 1900 to 2000 s. The performances of ESKF and ESKF-RBF were evaluated on a similar trajectory as test case 1.
The time complexity of ESKF and ESKF-RBF was tested by running the algorithms on an Intel I7 CPU with 4 GB RAM without any GPU. The testing software used was Matlab 2020b on Windows 7 platform. The timing comparison of both methods is given in Table 6.
On a high-speed microcontroller or field-programmable gate array FPGA, the execution time difference will be further reduced. Furthermore, the speed of surveying the type of underwater vehicle is in the range of 2 km/h to 10 km/h; this difference has no significant effect on navigation and localization.
The work tried to fill the gap by proposing a noval ESKF and RBF-augmented fusion solution, which was assessed in three different underwater cases. The ESKF performance strongly relies on the knowledge of the system models and noise properties, which was degraded by nonlinearity. The errors in the RBF-ESKF are smaller than the errors in the ESKF because of the recurcive learning of RBF. Moreover, the fusion algorithm based on RBF-ESKF with help from the aiding sensors was able to correct the drift problems in the INS with better accuracy.
Nevertheless, the proposed algorithm is not limited to underwater; it can also be used in other applications [66,67] such as improving aircraft navigation and tracking by using aerial sensors. Moreover, autonomous ground vehicles are another area where this method can be employed. Furthermore, by improving the Kalman filer response, our methodology can also enhance accuracy satellite attitude estimation [68].

7. Conclusions

This paper discusses the performance of the proposed algorithm RBF-ESKF for underwater vehicle localization. The primary aim of this work is to take advantage of the RBF neural network to improve the estimation performance of the conventional ESKF for the position, velocity, and attitude of an underwater vehicle. It contrasts outcomes with ESKF in three separate simulations, showing that RBF-ESKF performs better in estimating position, velocity, and attitude. Why the standard ESKF has an inferior performance is because it is designed by employing first-order Taylor series approximation in the error covariance matrix estimation, which results in a decrease in estimation accuracy under high nonlinearity. Thus, important information about the dynamic of underwater is lost because of this realization. However, RBF-ESKF efficiently handles nonlinearity due to its inherent capability for nonlinear function approximation and learning ability.
The research also compared robustness in cases when there is no available position information from the acoustic fix. Here, relative to ESKF, RBF-ESKF demonstrated better accuracy. When acoustic fix becomes available, RBF-ESKF converges quickly. In addition, when DVL fails due to short durations, RBF-ESKF also demonstrates less estimation error.
This work is part of an ongoing work on underwater vehicle localization. The novel proposed algorithm is intended for use as a state estimation for underwater seabed mapping application.
In the future, we would like to test and analyze our algorithm with a different set of sensors with different data rates in underwater environments. For instance, an enhanced version of the multi-sensor fusion algorithm could be designed by using stereo vision and underwater wireless sensor nodes to improve localization.

Author Contributions

All authors were involved in the research and preparation of this manuscript. Conceptualization, N.S., M.M., and P.O.; formal analysis, N.S. and M.M.; funding acquisition, P.O.; investigation, N.S.; methodology, M.J.I.; resources, M.J.I. and P.O.; software, N.S.; supervision, P.O.; writing—original draft, N.S. and A.A.; writing—review and editing, M.M. and P.O. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially funded by the Campus de Excelencia Internacional Andalucia Tech, University of Malaga, Malaga, Spain.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors express their gratitude to the Oceanic Engineering Research Institute, University of Malaga, Malaga, Spain.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Xu, C.; Xu, C.; Wu, C.; Qu, D.; Liu, J.; Wang, Y.; Shao, G. A novel self-adapting filter based navigation algorithm for autonomous underwater vehicles. Ocean Eng. 2019, 187, 106146. [Google Scholar] [CrossRef]
  2. Allotta, B.; Chisci, L.; Costanzi, R.; Fanelli, F.; Fantacci, C.; Meli, E.; Ridolfi, A.; Caiti, A.; Di Corato, F.; Fenucci, D. A comparison between EKF-based and UKF-based navigation algorithms for AUVs localization. In Proceedings of the OCEANS 2015—Genova, Genoa, Italy, 18–21 May 2015; pp. 1–5. [Google Scholar] [CrossRef]
  3. Chen, H.; Gong, Y.; Hong, X.; Chen, S. A Fast Adaptive Tunable RBF Network For Nonstationary Systems. IEEE Trans. Cybern. 2016, 46, 2683–2692. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Tomczyk, K.; Piekarczyk, M.; Sokal, G. Radial basis functions intended to determine the upper bound of absolute dynamic error at the output of voltage-mode accelerometers. Sensors 2019, 19, 4154. [Google Scholar] [CrossRef] [Green Version]
  5. Lu, S.; Başar, T. Robust nonlinear system identification using neural-network models. IEEE Trans. Neural Netw. 1998, 9, 407–429. [Google Scholar] [CrossRef] [Green Version]
  6. Li, D.M.; Li, F.C. Identification of chaotic systems with noisy data based on RBF neural networks. In Proceedings of the 2009 International Conference on Machine Learning and Cybernetics, Hebei, China, 12–15 July 2009; Volume 5, pp. 2578–2581. [Google Scholar] [CrossRef]
  7. Groves, P.D. Navigation using inertial sensors [Tutorial]. IEEE Aerosp. Electron. Syst. Mag. 2015, 30, 42–69. [Google Scholar] [CrossRef]
  8. Kepper, J.H.; Claus, B.C.; Kinsey, J.C. A Navigation Solution Using a MEMS IMU, Model-Based Dead-Reckoning, and One-Way-Travel-Time Acoustic Range Measurements for Autonomous Underwater Vehicles. IEEE J. Ocean. Eng. 2019, 44, 664–682. [Google Scholar] [CrossRef]
  9. Melo, J.; Matos, A. Survey on advances on terrain based navigation for autonomous underwater vehicles. Ocean Eng. 2017, 139, 250–264. [Google Scholar] [CrossRef] [Green Version]
  10. González-García, J.; Gómez-Espinosa, A.; Cuan-Urquizo, E.; García-Valdovinos, L.G.; Salgado-Jiménez, T.; Escobedo Cabello, J.A. Autonomous underwater vehicles: Localization, navigation, and communication for collaborative missions. Appl. Sci. 2020, 10, 1256. [Google Scholar] [CrossRef] [Green Version]
  11. Ullah, I.; Chen, J.; Su, X.; Esposito, C.; Choi, C. Localization and Detection of Targets in Underwater Wireless Sensor Using Distance and Angle Based Algorithms. IEEE Access 2019, 7, 45693–45704. [Google Scholar] [CrossRef]
  12. Paull, L.; Saeedi, S.; Seto, M.; Li, H. AUV Navigation and Localization: A Review. IEEE J. Ocean. Eng. 2014, 39, 131–149. [Google Scholar] [CrossRef]
  13. Qureshi, U.M.; Aziz, Z.; Shaikh, F.K.; Aziz, Z.; Shah, S.M.S.; Shah, S.M.S.; Sheikh, A.A.; Felemban, E.; Qaisar, S.B. RF path and absorption loss estimation for underwaterwireless sensor networks in differentwater environments. Sensors 2016, 16, 890. [Google Scholar] [CrossRef]
  14. Diversi, R.; Guidorzi, R.; Soverini, U. Kalman filtering in extended noise environments. IEEE Trans. Autom. Control 2005, 50, 1396–1402. [Google Scholar] [CrossRef]
  15. Almeida, J.; Matias, B.; Ferreira, A.; Almeida, C.; Martins, A.; Silva, E. Underwater localization system combining iusbl with dynamic sbl in ¡vamos! trials. Sensors 2020, 20, 4710. [Google Scholar] [CrossRef] [PubMed]
  16. Ko, N.Y.; Jeong, S.; Hwang, S.S.; Pyun, J.Y. Attitude estimation of underwater vehicles using field measurements and bias compensation. Sensors 2019, 19, 330. [Google Scholar] [CrossRef] [Green Version]
  17. Huang, H.; Chen, X.; Zhou, Z.; Xu, Y.; Lv, C. Study of the algorithm of backtracking decoupling and adaptive extended kalman filter based on the quaternion expanded to the state variable for underwater glider navigation. Sensors 2014, 14, 23041–23066. [Google Scholar] [CrossRef] [Green Version]
  18. Miller, A.; Miller, B.; Miller, G. On AUV control with the aid of position estimation algorithms based on acoustic seabed sensing and DOA measurements. Sensors 2019, 19, 5520. [Google Scholar] [CrossRef] [Green Version]
  19. Tal, A.; Klein, I.; Katz, R. Inertial navigation system/doppler velocity log (INS/DVL) fusion with partial dvl measurements. Sensors 2017, 17, 415. [Google Scholar] [CrossRef] [PubMed]
  20. Zhang, M.; Li, K.; Hu, B.; Meng, C. Comparison of Kalman Filters for Inertial Integrated Navigation. Sensors 2019, 19, 1426. [Google Scholar] [CrossRef] [Green Version]
  21. Sun, C.; Zhang, Y.; Wang, G.; Gao, W. A new variational bayesian adaptive extended kalman filter for cooperative navigation. Sensors 2018, 18, 2538. [Google Scholar] [CrossRef] [Green Version]
  22. Chen, L.; Fang, J. A Hybrid Prediction Method for Bridging GPS Outages in High-Precision POS Application. IEEE Trans. Instrum. Meas. 2014, 63, 1656–1665. [Google Scholar] [CrossRef]
  23. Jingsen, Z.; Wenjie, Z.; Bo, H.; Yali, W. Integrating Extreme Learning Machine with Kalman Filter to Bridge GPS Outages. In Proceedings of the 2016 3rd International Conference on Information Science and Control Engineering, ICISCE 2016, Beijing, China, 8–10 July 2016; pp. 420–424. [Google Scholar] [CrossRef]
  24. Huang, X.L.; Ma, X.; Hu, F. Editorial: Machine Learning and Intelligent Communications. Mob. Netw. Appl. 2018, 23, 68–70. [Google Scholar] [CrossRef] [Green Version]
  25. Tsiropoulou, E.E.; Mitsis, G.; Papavassiliou, S. Interest-aware energy collection & resource management in machine to machine communications. Ad Hoc Netw. 2018, 68, 48–57. [Google Scholar] [CrossRef]
  26. Zhang, X.; Mu, X.; Liu, H.; He, B.; Yan, T. Application of Modified EKF Based on Intelligent Data Fusion in AUV Navigation. In Proceedings of the 2019 IEEE Underwater Technology (UT), Kaohsiung, Taiwan, 16–19 April 2019; pp. 1–4. [Google Scholar] [CrossRef]
  27. Sabet, M.T.; Mohammadi Daniali, H.; Fathi, A.; Alizadeh, E. A Low-Cost Dead Reckoning Navigation System for an AUV Using a Robust AHRS: Design and Experimental Analysis. IEEE J. Ocean. Eng. 2018, 43, 927–939. [Google Scholar] [CrossRef]
  28. Allotta, B.; Caiti, A.; Chisci, L.; Costanzi, R.; Di Corato, F.; Fantacci, C.; Fenucci, D.; Meli, E.; Ridolfi, A. An unscented Kalman filter based navigation algorithm for autonomous underwater vehicles. Mechatronics 2016, 39, 185–195. [Google Scholar] [CrossRef]
  29. Julier, S.; Uhlmann, J.; Durrant-Whyte, H. A new method for the nonlinear transformation of means and covariances in filters and estimators. IEEE Trans. Autom. Control 2000, 45, 477–482. [Google Scholar] [CrossRef] [Green Version]
  30. Karimi, M.; Bozorg, M.; Khayatian, A.R. A comparison of DVL/INS fusion by UKF and EKF to localize an autonomous underwater vehicle. In Proceedings of the 2013 First RSI/ISM International Conference on Robotics and Mechatronics (ICRoM), Tehran, Iran, 13–15 February 2013; pp. 62–67. [Google Scholar] [CrossRef]
  31. Tsyganova, J.V.; Kulikova, M.V. SVD-Based Kalman Filter Derivative Computation. IEEE Trans. Autom. Control 2017, 62, 4869–4875. [Google Scholar] [CrossRef] [Green Version]
  32. Huang, G.; Mourikis, A.; Roumeliotis, S. On the complexity and consistency of UKF-based SLAM. In Proceedings of the 2009 IEEE International Conference on Robotics and Automation, Kobe, Japan, 12–17 May 2009; pp. 4401–4408. [Google Scholar] [CrossRef]
  33. Simon, D. Training radial basis neural networks with the extended Kalman filter. Neurocomputing 2002, 48, 455–475. [Google Scholar] [CrossRef] [Green Version]
  34. Wang, Y.; Chai, S.; Nguyen, H.D. Experimental and numerical study of autopilot using Extended Kalman Filter trained neural networks for surface vessels. Int. J. Nav. Archit. Ocean. 2020, 12, 314–324. [Google Scholar] [CrossRef]
  35. Kurban, T.; Beşdok, E. A Comparison of RBF Neural Network Training Algorithms for Inertial Sensor Based Terrain Classification. Sensors 2009, 9, 6312–6329. [Google Scholar] [CrossRef] [Green Version]
  36. Dong, X.; Wu, J.; Wang, S.; Chen, T. An improved CDKF algorithm based on RBF neural network for satellite attitude determination. In Proceedings of the 2012 International Conference on Image Analysis and Signal Processing, IASP 2012, Huangzhou, China, 9–11 November 2012; pp. 155–161. [Google Scholar] [CrossRef]
  37. Pesce, V.; Silvestrini, S.; Lavagna, M. Radial basis function neural network aided adaptive extended Kalman filter for spacecraft relative navigation. Aerosp. Sci. Technol. 2020, 96, 105527. [Google Scholar] [CrossRef]
  38. Chen, T.; Chen, H. Approximation capability to functions of several variables, nonlinear functionals, and operators by radial basis function neural networks. IEEE Trans. Neural Netw. 1995, 6, 904–910. [Google Scholar] [CrossRef] [Green Version]
  39. Markopoulos, A.P.; Georgiopoulos, S.; Manolakos, D.E. On the use of back propagation and radial basis function neural networks in surface roughness prediction. J. Ind. Eng. Int. 2016, 12, 389–400. [Google Scholar] [CrossRef] [Green Version]
  40. Farrell, J.A. Aided Navigation GPS with High Rate Sensors, 1st ed.; The McGraw-Hill Companies: New York, NY, USA, 2008; p. 553. [Google Scholar]
  41. Savage, P.G. Strapdown Inertial Navigation Integration Algorithm Design Part 2: Velocity and Position Algorithms. J. Guid. Control. Dyn. 1998, 21, 208–221. [Google Scholar] [CrossRef]
  42. Miller, P.A.; Farrell, J.A.; Zhao, Y.; Djapic, V. Autonomous Underwater Vehicle Navigation. IEEE J. Ocean Eng. 2010, 35, 663–678. [Google Scholar] [CrossRef]
  43. Karmozdi, A.; Hashemi, M.; Salarieh, H. Design and practical implementation of kinematic constraints in Inertial Navigation System-Doppler Velocity Log (INS-DVL)-based navigation. Navigation 2018, 65, 629–642. [Google Scholar] [CrossRef]
  44. Titterton, D.; Weston, J. Strapdown Inertial Navigation Technology, 2nd ed.; Institution of Engineering and Technology, The Institution of Engineering and Technology, Michael Faraday House: Hertfordshire, UK, 2004. [Google Scholar] [CrossRef]
  45. Valenti, R.G.; Dryanovski, I.; Xiao, J. Keeping a good attitude: A quaternion-based orientation filter for IMUs and MARGs. Sensors 2015, 15, 19302–19330. [Google Scholar] [CrossRef] [Green Version]
  46. Grewal, M.S.; Andrews, A.P.; Bartone, C.G. Global Navigation Satellite Systems, Inertial Navigation, and Integration; Wiley: New York, NY, USA, 2020. [Google Scholar] [CrossRef]
  47. Chatfield, A.B. Fundamentals Of High Accuracy Inertial Navigation; American Institute of Aeronautics and Astronautics: Reston, VA, USA, 1997. [Google Scholar] [CrossRef]
  48. Wang, Q.; Cui, X.; Li, Y.; Ye, F. Performance enhancement of a USV INS/CNS/DVL integration navigation system based on an adaptive information sharing factor federated filter. Sensor 2017, 17, 239. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  49. He, K.; Liu, H.; Wang, Z. A novel adaptive two-stage information filter approach for deep-sea USBL/DVL integrated navigation. Sensors 2020, 20, 6029. [Google Scholar] [CrossRef]
  50. Hegrenaes, O.; Ramstad, A.; Pedersen, T.; Velasco, D. Validation of a new generation DVL for underwater vehicle navigation. In Proceedings of the 2016 IEEE/OES Autonomous Underwater Vehicles (AUV), Tokyo, Japan, 6–9 November 2016; pp. 342–348. [Google Scholar] [CrossRef]
  51. Kang, Y.; Zhao, L.; Cheng, J.; Wu, M.; Fan, X. A Novel Grid SINS/DVL Integrated Navigation Algorithm for Marine Application. Sensors 2018, 18, 364. [Google Scholar] [CrossRef] [Green Version]
  52. Christ, R.D.; Wernli, R.L. Underwater Acoustics and Positioning. In The ROV Manual; Elsevier Ltd.: Amsterdam, The Netherland, 2007; pp. 81–124. [Google Scholar] [CrossRef]
  53. Healey, A.; An, E.; Marco, D. Online compensation of heading sensor bias for low cost AUVs. In Proceedings of the 1998 Workshop on Autonomous Underwater Vehicles (Cat. No.98CH36290), Cambridge, MA, USA, 21 August 1998; pp. 35–42. [Google Scholar] [CrossRef] [Green Version]
  54. Fanelli, F. Development and Testing of Navigation Algorithms for Autonomous Underwater Vehicles, 1st ed.; Springer Theses; Springer International Publishing: Cham, Sweden, 2020. [Google Scholar] [CrossRef]
  55. Roumeliotis, S.I.; Sukhatme, G.S.; Bekey, G.A. Circumventing Dynamic Modeling: Evaluation of the Error-State Kalman Filter applied to Mobile Robot Localization. In Proceedings of the IEEE International Conference on Robotics and Automation, Detroit, MI, USA, 10–15 May 1999. [Google Scholar] [CrossRef]
  56. Rogers, R.M. Applied Mathematics in Integrated Navigation Systems, 3rd ed.; American Institute of Aeronautics and Astronautics: Reston, VA, USA, 2007. [Google Scholar] [CrossRef] [Green Version]
  57. Foss, H.T.H.; Meland, E. Sensor Integration for Nonlinear Navigation System in Underwater Vehicles. Ph.D. Thesis, Norwegian University of Science and Technology, Trondheim, Norway, 2007. [Google Scholar]
  58. Emami, M.; Taban, M.R. A Low Complexity Integrated Navigation System for Underwater Vehicles. J. Navig. 2018, 71, 1161–1177. [Google Scholar] [CrossRef]
  59. Dinc, M.; Hajiyev, C. Integration of navigation systems for autonomous underwater vehicles. J. Mar. Eng. Technol. 2015, 14, 32–43. [Google Scholar] [CrossRef]
  60. Yu, H.; Xie, T.; Paszczynski, S.; Wilamowski, B.M. Advantages of Radial Basis Function Networks for Dynamic System Design. IEEE Trans. Ind. Electron. 2011, 58, 5438–5450. [Google Scholar] [CrossRef]
  61. Wu, Y.; Wang, H.; Zhang, B.; Du, K.L. Using Radial Basis Function Networks for Function Approximation and Classification. ISRN Appl. Math. 2012, 2012, 1–34. [Google Scholar] [CrossRef] [Green Version]
  62. Kalman, R.E. A New Approach to Linear Filtering and Prediction Problems. J. Basic Eng. 1960, 82, 35–45. [Google Scholar] [CrossRef] [Green Version]
  63. Solà, J. Quaternion kinematics for the error-state Kalman filter. arXiv 2017, arXiv:1711.02508. [Google Scholar]
  64. Bjaili, H.A.; Moinuddin, M.; Rushdi, A.M. A State-Space Backpropagation Algorithm for Nonlinear Estimation. Circuits Syst. Signal Process. 2019, 38, 3682–3696. [Google Scholar] [CrossRef]
  65. Zhao, Y.; Wang, D.; Wang, L. Convolution Accelerator Designs Using Fast Algorithms. Algorithms 2019, 12, 112. [Google Scholar] [CrossRef] [Green Version]
  66. Dul, F.; Lichota, P.; Rusowicz, A. Generalized Linear Quadratic Control for a Full Tracking Problem in Aviation. Sensors 2020, 20, 2955. [Google Scholar] [CrossRef] [PubMed]
  67. Wei, L.; Cappelle, C.; Ruichek, Y. Camera/Laser/GPS Fusion Method for Vehicle Positioning Under Extended NIS-Based Sensor Validation. IEEE Trans. Instrum. Meas. 2013, 62, 3110–3122. [Google Scholar] [CrossRef]
  68. Pham, M.D.; Low, K.S.; Goh, S.T.; Chen, S. Gain-scheduled extended kalman filter for nanosatellite attitude determination system. IEEE Trans. Aerosp. Electron. Syst. 2015, 51, 1017–1028. [Google Scholar] [CrossRef]
Figure 1. Frames of references: ECI, ECEF, NED and Body.
Figure 1. Frames of references: ECI, ECEF, NED and Body.
Sensors 21 01149 g001
Figure 2. Top level diagram of the RBF-ESKF multi-sensor fusion navigation architecture.
Figure 2. Top level diagram of the RBF-ESKF multi-sensor fusion navigation architecture.
Sensors 21 01149 g002
Figure 3. Complementary form representation of RBF-ESKF.
Figure 3. Complementary form representation of RBF-ESKF.
Sensors 21 01149 g003
Figure 4. (a) A comparison of the 2D trajectory of an underwater vehicle in the east and north directions and (b) a comparison of the 3D trajectory of an underwater vehicle.
Figure 4. (a) A comparison of the 2D trajectory of an underwater vehicle in the east and north directions and (b) a comparison of the 3D trajectory of an underwater vehicle.
Sensors 21 01149 g004
Figure 5. Simulation results of ESKF and RBF-ESKF for time 0–2500 s (x-axis) in case 1. (a,b) The RBF-ESKF estimation for the east positions and depth error is close to the actual error value. (c,d) Error velocity for down and north. (e,f) Euler angle errors for roll and pitch. It is evident from (cf) that ESKF has an oscillatory response that contributes to compromised accuracy compared to RBF-ESKF.
Figure 5. Simulation results of ESKF and RBF-ESKF for time 0–2500 s (x-axis) in case 1. (a,b) The RBF-ESKF estimation for the east positions and depth error is close to the actual error value. (c,d) Error velocity for down and north. (e,f) Euler angle errors for roll and pitch. It is evident from (cf) that ESKF has an oscillatory response that contributes to compromised accuracy compared to RBF-ESKF.
Sensors 21 01149 g005
Figure 6. Simulation results of ESKF and RBF-ESKF for time 0–3000 s (x-axis) in case 2. (ac) The position error significantly grows with a loss of acoustic fix. The proposed algorithm estimation is close to the actual error. (df) RBF-EKF has a minimal effect of acoustic fix loss on velocity estimation, and the response is much smoother than that of ESKF.
Figure 6. Simulation results of ESKF and RBF-ESKF for time 0–3000 s (x-axis) in case 2. (ac) The position error significantly grows with a loss of acoustic fix. The proposed algorithm estimation is close to the actual error. (df) RBF-EKF has a minimal effect of acoustic fix loss on velocity estimation, and the response is much smoother than that of ESKF.
Sensors 21 01149 g006
Figure 7. Simulation results of ESKF and RBF-ESKF for time 0–3000 s (x-axis) in case 3. As the error increases, nonlinearity in the error also increases. (ac) The velocity error for the north, east, and downward directions increase significantly with DVL loss. The proposed velocity error estimate of the algorithm is similar to real error. (df) RBF-ESKF has a smoother response and faster convergence of the full state velocity estimation.
Figure 7. Simulation results of ESKF and RBF-ESKF for time 0–3000 s (x-axis) in case 3. As the error increases, nonlinearity in the error also increases. (ac) The velocity error for the north, east, and downward directions increase significantly with DVL loss. The proposed velocity error estimate of the algorithm is similar to real error. (df) RBF-ESKF has a smoother response and faster convergence of the full state velocity estimation.
Sensors 21 01149 g007
Table 1. Mathematical notations for underwater vehicle kinematics.
Table 1. Mathematical notations for underwater vehicle kinematics.
NotationDescription
v b [u,v,w]Linear velocity in the body frame (surge, sway, and heave)
Θ n b [ ϕ , θ , ψ ] Attitude in Euler angles from the body to NED frame
q b n [ η , ϵ ] Attitude in quaternion from the body to NED frame
ω b [p,q,r]Angular velocity in the body frame (roll, pitch, and yaw)
p n [ n , e , d ] Position in the NED frame (north, east, and down)
v n [ v N , v E , v D ] Linear velocity in the NED frame (north, east, and down)
ω n [ ω N , ω E , ω D ] Angular velocity in the NED frame
p e [ x , y , z ] Position in the ECEF frame
v e [ v e x , v e y , v e z ] Linear velocity in the ECEF frame
p e [ ϕ l a t , λ l o n g , h d ] Position in the ECEF geoid (latitude, longitude, and depth)
R b e Rotation matrix from the body to ECEF frame
Ω Skew symmetric matrix of the angular velocity
Ω e Skew symmetric matrix in the ECEF frame
Ω b Skew symmetric matrix in the body frame
g e Earth gravity vector in the ECEF frame
g n Earth gravity vector in the NED frame
R M Radius of curvature of the prime vertical of Earth
R N Radius of curvature of the meridian of Earth
aSemi-major axis of the ellipsoidal Earth model
eEccentricity of the ellipse approximation of Earth
Table 2. RBF neural network structure with a Gaussian activation function.
Table 2. RBF neural network structure with a Gaussian activation function.
RBFNNNumbers
Input layer neurons20
Hidden layer neurons50
output layer neurons20
Learning rate of weights0.001
Learning rate of centers0.001
Table 3. ESKF and RBF-ESKF results with all sensors working in normal condition by running 100 Monte Carlo simulations.
Table 3. ESKF and RBF-ESKF results with all sensors working in normal condition by running 100 Monte Carlo simulations.
ESKFRBF-ESKF
North Position Max error1.15090.32578
East Position Max error0.82180.45685
Down Position Max error0.00817860.0071222
North Position RMSE0.4461240.1464
East Position RMSE0.31220.1844
Down Position RMSE0.00407190.002709
Sum Position RMSE0.762390.333509
North Velocity Max error0.0381850.01844
East Velocity Max error0.00459430.0037403
Down Velocity Max error0.00325070.0022476
North Velocity RMSE0.0394550.011407
East Velocity RMSE0.0336520.012919
Down Velocity RMSE0.00355260.020227
Sum Velocity RMSE0.07665960.0263487
Roll Max error0.132680.060874
Pitch Max error0.18090.16171
Yaw Max error0.466240.36001
Roll RMSE0.000395540.00019122
Pitch RMSE0.000555970.00036354
Yaw RMSE0.000498740.00035049
Sum Attitude RMSE0.001450250.00122578
Table 4. Performance Comparison of ESKF and RBF-ESKF with loss of acoustic fix for a short period by running 100 Monte Carlo simulations.
Table 4. Performance Comparison of ESKF and RBF-ESKF with loss of acoustic fix for a short period by running 100 Monte Carlo simulations.
ESKFRBF-ESKF
North Position Max error5.21151.3849
East Position Max error1.42750.4663
Down Position Max error0.0331670.016192
North Position RMSE0.940360.50629
East Position RMSE0.665110.40109
Down Position RMSE0.0056130.0039411
Sum Position RMSE1.6110830.9113211
North Velocity Max error0.0403760.029489
East Velocity Max error0.00365130.0030517
Down Velocity Max error0.00612160.0036592
North Velocity RMSE0.04393240.21818
East Velocity RMSE0.0657220.035722
Down Velocity RMSE0.00654510.0039131
Sum Velocity RMSE0.11619950.0614531
Roll Max error0.25670.1231
Pitch Max error0.415820.19178
Yaw Max error0.54940.46001
Roll RMSE0.00061350.00020485
Pitch RMSE0.000523580.00011917
Yaw RMSE0.000513630.0003413
Sum Attitude RMSE0.001690710.00117532
Table 5. Performance comparison of ESKF and RBF-ESKF with the Doppler velocity log (DVL) measurement unavailable for a short duration by running 100 Monte Carlo simulations.
Table 5. Performance comparison of ESKF and RBF-ESKF with the Doppler velocity log (DVL) measurement unavailable for a short duration by running 100 Monte Carlo simulations.
ESKFRBF-ESKF
North Position Max error1.94550.90707
East Position Max error0.916490.63511
Down Position Max error0.00841060.0072865
North Position RMS error0.6693240.41818
East Position RMS error0.537220.30722
Down Position RMS error0.00494550.0035131
Sum Position RMS error1.21148950.7289131
North Velocity Max error0.1057510.051603
East Velocity Max error0.0990180.061305
Down Velocity Max error0.01784540.00897
North Velocity RMS error0.1912960.09809
East Velocity RMS error0.153370.079982
Down Velocity RMS error0.00924550.005387
Sum Velocity RMS error0.35391150.183459
Roll Max error0.299350.171758
Pitch Max error0.519060.210442
Yaw Max error0.660410.42246
Roll RMSE0.000664810.000450075
Pitch RMSE0.000600170.00042717
Yaw RMSE0.000622050.000422102
Sum Attitude RMSE0.001887030.001299347
Table 6. Execution time (seconds).
Table 6. Execution time (seconds).
ESKFESKF-RBF
0.00280.0039
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Shaukat, N.; Ali, A.; Javed Iqbal, M.; Moinuddin, M.; Otero, P. Multi-Sensor Fusion for Underwater Vehicle Localization by Augmentation of RBF Neural Network and Error-State Kalman Filter. Sensors 2021, 21, 1149. https://doi.org/10.3390/s21041149

AMA Style

Shaukat N, Ali A, Javed Iqbal M, Moinuddin M, Otero P. Multi-Sensor Fusion for Underwater Vehicle Localization by Augmentation of RBF Neural Network and Error-State Kalman Filter. Sensors. 2021; 21(4):1149. https://doi.org/10.3390/s21041149

Chicago/Turabian Style

Shaukat, Nabil, Ahmed Ali, Muhammad Javed Iqbal, Muhammad Moinuddin, and Pablo Otero. 2021. "Multi-Sensor Fusion for Underwater Vehicle Localization by Augmentation of RBF Neural Network and Error-State Kalman Filter" Sensors 21, no. 4: 1149. https://doi.org/10.3390/s21041149

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop