Next Article in Journal
Assessing Temporal Trade-Offs of Ecosystem Services by Production Possibility Frontiers
Next Article in Special Issue
Performance Evaluation of Mangrove Species Classification Based on Multi-Source Remote Sensing Data Using Extremely Randomized Trees in Fucheng Town, Leizhou City, Guangdong Province
Previous Article in Journal
Remote Sensing Applications to Climate Change
Previous Article in Special Issue
Research on Self-Noise Suppression of Marine Acoustic Sensor Arrays
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Method for Long-Term Target Anti-Interference Tracking Combining Deep Learning and CKF for LARS Tracking and Capturing

1
School of Mechanical and Electrical Engineering, Guangzhou University, Guangzhou 510006, China
2
Guangdong Institute of Intelligent Unmanned System, Guangzhou 511458, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(3), 748; https://doi.org/10.3390/rs15030748
Submission received: 24 November 2022 / Revised: 20 January 2023 / Accepted: 24 January 2023 / Published: 28 January 2023

Abstract

:
Autonomous underwater vehicles (AUV) recycling in an underwater environment is particularly challenging due to the continuous exploitation of marine resources. AUV recycling via visual technology is the primary method. However, the current visual technology is limited by harsh sea conditions and has problems, such as poor tracking and detection. To solve these problems, we propose a long-term target anti-interference tracking (LTAT) method, which integrates Siamese networks, You Only Look Once (YOLO) networks and online learning ideas. Meanwhile, we propose using the cubature Kalman filter (CKF) for optimization and prediction of the position. We constructed a launch and recovery system (LARS) tracking and capturing the AUV. The system consists of the following parts: First, images are acquired via binocular cameras. Next, the relative position between the AUV and the end of the LARS was estimated based on the pixel positions of the tracking AUV feature points and binocular camera data. Finally, using a discrete proportion integration differentiation (PID) method, the LARS is controlled to capture the moving AUV via a CKF-optimized position. To verify the feasibility of our proposed system, we used the robot operating system (ROS) platform and Gazebo software to simulate the system for experiments and visualization. The experiment demonstrates that in the tracking process when the AUV makes a sinusoidal motion with an amplitude of 0.2 m in the three-dimensional space and the relative distance between the AUV and LARS is no more than 1 m, the estimated position error of the AUV does not exceed 0.03 m. In the capturing process, the final capturing error is about 28 mm. Our results verify that our proposed system has high robustness and accuracy, providing the foundation for future AUV recycling research.

Graphical Abstract

1. Introduction

AUV applications began in the 1990s. Since then, various types of AUV have been applied to marine science and technology investigations [1,2,3], seabed exploration [4], underwater pipeline inspection and maintenance [5], resource exploration [6], etc., due to advancements in power enhancement and energy utilization [7]. An AUV is loaded with energy for underwater operations [8], which means that an AUV needs to return to the mother ship in time for energy replenishment, data upload, equipment maintenance, and mission updates [9]. Reliable launch and recycling techniques are essential to ensure the dependability of AUV for recycling [10].
AUV recovery technology can be divided into water recovery and underwater recovery. The recovery of AUVs by watercraft is the current mainstream direction, and a large number of related devices have been successfully used [11,12]. Water recovery can be divided into hanging type and slide type, but the application scope and operability of the hanging type are more robust [13,14]. Therefore, we mainly study the technology of recovering AUVs by using hanging equipment on the water. In the past, AUV’s position in the recovery process of hanging equipment was mainly through human observation, which could have been more reliable and put the staff in a dangerous environment. Therefore, the close-range positioning technology of AUV is essential to research.
Compared with acoustic sensors, GPS, and radar, vision technologies are more suitable for motion measurement during the launch and recovery of AUV with close-range, real-time, and high-precision demand [15,16,17]. However, there are still relatively few examples of AUV recycling in extreme sea conditions. The position and orientation of AUVs frequently change due to harsh sea conditions. Challenges arise from large appearance variances caused by illumination, deformation, occlusion, and motion [18,19], which can impact tracking, position, and LARS control. The application scenarios of past visual technology to track and locate targets were relatively simple [20]. For example, Wu et al. [21] proposed an autonomous UAV tracking and landing system based on YOLO with high tracking effect and control accuracy levels; however, it depends on the pre-training of the YOLO network [22], making its universality limited. Liu et al. [23] proposed an AUV recovery technology based on the visual framework and verified the feasibility of autonomously recovering AUV using visual technology through experiments. The existing vision technology is mainly for underwater and air environment. There are few visual technologies for a surface ship to recover AUV.
Previous papers and methods have not proposed a system with high robustness, tracking, and real-time performance levels for LARS tracking and capturing [24,25]. To solve the above problems, our proposed system provides improvements for tracking and recycling the AUV in complex marine environments. The key is that the proposed system incorporates deep learning, CKF, and LARS with a complete process and reliable performance. The contributions of our work are:
  • We propose the LTAT method to achieve the designated AUV’s long-term pixel position information, integrating deep learning and online learning ideas. This effectively solves the inevitable tracking loss problem caused by complex sea conditions during long-term tracking.
  • We optimize and anticipate the AUV’s position via CKF to lessen the influence of interfering data on position.
  • We use pixel position information and binocular camera data to obtain the AUV’s coordinates in the camera coordinate system and estimate its orientation. We obtain the relative positions between the end of LARS and the AUV by calculating the coordinate transformation relationship.
  • We design the motion trajectory of the end of the LARS using a five-polynomial interpolation method. We use the discrete PID method to control the motion trajectory of the LARS. Based on our proposed system, the complete process of LARS tracking and capturing AUV is virtually verified via a physical simulation system.
The rest of this paper is structured as follows: In Section 2, we describe the framework of the LARS tracking and control system in detail, and we also demonstrate the LTAT method and the CKF in concrete terms. Section 3 shows the simulation of our primary method and system process. In Section 4, we analyze the advantages of our proposed system and its subsequent improvements based on the experimental findings. Finally, we provide our conclusions in Section 5.

2. Unmanned LARS Tracking and Capturing System

In this section, we show the complete flow of our proposed system and describe four main modules in detail, including pixel coordinate tracking, relative pose estimation, data optimization, and actuator control.
The system hardware includes the AUV, binocular camera, and three-axis mechanical arm. For the AUV, we selected the Explore-100 produced by Shenyang Institute of Automation Chinese Academy of Sciences for ocean exploration as the experimental model, which has been successfully applied to various sea trial experiments. The system software framework of this paper is shown in Figure 1. First, the AUV is specified in the image using the binocular camera, and the Siamese region proposal network (SiamRPN) is used to track the specified AUV and output the AUV’s pixel position. Tracking has the following two cases: The first is to continuously track the target without interference. The tracking pixel position can be accurately output, and the resulting tracking image output is added to the system database. In the second, when the target has interfered, the YOLO recognizes the potential presence of target regions in the image. The optimal region is filtered using an empirical threshold to update the SiamRPN template frames and pixel positions. Next, the AUV’s pose and position are calculated using the binocular stereo-matching principle based on the feature points’ pixel positions. Then we used CKF to optimize the positioning data. Finally, using the CKF-optimized position, the LARS is controlled via a discrete PID method to track and capture the AUV.

2.1. Pixel Coordinates Acquisition of AUV

An AUV’s position is the most important data to control LARS. Due to the complex and harsh marine environment, which has a considerable impact on system robustness and position accuracy, we propose the LTAT method for AUV tracking and use CKF to optimize the AUV position.
The SiamRPN provides excellent tracking performance for objects that abruptly change and deform. However, after the target is obscured, it cannot determine the target’s position and achieve unmanned re-tracking. The YOLOv5 algorithm has excellent detection speed and accuracy; however, it relies on pre-training and is not effective in detecting targets with deformations and mutations. Combining the advantages and disadvantages of the above two methods, our proposed LTAT method integrates SiamPRN and YOLOv5 with the idea of online learning to achieve highly robust tracking while improving generalizability.

2.1.1. Theoretical Background

SiamRPN is a Siamese network structure based on the RPN [26], which completes target tracking by matching the similarity of the template and test frames [27]. Unlike deep learning-related detection algorithms, it does not require pre-training regarding network structure weights. SiamRPN has universal applicability because it achieves target tracking through end-to-end offline training [28]. Moreover, the algorithm has a simple structure and solves interferences in the tracking process due to illumination, deformation, occlusion, and motion. SiamRPN consists of a Siamese subnetwork and a region-proposal subnetwork, and its network structure is shown in Figure 2. The network outputs the tracked target’s offset center coordinate, as well as the tracked area’s length and width offset.
The YOLO series algorithm is a single-stage detection method based on deep learning [29]. It divides the image into cells and predicts the bounding box parameters and confidence, as well as the cell categories [30]. YOLOv5 is the fifth generation update in the YOLO series. It has good detection accuracy and is widely used in engineering applications, where YOLOv5s is the network structure with the fastest operation rate when outputting the optimal predicted target’s center coordinates and the boxed region’s size.

2.1.2. Long-Term Target Anti-Interference Tracking Method

Because the YOLO relies on network weights trained on large-scale data, SiamRPN does not have a position-update function after target loss, which leads to missing positions when the target is obscured, resulting in a loss of field view. We propose the LTAT algorithm shown in Figure 3 by integrating the SiamRPN’s excellent tracking performance with YOLOv5’s high-precision detection effects, utilizing online learning ideas to obtain more accurate and robust target-pixel positions.
We used binocular cameras to capture photos; then, we filtered the image noise to improve clarity. Next, we specified the target to be tracked in the image as the template frame. We used the Siamese network to extract features from the template and detection frames, and then we input the features into the RPN’s classification and regression branches. The classification branch outputs the category prediction result of each anchor, and the regression branch outputs the center point and bounding box offset of each anchor. Finally, we obtained the center point and bounding box of the tracking target in the detection frame through non-maximum suppression. In order to accomplish the object’s continuous tracking effect, we also employed the detection frame’s tracking result as a template frame for the following detection frame.
The offline training of the detection algorithm based on the neural network requires a lot of effective data to adjust the network parameters; however, there are few training sets for AUV. In particular, AUVs do not have a unified appearance, and our study observes the AUV from above the water’s surface. The presence of the refraction phenomenon led to the AUV’s appearance being more unpredictable. All of the above reasons make it impossible to effectively track AUVs through offline training. Therefore, based on the idea of online learning, we utilized the excellent results of the initial tracking experiment for a period of time as system data to train the lightweight YOLOv5 network online. On the one hand, this solved the problem regarding the lack of training data sets. On the other hand, it makes the training data more targeted. Therefore, the online learning method effectively enhanced the system’s adaptability and anti-interference performance.
When the tracking target was disturbed and SiamRPN’s tracking performance was poor, we used the YOLOv5 to find a potential target region in the image. When the potential target region’s score reaches the empirical threshold, output the ideal region and update the SiamRPN tracking model. In conclusion, even when the target is significantly disturbed, the LTAT approach can still track it for a considerable amount of time and determine its pixel position.

2.2. AUV Position and Orientation Estimation

The positioning and tracking of AUV is based on three-dimensional space, so it is necessary to obtain the three-dimensional positioning information of AUV. We considered that AUV’s recycling needs to achieve close-range high-precision positioning in the marine environment. After analyzing the advantages and disadvantages of various sensors, we found that the binocular camera is the most suitable sensor in the cost, accuracy, and feasibility of the positioning sensors. The binocular camera positioning technology has good positioning accuracy and can realize dynamic positioning in the outdoor environment. Therefore, our research uses the binocular camera to obtain AUV positioning and uses CKF to optimize positioning data in positioning estimation.

2.2.1. Coordinate System Conversion Relationship

To control the LARS movement, the relative positions of AUV and LARS need to be calculated in real-time. We set the binocular camera to be installed at the end of the LARS with the lens vertically downward. For the purpose of estimating the relative positions of AUV and LARS, we defined six-coordinate systems, as shown in Figure 4. Set the base of LARS as the world coordinate system O w ( X w , Y w , Z w ) ; camera coordinate system O c ( X c , Y c , Z c ) ; LARS end coordinate system O t ( X t , Y t , Z t ) ; calibration plate coordinate system O g ( X g , Y g , Z g ) ; and the image coordinate system O p 1 (x, y) and pixel coordinate system O p 2 (u, v). The conversion relationship coordinate system is the key method to estimate relative pose, which is described as follows.
We obtained the image pixel coordinates (u, v) of the target using feature point extraction. Generally, the left camera’s optical center is set as the origin of the camera coordinate system, and the camera coordinate system is parallel to the image coordinate system O p 1 . So, we first obtained the conversion relationship from the pixel coordinate system to the image coordinate system. Let dx and dy denote the width and length of the unit pixel, respectively, in the image coordinate system and ( u 0 , v 0 ) denote the center point of the pixel coordinate system O p 2 . The conversion relationship between the pixel coordinate system and the image coordinate system is:
u v 1 = 1 d x 0 u 0 0 1 d y v 0 0 0 1 x y 1
We used a binocular camera to obtain the image of the target. The disparity of the feature points in the binocular image can be used to determine the coordinates of the target in the camera coordinate system O c [31]. We set the coordinates of matching feature points on binocular images to ( x L , y L ), ( x R , y R ). Let f denote the focal length of the left and right cameras and B denote the binocular baseline distance. Furthermore, the conversion relationship between image coordinate (x, y) and camera coordinate ( X c , Y c , Z c ) is:
X c = x L x L x R B Y c = y L x L x R B Z c = f x L x R B
The coordinate system’s conversion from pixel coordinate system to camera coordinate system is realized above. Because the camera is set on LARS, the conversion relationship between the camera and the LARS end coordinate system needs to be obtained through hand-eye calibration to realize the capturing of LARS. Let t b H denote the conversion relationship between the camera and LARS end coordinate system, and let g c H denote the conversion relationship between the camera and calibration plate coordinate system, the conversion relationship c t H between the camera and LARS end coordinate system is:
( t b H j ) 1 × t b H i × c t H = c t H j × ( g c H j ) 1 × g c H i
In the above equation, t b H can be obtained using external parameters of camera calibration, while g c H can be obtained using forward kinematics.
Using the above conversion relationship, the AUV position can be converted from the pixel coordinate system to the LARS base coordinate system. This means that we can locate the AUV according to the feature point information of the AUV and control the LARS tracking and capturing based on the accurate position.

2.2.2. Orientation Measurement Method

Feature point extraction and matching of images is a crucial step for the system to position AUV. For the purpose of realizing the function of accurate real-time position, we designed an identifier and adopted the method of sparse stereo matching. Because the triangle’s corner-point detection is more distinguished and has fewer feature points, it can meet the demands of practical engineering for tracking and position. Therefore, we designed an isosceles triangle with side lengths of 0.15 m, 0.4 m and 0.4 m on the surface of the AUV as the marker for tracking the target, as shown in Figure 5.
The appearance of our design is based on the exploration-100 AUV, and its specific physical parameters are shown in Table 1.
The AUV’s pitch and yaw angles can be calculated by detecting the target area’s three corner points. Let A 1 ( x 1 , y 1 , z 1 ), A 2 ( x 2 , y 2 , z 2 ), and A 3 ( x 3 , y 3 , z 3 ) denote the triangular feature points and A 4 ( x 4 , y 4 , z 4 ) denote the midpoints in the image coordinate system. The triangular feature points can be accessed through the position estimation method, and the midpoint is as follows:
x 4 = x 1 + x 2 2 y 4 = y 1 + y 2 2 z 4 = z 1 + z 2 2 + D
where D is the fixed error caused by the marking on the cylindrical surface. The AUV’s yaw and pitch angles can be calculated via the midpoint as follows:
P i t c h = a r c t a n ( z 3 z 4 ) ( x 3 x 4 ) 2 + ( y 3 y 4 ) 2
Y a w = a r c t a n x 3 x 4 y 3 y 4
In the above equation, Pitch is the pitch angle of the AUV, and Yaw is the yaw angle for straight navigation.
By knowing the pitch and yaw angles, collisions between the LARS and AUV can be effectively avoided during the LARS movement, avoiding unnecessary risk and financial loss.

2.3. Position Optimization Based on CKF

There are disturbances, such as wave masking, light changes, and wave fluctuations, in the marine environment. These disturbances can cause errors and offsets in the positioning data. Kalman filter is a widely used data optimization method in the industry [32]. Considering that the AUV’s motion system is nonlinear, Kalman filters, such as Extended Kalman (EKF), Unscented Kalman (UKF), and CKF can be used. The specific implementation of EKF is to linearize the nonlinear system [33], which has a large error probability. CKF has a more rigorous and stable theoretical basis than UKF, and when the system state dimension is greater than three, the reliability factor of UKF increases with the increase in dimension, while the reliability and filtering accuracy are greatly reduced. At the same time, the CKF algorithm has higher execution efficiency and lower time complexity, which meets the real-time requirements of our experiment [34,35,36]. Therefore, position accuracy and anti-interference performance are considerably enhanced in our study by using CKF to optimize and anticipate the position.
We set the position and velocity of the AUV as the state vector of CKF, and the state vector at moment k is formed as follows:
X k = ( x k , y k , z k , v k x , v k y , v k z )
where η = ( x k , y k , z k ) is the position of AUV relative to the end of LARS at the moment k, and γ = ( v k x , v k y , v k z ) is the velocity of AUV at the moment k. In our study, we make the assumption that the AUV moves uniformly along the X-axis in a straight line, stays stationary along the Y-axis, and moves in a sine wave along the Z-axis. Consequently, the AUV’s state–space model is given as:
x k y k z k v k x v k y v k z = x k 1 + T v k 1 x y k 1 z k 1 + C T v k 1 z v k 1 x v k 1 y v k 1 z C T z k 1 + n k
where T is the system’s sampling interval time, C is determined by the sine wave period, and n k is shown as the process noise.
The mother ship obtains the AUV’s speed and position information through visual sensors. Let Z k = (x, y, z) denote the measurement vector and m k denote the measurement noise. The linear observation model is:
Z k = H X k + m k H = 1   0   0   0   0   0 0   1   0   0   0   0 0   0   1   0   0 0
Combing Equations (4)–(6), the discrete dynamic alignment model is written as:
X k = f ( X k 1 ) + n k Z k = h ( X k ) + m k
where f ( ) is the nonlinear function and h ( ) is the nonlinear measurement transition function. In the non-Gaussian filter estimation process, the key is to solve the multi-dimensional weighted integral [37], which can be formed as follows:
I ( f ) = R n f ( x ) W ( x ) d x
where R n is the region of the integration and W( x ) is the weight function that W( x ) ≥ 0. If the solution to the above integral is difficult to obtain, we use numerical integration methods instead. The basic task of numerically computing Equation (10) is to find a set of points ξ i and weights ω i that approximates the integral I ( f ) by a weighted sum of function evaluations:
I ( f ) i = 1 m ω i f ( ξ i )
The spherical-radial rule is used by the CKF to determine the points and weights [38]. The dimension of the random variable equals n, and the third-degree spherical-radial rule requires a total of 2n cubature points. The cubature points and associated weights are as follows:
ξ i = m 2 [ 1 ] i = m 2 1 0 , , 0 1 , 1 0 , , 0 1
ω i = 1 m , i = 1 , 2 , , m = 2 n
On the basis of the above, the steps of CKF are divided into two parts [39]: prediction and measurement updates. The state variable x ^ k , error covariance matrix P k , process noise covariance matrix Q k , and measurement noise covariance matrix R k are initialized at first. Then, the state matrix’s cubature point is calculated with the equation:
P k = S k S k T
x k i = S k ξ i + x ^ k i = 1 , 2 , , 2 n
where S k is the square root of the covariance matrix. Then, spread cubature points, calculate state prediction, and predict error covariance as follows:
x k + 1 | k i = f ( x k i )
x ^ k + 1 | k = ω i i = 1 m x k + 1 | k i
P k + 1 | k = ω i i = 1 m x k + 1 | k i ( x k + 1 | k i ) T x ^ k + 1 | k ( x ^ k + 1 | k ) T + Q k
Here are the procedures for the measurement update section. The equation to recalculate the cubature points and re-spread cubature points will be given as:
P k + 1 | k = S k + 1 | k S k + 1 | k T
X k + 1 | k i = S k + 1 | k ξ i + x ^ k + 1 | k
z k + 1 i = h ( X k + 1 | k i )
The measurement prediction, estimate error covariance, and cross-covariance matrix should all be calculated as follows:
z ^ k + 1 = ω i i = 1 m z k + 1 i
P k + 1 z = ω i i = 1 m z k + 1 i ( z k + 1 i ) T z ^ k + 1 ( z ^ k + 1 ) T + R k
P k + 1 x z = ω i i = 1 m X k + 1 | k i ( z k + 1 i ) T x ^ k + 1 | k ( z ^ k + 1 ) T
Finally, calculate Kalman gain, update state and error covariance matrix as follows:
K k + 1 = P k + 1 x z ( P k + 1 z ) 1
x ^ k + 1 = x ^ k + 1 | k + K k + 1 ( z k + 1 z ^ k + 1 )
P k + 1 = P k + 1 | k K k + 1 P k + 1 z K k + 1 T
Loss of vision is a relatively regular occurrence because of the harsh sea environment. Using CKF can prevent interference to position data due to the loss of vision. The anti-interference capabilities of the CKF-optimized position data are substantially enhanced, and it is closer to the true AUV trajectory.

2.4. LARS Control

Inverse kinematics can be used to determine the appropriate orientation of the LARS based on the relative position between the AUV and LARS base. We used three-degree-of-freedom LARS in our experiment [40]; therefore, it is necessary to plan the moving trajectory according to the expected pose and calculate the rotation angle, rotation angular velocity, and rotation angular acceleration of each joint. We adopted the PD control method in dynamic control to control the LARS. Through the link parameters, we can calculate the torque M (q) required to overcome the LARS acceleration inertia and its influence between different connecting rods, in addition to calculating torque C (q), which is required to overcome Coriolis and centrifugal forces, and torque G (q) which calculated to overcome the Earth’s gravity. The dynamic equation of the LARS is:
τ = M ( q ) q . . + C ( q , q . ) + G ( q )
In the above equation, τ is the control moment vector, q is the rotation angle vector of the system, q . is shown as the angular velocity vector of the system, and q . . is shown as the angular acceleration vector of the system.
We used a control method with negative feedback to lessen the inaccuracy levels because LARS’s estimated physical parameter values are inaccurate. Our PD control method is given as:
e . . = K d e . + K p e e = q d q
In the above equation, e is the position error, q d is the desired joint rotation angle, and q is the system’s control output rotation angle. K d is shown as the differential coefficient, which can reflect the variation trend of the deviation signal to reduce the control system’s oscillation. K p is shown as the proportional coefficient, which can proportionally reflect the control system’s deviation signal to speed up its dynamic response. Moreover, we assigned no integral coefficient because the LARS is in the dynamic tracking stage [41].

3. Simulation and Experimental Results

In this section, we show the results of the experiments we conducted using the LTAT method and position estimation. We verified the robustness of LTAT and the accuracy of position estimation under a variety of sea conditions.

3.1. Experimental Design of Simulation

In this paper, we used the ROS framework and Gazebo software to construct the physical simulation environment to realize the modeling, simulation, and visualization of the system’s complete process. In the simulation environment, we used LARS as the executive mechanism, and we completed the AUV’s model construction and appearance design using Solidworks and software Blender. We used the PyTorch framework and the OpenCV library to implement the LTAT method, where the GPU uses NVIDIA RTX3060. We verified the control method through Simulink on the MatLab platform. Our experimental procedure simulates the mother ship approaching the AUV while maintaining the same forward speed as the AUV; then, it catches the AUV in a reasonably stable scenario. The experimental steps are as follows:
  • First, we set LARS to be in an initial state, and we executed the end of LARS to obtain the image data. The image data’s simulated scene includes normal driving and bad sea conditions and similar interference. We mainly simulated severe sea conditions for the most common wave coverage interference. By reviewing the relevant literature [42,43], we found that the primary interference in the recovery process was the smaller ships during the simultaneous advance of the two ships, which exists in the direction of heave. Therefore, we set the AUV to perform sine wave movement on the Z axis, make a uniform linear motion on the X axis, and ensure that it is in a fixed position on the Y axis;
  • Second, we used a binocular camera to obtain ocean images, and then we used the LTAT method to track the specified AUV in the image. We used the tracking results to calculate the AUV’s position, as well as its pitch and yaw angles;
  • Finally, we regulated the end of LARS to follow the AUV’s position trajectory in accordance with the position results, avoiding collision with the AUV in the process. We operated the LARS to capture the AUV when it sinks on the Z axis.

3.2. Long-Term Target Anti-Interference Tracking Experiment

To test the LTAT method, we set the AUV trajectory using Gazebo. We simulated the most common AUV recycling states at sea, including the normal driving and wave-covered states, as well as the interference state of the analogs. To validate the generality of the LTAT, we collected 300 images of AUVs under wave coverage via Gazebo as the tracking data, and we used the labeling software to calibrate the tracking data’s true position. Using one-pass evaluation (OPE), we used the first true position as the LTAT method’s tracking target. We obtained the LTAT method’s success rate and accuracy by comparing the tracking value with the true value. The accuracy and success rates of SiamRPN and LTAT are shown in Figure 6. The LTAT method is superior to the SiamRPN in terms of accuracy and success rate.
The visual tracking results for AUVs interfered with by multi-target analogs are shown in Figure 7. The yellow box in Figure 7 is the AUV position obtained through LTAT tracking, and the blue box is the true position of the AUV. Figure 7a,b show the results of the tracking start and end, respectively. The results show that the LTAT method can keep tracking the AUV for a long time when it is disturbed by multi-target analogues.
The visual tracking results of AUVs interfered with by wave cover are shown in Figure 8. Figure 8a shows that the AUV begins to be affected by waves, which will cause the main part of the AUV to sink underwater. Figure 8b shows the subsequent loss of the AUV’s visual field under severe wave interference. Figure 8c shows the LTAT method’s automatic correction performance for the position after the AUV’s main body returns to view. Figure 8 shows that the LTAT method can autonomously correct the position and achieve long-term anti-interference tracking when the AUV is obscured by waves.

3.3. Relative Position and Orientation Estimation Experiments

On the X, Y, and Z axes, we examined trajectory position and position error. We used both severe and normal sea conditions for our experiments. Under normal sea conditions, AUV floats on the water surface without any external factors, and AUV is in an ideal state. AUV is an underactuated model, so it is impossible to independently adjust the depth of the level. To ensure the visual tracking and position of the AUV, we set its buoyancy to be sufficient to support its upper surface above the water surface. The severe sea condition in this paper refers to the situation that AUV is covered by waves. In the experiment, the error we set in the data is caused by the wave cover. Wave masking AUV will lead to an offset of the positioning data.
Figure 9 shows our comparison of the measured, true, and CKF trajectories of AUV motion in a normal sea state. Our comparisons of trajectories for the X and Z axes are shown in Figure 9a,c, respectively. The fact that the three trajectory curves almost overlap demonstrates that there are fewer estimation errors in the X and Z axes. After using CKF, the average value of relative error decreased. Our trajectory comparison of the Y-axis is shown in Figure 9b. There are a lot of similar data in the estimated trajectory and the scatter plot cannot be displayed visually; therefore, we sampled and fitted the estimated trajectory. The position error varies substantially, while the true value of the Y-axis is constant. However, employing the CKF greatly reduced the average position error value.
To make the position error more intuitive, we calculate the average position error of every 25 data and show the size and trend of the error through a histogram. The error histogram of the X-axis is shown in Figure 10a, the maximum measurement error is 30 mm, and the maximum error after using CKF is 20 mm. Figure 10b is the error histogram of the Y-axis, the maximum measurement error is 50 mm, and the maximum error after using CKF is 5 mm, which has been greatly improved. Figure 10c is the Z-axis error histogram, the maximum measurement error is 20 mm. After utilizing CKF, the maximum error remains unchanged, but the overall error value is decreased.
Figure 11 shows the comparison of measured trajectory, true trajectory, and CKF trajectory of AUV motion in severe sea conditions, which demonstrates that the position is seriously disturbed due to the influence of sea conditions. However, when the position is disturbed, the use of CKF can effectively suppress the influence of interference data. As shown in Figure 11a–c, CKF can efficiently filter the interference point and properly estimate the position of the AUV when the position is disrupted.
The error histogram of the measured and CKF positions in severe sea conditions is shown in Figure 12. Tough sea conditions have a great influence on the measurement position. After using CKF, the interference position error considerably improved on the X, Y, and Z axes. The CKF’s optimization effect is consistent with the prior in a normal sea state. Figure 11 and Figure 12 can be compared to demonstrate how effectively CKF inhibits interference positions with large mistakes.
Figure 13 shows a more intuitive three-dimensional position trajectory. Figure 13 shows that there is a non-negligible deviation between the measured and true trajectory. Figure 13a is the trajectory comparison under normal sea conditions, and the error of the measured trajectory is improved after using CKF. Figure 13b shows a trajectory comparison under severe sea conditions, in which the interference points of the measured trajectory are effectively suppressed after using the CKF. The CKF trajectory basically coincides with the true trajectory, showing that our position estimation has high accuracy and strong robustness.
We collected images of the AUV at multiple pitch and yaw angles. The AUV’s partial orientation measurement results are shown in Table 2 and Table 3, showing that our method can accurately estimate the AUV’s pitch and yaw angles.

3.4. LARS Tracking Trajectory

Because the position trajectory under severe and normal sea conditions is basically the same after using the CKF, we used the positioning trajectory under normal sea conditions as the LARS tracking target. We set the LARS motion to an initial state to be close to the AUV before the LARS starts tracking. Considering that the LARS motion rate and reaction time are lower than the frame rate of visual tracking, we sampled the trajectory for a short time. We performed fifth-order polynomial interpolation trajectory planning between the sampling points to obtain the desired LARS motion trajectory. We obtained the expected rotation angle of each joint by inverse kinematics. Then, we input the expected rotation angle into the control system as the control target; therefore, we obtained the true LARS motion trajectory. In the true control, we set the control system parameters K p = 160 and K d = 20 according to experience, under this parameter control effect is best.
The comparison of the positioning trajectories on the three axes during LARS tracking is shown in Figure 14. During the tracking process, we kept the LARS end approximately 0.8 m above the AUV to avoid collisions. Furthermore, we controlled the LARS to approach the AUV on the X and Y axes to ensure that the AUV is within the capturing range of the end actuator. Then, we analyzed the AUV’s acceleration on the Z axis. When the AUV began to sink according to the positioning trajectory, we controlled the LARS to quickly approach the AUV on the Z-axis and complete the capture.
The relative distance between the LARS end and the AUV on the X, Y, and Z axes is shown in Figure 15. Figure 15a shows that the maximum control error on the X-axis is 45 mm. However, because the AUV advances at a constant speed on the X-axis, and because of the AUV’s cylindrical shape, the error on the X-axis has little effect on the capturing. The positioning errors on the X and Y axes also decreased during the capture process, and the maximum control error on the Y axis is less than 22 mm as shown in Figure 15b. In the tracking stage, because LARS is kept far from the AUV on the Z-axis, we primarily took its final capturing error into account. Figure 15c shows that the final control error on the Z-axis is about 28 mm. The inaccuracies in the three tracking-stage dimensions and the capture stage are within the real engineering requirements.
The results of the three-dimensional visualization process are shown in Figure 16. When the AUV drops on the Z-axis, the LARS is controlled to quickly approach the AUV and capture it through the recycling device on the LARS end. Because the AUV shows a descending trend, the LARS will not collide with the AUV during the catching process. The control inaccuracy is roughly 30 mm when the capture is finished. Our experiments show that the system in this paper can accurately recycle the AUV in various sea conditions.

4. Discussion

The simulation results show that the system can recycle AUVs in extreme sea conditions. Compared with the previous AUV recycling system, our proposed system is based on the LTAT method, CKF, and LARS has a better effect on position accuracy and anti-interference. In this section, we analyze the system’s tracking effect, position accuracy, and control method based on the experimental data. Moreover, we provide recommendations for improvement after performing an analysis of the influencing factors on the system’s performance.

4.1. Performance Analysis of LTAT Method

When there is similar target interference and visual loss, the LTAT method can still accurately track the AUV. The LTAT method can swiftly and autonomously identify the AUV’s pixel position without manually correcting the tracking error, especially when it reappears elsewhere in the image after missing its field of view due to wave interference; therefore we verified the feasibility of updating the database to track non-standard objects step by step. Furthermore, the LTAT method’s tracking rate can reach 48 frames per second at an image resolution of 1280 × 960, which is much higher than the 25 frames per second required for real-time tracking.
The deployment platform determines how frequently an online learning module is updated. On the platform in our experiment, the YOLO network’s weight updates takes roughly 20 s. The learning module is mainly designed for long-term tracking, and its effect cannot be fully reflected in instantaneous tracking.
The visual sensor’s imitations make it impossible to improve the interference position caused by the loss of the visual field from visual technology. Multi-sensor fusion can be utilized for position to make up for the limitations of visual sensors in order to address the position impact brought on by interference.

4.2. Analysis of Position Estimation Method

We used binocular images to measure the AUV’s position; then, we used a CKF to optimize the measurement position. We carried out an orientation estimation experiment of AUVs in multiple orientations. The position errors on the X and Z axes were smaller when assessing AUV position through binocular pictures; however, the position error on the Y-axis is bigger, as indicated in Figure 9. On the one hand, because the AUV is in a fixed position on the Y-axis when the pixel position changes slightly, it will have a greater impact on the position error of the Y-axis. On the other hand, the position error on the Y-axis is due to floating errors in the feature point coordinates. According to Equation (2), the camera coordinate system’s Y-axis position is related to the pixel coordinate system’s Y-axis position; therefore, the error on the Y-axis is significant.
It is critical to optimize the AUV’s position because its measurement position is strongly affected by interference. In our experiment, we used a CKF to optimize the position, providing a position with stronger anti-interference and higher precision. The CKF can effectively suppress the interference of position with large mistakes, as demonstrated in Figure 11. We optimized the average position error to be below 10 mm, and the maximum position error is optimized to be below 20 mm after using the CKF.
In summary, the factors affecting position accuracy include the extraction method of feature points, external interference factors, and image resolution. A more reliable feature point extraction technique can result in a more precise position. In order to increase position accuracy, the Kalman filter can also be used to combine data from the multi-sensor.

4.3. Constraint and Improvement of Control Method

Between the LARS end trajectory and the true AUV trajectory, there is a divergence. One reason is that there are unavoidable measurement inaccuracies in the measured position. On the other hand, as our work makes use of a true physical model, there will be a variety of disturbances in the LARS control. The control method in this paper is to reduce the influence of these disturbances on the end of the LARS. The control error of LARS is about 2% relative to the measurement position and about 5% relative to the true position. This error is acceptable in the capturing process because the capturing range of the capturing mechanism at the end of LARS is wide enough for AUV recovery. In our experiment, it is possible to reliably capture an AUV through LARS.
However, our system does not consider the influence of waves on LARS, although it can compensate for the influence of waves by a six-degree-of-freedom parallel manipulator platform to ensure that LARS is in a relatively stable state. LARS’s motion can also be managed using a more sophisticated control method to compensate for wave interference.

5. Conclusions

Optimizing the positioning and capturing of AUV recycling system are the main objectives of our research. We construct a complete AUV recycling system in our experiment. The system can adapt to the challenging marine environment and complete the positioning and capturing of the specified AUV via LARS under various disturbances.
Specifically, we used the LTAT method, CKF, and LARS to achieve the specified AUV positioning and capturing. At the same time, we calculated the AUV’s yaw and pitch angles by using position data. Our experiments demonstrate that our proposed system can achieve an AUV position error of less than 0.02 m and a control error of LARS of less than 0.03 m when the AUV fluctuates with waves at 0.2 m; furthermore, the initial relative distance between AUV and LARS does not exceed 1 m. This satisfies the performance criteria for precise and real-time AUV recovery.
However, our proposed system still has room for improvement. In our research, our proposed system is still in the laboratory research stage and cannot be verified by sea trials. Subsequently, we plan to verify the system in the sea and develop improvements during the experiment. By continuously improving the sensor fusion and LARS control methods, as well as the mechanical structure of the actuator, an autonomous and intelligent AUV unmanned recycling system can be realized in the future.

Author Contributions

Conceptualization, T.Z., W.S. and W.Y.; methodology, T.Z., W.S.; software, W.S. and W.Z.; validation, W.S.; formal analysis, T.Z., W.S. and W.Y.; investigation, T.Z., W.S., W.Y., W.Z. and Y.W.; resources, T.Z.; data curation, W.S.; writing—original draft preparation, T.Z., and W.S.; writing—review and editing, T.Z., W.S., W.Y., W.Z. and Y.W.; visualization, W.S.; supervision, W.Y.; project administration, T.Z. and W.Y.; funding acquisition, T.Z. and W.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the National Natural Science Foundation of China [Grant number 52171331]; the International Science and Technology Cooperation Project of Guangdong Province [Grant number 2022A0505050027]; Guangdong Basic and Applied Basic Research Foundation [Grant number 2021A1515012552].

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AUVAutonomous underwater vehicles
CKFCubature Kalman filter
EKFExtended Kalman filter
GPUGraphics processing unit
INSInertial navigation system
IMUInertial Measurement Unit
LARSLaunch and recovery system
LTATLong-term target anti-interference tracking
OPEOne-pass evaluation
PIDProportion integration differentiation
ROSRobot operating system
SiamRPNSiamese region proposal network
UKFUnscented Kalman filter
UWBUltra-Wide Band
VIOVisual-Inertial Odometry
YOLOYou Only Look Once

References

  1. Caruso, F.; Tedesco, P.; Della Sala, G.; Palma Esposito, F.; Signore, M.; Canese, S.; Romeo, T.; Borra, M.; Gili, C.; de Pascale, D. Science and Dissemination for the UN Ocean Decade Outcomes: Current Trends and Future Perspectives. Front. Mar. Sci 2022, 9, 863647. [Google Scholar] [CrossRef]
  2. Mabus, R. Autonomous Undersea Vehicle Requirement for 2025; Chief of Naval OPERATIONS Undersea Warfare Directorate: Washington, DC, USA, 2016. [Google Scholar]
  3. Zhang, F.; Marani, G.; Smith, R.N.; Choi, H.T. Future trends in marine robotics [tc spotlight]. IEEE Robot. Autom. Mag. 2015, 22, 14–122. [Google Scholar] [CrossRef]
  4. Zacchini, L.; Ridolfi, A.; Allotta, B. Receding-horizon sampling-based sensor-driven coverage planning strategy for AUV seabed inspections. In Proceedings of the 2020 IEEE/OES Autonomous Underwater Vehicles Symposium (AUV), St. Johns, NL, Canada, 30 September–2 October 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–6. [Google Scholar]
  5. Mattioli, M.; Bernini, T.; Massari, G.; Lardeux, M.; Gower, A.; De Baermaker, S. Unlocking Resident Underwater Inspection Drones or AUV for Subsea Autonomous Inspection: Value Creation between Technical Requirements and Technological Development. In Proceedings of the Offshore Technology Conference, Houston, TX, USA, 2–5 May 2022. [Google Scholar]
  6. Wang, C.; Mei, D.; Wang, Y.; Yu, X.; Sun, W.; Wang, D.; Chen, J. Task allocation for Multi-AUV system: A review. Ocean Eng. 2022, 266, 112911. [Google Scholar] [CrossRef]
  7. Liang, X.; Dong, Z.; Hou, Y.; Mu, X. Energy-saving optimization for spacing configurations of a pair of self-propelled AUV based on hull form uncertainty design. Ocean Eng. 2020, 218, 108235. [Google Scholar] [CrossRef]
  8. Zhang, B.; Xu, W.; Lu, C.; Lu, Y.; Wang, X. Review of low-loss wireless power transfer methods for autonomous underwater vehicles. IET Power Electron. 2022, 15, 775–788. [Google Scholar] [CrossRef]
  9. Heo, J.; Kim, J.; Kwon, Y. Technology development of unmanned underwater vehicles (UUVs). J. Comput. Commun. 2017, 5, 28. [Google Scholar] [CrossRef] [Green Version]
  10. Rong, Z.; Chuanlong, X.; Zhong, T.; Tao, S. Review on the Platform Technology of Autonomous Deployment of AUV by USV. Acta Armamentarii 2020, 41, 1675. [Google Scholar]
  11. Meng, L.; Lin, Y.; Gu, H.; Su, T.C. Study on the mechanics characteristics of an underwater towing system for recycling an Autonomous Underwater Vehicle (AUV). Appl. Ocean Res. 2018, 79, 123–133. [Google Scholar] [CrossRef]
  12. Sarda, E.I.; Dhanak, M.R. Launch and recovery of an autonomous underwater vehicle from a station-keeping unmanned surface vehicle. IEEE J. Ocean. Eng. 2018, 44, 290–299. [Google Scholar] [CrossRef]
  13. Sarda, E.I.; Dhanak, M.R. A USV-based automated launch and recovery system for AUVs. IEEE J. Ocean. Eng. 2016, 42, 37–55. [Google Scholar] [CrossRef]
  14. Haiting, Z.; Haitao, G.; Yang, L.; Guiqiang, B.; Lingshuai, M. Design and Hydrodynamic Analysis of Towing Device of the Automated Recovery of the AUV by the USV. In Proceedings of the 2018 IEEE International Conference on Information and Automation (ICIA), Wuyishan, China, 11–13 August 2018; pp. 416–421. [Google Scholar]
  15. Zhang, G.; Tang, G.; Huang, D.; Huang, Y. Research on AUV Recovery by Use of Manipulator Based on Vision Servo. In Proceedings of the The 32nd International Ocean and Polar Engineering Conference, Shanghai, China, 5–10 June 2022. [Google Scholar]
  16. Kebkal, K.; Mashoshin, A. AUV acoustic positioning methods. Gyroscopy Navig. 2017, 8, 80–89. [Google Scholar] [CrossRef]
  17. Maurelli, F.; Krupiński, S.; Xiang, X.; Petillot, Y. AUV localisation: A review of passive and active techniques. Int. J. Intell. Robot. Appl. 2022, 6, 246–269. [Google Scholar] [CrossRef]
  18. Wu, Y.; Lim, J.; Yang, M.H. Online Object Tracking: A Benchmark. In Proceedings of the Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), New York, NY, USA, 15–17 June 2013. [Google Scholar]
  19. Zhu, Z.; Wu, W.; Zou, W.; Yan, J. End-to-End Flow Correlation Tracking With Spatial-Temporal Attention. In Proceedings of the Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), New York, NY, USA, 15–17 June 2018. [Google Scholar]
  20. Wang, T.; Zhao, Q.; Yang, C. Visual navigation and docking for a planar type AUV docking and charging system. Ocean Eng. 2021, 224, 108744. [Google Scholar] [CrossRef]
  21. Wu, D.; Zhu, H.; Lan, Y. A Method for Designated Target Anti-Interference Tracking Combining YOLOv5 and SiamRPN for UAV Tracking and Landing Control. Remote Sens. 2022, 14, 2825. [Google Scholar] [CrossRef]
  22. Pham, M.T.; Courtrai, L.; Friguet, C.; Lefèvre, S.; Baussard, A. YOLO-Fine: One-stage detector of small objects under various backgrounds in remote sensing images. Remote Sens. 2020, 12, 2501. [Google Scholar] [CrossRef]
  23. Liu, S.; Xu, H.; Lin, Y.; Gao, L. Visual Navigation for Recovering an AUV by Another AUV in Shallow Water. Sensors 2019, 19, 1889. [Google Scholar] [CrossRef] [Green Version]
  24. Meng, L.; Lin, Y.; Gu, H.; Su, T.C. Study on dynamic docking process and collision problems of captured-rod docking method. Ocean Eng. 2019, 193, 106624. [Google Scholar] [CrossRef]
  25. Szczotka, M. AUV launch & recovery handling simulation on a rough sea. Ocean Eng. 2022, 246, 110509. [Google Scholar]
  26. Li, B.; Yan, J.; Wu, W.; Zhu, Z.; Hu, X. High performance visual tracking with siamese region proposal network. In Proceedings of the IEEE conference on computer vision and pattern recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8971–8980. [Google Scholar]
  27. Chen, X.; He, K. Exploring simple siamese representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 15750–15758. [Google Scholar]
  28. Zuo, M.; Zhu, X.; Chen, Y.; Yu, J. Survey of Object Tracking Algorithm Based on Siamese Network. J. Phys. 2022, 2203, 12035. [Google Scholar] [CrossRef]
  29. Benjumea, A.; Teeti, I.; Cuzzolin, F.; Bradley, A. YOLO-Z: Improving small object detection in YOLOv5 for autonomous vehicles. arXiv 2021, arXiv:2112.11798. [Google Scholar]
  30. Jiang, P.; Ergu, D.; Liu, F.; Cai, Y.; Ma, B. A Review of Yolo algorithm developments. Procedia Comput. Sci. 2022, 199, 1066–1073. [Google Scholar] [CrossRef]
  31. Zhou, Y.; Xu, C.; Dai, Y.; Feng, X.; Ma, Y.; Li, Q. Dual-View Stereovision-Guided Automatic Inspection System for Overhead Transmission Line Corridor. Remote Sens. 2022, 14, 4095. [Google Scholar]
  32. Auger, F.; Hilairet, M.; Guerrero, J.M.; Monmasson, E.; Orlowska-Kowalska, T.; Katsura, S. Industrial applications of the Kalman filter: A review. IEEE Trans. Ind. Electron. 2013, 60, 5458–5471. [Google Scholar] [CrossRef] [Green Version]
  33. Kim, T.; Park, T.H. Extended Kalman filter (EKF) design for vehicle position tracking using reliability function of radar and lidar. Sensors 2020, 20, 4126. [Google Scholar] [CrossRef]
  34. Jin, Z.; Zhao, J.; Chakrabarti, S.; Ding, L.; Terzija, V. A hybrid robust forecasting-aided state estimator considering bimodal Gaussian mixture measurement errors. Int. J. Electr. Power Energy Syst. 2020, 120, 105962. [Google Scholar] [CrossRef]
  35. Mazaheri, A.; Radan, A. Performance evaluation of nonlinear Kalman filtering techniques in low speed brushless DC motors driven sensor-less positioning systems. Control Eng. Pract. 2017, 60, 148–156. [Google Scholar] [CrossRef]
  36. Kulikov, G.Y.; Kulikova, M.V. Do the cubature and unscented Kalman filtering methods outperform always the extended Kalman filter? IFAC-PapersOnLine 2017, 50, 3762–3767. [Google Scholar] [CrossRef]
  37. Wan, M.; Li, P.; Li, T. Tracking maneuvering target with angle-only measurements using IMM algorithm based on CKF. In Proceedings of the 2010 International Conference on Communications and Mobile Computing, Shenzhen, China, 12–14 April 2010; IEEE: Piscataway, NJ, USA, 2010; Volume 3, pp. 92–96. [Google Scholar]
  38. Arasaratnam, I.; Haykin, S. Cubature kalman filters. IEEE Trans. Autom. Control 2009, 54, 1254–1269. [Google Scholar] [CrossRef] [Green Version]
  39. Shen, C.; Zhang, Y.; Tang, J.; Cao, H.; Liu, J. Dual-optimization for a MEMS-INS/GPS system during GPS outages based on the cubature Kalman filter and neural networks. Mech. Syst. Signal Process. 2019, 133, 106222. [Google Scholar] [CrossRef]
  40. Naing, W.W.; Thanlyin, M.; Aung, K.Z.; Thike, A. Position Control of 3-DOF Articulated Robot Arm using PID Controller. Int. J. Sci. Eng. Appl. 2018, 7, 254–259. [Google Scholar]
  41. Bi, M. Control of Robot Arm Motion Using Trapezoid Fuzzy Two-Degree-of-Freedom PID Algorithm. Symmetry 2020, 12, 665. [Google Scholar] [CrossRef] [Green Version]
  42. Xie, N.; Iglesias, G.; Hann, M.; Pemberton, R.; Greaves, D. Experimental study of wave loads on a small vehicle in close proximity to a large vessel. Appl. Ocean Res. 2019, 83, 77–87. [Google Scholar] [CrossRef]
  43. Kashiwagi, M.; Endo, K.; Yamaguchi, H. Wave drift forces and moments on two ships arranged side by side in waves. Ocean Eng. 2005, 32, 529–555. [Google Scholar] [CrossRef]
Figure 1. LARS tracking and capturing flow chart.
Figure 1. LARS tracking and capturing flow chart.
Remotesensing 15 00748 g001
Figure 2. SiamRPN main framework: left side is the Siamese network; right side is the region proposal subnetwork; ⋆ denotes correlation operator.
Figure 2. SiamRPN main framework: left side is the Siamese network; right side is the region proposal subnetwork; ⋆ denotes correlation operator.
Remotesensing 15 00748 g002
Figure 3. LTAT method structure.
Figure 3. LTAT method structure.
Remotesensing 15 00748 g003
Figure 4. Schematic diagram of system coordinate transformation.
Figure 4. Schematic diagram of system coordinate transformation.
Remotesensing 15 00748 g004
Figure 5. Appearance design of AUV.
Figure 5. Appearance design of AUV.
Remotesensing 15 00748 g005
Figure 6. Performance comparison of tracking algorithms: (a) OPE success rate; (b) OPE accuracy.
Figure 6. Performance comparison of tracking algorithms: (a) OPE success rate; (b) OPE accuracy.
Remotesensing 15 00748 g006
Figure 7. Tracking results under interference from similar targets: (a) Tracking start frame; (b) Tracking end frame.
Figure 7. Tracking results under interference from similar targets: (a) Tracking start frame; (b) Tracking end frame.
Remotesensing 15 00748 g007
Figure 8. Tracking results under interference from waves: (a) Start affected frame; (b) Tracking lost frame; (c) Tracking end frame.
Figure 8. Tracking results under interference from waves: (a) Start affected frame; (b) Tracking lost frame; (c) Tracking end frame.
Remotesensing 15 00748 g008
Figure 9. Measured, true, and CKF trajectories of AUV motion in normal sea state. (a) X-axis, (b) Y-axis, and (c) Z-axis trajectory comparisons.
Figure 9. Measured, true, and CKF trajectories of AUV motion in normal sea state. (a) X-axis, (b) Y-axis, and (c) Z-axis trajectory comparisons.
Remotesensing 15 00748 g009
Figure 10. Error between true and estimated position in normal sea state: (a) X-axis, (b) Y-axis, and (c) Z-axis error comparisons.
Figure 10. Error between true and estimated position in normal sea state: (a) X-axis, (b) Y-axis, and (c) Z-axis error comparisons.
Remotesensing 15 00748 g010
Figure 11. Measured, true, and CKF trajectories of AUV motion in severe sea state. (a) X-axis, (b) Y-axis, and (a) Z-axis trajectory comparisons.
Figure 11. Measured, true, and CKF trajectories of AUV motion in severe sea state. (a) X-axis, (b) Y-axis, and (a) Z-axis trajectory comparisons.
Remotesensing 15 00748 g011
Figure 12. Error between true and estimated position in severe sea state: (a) X-axis, (b) Y-axis, and (c) Z-axis error comparisons.
Figure 12. Error between true and estimated position in severe sea state: (a) X-axis, (b) Y-axis, and (c) Z-axis error comparisons.
Remotesensing 15 00748 g012
Figure 13. 3D trajectories of AUV motion in different sea state: (a) 3D trajectory under normal sea state; (b) 3D trajectory under severe sea state.
Figure 13. 3D trajectories of AUV motion in different sea state: (a) 3D trajectory under normal sea state; (b) 3D trajectory under severe sea state.
Remotesensing 15 00748 g013
Figure 14. Trajectory comparison of AUV and LARS end position in normal sea state: (a) X-axis, (b) Y-axis, and (c) Z-axis trajectory comparisons.
Figure 14. Trajectory comparison of AUV and LARS end position in normal sea state: (a) X-axis, (b) Y-axis, and (c) Z-axis trajectory comparisons.
Remotesensing 15 00748 g014
Figure 15. Relative position between AUV and LARS end in normal sea state: (a) X-axis, (b) Y-axis, and (c) Z-axis relative position.
Figure 15. Relative position between AUV and LARS end in normal sea state: (a) X-axis, (b) Y-axis, and (c) Z-axis relative position.
Remotesensing 15 00748 g015
Figure 16. Three−dimensional tracking trajectory of LARS.
Figure 16. Three−dimensional tracking trajectory of LARS.
Remotesensing 15 00748 g016
Table 1. Physical parameters of AUV.
Table 1. Physical parameters of AUV.
WeightDiameterLength
47 kg200 mm1.8 m
Table 2. Estimates and true values of AUV pitch angles.
Table 2. Estimates and true values of AUV pitch angles.
Set the Angle152030
Pitch angle14.789120.818833.4522
Table 3. Estimates and true values of AUV yaw angles.
Table 3. Estimates and true values of AUV yaw angles.
Set the Angle−30030
Yaw angle−29.5647−0.0124−29.6427
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zou, T.; Situ, W.; Yang, W.; Zeng, W.; Wang, Y. A Method for Long-Term Target Anti-Interference Tracking Combining Deep Learning and CKF for LARS Tracking and Capturing. Remote Sens. 2023, 15, 748. https://doi.org/10.3390/rs15030748

AMA Style

Zou T, Situ W, Yang W, Zeng W, Wang Y. A Method for Long-Term Target Anti-Interference Tracking Combining Deep Learning and CKF for LARS Tracking and Capturing. Remote Sensing. 2023; 15(3):748. https://doi.org/10.3390/rs15030748

Chicago/Turabian Style

Zou, Tao, Weilun Situ, Wenlin Yang, Weixiang Zeng, and Yunting Wang. 2023. "A Method for Long-Term Target Anti-Interference Tracking Combining Deep Learning and CKF for LARS Tracking and Capturing" Remote Sensing 15, no. 3: 748. https://doi.org/10.3390/rs15030748

APA Style

Zou, T., Situ, W., Yang, W., Zeng, W., & Wang, Y. (2023). A Method for Long-Term Target Anti-Interference Tracking Combining Deep Learning and CKF for LARS Tracking and Capturing. Remote Sensing, 15(3), 748. https://doi.org/10.3390/rs15030748

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop