A Novel Underwater Simultaneous Localization and Mapping Online Algorithm Based on Neural Network

The navigation and localization of autonomous underwater vehicles (AUVs) in seawater are of the utmost importance for scientific research, petroleum engineering, search and rescue, and military missions concerning the special environment of seawater. However, there is still no general method for AUVs navigation and localization, especially in the featureless seabed. The reported approaches to solving AUVs navigation and localization problems employ an expensive inertial navigation system (INS), with cumulative errors and dead reckoning, and a high-cost long baseline (LBL) in a featureless subsea. In this study, a simultaneous localization and mapping (AMB-SLAM) online algorithm, based on acoustic and magnetic beacons, was proposed. The AMB-SLAM online algorithm is based on multiple randomly distributed beacons of low-frequency magnetic fields and a single fixed acoustic beacon for location and mapping. The experimental results show that the performance of the AMB-SLAM online algorithm has a high robustness. The proposed approach (the AMB-SLAM online algorithm) provides a low-complexity, low-cost, and high-precision online solution to the AUVs navigation and localization problem in featureless seawater environments. The AMB-SLAM online solution could enable AUVs to autonomously explore or autonomously intervene in featureless seawater environments, which would enable AUVs to accomplish fully autonomous survey missions.


Introduction
Autonomous underwater vehicles (AUVs) play a crucial role in many subsea missions and are widely used in scientific research, search and rescue, the military, petroleum engineering, and civilian applications [1,2]. To execute these missions, precise navigation and localization are the most important technologies for pursuing a preprogrammed trajectory and acquiring accurate geographic data. Generally speaking, above water, most unmanned aerial vehicles (UAVs) and unmanned surface vessels (USVs) are remotely operated by radio and navigated by an inertial navigation system (INS) or global positioning system (GPS) [3]. Nevertheless, under the sea, GPS and radio signals are unavailable, because electromagnetic radiation cannot adequately penetrate the unstructured nature of the undersea environment. Furthermore, underwater communication is highly restricted by the environmental factors of seawater, such as the temperature, pressure, or salinity of the seawater. While there are ISPRS Int. J. Geo-Inf. 2020, 9, 5 4 of 28 The remainder of the paper is organized as follows. Section 2 introduces the theoretical framework concerning the positioning principle of the magnetic beacon, artificial neural network (ANN), and the AMB-SLAM online algorithm. In Section 3, the simulation experiment is described in detail, and the results of applying the AMB-SLAM algorithm and a performance comparison with the MB-SLAM algorithm are presented. Finally, in Section 4, we discuss the AMB-SLAM method and summarize our major findings, provide a conclusion of the whole work, and give suggestions for future research.

Description of the Positioning Principle of the Beacon of the Low-Frequency Magnetic Field
As mentioned in the previous section, in the field of modern underwater navigation, no positioning method is a perfect solution on account of the particularity of the seawater environment. In simultaneous localization and mapping (SLAM) research, no one utilized magnetic field information as geophysical information for underwater localization. Davis [70] once measured the magnetic field intensity of the target point and then calculated the relative position between the target point and the magnetic beacon by taking advantage of the GPS positioning principle and the penetrability of the magnetic near field. However, in his research, only a theoretical formula is given, which is not feasible in practical experiments.
The method of magnetic field localization based on the fundamental principle of the Biot-Savart law is to use the low-frequency alternating magnetic field as the signal source. The Biot-Savart law explains the magnetic field produced by a current carrying segment. As shown in the Figure 1, a wire carries a specific direction current I, taking a small element of the wire of length d. The direction of this element is along that of the current so that it forms a vector Id → l . The magnitude of the magnetic field d → B at a distance r from a current carrying element dl is found to be proportional to I and to the length dl and is inversely proportional to the cubic of the distance r. The Biot-Savart law is formulated as follows: Id → l × → r r 3 (1) where d → B is the magnetic field at field point P, which is produced due to the current element Id → l ; d → l is the infinitesimal length of the conductor carrying the electric current I; → r is the distance vector that specifies the direction of the distance r from the current element to the field point; L is the integration path; and µ 0 is the permeability of the vacuum, whose value is equal to 4π × 10 −7 Tm/A. r is defined as follows: The Biot-Savart law is also true for superimposed magnetic fields. According to Formulas (1) and (2), if we know the magnitude of the magnetic field B and the current element Idl  , we can get the value of r , but we cannot measure the magnitude of the current element Idl  , so it is impossible to solve r by the Biot-Savart law in the experiment. Therefore, in the present study, we use the artificial neural network (ANN) to solve r .

Description of the Artificial Neural Network
The artificial neural network (ANN) can solve the optimization problem of the super nonlinear equation. Therefore, we use ANN to solve the nonlinear problem of magnetic field positioning. It simulates the human brain and nervous system from the perspective of information processing, establishes a mathematical model, and consists of a large number of complex interconnected neurons. Each neuron represents a specific output function, which is called an activation function. Each connection between two neurons represents a weighted value, which is equivalent to the memory of the artificial neural network. Learning in the human brain requires adjustments to the synaptic relationship between and among neurons, which is similar to the learning in ANN [71]. In order to circumvent the complicated nonlinear problem, multi-layer perceptrons (MLP), which are feedforward neural networks, were proposed [71]. Thus, in this study, three hidden layers of MLP, in combination with the supervised learning method, called the error back propagation (BP) algorithm, were applied. The architecture of the selected MLP ANN is shown in Figure 2. The Biot-Savart law is also true for superimposed magnetic fields. According to Formulas (1) and (2), if we know the magnitude of the magnetic field B and the current element Id → l , we can get the value of r, but we cannot measure the magnitude of the current element Id → l , so it is impossible to solve r by the Biot-Savart law in the experiment. Therefore, in the present study, we use the artificial neural network (ANN) to solve r.

Description of the Artificial Neural Network
The artificial neural network (ANN) can solve the optimization problem of the super nonlinear equation. Therefore, we use ANN to solve the nonlinear problem of magnetic field positioning. It simulates the human brain and nervous system from the perspective of information processing, establishes a mathematical model, and consists of a large number of complex interconnected neurons. Each neuron represents a specific output function, which is called an activation function. Each connection between two neurons represents a weighted value, which is equivalent to the memory of the artificial neural network. Learning in the human brain requires adjustments to the synaptic relationship between and among neurons, which is similar to the learning in ANN [71]. In order to circumvent the complicated nonlinear problem, multi-layer perceptrons (MLP), which are feed-forward neural networks, were proposed [71]. Thus, in this study, three hidden layers of MLP, in combination with the supervised learning method, called the error back propagation (BP) algorithm, were applied. The architecture of the selected MLP ANN is shown in Figure 2. In the present study, it is assumed that the number of input layer neurons is n , the number of hidden layer neurons is l , and the number of output layer neurons is m . The forward propagation of MLP ANN is mainly composed of three parts. Firstly, MLP computes a linear weighted sum of its n inputs mathematically: where j x is the linear weighted sum of all inputs, ij w is the synapse weight between the i th neuron and jth neuron, i x is the input vector, j a is the bias value from the input layer to the hidden layer, and the subscript n denotes the number of input layer neurons. Then, the output of the hidden layer neurons can be represented as: where j H is the output of hidden layer neurons, and ( ) f  is the activation function (see Formula (6)). In the present study, the chosen activation function is a hyperbolic tangent activation function: Finally, the output of the output layer neuron of MLP-ANN can be represented as: In the present study, it is assumed that the number of input layer neurons is n, the number of hidden layer neurons is l, and the number of output layer neurons is m. The forward propagation of MLP ANN is mainly composed of three parts. Firstly, MLP computes a linear weighted sum of its n inputs mathematically: where x j is the linear weighted sum of all inputs, w ij is the synapse weight between the ith neuron and jth neuron, x i is the input vector, a j is the bias value from the input layer to the hidden layer, and the subscript n denotes the number of input layer neurons. Then, the output of the hidden layer neurons can be represented as: where H j is the output of hidden layer neurons, and f (·) is the activation function (see Formula (6)).
In the present study, the chosen activation function is a hyperbolic tangent activation function: Finally, the output of the output layer neuron of MLP-ANN can be represented as: where w jk is the connection weight between the jth neuron and the kth neuron, b j is the bias value from the hidden layer to output layer, and the subscript l denotes the number of hidden layer neurons.

Optimization of MLP-ANN
In the present study, MLP-ANN is optimized by the unsupervised optimization method (or learning method), which is called the error backward propagation (BP) algorithm. In the backward propagation of MLP-ANN, three processes are required: data preprocessing, training data, and validating data.

Data Preprocessing
In the present study, a feed-forward neural network (see Figure 2), with the supervised learning method, called the error back-propagation (BP) algorithm, was used for learning. In our work, we chose the hyperbolic tangent activation function as the activation function.
The magnetic field data and relative position between the magnetic beacons and AUV of the magnetic beacon, which were collected in the field, were taken as the sample data of the neural network. When limited sample data are available, it might be difficult to maximize the utilization of the representative input data set. Meanwhile, it is also essential that the training data and validation data are representative of the same population. Therefore, we adopted the holdout method [72,73], splitting the input data set into two subsets: 10% of the data for validation and the remaining data for training the network. In addition, different variables span different ranges, and not all sample data could receive the same attention during the training steps (see below). Thus, in this experiment, it is essential to scale all the sample data within a specified range using Equation (8). In addition, the scaling of the input sets must be compatible with the boundary of the activation function used in the output layer or hidden layer [74]. The standardization formula is defined as follows: where x min and x max are the minimum and maximum value of the input sets, respectively, f min and f max are the minimum and maximum value of the activation function, respectively, and the subscript i denotes the input data number. The value of the hyperbolic tangent transfer function is between −1 and 1, therefore, the sample data are scaled within the range of −0.9 to 0.9. Thus, Equation (8) is given as:

Training
After the preparation of the sample data set, the magnitude of the magnetic field B was set as the input and the relative position of the magnetic beacon and AUV r were set as the outputs.
The training process of ANN involves optimizing the neural weights by adjusting the internal parameters of MLP-ANN (such as the number of hidden layers, their neurons, and so on, as shown in Figure 3) by trial and error, which is also called 'learning'. The purpose of learning is to find an optimal global solution for the highly nonlinear objective equation [75] and minimize the value of the error function. In the present study, the mean squared error (MSE) over the training samples was the error function of the ANN performance evaluation to be optimized, and the minimization of the error function is performed by iterative techniques. The performance function is as follows: where Y k represents the actual output, O k represents the actual outputs, m denotes the number of output nodes, and e k is the mth output residual. In order to attain the optimum architecture, different steps have been followed, according to the work flow diagram shown in Figure 3. The weight was optimized by the gradient-descent method. At the beginning of the training, the weights were initialized to zero-mean random values. This weight update equation is drawn from [76]: where w j+1 denotes the updated weights value, w j denotes the initial weights value η j is the learning rate (η j is proportional to the size of the steps), d j is the direction of the descent, and the subscript n denotes the update time. The direction of descent d j is the negative of the error gradient. The final weight update formula is where ∇ wj E is the error gradient.

Description of the Proposed AMB-SLAM Algorithm
In this section, we will explain how the AMB-SLAM algorithm can map and locate AUVs in underwater environments without features. Figure 4 provides a basic schematic diagram of the AMB-SLAM algorithm. Before the AUV performs its mission, we randomly deploy a sufficient number of beacons of a low-frequency magnetic field to the area of interest. When the AUV observes the  The architecture of MLP-ANN determines the number of connection weights and the number of hidden layers and neurons in each hidden layer. These hyper-parameters are normally determined through a trial and error method by comparison with the performance of different types of MLP-ANN architecture, as shown in the training flow in Figure 3. In addition, in order to prevent overfitting, the number of training samples is at least 30 times the number of connection weights [77] in this study. Similarly, the effect of the number of epochs on the performance of ANN was equally important. In this study, the number of epochs was set as the number of training samples [78]. Finally, the best MLP-ANN network is selected based on statistical criteria (MSE).

Description of the Proposed AMB-SLAM Algorithm
In this section, we will explain how the AMB-SLAM algorithm can map and locate AUVs in underwater environments without features. Figure 4 provides a basic schematic diagram of the AMB-SLAM algorithm. Before the AUV performs its mission, we randomly deploy a sufficient number of beacons of a low-frequency magnetic field to the area of interest. When the AUV observes the magnetic flux density of the No. 1, No. 2, and No. 3 magnetic beacon, it takes these values as the input values of the optimal MLP-ANN model to get the relative position ∆x, ∆y, ∆z (see Figure 5) between these magnetic beacons and the AUV. The angle between the beacon and AUV are also available. In addition, the fixed acoustic beacon can constrain the AUV position in real time. If the absolute position of the acoustic beacon is known in the global coordinate system, we can obtain the position of the AUV through the coordinate conversion. The detailed flow chart of the AMB-SLAM algorithm, based on the extended Kalman filter (EKF), is shown in Figure 5.  When the AUV is in an unknown underwater environment, it conducts incremental environment modeling using the available geophysical information, which is gathered by an acoustic or optical sensor, and it uses the environment map to estimate its position. In this study, we assume that the initial position of the vehicle is (0 m, 0 m, 0 m), the relative position between the magnetic beacon and AUV can be estimated using the information from the observed magnetic beacons, and the position of the AUV can be updated by reobserving the beacons. The overall procedures of the AMB-SLAM, based the EKF filter navigation algorithm, are illustrated in Figure 5. When the AUV is in an unknown underwater environment, it conducts incremental environment modeling using the available geophysical information, which is gathered by an acoustic or optical sensor, and it uses the environment map to estimate its position. In this study, we assume that the initial position of the vehicle is (0 m, 0 m, 0 m), the relative position between the magnetic beacon and AUV can be estimated using the information from the observed magnetic beacons, and the position of the AUV can be updated by reobserving the beacons. The overall procedures of the AMB-SLAM, based the EKF filter navigation algorithm, are illustrated in Figure 5. environment modeling using the available geophysical information, which is gathered by an acoustic or optical sensor, and it uses the environment map to estimate its position. In this study, we assume that the initial position of the vehicle is (0 m, 0 m, 0 m), the relative position between the magnetic beacon and AUV can be estimated using the information from the observed magnetic beacons, and the position of the AUV can be updated by reobserving the beacons. The overall procedures of the AMB-SLAM, based the EKF filter navigation algorithm, are illustrated in Figure 5. Now, supposing that these magnetic beacons are static landmarks with zero-input disturbance, the model of the i th beacon can be simplified as follows: where n i b (k) is the location of the i th beacon at time k . Now, supposing that these magnetic beacons are static landmarks with zero-input disturbance, the model of the ith beacon can be simplified as follows: where b n i (k) is the location of the ith beacon at time k. As soon as the online feature extraction and the data association are in place, the AMB-SLAM algorithm includes the following three steps. Firstly, the AUV current state is predicted using the output values of MLP ANN. Secondly, the estimated state is updated from the reobserved landmark positions. Finally, the landmark map is updated.

State Equation
When the kinematics model of the AUV is known, the relative position of the magnetic beacon from the vehicle can be obtained through the optimal MLP-ANN model, and the augmented mathematical model of the state estimation can be established. The state is estimated using the EKF filter, and the AUV state is described by its local coordinates and heading angle as where X v (k) is the motion state of the AUV at time k, p v (k) = x v (k) y v (k) represents the position of the AUV in the local coordinate system, and φ v (k) represents the AUV's heading angle at time k.
The coordinate of the nth feature is denoted as x m (k) = (x b (k), y b (k)) T , and the environmental landmarks can be denoted as and its covariance matrix P m is The non-diagonal elements of the covariance matrix P m are the cross-correlation information between different landmarks. We assume that the landmarks are stationary, and that their positions do not change over time. These correlations will be enhanced with each reobservation. The landmark map is represented by an augmented state vector: If the change value of the AUV state is represented by the course change at the adjacent time, the distance change in the course direction and the vertical course direction can be written as ∆X = ∆x ∆y ∆φ T , and the position of the AUV can be described as: According to the above state equation, the Jacobian matrix is represented as follows:

Observation Equation
In the present study, the observation value is the relative position and the angle between the magnetic beacon and AUV. If an existing characteristic of the observed quantity and a state variable is determined, x b i , y b i is the position of the magnetic beacon in the map. Then, according to the estimated position of the AUV and the beacon b i , we can get the estimated observation equation: In addition, we deploy an acoustic beacon on the seafloor. Thus, the AUV observes the distance and angle between the AUV and the fixed acoustic beacon, which is a constraint when the AUV searches for the features, especially when there is no feature in the underwater, and it is the only useful information. To some extent, the positioning accuracy can be improved in this way.
The Jacobian matrix was obtained for the above observation equation: where r represents the distance between the beacon and AUV, and Therefore, the problem of the position of the random magnetic beacon can be transformed into the problem of the standard nonlinear filtering problem. Since the Jacobian matrix of the state equation and the measurement equation have been obtained, the extended Kalman filter can be directly performed.
We define R to be the measurement covariance, and innovation v represents the difference between the true value and the predicted value: The covariance matrix of the increment is The Kalman gain W is expressed as The updated location and covariance can be described aŝ Through the above formula, the neural network library and observation can be used to complete the synchronous update process of the vehicle and characteristic state. If no new features are observed, it can complete the iterative recursive process of prediction and updating in the next moment.

Updating
In the movement process of the AUV, it continuously explores new environments and discovers new beacons, so after completing the existing beacon position update, it needs to enlarge the dimensional state variables, add new beacons to the state vector, establish a craft condition, and map the relationship between the existing and new characteristics in order to construct an accurate map by merging these features.
Assuming that the sure measure z = r φ T original from a new feature, (x, y) represents the position of the AUV, and r, φ represent the relative position and angle between the AUV and the new beacon, so the new beacon can be described as The new state is By adding new beacons, the algorithm realizes a step-by-step automatic map extension.

Data Sets
In order to train the MLP-ANN model for the best performance, one set of sample data was acquired in the waters of Weihai city, China (as shown in the Figure 6). The size of the field acquisition area is a square, with a side length of 500 m, and its water depth ranges from 14.6 m to 15.4 m. In our experiment, the beacons of a low-frequency magnetic field are composed of two orthodox solenoids (see Figure 7a), and the model of the triaxial magnetic component magnetometer model is MS3A-02 (see Figure 7b), with a measurement range of −100 µT ∼ 100 µT and measurement accuracy of 0.2 nT. In this experiment, the AUV maintains a speed of 5 m per second and is about 6 m above the seafloor.

Optimal MLP-ANN Configuration
As shown in the above section, the number of hidden layers and the number of neurons in each hidden layer determine the complexity of the geometric structure of the MLP-ANN. While a simpler architecture of the neural network structure has a better generalization, it allows the local minima to be produced more easily. Similarly, a more complex architecture increases the ability of the neural network to avoid local minima, but it generally requires more training samples to realize a good generalization ability [79]. However, there is no empirical formula to establish the optimal architecture of MLP-ANN. Therefore, a variety of MLP-ANN parameters were evaluated by a trial and error procedure in order to optimize the network architecture.
In the present study, we set the hyperbolic tangent activation function as the activation function; the learning algorithm is BP; the number of epochs is equal to 1000; the goal accuracy is equal to

Optimal MLP-ANN Configuration
As shown in the above section, the number of hidden layers and the number of neurons in each hidden layer determine the complexity of the geometric structure of the MLP-ANN. While a simpler architecture of the neural network structure has a better generalization, it allows the local minima to be produced more easily. Similarly, a more complex architecture increases the ability of the neural network to avoid local minima, but it generally requires more training samples to realize a good generalization ability [79]. However, there is no empirical formula to establish the optimal architecture of MLP-ANN. Therefore, a variety of MLP-ANN parameters were evaluated by a trial and error procedure in order to optimize the network architecture.
In the present study, we set the hyperbolic tangent activation function as the activation function; the learning algorithm is BP; the number of epochs is equal to 1000; the goal accuracy is equal to

Optimal MLP-ANN Configuration
As shown in the above section, the number of hidden layers and the number of neurons in each hidden layer determine the complexity of the geometric structure of the MLP-ANN. While a simpler architecture of the neural network structure has a better generalization, it allows the local minima to be produced more easily. Similarly, a more complex architecture increases the ability of the neural network to avoid local minima, but it generally requires more training samples to realize a good generalization ability [79]. However, there is no empirical formula to establish the optimal architecture of MLP-ANN. Therefore, a variety of MLP-ANN parameters were evaluated by a trial and error procedure in order to optimize the network architecture.
In the present study, we set the hyperbolic tangent activation function as the activation function; the learning algorithm is BP; the number of epochs is equal to 1000; the goal accuracy is equal to 8 × 10 −4 ; and the learning rate is equal to 0.01, before optimizing the neural network. The performance of MLP-ANN was compared by changing the number of hidden layers and the number of neurons in each hidden layer. The best network architecture is selected based on several statistical criteria, including MSE and the coefficient of determination (R 2 ).
where n is the number of sample data; and x, y i and z i are average values of the sample data. The optimal architecture of the developed ANN is shown in Figure 2. Finally, the optimal MLP ANN architecture includes two hidden layers of 40 and 20 neurons. Figure 8 is a variation of the training error of MLP-ANN, whose architecture is optimal. It shows that the training error of MLP-ANN decreased as the number of epochs decreased. When the training procedure is at the 3950th epoch, a small tolerance is about 7.9 × 10 −4 , which means that the MSE on each training data achieves the training accuracy. Therefore, the neural network is well trained.
where n is the number of sample data; and x , i y and i z are average values of the sample data.
The optimal architecture of the developed ANN is shown in Figure 2. Finally, the optimal MLP ANN architecture includes two hidden layers of 40 and 20 neurons. Figure 8 is a variation of the training error of MLP-ANN, whose architecture is optimal. It shows that the training error of MLP-ANN decreased as the number of epochs decreased. When the training procedure is at the 3950th epoch, a small tolerance is about 4 7.9 10 -, which means that the MSE on each training data achieves the training accuracy. Therefore, the neural network is well trained.
As stated in the previous section, 10% of the sample data were used as a test set for the optimal MLP-ANN model by trial and error. Figure 9 shows the MLP-ANN predicted values of the relative position in comparison with the true position of the sample data set. As can be seen from Figure 9, the predicted data of MLP ANN are highly consistent with the real location data.  As stated in the previous section, 10% of the sample data were used as a test set for the optimal MLP-ANN model by trial and error. Figure 9 shows the MLP-ANN predicted values of the relative position in comparison with the true position of the sample data set. As can be seen from Figure 9, the predicted data of MLP ANN are highly consistent with the real location data. Finally, in order to verify the correctness and effectiveness of the neural network method proposed in our paper, we conducted an experiment, which was a 10-fold cross-validation. We divided all samples into 10 groups (every group is composed of 100 samples) in Table 1 and adopted the cyclic approach. We selected nine groups as the training set and one group as the test set in Table  2. Each experiment was repeated 10 times, so we carried out a total of 100 experiments. Table 3 is the experimental results of cross-validation.
The division of the groups was as follows:  Finally, in order to verify the correctness and effectiveness of the neural network method proposed in our paper, we conducted an experiment, which was a 10-fold cross-validation. We divided all samples into 10 groups (every group is composed of 100 samples) in Table 1 and adopted the cyclic approach. We selected nine groups as the training set and one group as the test set in Table 2. Each experiment was repeated 10 times, so we carried out a total of 100 experiments. Table 3 is the experimental results of cross-validation.
The division of the groups was as follows:  Therefore, the result of the experiment is the average result of 100 repetitions of the experiment, and the final test value of the correlation coefficient is 0.90362 in the direction of the X-axis, 0.87182 in the direction of the Y-axis, and 0.90904 in the direction of the Z-axis, which means that the ANN in our paper is effective.
As can be observed from Figure 10, all data points are scattered around the diagonal line. All the statistical analyses revealed that the proposed MLP-ANN model is a valid and accurate model for AUV position prediction. Finally, the statistical results are shown below: Therefore, the result of the experiment is the average result of 100 repetitions of the experiment, and the final test value of the correlation coefficient is 0.90362 in the direction of the X-axis, 0.87182 in the direction of the Y-axis, and 0.90904 in the direction of the Z-axis, which means that the ANN in our paper is effective.
As can be observed from Figure 10, all data points are scattered around the diagonal line. All the statistical analyses revealed that the proposed MLP-ANN model is a valid and accurate model for AUV position prediction.

AMB-SLAM Loop Map Simulation Setups
The following MATLAB simulation experiment is carried out for AMB-SLAM closed-loop mapping, which is performed in the context of a generic AUV observing the surrounding magnetic beacons with a range-bearing sensor in a two-dimensional area. The values of the different parameters could be changed according to the number of detected magnetic beacons and their positions in the environment, the actual speed of the robot, and the maximum observation distance of the selected sonar sensor.
The simulation environment for the AMB-SLAM closed-loop map is defined as an area of 500 m × 500 m (from 0 m to 500 m), containing 10 AUV waypoints connected in a serpentine trajectory and

AMB-SLAM Loop Map Simulation Setups
The following MATLAB simulation experiment is carried out for AMB-SLAM closed-loop mapping, which is performed in the context of a generic AUV observing the surrounding magnetic beacons with a range-bearing sensor in a two-dimensional area. The values of the different parameters could be changed according to the number of detected magnetic beacons and their positions in the environment, the actual speed of the robot, and the maximum observation distance of the selected sonar sensor.
The simulation environment for the AMB-SLAM closed-loop map is defined as an area of 500 m × 500 m (from 0 m to 500 m), containing 10 AUV waypoints connected in a serpentine trajectory and 20 magnetic beacons as features distributed near the robot trajectory, as shown in Figure 11. The performance of AMB-SLAM was evaluated using the field sample data. The AUV starts off at (0 m, 0 m, 0 m), and the other essential control parameters are listed in Table 4. All the simulation experiments were implemented in MATLAB R2017a.

AMB-SLAM Simulation Results and Discussion
In this study, we compared the performance of AMB-SLAM using multiple randomly distributed beacons of a low-frequency magnetic field with single fixed acoustic beacons. The estimated AUV trajectory and features achieved by AMB-SLAM are shown in Figure 11. Figure 11 demonstrates the 2D landmark map, which was obtained using the AMB-SLAM online algorithm. The positions of the sensor scans for the point landmarks are clearly visible, and few of these sightings are rejected based on statistical outlier rejection techniques [80], since they are out of the observation range of the AUV. The rest are believed to represent landmarks and are added to the 2D map. The true landmarks are shown as blue solid stars (' '), and the red crosses ('+') are the estimated landmark positions. The ellipses around each red cross ('+') illustrate the uncertainty covariances for the landmark coordinate estimations, and the larger they are, the more uncertain the estimated landmark horizontal positions. The estimated AUV trajectory is depicted as the solid red line, and the blue line shows the true AUV trajectory. The solid red triangle represents the acoustic beacon. A closed loop has a close relation with "positioning" and "mapping". In fact, the main meaning of the map is to inform the AUV about what is happening. At the same time, if the error is cumulative when the AUV reaches the same position, the improved algorithm loses its true meaning. As shown in Figure 11, we conducted closed-loop detection for the following scene, and the loop number here is two. In addition, we calculate the error of closed-loop detection and open-loop detection in the X and Y-axis. The error is shown in Figure 12. Figure 11 shows that the estimated trajectory path of AMB-SLAM is much closer to the real trajectory, and the estimated magnetic beacon position of AMB-SLAM is closer to the true feature point position. Thus, the AMB-SLAM method efficiently improves the accuracy of AUV localization and the feature map. We can see from Figure 12a   Next, under the same simulation conditions as those tested with the AMB-SLAM method, we study the performance of the MB-SLAM algorithm, which only utilizes magnetic beacons as landmarks. Figure 13 shows the 2D trajectory map of the MB-SLAM algorithm, and Figure 14 shows the error comparison diagram of the X-axis and Y-axis of the two algorithms. We can see from Figure  14a    Next, under the same simulation conditions as those tested with the AMB-SLAM method, we study the performance of the MB-SLAM algorithm, which only utilizes magnetic beacons as landmarks. Figure 13 shows the 2D trajectory map of the MB-SLAM algorithm, and Figure 14 shows the error comparison diagram of the X-axis and Y-axis of the two algorithms. We can see from Figure 14a  Next, under the same simulation conditions as those tested with the AMB-SLAM method, we study the performance of the MB-SLAM algorithm, which only utilizes magnetic beacons as landmarks. Figure 13 shows the 2D trajectory map of the MB-SLAM algorithm, and Figure 14 shows the error comparison diagram of the X-axis and Y-axis of the two algorithms. We can see from Figure  14a    Finally, in this study, we verified the influence of the distribution density of magnetic beacons on the performance of the AMB-SLAM and MB-SLAM algorithms. In this study, magnetic beacons are evenly distributed in the first half of the AUV path, while no magnetic beacons are distributed in the second half of the AUV planned path, which is the same as the control parameters in the previous simulation environment. The experimental results are shown in Figure 15 and Figure 16. As can be seen from these figures, the density of the magnetic beacons has a disproportional influence on both the AMB-SLAM and MB-SLAM algorithms, but the influence on the MB-SLAM algorithm is much greater than that on MB-SLAM. In addition, we can also see from the error diagram ( Figure 17) that the density of the magnetic beacons has an impact on both the AMB-SLAM and MB-SLAM algorithms, and the error in the X-axis direction of the MB-SLAM algorithm can reach −60 to 90 m. From Figure 17b, we can see that when the time is about 4200, the error in the Y-axis of the AMB-SLAM algorithm suddenly goes from 2 m to about −5 m. From this, we can draw a conclusion that the density of the magnetic beacon has an effect on both AMN-SLAM and MB-SLAM. Finally, in this study, we verified the influence of the distribution density of magnetic beacons on the performance of the AMB-SLAM and MB-SLAM algorithms. In this study, magnetic beacons are evenly distributed in the first half of the AUV path, while no magnetic beacons are distributed in the second half of the AUV planned path, which is the same as the control parameters in the previous simulation environment. The experimental results are shown in Figures 15 and 16. As can be seen from these figures, the density of the magnetic beacons has a disproportional influence on both the AMB-SLAM and MB-SLAM algorithms, but the influence on the MB-SLAM algorithm is much greater than that on MB-SLAM. In addition, we can also see from the error diagram ( Figure 17) that the density of the magnetic beacons has an impact on both the AMB-SLAM and MB-SLAM algorithms, and the error in the X-axis direction of the MB-SLAM algorithm can reach −60 to 90 m. From Figure 17b, we can see that when the time is about 4200, the error in the Y-axis of the AMB-SLAM algorithm suddenly goes from 2 m to about −5 m. From this, we can draw a conclusion that the density of the magnetic beacon has an effect on both AMN-SLAM and MB-SLAM.

Conclusions
In this study, the simultaneous localization and mapping (AMB-SLAM) algorithm for acoustic and magnetic beacons, based on MLP ANN, was proposed to allow AUVs to localize and map in a featureless seafloor environment. In our case, the 2D landmark map was chosen to represent the underwater exploring region. In order to verify the correctness and effectiveness of the neural network method in our study, the 10-fold cross-validation method was used. The results show that the presented approaches significantly improved the navigation performance of the AUV in cases where no features in the underwater field of interest are available. Firstly, the positioning accuracy of AMB-SLAM can reach approximately 1-2 m, and the positioning accuracy of the AMB-SLAM method is much higher than that of MB-SLAM. Secondly, the density of the magnetic beacons affects the positioning accuracy of AMB-SLAM and MB-SLAM. Compared with the MB-SLAM method, the AMB-SLAM algorithm has a set of redundant data regarding the relative range between the AUV and acoustic beacon, so its positioning accuracy is higher.
Compared with the existing traditional acoustic method, the INS method, the proposed method has many advantages. (1) By randomly distributing magnetic beacons, the AUV can localize and map in a featureless seafloor. (2) There is no need to deploy acoustic sensors on the seafloor and georeference them before the AUV performs its mission, as in the acoustic navigation method, which therefore allows for a low-complexity, low-cost, and high-precision solution to the AUV navigation and localization problem in featureless seawater environments. (3) The AMB-SLAM algorithm can be used for online positioning and mapping, which enables AUVs to autonomously explore or autonomously intervene in featureless seawater environments.
Compared with the existing sonar SLAM and optical SLAM algorithm, this is the first time that the SLAM method has been used to locate and map using the magnetic beacon. What is more, the computational cost of our proposed method is much lower, compared to the sonar SLAM and optical SLAM, which have to extract high-quality features from a large number of low-quality sonar or optical images. Thus, AMB-SLAM directly skips the process of extracting features as in the traditional SLAM algorithm, which allows a great deal of time to be saved and the data fusion complexity to be reduced. Moreover, the presented AMB-SLAM algorithm has a much lower mission cost and much

Conclusions
In this study, the simultaneous localization and mapping (AMB-SLAM) algorithm for acoustic and magnetic beacons, based on MLP ANN, was proposed to allow AUVs to localize and map in a featureless seafloor environment. In our case, the 2D landmark map was chosen to represent the underwater exploring region. In order to verify the correctness and effectiveness of the neural network method in our study, the 10-fold cross-validation method was used. The results show that the presented approaches significantly improved the navigation performance of the AUV in cases where no features in the underwater field of interest are available. Firstly, the positioning accuracy of AMB-SLAM can reach approximately 1-2 m, and the positioning accuracy of the AMB-SLAM method is much higher than that of MB-SLAM. Secondly, the density of the magnetic beacons affects the positioning accuracy of AMB-SLAM and MB-SLAM. Compared with the MB-SLAM method, the AMB-SLAM algorithm has a set of redundant data regarding the relative range between the AUV and acoustic beacon, so its positioning accuracy is higher.
Compared with the existing traditional acoustic method, the INS method, the proposed method has many advantages. (1) By randomly distributing magnetic beacons, the AUV can localize and map in a featureless seafloor. (2) There is no need to deploy acoustic sensors on the seafloor and georeference them before the AUV performs its mission, as in the acoustic navigation method, which therefore allows for a low-complexity, low-cost, and high-precision solution to the AUV navigation and localization problem in featureless seawater environments. (3) The AMB-SLAM algorithm can be used for online positioning and mapping, which enables AUVs to autonomously explore or autonomously intervene in featureless seawater environments.
Compared with the existing sonar SLAM and optical SLAM algorithm, this is the first time that the SLAM method has been used to locate and map using the magnetic beacon. What is more, the computational cost of our proposed method is much lower, compared to the sonar SLAM and optical SLAM, which have to extract high-quality features from a large number of low-quality sonar or optical images. Thus, AMB-SLAM directly skips the process of extracting features as in the traditional SLAM algorithm, which allows a great deal of time to be saved and the data fusion complexity to be reduced. Moreover, the presented AMB-SLAM algorithm has a much lower mission cost and much higher navigation precision than sonar SLAM and optical SLAM. Extracting available features is not required, as in the acoustic-based SLAM or optical-based SLAM. In conclusion, the presented AMB-SLAM method could save a lot of time and money associated with AUV missions and has a high positioning accuracy. In addition, AMB can realize real-time online positioning and mapping in featureless underwater environments.

Future Work
In the next stage of this research, in order to realize the goal of an artificially intelligent AUVs, the AMB-SLAM method that we proposed will be tested and validated with the help of the MLP ANN in the Bohai Sea, China in a featureless seawater environment, with a working area of 2000 × 2000 m 2 .
As for further improvements to the current study, the following will be considered in future research.
(1) The proposed algorithm AMB-SLAM does have some limitations. In this study, we assume that the effect of the geomagnetic field on the magnetic field of the magnetic beacon is the same in a small area. Future research needs to prove the influence of the geomagnetic field on the magnetic beacon field in order to validate this hypothesis.
(2) In our current study, the experiments were conducted under ideal conditions. In fact, the marine environment is so complex and changeable that we need a more realistic environment to verify the stability and applicability of this method. For example, the experiments were conducted in extreme submarine environments, with a low temperature (below 0 • C), high flow velocity, and high pressure (>10 atm), etc.