Next Article in Journal
A Numerical Simulation-Based Study on the Impact of Changes in Flow Rate of a Typical River Emptying into the Northern Yellow Sea on Water Environment of the River Estuary and Coastal Waters
Previous Article in Journal
Design of Portable Reefs to Protect Young Mangroves
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimized Deep Reinforcement Learning for Dual-Task Control in Deep-Sea Mining: Path Following and Obstacle Avoidance

1
State Key Laboratory of Ocean Engineering, Shanghai Jiao Tong University, 800 Dongchuan Road, Minhang District, Shanghai 200240, China
2
Yazhou Bay Institute of Deepsea SCI-TECH, Shanghai Jiao Tong University, Sanya 572024, China
3
Institute of Marine Equipment, Shanghai Jiao Tong University, Shanghai 200240, China
*
Author to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2025, 13(4), 735; https://doi.org/10.3390/jmse13040735
Submission received: 17 March 2025 / Revised: 30 March 2025 / Accepted: 3 April 2025 / Published: 6 April 2025
(This article belongs to the Section Ocean Engineering)

Abstract

:
This study investigates the dual-task control challenge of path following and obstacle avoidance for deep-sea mining robots operating in complex, unstructured environments. To address the limitations of traditional training strategies, we propose an optimized training framework that integrates environmental design enhancements and algorithmic advancements. Specifically, we develop a Dual-Task Training Environment by combining the Random Obstacle Environment with a newly proposed Obstructed Path Environment, ensuring a balanced learning approach. While agents trained solely in the Random Obstacle Environment exhibit unilateral obstacle avoidance strategies and achieve a 0% success rate in randomized obstacle scenarios, those trained in the Dual-Task Environment demonstrate 85.4% success under identical test conditions and acquire more complex bilateral avoidance strategies. Additionally, we introduce a Dynamic Multi-Step Update mechanism, which integrates immediate rewards with long-term returns to enhance deep reinforcement learning (Twin Delayed Deep Deterministic Policy Gradient, TD3) performance without increasing computational complexity. Under the optimal multi-step setting (n = 5), the Dynamic Multi-Step Update mechanism significantly improves path following accuracy, reducing trajectory deviations to 0.128 m on straight paths and 0.195 m on S-shaped paths, while achieving nearly 100% success in multi-directional obstacle avoidance tests. These improvements collectively enhance the adaptability, robustness, and operational performance of deep-sea mining robots, advancing intelligent control strategies for autonomous deep-sea exploration and resource extraction.

1. Introduction

The global demand for critical metals such as nickel, copper, cobalt, and rare earth elements continues to increase, while terrestrial resources are depleting, driving the shift toward deep-sea mineral resources [1]. Although deep-sea mining holds great promise, it faces significant challenges due to the extreme conditions of high pressure, low temperature, and complex seafloor terrain [2]. Mining robots, as illustrated in Figure 1, are central to this process, responsible for both mineral extraction and environmental monitoring. Deep-sea mining robots must possess highly accurate path following and real-time obstacle avoidance capabilities, effectively achieving dual-task control to operate efficiently and safely in such demanding environments [3,4].
In the field of path following for marine unmanned vehicles, Londhe [5] proposed a robust task-space control method based on PID-like fuzzy control, incorporating feed-forward terms and disturbance estimators to handle external disturbances and ensure accurate path following. On the other hand, Dai [6] combined Computational Fluid Dynamics (CFD) with PID control to develop a path following control model for underactuated unmanned underwater vehicles, successfully optimizing path following performance under complex ocean current conditions. Li [7] applied a Model Predictive Control (MPC) method to address track slippage issues, introducing an enhanced path following control strategy that ensures smooth following and stability in challenging environments such as soft sediments. Yan [8] proposed a robust MPC-based path following method utilizing a Finite-Time Extended State Observer (FTESO) to estimate and compensate for system uncertainties, improving the AUV’s disturbance rejection in dynamic environments. Chen [9] introduced a path following controller for deep-sea mining vehicles based on the Improved Deep Deterministic Policy Gradient (IDDPG) algorithm, incorporating slip control and random resistance. The controller demonstrated superior path following performance in simulations, better adapting to the complexities of deep-sea mining environments.
Although significant advancements have been made in path following control methods, a critical challenge remains in practical operations: how to effectively avoid obstacles while maintaining high following accuracy. To address this challenge, many scholars have proposed various path planning and obstacle avoidance algorithms aimed at optimizing underwater robots’ movement control in complex environments. Zhou [10] proposed an autonomous recovery path planning method for unmanned surface vehicles (USVs) based on the 3D-Sparse A* algorithm, optimizing waypoint generation to reduce computational load and improving path smoothness and accuracy during dynamic obstacle avoidance. Lyridis [11] introduced an improved Ant Colony Optimization (ACO) algorithm enhanced by fuzzy logic for local path planning in USVs, significantly improving the efficiency and accuracy of dynamic obstacle avoidance in complex environments. Lu [12] proposed a path planning method for deep-sea mining vehicles based on an improved Particle Swarm Optimization (IPSO) algorithm, which specifically addresses obstacle avoidance, path optimization, and crawler slippage, achieving efficient and safe navigation in complex seabed environments. Wang [13] introduced a speed-adaptive robust obstacle avoidance method based on Deep Reinforcement Learning (DRL) for environmentally driven unmanned surface vehicles (USVs). This approach optimizes the avoidance strategy, significantly improving the USV’s obstacle avoidance capability and adaptability in large-scale, uncertain environments. Li [14] proposed a path planning method for underwater obstacle avoidance in 3D unknown environments based on the Constrained Artificial Potential Field (C-APF) and Twin Delayed Deep Deterministic Policy Gradient algorithm. By combining classical path planning techniques with deep reinforcement learning, this method significantly enhances the obstacle avoidance capability of underwater robots in dynamic and complex environments.
While significant achievements have been made in the single tasks of path following and obstacle avoidance for marine unmanned robots, the challenge of effectively optimizing these dual tasks collaboratively remains unsolved. Wu [15] proposed a trajectory following method for deep-sea mining vehicles based on MPC, addressing track slippage and environmental disturbances, enabling precise path following and effective obstacle avoidance in dynamic deep-sea environments. However, the study was only tested in an environment with a single obstacle and did not consider dual-task collaborative control in dense obstacle scenarios, which limits its applicability in more complex environments. Nevertheless, the study provides a solid technical foundation for solving the dual-task collaborative control problem and lays the groundwork for further algorithm optimization and practical applications.
Although MPC has achieved some results in dual-task control for path following and obstacle avoidance, its application to dual-task collaboration still faces limitations. First, MPC relies on accurate environmental modeling and state estimation, and in dynamic and uncertain environments like the deep sea, the accuracy and real-time computational capacity of the model may be constrained. Additionally, MPC requires trade-offs between different objectives and relies on predefined models and control strategies, making it difficult to adapt to rapidly changing complex environments. In contrast, DRL, as a data-driven approach, overcomes these limitations by learning and optimizing through interaction with the environment. DRL does not rely on precise environmental models and can adjust control strategies in real time, effectively coordinating the trajectory-following and obstacle avoidance tasks. Especially in complex deep-sea mining environments, DRL offers stronger adaptability and robustness. By designing appropriate reward functions, DRL can dynamically optimize the collaboration between the two tasks, achieving more efficient and stable control.
To provide a clearer comparison of existing approaches in marine unmanned robots, Table 1 summarizes the main methods used for path following and obstacle avoidance. The comparison highlights their respective strengths and limitations, particularly in terms of adaptability, dual-task collaboration, and suitability for deep-sea mining applications.
While progress has been made in single-task control, research on the collaborative control of path following and obstacle avoidance is still limited. Therefore, the primary objective of this study is to develop a dual-task control algorithm that simultaneously optimizes both path following precision and obstacle avoidance performance, addressing the operational demands of deep-sea mining robots in complex and dynamic environments. This approach aims to enhance the overall efficiency and reliability of robotic operations under challenging conditions.
Distinguishing itself from previous works, this study leverages DRL to improve adaptability in the highly nonlinear, unpredictable nature of deep-sea mining operations. While DRL has demonstrated effectiveness in individual tasks of path following and obstacle avoidance, its application to an integrated dual-task framework remains underexplored. To bridge this gap, we propose a dual-task collaborative control framework that optimizes both objectives simultaneously, achieving superior performance within the same training time and computational budget. Furthermore, we introduce a Dynamic Multi-Step Update mechanism, which significantly enhances dual-task coordination while maintaining computational efficiency. By refining the update strategy without introducing substantial overhead, the Dynamic Multi-Step Update mechanism enables more stable and effective policy learning, leading to a marked improvement in the overall operational effectiveness of deep-sea mining robots.
The remainder of the paper is organized as follows: Notations frequently used in this paper are summarized in Table 2. Section 2 outlines the methodology, including the optimization of dynamics, noise incorporation, and reward design. Section 3 discusses training environments, highlighting the advantages of the Dual-Task Environment. Section 4 details the TD3 algorithm enhanced with Dynamic Multi-Step Update and its performance evaluation. Finally, Section 5 concludes the study and suggests future research directions.

2. Methodology

This section introduces the development methodology of the dual-task control system for deep-sea mining, focusing on two core aspects: modeling and controller design. Section 2.1 presents the simulation model of the deep-sea mining robot and optimizes its dynamic response, providing foundational support for controller development. Section 2.2 details the deep reinforcement learning-based control strategy, including the core principles of the TD3 algorithm, the incorporation of Gaussian noise in input–output design, and the reward function tailored for the dual tasks of path following and obstacle avoidance. All the work was conducted using Python 3.9.18.

2.1. Modeling

This study builds upon the system model of the four-tracked deep-sea mining robot “Pioneer II” developed by Chen [16], with its three-dimensional representation shown in Figure 2. Chen [16] systematically described the kinematic, dynamic, and slippage characteristics of the robot. The four-tracked design overcomes the limitations of two-tracked robots in navigating complex terrains, significantly enhancing the robot’s obstacle-surmounting capability under challenging seabed conditions. This improvement enables the robot to better meet the demanding operational requirements of deep-sea mining. To provide a comprehensive understanding of the robot’s design, Table 3 presents the key specifications of “Pioneer II”. These specifications serve as the foundation for analyzing the robot’s dynamic behavior and evaluating its control signal response. Building on this framework, we conducted an in-depth analysis of the control signal response characteristics of “Pioneer II”.
In the model proposed by Chen [16], the robot is assumed to adjust its track angular velocity at a uniform rate after receiving control signals, gradually reaching the target angular velocity within a 0.5 s control cycle. However, actual dynamic behavior during deep-sea operations reveals that “Pioneer II” exhibits a much faster response to control signals, typically approaching the target angular velocity within the first-time step (0.05 s) after the signal is issued, with subsequent velocity adjustments becoming minimal or stabilizing. To better align with this observed behavior, we introduced adaptive modifications to the model, enabling the track angular velocity to rapidly approach the target value during the initial response, rather than adjusting at a uniform rate. This modification provides a more accurate representation of the robot’s real-world dynamic characteristics.
Additionally, to further enhance the model’s realism and adaptability, a randomization mechanism was proposed to introduce variability in the track angular velocity response during each control time step. We formulate the refined model as Equation (1).
ω i t + Δ t = ω i t + ξ u i T ω i t ,
where ω i ( t ) denotes the angular velocity of the track driving wheel at time t , u i ( T ) represents the control signal, corresponding to the target angular velocity, and ξ is a random coefficient governing the rate of angular velocity change during the response time step, with values ranging from 0.8 to 1.0. The variable Δ t represents the minimum time step in the robot’s response cycle to the control signal, set to 0.05 s, while T denotes the time at which the control signal is issued.
This improved model not only enhances the dynamic adaptability of the control signal response, but also effectively simulates the nonlinear characteristics induced by complex terrains or environmental disturbances, making the model more representative of the actual dynamic behavior of deep-sea mining robots. However, it should be noted that the model does not account for extreme dynamics caused by sudden track obstructions or highly uneven terrain, which may result in an underestimation of velocity adjustments under such conditions. Future studies will aim to introduce more comprehensive adaptive mechanisms to further refine the model, ensuring greater alignment with real-world deep-sea operational scenarios.

2.2. Controller Design

Building on the previous section’s advancements in mining robot modeling and optimization, this section presents a DRL-based control framework for deep-sea mining robots. The framework employs the Twin Delayed Deep Deterministic Policy Gradient algorithm to achieve stable path following and real-time obstacle avoidance. To account for sensor uncertainties, Gaussian noise is introduced to the robot, target, and obstacle coordinates, enhancing robustness. Additionally, a multi-dimensional reward function is designed to balance path following and obstacle avoidance, optimizing the learning process.

2.2.1. DRL Algorithm

In dual-task control for path following and obstacle avoidance in deep-sea mining robots, the challenging operational environment and task demands place high requirements on the control system. Specifically, the robots need to make real-time decisions in dynamic environments to follow target points while avoiding obstacles. Additionally, the uncertainty of the deep-sea environment, such as sensor measurement errors and path disturbances, further increases the complexity of the control task. Traditional model-based control methods often struggle to address these challenges effectively.
In contrast, DRL offers significant potential in solving dual-task control problems by leveraging its powerful capability for online policy learning and adaptability to complex dynamic systems. By employing a data-driven approach, DRL transforms the problem from direct control to policy optimization, improving robustness against sensor noise and environmental disturbances. Moreover, DRL excels in high-dimensional state spaces, offering more efficient and intelligent solutions tailored for the operational demands of deep-sea mining robots.
This study adopts the TD3 algorithm as the core control framework due to its key advantages, as outlined below.
Smooth Action Generation and Stability: TD3 is based on the Actor–Critic framework and generates smooth and stable control signals for continuous tasks. The Actor network learns the policy by maximizing the estimated Q-value, and its gradient update is formulated as in Equation (2).
ϕ J ( ϕ ) = E a Q θ 1 ( s , a ) a = π ϕ ( s ) ϕ π ϕ ( s ) ,
where the objective function J ( ϕ ) is optimized to improve policy performance. The estimated Q-value Q θ 1 ( s , a ) , provided by the first Critic network, evaluates how beneficial an action is in a given state. The policy function π ϕ ( s ) , parameterized by ϕ , determines the action selection strategy. The gradient a Q θ 1 ( s , a ) indicates how the Q-value changes with respect to the action, guiding policy updates, while ϕ π ϕ ( s ) represents the sensitivity of the policy function to its parameters. By incorporating these gradients, TD3 ensures that the Actor network progressively refines the policy toward actions that yield higher expected rewards, ultimately enhancing control smoothness and stability in complex deep-sea environments [17].
High Stability and Sample Efficiency: By utilizing dual Critic networks and delayed policy updates, TD3 effectively mitigates instability during training and accelerates policy convergence. Specifically, the Critic networks estimate the Q-values, and the target Q-value is computed as formulated in Equation (3).
y = r + γ m i n i = 1 , 2 Q θ i s , π ϕ s + ϵ ,
where y represents the target Q-value used to update the Critic network. The reward r is obtained after executing action a , while γ is the discount factor that determines the importance of future rewards. The function Q θ i s , a represents the estimated Q-value given by the target Critic networks, with two separate networks indexed by i = 1 , 2 , ensuring the selection of the minimal value to reduce overestimation bias. The next action a is determined by the target policy π ϕ s , which adds a clipped noise term ϵ ~ c l i p ( N ( 0 , σ ) , - c , c ) to smooth policy updates and enhance training stability. The parameters θ i and ϕ belong to the target networks, which are updated using a soft update mechanism discussed later. This approach significantly improves training stability and sample efficiency, making it particularly suitable for environments with high uncertainty, such as deep-sea mining operations [18].
Environmental Adaptability and Robustness: TD3 improves adaptability to sensor noise and environmental disturbances through the minimum target value selection strategy and delayed policy update mechanism. These enhancements ensure stable decisions and the reliable execution of dual-task operations in complex deep-sea environments. To ensure stable training, soft updates are applied to both the Actor and Critic target networks, as formulated in Equation (4).
θ i τ θ i + ( 1 τ ) θ i ϕ τ ϕ + ( 1 τ ) ϕ ,
where θ i and ϕ represent the parameters of the target Critic and target Actor networks, respectively, while θ i and ϕ correspond to the parameters of the main Critic and Actor networks. The soft update coefficient τ controls the rate at which the target networks are updated, preventing abrupt changes in the target network, contributing to stable convergence and better performance in dynamic underwater environments [18].
These characteristics make TD3 particularly suited for dual-task control in deep-sea mining robots, providing a robust foundation for achieving efficient and safe operations in underwater conditions. The values of the relevant parameters in the algorithm are listed in Table 4. The training process of this study is based on the standard TD3 algorithm, with the hyperparameters in Table 4 meticulously fine-tuned from its baseline values to enhance the algorithm’s adaptability and performance in deep-sea mining path tracking and obstacle avoidance.
The design of state and action spaces, as well as the reward function tailored for dual-task control of the deep-sea mining robot, is detailed in Section 2.2.2 and Section 2.2.3.
To ensure the comparability of experimental results in Section 3 and Section 4, a fixed random seed was introduced in this study. The use of a random seed ensures consistency in network initialization, significantly reducing performance fluctuations caused by random variations during initialization. This approach effectively eliminates the influence of stochastic factors, enabling the experimental results to better reflect the actual effects of environment design and algorithm optimization. Consequently, this setup enhances the stability of experimental outcomes and the reliability of the research conclusions.

2.2.2. Robust Noise Injection and Input/Output

In deep-sea mining operations, robots are required to simultaneously perform path following and obstacle avoidance tasks. However, the complex working environment and sensor measurement errors may lead to control system instability. To enhance the controller’s adaptability to sensor errors, this study introduces Gaussian noise to the coordinates of the robot, target point, and obstacles when calculating their relative positions. This approach effectively simulates the impact of sensor measurement errors on positioning accuracy, allowing the controller to maintain stable performance even in the presence of localization deviations. As a result, the reliability and robustness of path following and obstacle avoidance tasks are significantly improved.
When calculating the distance and angular deviation between the robot, the target point, and obstacles, Gaussian noise is applied to the robot’s coordinates x R , y R , the target point’s coordinates x T , y T , and the obstacle’s coordinates x O , y O . The noise-affected coordinates, indicated by a superscript, represent the positions after noise is added, as introduced in Equation (5).
x R n o i s e = x R + μ c l i p N 1 0 , 1 , 1 , 1 y R n o i s e = y R + μ c l i p N 2 0 , 1 , 1 , 1 x O n o i s e = x O + μ c l i p N 3 0 , 1 , 1 , 1 y O n o i s e = y O + μ c l i p N 4 0 , 1 , 1 , 1 x T n o i s e = x T + μ c l i p N 5 0 , 1 , 1 , 1 y T n o i s e = y T + μ c l i p N 6 0 , 1 , 1 , 1 ,
where N i 0 , 1 i = 1 , 2 , , 6 represents random variables sampled from a standard normal distribution with mean 0 and variance 1, and μ is a scaling factor, set to 0.5 in this study. To prevent excessive noise that could cause data to deviate beyond reasonable bounds, a clipping operation c l i p is applied. This ensures that noise does not overly interfere with the localization data while retaining a degree of randomness to simulate real-world sensor errors. After incorporating Gaussian noise, the distance deviations between the robot and the target point and obstacles, D R T and D R O , can be calculated as shown in Equation (6).
D R T n o i s e = x T n o i s e x R n o i s e 2 + y T n o i s e y R n o i s e 2 D R O n o i s e = x O n o i s e x R n o i s e 2 + y O n o i s e y R n o i s e 2 R O ,
where R O is the radius of the obstacle. In this study, R O ranges from 2 m to 15 m.
Similarly, the heading angle deviations of the robot with respect to the target point and obstacles, φ R T and φ R O , are computed based on the noisy coordinates as shown in Equation (7).
φ R T n o i s e = θ h e a d i n g a r c t a n 2 y T n o i s e - y R n o i s e , x T n o i s e - x R n o i s e φ R O n o i s e = θ h e a d i n g a r c t a n 2 y O n o i s e - y R n o i s e , x O n o i s e - x R n o i s e ,
where θ h e a d i n g is the heading angle of the deep-sea mining robot.
To ensure that the angle deviations remain within a valid range, they are restricted to π , π , as formulated in Equation (8).
φ R T n o i s e = φ R T n o i s e 2 π r o u n d φ R T n o i s e / 2 π φ R O n o i s e = φ R O n o i s e 2 π r o u n d φ R T n o i s e / 2 π
These calculations ensure that, even with the introduction of random noise, the controller can accurately perceive relative positions and angular deviations, thereby enhancing the system’s robustness in complex operational environments.
To enhance the training efficiency and stability of the deep reinforcement learning (DRL) algorithm, the input states and output actions of the controller were normalized to the range 1 , 1 . This normalization not only eliminates the impact of varying data scales on training, but also accelerates convergence and improves algorithm performance. Based on this setup, the controller’s inputs consist of 12-dimensional observational data, while the outputs are 4-dimensional track angular velocity adjustments.
Observation Inputs:
i n p u t = c l i p D R T n o i s e / 50 1 , 1 , 1 c l i p D R O n o i s e / 15 1 , 1 , 1 c l i p φ R T n o i s e / π , 1 , 1 c l i p φ R O n o i s e / π , 1 , 1 c l i p V L / 0.5 1 , 1 , 1 c l i p V R / 0.5 1 , 1 , 1 c l i p A x + 0.4 / 4.8 , 1 , 1 c l i p A ω / 2.2 , 1 , 1 c l i p i L 1 / 1.0 , 1 , 1 c l i p i L 2 / 1.0 , 1 , 1 c l i p i R 1 / 1.0 , 1 , 1 c l i p i R 2 / 1.0 , 1 , 1 ,
where V L and V R denote the velocities of the left and right tracks of the deep-sea mining robot, respectively. Additionally, A x and A ω denote the forward acceleration and rotational acceleration, respectively, while i L 1 , i L 2 , i R 1 , i R 2 represent the slippage rates of the four tracks [16].
Notably, during training, the initial distance between the robot and the target point is fixed at 100 m, and the robot’s detection radius is set to 30 m. When no obstacle is detected, φ R O n o i s e is set to π / 2 , indicating that the obstacle is assumed to always be to the robot’s right, thereby not affecting the robot’s path following.
Action Outputs:
ω L 1 s i g n a l = ω m a x o u t p u t 1 + 1 / 2 ω L 2 s i g n a l = ω m a x o u t p u t 2 + 1 / 2 ω R 1 s i g n a l = ω m a x o u t p u t 3 + 1 / 2 ω R 2 s i g n a l = ω m a x o u t p u t 4 + 1 / 2 ,
where ω max represents the maximum allowable angular velocity for the tracks, with a value of 2.8274 rad/s in this study [16]. The Deep Reinforcement Learning (DRL) model generates a set of actions, denoted as output i for i = 1 , 2 , 3 , 4 , which serve as the raw control commands. These outputs are subsequently processed and transformed into the actual control signals, ω L 1 s i g n a l , ω L 2 s i g n a l , ω R 1 s i g n a l , ω R 2 s i g n a l , ensuring the precise execution of the control strategy.

2.2.3. Reward Design for Dual-Task Control

The design of reward functions is a critical challenge in reinforcement learning, essential for guiding agents to learn effective strategies. Research highlights the importance of balancing informativeness and sparsity in reward design to accelerate learning and maintain simplicity [19,20]. Explicable rewards shaped by potential functions enhance learning efficiency without altering optimal policies [20], while complex tasks can be addressed by decomposing them into subtasks with dedicated reward functions, improving robustness and adaptability [19,21]. Drawing on these principles and prior designs in similar scenarios [16], this study develops a multi-dimensional reward function system. Grounded in the TD3 algorithm and input–output framework, the system allocates reward signals to dynamically balance strategies for path following and obstacle avoidance tasks.
The reward functions are grounded in environmental feedback, avoiding the use of estimated or indirectly observed state information to ensure reliability and practical significance during training.
The reward framework is designed around the core objectives of dual-task control for deep-sea mining robots, integrating key parameters and logical conditions to effectively guide the robot in optimizing its decision-making strategy during training. The path following reward function r T F evaluates the robot’s angular deviation φ R T from the target in real time, as formulated in Equation (11). For small deviations, an exponentially decaying reward is applied to encourage rapid directional adjustments; for deviations exceeding π 2 , a fixed penalty is imposed to prevent loss of the target or excessive directional deviation. The parameter K 1 = 5 is introduced to modulate the magnitude of positive rewards in r T F , ensuring a balanced trade-off between directional correction and stable path-following behavior.
r T F = e K 1 φ R T 2 ,   if   φ R T π 2 1 ,   else
The obstacle avoidance reward function r O A monitors the minimum distance D R O between the robot and obstacles, implementing a dynamic penalty mechanism within the safety threshold, as formulated in Equation (12). When the robot approaches an obstacle closer than R e f f + D s a f e (the specific definitions of the effective radius R e f f and the safety distance D s a f e are illustrated in Figure 3), a penalty is applied to mitigate potential collision risks. Within the safe range, no reward or penalty is issued, maintaining the simplicity and stability of the reward function to support regular cruising behavior.
r O A = 1 ,   if   D R O < R e f f + D s a f e 0 ,   else
The velocity control reward function r V aims to encourage the robot to maintain an appropriate operating speed during tasks, preventing efficiency losses caused by low-speed operation. In deep-sea mining tasks, the robot’s speed range is defined as 0-1 m/s, with a minimum allowable forward velocity of 0.5 m/s. The reward value is linearly correlated with the velocity deviation, where K 2 = 2 . This parameter is selected such that the reward reaches its maximum value of +1 when the speed is 1.0 m/s and decreases linearly to 0 at the minimum allowable speed of 0.5 m/s. A fixed penalty is applied for speeds below this threshold, thereby improving overall task efficiency, as formulated in Equation (13).
r V = K 2 V h e a d i n g 0.5 ,   if   V h e a d i n g 0.5 1 ,   else
The slip reward function r S l i p monitors the slip rates of all four tracks to ensure stable operation on complex terrains. A penalty is imposed when the slip rate of any track exceeds 50% [16], reducing the risk of path deviation or task failure, as defined in Equation (14).
r S l i p = 1 ,   if max i L 1 , i L 2 , i R 1 , i R 2 > 0.5 0 ,   else
The single-step reward function r S t e p serves as a cornerstone of the dual-task control strategy for deep-sea mining robots, playing a pivotal role in balancing the priorities between path following and obstacle avoidance during reinforcement learning training. This function integrates multiple sub-tasks through a multi-dimensional weighting mechanism, guiding the robot to execute operations efficiently and reliably in complex deep-sea environments. The reward function design holistically incorporates four critical factors: path following, obstacle avoidance, velocity control, and slip detection, establishing a task-oriented and dynamically adjustable multi-task reinforcement learning framework, as formulated in Equation (15).
r S t e p = 5 ,   if   n s t e p 600 60 ,   elif   D R T L / 2 120 ,   elif   D R O R e f f ε 1 r T F + ε 2 r V + ε 3 r O A + ε 4 r S l i p ε 1 + ε 2 ,   else
The reward design incorporates direct rewards and penalties, focusing on balancing goal achievement and obstacle avoidance safety. Penalties are imposed if the task exceeds 600 steps without reaching the goal, preventing inefficient exploration. Rewards are provided when the robot approaches the target within half its body length, encouraging task completion [16]. Additionally, strong penalties are applied if the robot’s distance to an obstacle falls below the inflated radius R e f f , ensuring timely avoidance and enhancing the robustness of the obstacle avoidance strategy and overall operational safety.
During regular operations, the step reward function dynamically balances sub-task rewards for path following, obstacle avoidance, speed control, and slip detection through weighted integration. The specific weights are assigned as follows: ε 1 = 2 for path following, ε 2 = 1 for speed, ε 3 = ε 1 + ε 2 = 3 for obstacle avoidance, and ε 4 = 1 for slip penalties. By normalizing the positive rewards, the maximum step reward is capped at + 1 , ensuring a rational distribution of reward signals. This approach not only reflects the priority of individual sub-tasks, but also enables the dynamic adjustment of policy tendencies during reinforcement learning, achieving a synergy of safety and efficiency for the robot in complex deep-sea operational environments.

3. Environment Optimization

This section presents a training environment optimization framework aimed at enhancing the robustness and adaptability of the deep-sea mining robot in complex underwater conditions. To this end, a Dual-Task Environment (DTE) is developed, integrating the Random Obstacle Environment (ROE) for generalization and the Obstructed Path Environment (OPE) for structured navigation challenges. By dynamically alternating between these two settings, the DTE ensures a balanced training approach that strengthens both path following accuracy and obstacle avoidance capabilities. The following sections detail the environment design, training methodology, and simulation-based performance evaluation.

3.1. Environment Set Up for Dual-Task

Deep-sea mining robots must simultaneously achieve path following and obstacle avoidance, requiring extensive training across diverse environments to enhance their overall adaptability. This study builds on the conventional Random Obstacle Environment and proposes a novel Obstructed Path Environment. Additionally, a Dual-Task Environment is developed by integrating ROE and OPE to provide tailored training support for reinforcement learning, as illustrated in Figure 4.
The ROE, shown in Figure 4a, is a commonly used setup to evaluate the path following and obstacle avoidance capabilities of intelligent agents. The specific configurations are as follows:
Robot Initial Position: The robot’s initial position is randomly generated within a circular area centered at the origin (0, 0), with an initial heading angle θ H e a d i n g uniformly distributed between 0 and 180 .
Target Position: The target is randomly placed on a circular arc 100 m away from the origin, with an angular span θ T in the range of 0 , 180 .
Obstacle Placement: Obstacles are randomly positioned in the potential trajectory area between the robot and the target. The obstacles are distributed along a circular arc with a central angle θ O spanning [0°, 180°]. To ensure sufficient clearance, obstacles are placed at least 100 m away from both the origin and the target. Their radii range between 2 m and 15 m. While the obstacle placement is random, it rarely blocks the direct path between the robot and the target, occasionally creating hindrances.
ROE provides a simplified environment widely adopted in reinforcement learning for deep-sea mining tasks. While effective for training basic path following strategies, it often lacks the complexity to sufficiently simulate dual-task scenarios, where simultaneous path following and dynamic obstacle avoidance are critical.
To address the limitations of the ROE in training effective obstacle avoidance strategies, this study introduces the OPE. The core idea is to strategically place obstacles in positions that are more likely to directly block the robot’s path to the target, thereby enhancing the robot’s learning potential in complex avoidance tasks. As shown in Figure 4b, the OPE imposes additional constraints on obstacle distribution based on the initial positions of the robot and the target. The valid angular range for obstacle placement is rigorously constrained according to Equation (16).
φ m i n = θ T a r c s i n R O / D O φ m a x = θ T + a r c s i n R O / D O
Here, obstacle centers are randomly generated within the range, ensuring that the obstacles intersect with the direct path between the robot and the target.
The OPE enhances the specificity and complexity of avoidance strategy training by carefully controlling obstacle placement. Under OPE conditions, the robot must overcome direct obstructions, thereby improving its ability to handle challenging avoidance scenarios. However, while the OPE increases the complexity and difficulty of avoidance tasks, it may reduce the success rate of path following tasks.
To balance these trade-offs, this study combines ROE and OPE into a unified framework termed the DTE. The DTE is a comprehensive training framework designed to balance the training requirements for both path following and obstacle avoidance. In the DTE, each training episode alternates randomly between ROE and OPE, ensuring that the robot acquires the adaptability and task versatility required for different scenarios. Specifically, the DTE dynamically adjusts obstacle distribution characteristics (as in the OPE) and obstacle randomness (as in the ROE) throughout the training process, enabling the robot to achieve more holistic and robust training outcomes.
The DTE aims to balance the agent’s performance in path following and obstacle avoidance. Section 3.2 will evaluate its effectiveness through training analysis across different environments.

3.2. Training Process

This section presents a comprehensive analysis of the training process across three distinct environments: ROE, OPE, and DTE. The evaluation is conducted using two key performance metrics: cumulative reward and success rate, as illustrated in Figure 5. A rigorous assessment of these metrics provides deeper insights into the learning efficiency and adaptability of the proposed approach in diverse operational scenarios.
In the Random Obstacle Environment (ROE), the agent demonstrated rapid convergence, with a stable increase in cumulative reward. After approximately 2000 episodes, the cumulative reward plateaued at around 230, and the success rate surpassed 80%. This suggests that the relatively straightforward nature of the path following task in the ROE facilitated the agent’s ability to quickly develop effective navigation strategies. However, the limited complexity of obstacles in this environment may have restricted the agent’s capacity to acquire advanced obstacle avoidance skills.
In contrast, the OPE presented significantly greater challenges, as reflected by the instability of cumulative rewards throughout the training process. During the initial training phase, the cumulative reward increased rapidly but soon entered a prolonged stagnation period, during which it remained negative. It was not until approximately 2000 episodes that the reward began to rise again, turning positive at around 2300 episodes. However, due to the heightened complexity of the OPE, the cumulative reward exhibited persistent fluctuations throughout the training process, oscillating between −50 and the stabilized value of 230 observed in the ROE. This instability underscores the increased difficulty posed by the dynamic and obstructed nature of the environment, which required the agent to continuously adjust its strategies. Moreover, the overall success rate declined during the reward stagnation period and only began to recover once the cumulative reward started increasing again. Nevertheless, the final success rate remained below 50%, highlighting the considerable challenges introduced by the intricate obstacle arrangements in the OPE.
The DTE, designed as a hybrid training strategy integrating the characteristics of the ROE and the OPE, enabled the agent to concurrently learn both path following and obstacle avoidance tasks. In the early training phase, the cumulative reward increased rapidly but then entered a stagnation period before 1500 episodes—a characteristic similar to that observed in the OPE. However, due to the mixed training approach, this stagnation phase was significantly shorter, and the cumulative reward remained predominantly positive during this period. As training progressed, the cumulative reward continued to rise, eventually oscillating around 200 after surpassing the stabilized value of 230 observed in the ROE. This indicates that the DTE exhibited greater stability than the OPE while still being slightly less stable than the ROE, with a notably slower convergence rate than the latter.
The success rate in the DTE initially followed a trend similar to that in the OPE, declining during the cumulative reward stagnation period and only beginning to recover once the reward started increasing again. However, a key distinction was that the success rate in the DTE continued to rise steadily throughout the later stages of training, mirroring the trend observed in the ROE. Ultimately, after 8000 episodes, the success rate stabilized at approximately 80%, demonstrating the agent’s ability to adapt to dual-task scenarios. While the training process was inherently more complex, the DTE provided a structured learning framework that enhanced the agent’s capacity to balance path following and obstacle avoidance behaviors under dynamic conditions, contributing to its overall performance across diverse scenarios.
In summary, the ROE facilitated foundational training for path following, while the OPE significantly enhanced the agent’s obstacle avoidance capabilities. By integrating these characteristics, the DTE serves as a robust platform for comprehensive dual-task evaluation. Section 3.3 will further validate the agent’s performance through simulated testing across these environments.

3.3. Dual-Task Simulation Testing

This section evaluates the performance of controllers trained in three environments (ROE, OPE, and DTE) to assess their capabilities in path following and obstacle avoidance tasks. The evaluation consists of two parts: pure path following tests (without obstacles) and randomized obstacle avoidance tests. Each controller was subjected to 500 tests in its respective environment, with each test independently initialized using a randomized starting state and obstacle configuration. This ensures a diverse and unbiased assessment of the controller’s generalization ability across different scenarios.
Path following performance across different training environments: The evaluation of path following tasks was conducted on straight and S-shaped paths in three training environments: ROE, OPE, and DTE. In the context of path following tasks for deep-sea mining operations, the selected path designs aim to rigorously evaluate the robot’s capabilities in both straight-line and turn navigation. The straight path represents stable movement in constrained or relatively flat seabed environments, whereas the S-shaped path effectively simulates the robot’s ability to maneuver through complex terrain and avoid obstacles in operational settings. The performance metrics, including trajectory deviation and average speed, are summarized in Table 5 and illustrated in Figure 6.
To further evaluate the effectiveness of controllers trained in different environments, trajectory deviation and speed performance were assessed across both straight and S-shaped path following tasks.
For the straight path scenario, the controller trained in the ROE exhibited the highest precision, with the smallest trajectory deviation of 0.169 m. This indicates that training in a relatively structured environment optimized for path following tasks enabled the agent to achieve superior tracking accuracy. The DTE yielded a moderate trajectory deviation of 0.409 m, reflecting a balanced trade-off between precision and adaptability. In contrast, the controller trained in the OPE exhibited the largest trajectory deviation of 0.685 m, likely due to the increased complexity and dynamic obstacle arrangements encountered during training, which may have prioritized obstacle avoidance over strict trajectory adherence.
A similar trend was observed in the more challenging S-shaped path task. The ROE-trained controller maintained a trajectory deviation of 0.432 m, slightly outperforming the DTE-trained controller, which exhibited a deviation of 0.604 m. Meanwhile, the OPE-trained controller demonstrated the highest deviation at 0.734 m, indicating the heightened difficulty in maintaining precise path tracking under curved and complex trajectories.
Despite these variations in trajectory deviation, the average speed remained consistent at approximately 0.8 m/s across all environments. This suggests that differences in training strategies primarily influenced trajectory precision rather than the controller’s capability to maintain a stable velocity. The results highlight the trade-offs between training environments, with the ROE favoring precision, OPE emphasizing adaptability to complex conditions, and DTE achieving a balance between the two.
Obstacle avoidance performance across different training environments: The obstacle avoidance performance was evaluated by analyzing the completion rate across 500 independently initialized, randomly generated obstacle scenarios for each environment, as detailed in Table 5, with one such scenario illustrated in Figure 7. The completion rate is defined as the proportion of successful target-reaching instances out of 500 independently randomized initialization tests. To ensure a diverse set of obstacle configurations, each environment was uniquely initialized, thereby more accurately simulating the unpredictability and complexity of real-world deep-sea mining operations. This randomized initialization enables a comprehensive assessment of the algorithm’s adaptability to varying environmental conditions. The results indicate substantial performance differences across the three environments, underscoring the distinct challenges posed by different randomized obstacle distributions.
In the ROE, the controller failed to complete any obstacle avoidance tasks, with a completion rate of 0%, highlighting the inadequacy of this setting for training robust obstacle avoidance capabilities. The OPE-trained controller achieved a completion rate of only 6.6%, likely due to constrained obstacle placements increasing task difficulty. Both the OPE and ROE utilize unilateral obstacle avoidance strategies, which limit their effectiveness in handling more complex scenarios. In contrast, the DTE-trained controller significantly outperformed both, achieving a completion rate of 85.4%. This demonstrates the effectiveness of the bilateral obstacle avoidance strategies employed in the DTE, enabling dynamic adaptability to complex obstacle configurations and successful task completion.

3.4. Summary

This section presents a comprehensive investigation into the path following and obstacle avoidance capabilities of deep-sea mining robots, focusing on the design, training, and simulation testing of three distinct training environments: ROE, OPE, and DTE. The ROE is well suited for fundamental path following tasks, demonstrating robust stability and trajectory accuracy. However, its random obstacle distribution limits its training efficacy for obstacle avoidance. The OPE addresses this limitation by refining obstacle placement, significantly enhancing the robot’s obstacle avoidance capabilities. Nonetheless, its support for path following is weaker, and the high complexity of obstacle avoidance tasks reduces overall training efficiency. To achieve a balanced optimization of dual-task performance, this study introduces the DTE, which dynamically integrates the strengths of the ROE and OPE, enabling the robot to achieve a well-rounded performance in both tasks.
During training, the DTE exhibited superior adaptability, achieving relatively high cumulative rewards and success rates among the three environments, underscoring its potential for complex tasks. In path following, the DTE closely matched the ROE in trajectory accuracy and speed. In obstacle avoidance, its success rate far exceeded that of the other environments. Overall, the DTE provides robust support for the stability and operational efficiency of deep-sea mining robots in complex operational scenarios. In the next section, the study will further optimize the algorithm design based on the DTE to address more intricate and dynamic operational demands.

4. Algorithm Optimization

This section introduces an optimization methodology for enhancing the adaptability and decision-making efficiency of the TD3 algorithm in complex dual-task scenarios. To this end, a Dynamic Multi-Step Update mechanism is proposed, which extends the target Q-value estimation to multi-step returns, integrating both immediate rewards and long-term returns to improve policy learning. The mechanism dynamically adjusts trajectory selection during the multi-step return calculation, ensuring a balance between computational efficiency and optimization performance. To evaluate its effectiveness, training performance under different DMSU parameter settings is analyzed, followed by a comprehensive assessment of its impact on path following accuracy, strategy stability in obstacle avoidance, and robustness in multi-directional obstacle scenarios.

4.1. The Principle of Dynamic Multi-Step Update

To further enhance the adaptability and performance of the TD3 algorithm in complex dual-task scenarios, this study introduces the Dynamic Multi-Step Update mechanism. The DMSU extends the calculation range of target Q-values to multi-step returns, integrating immediate rewards with long-term returns to significantly improve the agent’s policy learning capability. Moreover, the DMSU dynamically adjusts the trajectory selection during the multi-step return calculation without imposing additional constraints, ensuring a balance between computational efficiency and effectiveness.
The target Q-value is calculated as shown in Equation (17).
Q t a r g e t s k , a k = t = k k + n 1 γ t k r t + γ n m i n Q θ 1 s k + n , a k + n , Q θ 2 s k + n , a k + n ,
where s k and a k denote the state and action at time step k ; n is the hyperparameter of the DMSU, defining the maximum number of steps for multi-step returns; Q θ 1 and Q θ 2 are the two target Critic networks that evaluate the state–action pairs at time step k + n .
During training, the DMSU adjusts the target Q-value calculation based on the trajectory length.
Complete trajectory calculation: When the remaining trajectory steps are greater than or equal to n , the full n-step return is computed as shown in Equation (17).
Incomplete trajectory calculation: When the remaining trajectory steps are fewer than n , all remaining step rewards are calculated, and the target Q-value for steps beyond n is set to zero. The simplified target Q-value is given by Equation (18).
Q t a r g e t s k , a k = t = k k + s 1 γ t k r t
This dynamic adjustment mechanism ensures the rationality and stability of target Q-value computation even in the presence of incomplete trajectories, thereby improving the algorithm’s generalization ability.
To visualize the integration of the DMSU into the TD3 algorithm, the improved algorithmic framework is depicted in Figure 8. The framework retains the dual-Critic structure of TD3 while enabling dynamic trajectory adjustment through the DMSU. The trajectories stored in the experience replay buffer are dynamically adapted to the logic of the DMSU, ensuring full utilization of trajectory information and stable training. The detailed algorithmic procedure is presented in Algorithm 1.
Algorithm 1: Optimized TD3 (reducing to the original TD3 algorithm when n = 1 )
Input:initial critic networks Q θ 1 , Q θ 2 and actor network π ϕ with random parameters θ 1 , θ 2 , ϕ
Input:initial target networks θ 1 θ 1 , θ 2 θ 2 , ϕ ϕ
Input:initial replay buffer B
Input:initial hyperparameters: actor network learning rate l α , critic network learning rate l β , target network soft update rate τ , mini-batch size k b a t c h , memory pool size N m a x , update actor interval N i n t , discount factor γ , learning time step t , standard deviation σ 1 , σ 2 , target action noise clipping c , exploration noise decaying coefficient λ e x p , policy noise ξ p o l i c y = c l i p ( n o r m a l ( 0 , σ 1 ) ) , c , c ) , exploration noise ξ e x p = n o r m a l ( 0 , σ 2 λ e x p t 5000 ) , total episodes T , random seed 24, max step length n
for  t = 1  to  T  do
Select action a t π ϕ ( s ) + ξ e x p
Observe: reward r t , new state s t + 1 , Store trajectory ( s t , a t , r t , s t + 1 ) in B with max size N m a x
Sample: mini-batch of k b a t c h trajectories ( s t , a t , k = t t + n 1 γ k t r k , s t + n ) from B , then update critic using:
     a t + n π ϕ ( s t + n ) + ξ p o l i c y ,
     y t = k = t t + n 1 γ k t r k + γ n m i n i = 1 , 2 Q θ i ( s t + n , a t + n ) ,
     θ i a r g m i n θ i k 1 ( y t Q θ i ( s t , a t ) ) , i = 1 , 2 .
if  t mod   N i n t = 0
    Update ϕ by the deterministic policy gradient:
     ϕ J ( ϕ ) = k 1 a t Q θ 1 ( s t , a t ) | a t = π ϕ ( s t ) ϕ π ϕ ( s t )
    Update target networks by:
     θ 1 ( 1 τ ) θ 1 + τ θ 1 ,
     θ 2 ( 1 τ ) θ 2 + τ θ 2 ,
     ϕ ( 1 τ ) ϕ + τ ϕ .
end if
end for
The Dynamic Multi-Step Update (DMSU) enhances the agent’s ability to optimize long-term objectives by incorporating multi-step returns while ensuring stable target Q-value computation under varying trajectory lengths. This mechanism mitigates the stochasticity of single-step rewards, leading to smoother policy updates and improved adaptability to complex task scenarios. Additionally, the flexibility and compatibility of the DMSU allow it to seamlessly integrate into the TD3 framework, preserving the stability of TD3 while significantly improving its learning efficiency in dual-task scenarios. As the core parameter of the DMSU, the step limit n plays a critical role in balancing short-term and long-term returns. Therefore, subsequent sections investigate the training and testing performance of agents under different n values, providing empirical evidence for optimizing algorithms in deep-sea mining dual-task control scenarios.

4.2. Training Process

In analyzing the impact of different multi-step settings ( n = 1 , 2 , 3 , 4 , 5 , 6 ) on the training process of the Dynamic Multi-Step Update (DMSU), significant differences in performance were observed through cumulative reward trends and success rates, as illustrated in Figure 9 and Figure 10. The evaluation was conducted over 8000 training episodes for each setting. Figure 9 presents the smoothed reward curves and their envelopes across iterations, while Figure 10 compares the sliding average success rates and overall success rates.
As n increased, the training performance exhibited noticeable variations. For n = 1 , although the cumulative rewards grew slowly in the early stages, the short-term update strategy struggled to adapt to complex tasks, leading to higher reward fluctuations and significantly lower final success rates compared to other settings. When n = 2 , 3 , 4 , both reward growth rates and success rates improved noticeably, but some fluctuations persisted during training, suggesting that multi-step updates in this range strike a moderate balance between short-term gains and long-term learning outcomes. At n = 5 , the training achieved its best performance. The reward curve stabilized, and the final success rate reached its peak, indicating outstanding training stability and efficiency in both path following and obstacle avoidance tasks. However, for n = 6 , the increased step count likely introduced more noise due to computational complexity, resulting in larger reward fluctuations and slightly lower final success rates compared to n = 5 .
Overall, n = 5 emerged as the optimal multi-step update setting, demonstrating a superior balance between stability and efficiency while ensuring long-term benefits for path following and obstacle avoidance tasks. Compared to traditional single-step updates, the DMSU effectively mitigates the impact of single-step reward fluctuations and dynamically enhances the adaptability of the agent to complex task scenarios. In the upcoming Section 4.3, simulation tests will further evaluate the impact of different multi-step settings on control performance and assess the practical effectiveness of the improved DMSU algorithm in path following and obstacle avoidance tasks, providing additional validation for its efficacy.

4.3. Dual-Task Simulation Testing

4.3.1. Path Following Without Obstacles

To comprehensively evaluate the performance of the DMSU strategy in path following tasks, this section conducts detailed simulation tests on controllers with varying multi-step update settings. The test paths include a straight path and an S-shaped path. Each controller was subjected to 500 repetitions on each test path to comprehensively assess following accuracy and speed performance in scenarios with varying path complexities. The results are summarized in Table 6, and the following trajectories are visualized in Figure 11.
For the straight path tests, the controller with n = 5 demonstrated the smallest trajectory deviation of 0.128 m, indicating optimal accuracy. This was closely followed by n = 4 , with a deviation of 0.144 m. In contrast, n = 1 (representing the original TD3 algorithm) showed a much higher deviation of 0.413 m, significantly underperforming. These results validate the notable improvement in following accuracy achieved through the Dynamic Multi-Step Update strategy. Additionally, the average speed across all settings remained close to 0.8 m/s, with n = 6 achieving the highest average speed of 0.826 m/s, slightly exceeding n = 5 at 0.822 m/s.
For the more complex S-shaped path tests, the controller with n = 4 achieved the lowest trajectory deviation of 0.179 m, showcasing the best following accuracy, followed by n = 5 with a deviation of 0.195 m, which also falls within acceptable limits. However, the deviation for n = 6 increased to 0.282 m, likely due to excessive update steps causing instability or disturbances. In terms of speed, all configurations maintained an average speed near 0.8 m/s. The highest average speed of 0.823 m/s was observed for n = 6 , but its reduced accuracy indicates potential trade-offs in performance.
Overall, the results from both the straight and S-shaped paths demonstrate that the Dynamic Multi-Step Update strategy significantly enhances following accuracy in path following tasks. Notably, the configurations with n = 4 and n = 5 outperformed the others. n = 4 achieved the lowest trajectory deviation in complex path scenarios, while n = 5 delivered the best balance of accuracy and speed, exhibiting superior overall performance. By comparison, n = 1 (original TD3 algorithm) showed markedly inferior results, further validating the effectiveness of the Dynamic Multi-Step Update strategy. In conclusion, n = 5 is identified as the optimal multi-step update configuration, providing a solid foundation for high-performing controllers in subsequent tasks. The next section will further evaluate the multi-step update strategy’s performance in obstacle avoidance tasks to comprehensively validate its practical applicability.

4.3.2. Validation of Strategy Consistency and Stability in Obstacle Avoidance

To verify the consistency and stability of the agent’s strategy in obstacle avoidance tasks, we conducted repeated simulations under identical random obstacle distribution environments for different step-length constraints ( n = 1 , 2 , 3 , 4 , 5 , 6 ). For each step-length setting, three trials were performed using a fixed set of randomly generated obstacles in the environment, and their trajectories were recorded to evaluate repeatability and consistency. The results are presented in Figure 12.
As shown in Figure 12, under identical obstacle distribution environments, the three repeated trajectories for different step-length settings exhibited high overlap on the overall path. Particularly for n = 4 and n = 5 , the trajectories across three trials were nearly identical, demonstrating the agent’s highly stable execution capability in obstacle avoidance tasks under consistent conditions. Even in regions with dense or complex obstacles, the overlap of trajectories remained high, indicating reliable repeatability in complex environments. For n = 1 , the limited step-length constraint reduced the generalization capability of the agent’s strategy, leading to slight deviations in certain regions. However, the overall repeatability remained within an acceptable range, further validating the significant impact of step-length parameters on the stability of the agent’s learning strategy.
The experimental results clearly demonstrate that the agent’s strategies in obstacle avoidance tasks possess a high degree of repeatability and consistency. This underscores the rationality and effectiveness of the training method and experimental design. High trajectory repeatability not only enhances the reliability of the experimental results, but also lays a solid foundation for subsequent tests in more complex environments. The next subsection will further evaluate the robustness and adaptability of the agent’s obstacle avoidance strategies in randomized dynamic obstacle environments.

4.3.3. Robustness Test in Multi-Directional Obstacle Avoidance

To evaluate the robustness and adaptability of the DMSU strategy in complex multi-directional obstacle environments, we conducted comprehensive tests under eight distinct directional obstacle scenarios, each with varying step-length constraints. For each step-length setting, 500 simulation trials were performed for each direction, with obstacles independently and randomly initialized in every trial. In total, 8 × 500 unique obstacle environments were generated to ensure a diverse and comprehensive assessment of the deep-sea mining robot’s obstacle avoidance capability.
Multi-directional obstacle avoidance evaluation is essential due to the inherent unpredictability of deep-sea mining operations. Variations in environmental disturbances, local flow conditions, and obstacle distributions can lead to non-uniform navigation challenges across different movement directions. A robust obstacle avoidance strategy must maintain consistent performance across diverse orientations to ensure operational reliability. Additionally, unpredictable external perturbations, such as fluctuating water flow, may introduce asymmetries in obstacle encounters, further necessitating a comprehensive multi-directional evaluation.
By systematically analyzing performance across a large set of randomly generated directional scenarios, this study ensures that the proposed method remains effective under varying obstacle configurations and environmental conditions, reinforcing its reliability in practical deep-sea mining applications.
Table 7 presents the obstacle avoidance completion rates across eight directions under different step-length constraints. The completion rate is defined as the proportion of successful target-reaching instances in 500 independently randomized initialization tests for a given direction and step-length constraint. The results indicate that when the step length was constrained to n = 5 , the completion rate reached 99.98%, demonstrating exceptional robustness and stability. This highlights that a reasonable step-length configuration can effectively optimize the agent’s adaptability to complex dynamic environments. However, for n = 6 , while the overall completion rate remained high, a slight decrease in certain directions (e.g., 135° and 180°) suggests that excessive step lengths might introduce computational complexity, reducing local path-planning efficiency.
The obstacle avoidance trajectories under step-length constraints of n = 1 and n = 5 are illustrated in Figure 13 and Figure 14, respectively. For n = 1 , the agent’s trajectories in certain directions, such as −90° and 180°, showed significant deviations from the target path, indicating that shorter step lengths limited the agent’s ability to effectively plan continuous paths. In contrast, for n = 5 , the trajectories across all directions were highly consistent and closely aligned, demonstrating superior path consistency and planning capabilities. This further validates that the controller under n = 5 exhibits exceptional path-planning stability and robustness in complex obstacle environments.
These findings lead to the conclusion that a well-balanced step-length setting (e.g., n = 5 ) effectively ensures both continuity and stability in path planning, enabling the agent to exhibit remarkable adaptability and robustness in complex multi-directional environments. In contrast, excessively short step lengths (e.g., n = 1 ) may restrict the agent’s global path-planning capabilities, while excessively large step lengths (e.g., n = 6 ) could introduce local inefficiencies due to increased computational complexity. The results underscore the critical role of step-length optimization in enhancing dynamic obstacle avoidance performance, providing a strong foundation for future algorithmic applications in complex environments.

5. Conclusions

This study introduces an optimized Deep Reinforcement Learning algorithm based on the Dynamic Multi-Step Update to systematically investigate and validate dual-task control for path following and obstacle avoidance in deep-sea mining vehicles under complex dynamic environments. By designing three training environments (ROE, OPE, and DTE), the research focuses on analyzing the performance of controllers across different scenarios, ultimately selecting the DTE as the optimal training environment. Under the DTE, the robot demonstrated enhanced capabilities in both path following and obstacle avoidance, overcoming the limitations of single-task environments and exhibiting remarkable balance and adaptability. Furthermore, the step length parameter, as a critical hyperparameter in the DMSU, played a decisive role in improving training efficiency, control precision, and algorithm stability. Systematic simulation tests confirmed that the controller achieved optimal comprehensive performance in path following and obstacle avoidance when the step length was set to a specific value.
In the path following tests, the simulation results under different step-length settings revealed that the controller with the specific step length achieved one of the lowest trajectory deviations on both straight and S-shaped paths, demonstrating the significant impact of the Dynamic Multi-Step Update mechanism on improving following accuracy. Specifically, it achieved the lowest trajectory deviation in the straight path test and the second lowest in the S-shaped path test. In randomized obstacle avoidance tests, the controller with the specific step length achieved a nearly 100% obstacle avoidance success rate across eight directions, with stable and repeatable trajectory performance, showcasing exceptional robustness and task execution capabilities. A comprehensive analysis of the simulation results validated the effectiveness and reliability of the improved algorithm in complex dynamic environments.
While the proposed method demonstrates promising performance in simulation, several aspects require further investigation to bridge the gap between the current study and real-world deep-sea mining operations.
Future research will focus on extending the control algorithm to real-world three-dimensional deep-sea environments, where the presence of complex and irregular seabed topographies poses significant challenges for obstacle avoidance. Additionally, the current study employs a simplified dynamic model with limited consideration of external disturbances, such as ocean currents and sediment dynamics. To enhance the system’s robustness, future work will incorporate more comprehensive hydrodynamic modeling and adaptive disturbance rejection mechanisms.
Furthermore, the proposed framework currently operates in an open-loop manner, lacking real-time perception and environmental feedback. To address this limitation, integrating sonar-based perception and simultaneous localization and mapping (SLAM) techniques will be a key focus. This enhancement will enable the deep-sea mining robot to construct a dynamic environmental representation, facilitating real-time decision-making and adaptive control in unstructured and evolving conditions.
By addressing these limitations, future studies aim to develop a more intelligent and autonomous deep-sea mining robot capable of operating efficiently in complex and unpredictable underwater environments, ultimately contributing to the advancement of deep-sea resource exploration and utilization.

Author Contributions

Y.X.: Conceptualization, methodology, software, data curation, investigation, writing—original draft, writing—review and editing, validation, visualization. J.Y.: Methodology, supervision, project administration, funding acquisition. Q.C.: Conceptualization, methodology, investigation, writing—review and editing. J.M.: Conceptualization, methodology, writing—review and editing. W.X.: Methodology, writing—review and editing. C.L.: Conceptualization, writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the Yazhou Bay Institute of Deepsea SCI-TECH, the State Key laboratory of Ocean Engineering, and the Institute of Marine Equipment. This work was also supported by the Major Projects of Strategic Emerging Industries in Shanghai (BH3230001) and the Science and Technology Committee Shanghai Municipality (19DZ1207300).

Data Availability Statement

The original contributions presented in this study are included in the article. Should there be any further inquiries, please feel free to contact the corresponding author.

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Sakellariadou, F.; Gonzalez, F.J.; Hein, J.R.; Rincón-Tomás, B.; Arvanitidis, N.; Kuhn, T. Seabed mining and blue growth: Exploring the potential of marine mineral deposits as a sustainable source of rare earth elements (MaREEs) (IUPAC Technical Report). Pure Appl. Chem. 2022, 94, 329–351. [Google Scholar] [CrossRef]
  2. Sharma, R. Deep-Sea Mining and the Water Column: Advances, Monitoring and Related Issues; Springer: Cham, Switzerland, 2024; pp. 3–40. [Google Scholar]
  3. Zhang, Q.; Chen, X.; Luan, L.; Sha, F.; Liu, X. Technology and equipment of deep-sea mining: State of the art and perspectives. Earth Energy Sci. 2025, 1, 65–84. [Google Scholar] [CrossRef]
  4. Leng, D.; Shao, S.; Xie, Y.; Wang, H.; Liu, G. A brief review of recent progress on deep sea mining vehicle. Ocean Eng. 2021, 228, 108565. [Google Scholar] [CrossRef]
  5. Londhe, P.S.; Mohan, S.; Patre, B.M.; Waghmare, L.M. Robust task-space control of an autonomous underwater vehicle-manipulator system by PID-like fuzzy control scheme with disturbance estimator. Ocean Eng. 2017, 139, 1–13. [Google Scholar] [CrossRef]
  6. Dai, Y.; Yin, W.; Ma, F. Nonlinear multi-body dynamic modeling and coordinated motion control simulation of deep-sea mining system. IEEE Access 2019, 7, 86242–86251. [Google Scholar] [CrossRef]
  7. Li, Y.; He, D.; Ma, F.; Liu, P.; Liu, Y. MPC-based trajectory tracking control of unmanned underwater tracked bulldozer considering track slipping and motion smoothing. Ocean Eng. 2023, 279, 114449. [Google Scholar] [CrossRef]
  8. Yan, Z.; Yan, J.; Cai, S.; Yu, Y.; Wu, Y. Robust MPC-based trajectory tracking of autonomous underwater vehicles with model uncertainty. Ocean Eng. 2023, 286, 115617. [Google Scholar] [CrossRef]
  9. Chen, Q.; Yang, J.; Mao, J.; Liang, Z.; Lu, C.; Sun, P. A path following controller for deep-sea mining vehicles considering slip control and random resistance based on improved deep deterministic policy gradient. Ocean Eng. 2023, 278, 114069. [Google Scholar] [CrossRef]
  10. Zhou, L.; Ye, X.; Yang, X.; Shao, Y.; Liu, X.; Xie, P.; Tong, Y. A 3D-Sparse A* autonomous recovery path planning algorithm for Unmanned Surface Vehicle. Ocean Eng. 2024, 301, 117565. [Google Scholar] [CrossRef]
  11. Lyridis, D.V. An improved ant colony optimization algorithm for unmanned surface vehicle local path planning with multi-modality constraints. Ocean Eng. 2021, 241, 109890. [Google Scholar] [CrossRef]
  12. Lu, C.; Yang, J.; Leira, B.J.; Chen, Q.; Wang, S. Three-dimensional path planning of Deep-Sea Mining vehicle based on improved particle swarm optimization. J. Mar. Sci. Eng. 2023, 11, 1797. [Google Scholar] [CrossRef]
  13. Wang, P.; Liu, R.; Tian, X.; Zhang, X.; Qiao, L.; Wang, Y. Obstacle avoidance for environmentally-driven USVs based on deep reinforcement learning in large-scale uncertain environments. Ocean Eng. 2023, 270, 113670. [Google Scholar] [CrossRef]
  14. Li, X.; Yu, S. Obstacle avoidance path planning for AUVs in a three-dimensional unknown environment based on the C-APF-TD3 algorithm. Ocean Eng. 2025, 315, 119886. [Google Scholar] [CrossRef]
  15. Wu, H.; Chen, Y.; Qin, H. MPC based trajectory tracking for an automonous deep-sea tracked mining vehicle. ASP Trans. Internet Things 2021, 1, 1–13. [Google Scholar] [CrossRef]
  16. Chen, Q.; Yang, J.; Zhao, W.; Tao, L.; Mao, J.; Li, Z. Algorithms for dynamic control of a deep-sea mining vehicle based on deep reinforcement learning. Ocean Eng. 2024, 298, 117199. [Google Scholar] [CrossRef]
  17. Silver, D.; Lever, G.; Heess, N.; Degris, T.; Wierstra, D.; Riedmiller, M. Deterministic policy gradient algorithms. In Proceedings of the International conference on machine learning, Beijing, China, 22–24 June 2014; pp. 387–395. [Google Scholar]
  18. Fujimoto, S.; Hoof, H.; Meger, D. Addressing function approximation error in actor-critic methods. In Proceedings of the 35th International conference on machine learning, Stockholm, Sweden, 10–15 July 2018; pp. 1587–1596. [Google Scholar]
  19. Devidze, R.; Radanovic, G.; Kamalaruban, P.; Singla, A. Explicable reward design for reinforcement learning agents. Adv. Neural Inf. Process. Syst. 2021, 34, 20118–20131. [Google Scholar]
  20. Eschmann, J. Reward function design in reinforcement learning. In Reinforcement Learning Algorithms: Analysis and Applications; Springer: Berlin/Heidelberg, Germany, 2021; pp. 25–33. [Google Scholar]
  21. Ratner, E.; Hadfield-Menell, D.; Dragan, A.D. Simplifying reward design through divide-and-conquer. arXiv 2018, arXiv:1806.02501. [Google Scholar]
Figure 1. Schematic diagram of deep-sea mining operations.
Figure 1. Schematic diagram of deep-sea mining operations.
Jmse 13 00735 g001
Figure 2. Three-dimensional model of the DSM robot (“Pioneer II”).
Figure 2. Three-dimensional model of the DSM robot (“Pioneer II”).
Jmse 13 00735 g002
Figure 3. Effective radius R e f f and safe distance D s a f e : The effective radius ( R e f f = 3.355   m ) is calculated for the deep-sea mining robot (length L = 6   m , width W = 3   m ) to ensure safe navigation. The safe distance ( D s a f e ) is set to 3 m to prevent collisions.
Figure 3. Effective radius R e f f and safe distance D s a f e : The effective radius ( R e f f = 3.355   m ) is calculated for the deep-sea mining robot (length L = 6   m , width W = 3   m ) to ensure safe navigation. The safe distance ( D s a f e ) is set to 3 m to prevent collisions.
Jmse 13 00735 g003
Figure 4. Deep-sea mining robot environments.
Figure 4. Deep-sea mining robot environments.
Jmse 13 00735 g004
Figure 5. Cumulative rewards and success rates during training in different environments: The cumulative reward represents the moving average of cumulative rewards over the past 100 episodes. The left axis indicates cumulative rewards, while the right axis corresponds to success rates.
Figure 5. Cumulative rewards and success rates during training in different environments: The cumulative reward represents the moving average of cumulative rewards over the past 100 episodes. The left axis indicates cumulative rewards, while the right axis corresponds to success rates.
Jmse 13 00735 g005
Figure 6. Representative path following test results across different training environments (selected from 500 randomized trials): The left figure shows the path following performance on a straight path, while the right figure illustrates the path following performance on an S-shaped path.
Figure 6. Representative path following test results across different training environments (selected from 500 randomized trials): The left figure shows the path following performance on a straight path, while the right figure illustrates the path following performance on an S-shaped path.
Jmse 13 00735 g006
Figure 7. Representative obstacle avoidance path comparison in different environments (selected from 500 randomized trials). The red rectangle represents the endpoint of the ROE path, while the purple rectangle signifies the endpoint of the OPE path.
Figure 7. Representative obstacle avoidance path comparison in different environments (selected from 500 randomized trials). The red rectangle represents the endpoint of the ROE path, while the purple rectangle signifies the endpoint of the OPE path.
Jmse 13 00735 g007
Figure 8. Algorithmic framework of OTD3 for dual-task control. When the red parameter n in the figure is set to 1, the control flow degenerates to the standard TD3 algorithm.
Figure 8. Algorithmic framework of OTD3 for dual-task control. When the red parameter n in the figure is set to 1, the control flow degenerates to the standard TD3 algorithm.
Jmse 13 00735 g008
Figure 9. Smoothed rewards and reward envelopes for different multi-step updates ( n = 1 to n = 6 ). The smoothed reward represents the moving average of cumulative rewards over the past 100 episodes, while the reward envelope indicates the range of variation during training for different n settings.
Figure 9. Smoothed rewards and reward envelopes for different multi-step updates ( n = 1 to n = 6 ). The smoothed reward represents the moving average of cumulative rewards over the past 100 episodes, while the reward envelope indicates the range of variation during training for different n settings.
Jmse 13 00735 g009
Figure 10. Success rate comparison under different multi-step settings. (a) Sliding average of success rate over episodes; (b) overall success rate over episodes.
Figure 10. Success rate comparison under different multi-step settings. (a) Sliding average of success rate over episodes; (b) overall success rate over episodes.
Jmse 13 00735 g010
Figure 11. Representative path following test results under different multi-step settings (selected from 500 randomized trials). The left figure shows the following performance on a straight path, while the right figure illustrates the following performance on an S-shaped path.
Figure 11. Representative path following test results under different multi-step settings (selected from 500 randomized trials). The left figure shows the following performance on a straight path, while the right figure illustrates the following performance on an S-shaped path.
Jmse 13 00735 g011
Figure 12. Obstacle avoidance performance comparison between different multi-step settings with repeatability validation.
Figure 12. Obstacle avoidance performance comparison between different multi-step settings with repeatability validation.
Jmse 13 00735 g012
Figure 13. Representative multi-directional obstacle avoidance test results under step limit n = 1 (selected from 500 randomized trials). Different colored rectangular blocks correspond to the endpoints of their respective paths.
Figure 13. Representative multi-directional obstacle avoidance test results under step limit n = 1 (selected from 500 randomized trials). Different colored rectangular blocks correspond to the endpoints of their respective paths.
Jmse 13 00735 g013
Figure 14. Representative multi-directional obstacle avoidance test results under step limit n = 5 (selected from 500 randomized trials). Different colored rectangular blocks correspond to the endpoints of their respective paths.
Figure 14. Representative multi-directional obstacle avoidance test results under step limit n = 5 (selected from 500 randomized trials). Different colored rectangular blocks correspond to the endpoints of their respective paths.
Jmse 13 00735 g014
Table 1. Comparison of path following and obstacle avoidance methods for marine unmanned robots.
Table 1. Comparison of path following and obstacle avoidance methods for marine unmanned robots.
Method CategoryRepresentative WorksPath
Following
Obstacle
Avoidance
Dual-Task
Collaboration
Adaptability to Dynamic
Environments
Suitability for Deep-Sea Mining
PID-BasedLondhe (PID-Fuzzy)
Dai (CFD + PID)
BasicNoneNoneLow–MediumLimited
MPC-BasedLi (MPC for Slippage)
Yan (MPC + FTESO)
StrongNoneNoneMedium–HighModerate
Wu (MPC for Dual-Task)StrongLimitedPartialMedium–HighModerate
Heuristic Path PlanningZhou (3D-Sparse A*)
Lyridis (ACO-Fuzzy)
Lu (IPSO)
NoneStrongNoneMedium–HighModerate
DRL-BasedChen (IDDPG-based)
Wang (DRL for USVs)
StrongStrongLimitedHighStrong
This paperStrongStrongFullHighStrong
Table 2. Nomenclature.
Table 2. Nomenclature.
SymbolsDescription
DRLDeep Learning Reinforcement
TD3Twin Delayed Deep Deterministic Policy Gradient
ROERandom Obstacle Environment
OPEObstructed Path Environment
DTEDual-Task Environment
DMSUDynamic Multi-Step Update Mechanism
PIDProportional–Integral–Derivative
CFDComputational Fluid Dynamics
MPCModel Predictive Control
FTESOFinite-Time Extended State Observer
AUVAutonomous Underwater Vehicle
USVUnmanned Surface Vehicle
ACOAnt Colony Optimization
IDDPGImproved Deep Deterministic Policy Gradient
IPSOimproved Particle Swarm Optimization
C-APFConstrained Artificial Potential Field
Table 3. The specifications of the four-tracked deep-sea mining robot.
Table 3. The specifications of the four-tracked deep-sea mining robot.
SymbolValueDescription
M 7000 kgTotal mass of the deep-sea mining robot underwater
I 34,096 kg · m2Rotational inertia of the deep-sea mining robot
D 1.88 mCenter distance of the tracks on both sides of the deep-sea mining robot
L 6 mThe length of the deep-sea mining robot
W 3 mThe width of the deep-sea mining robot
H 2.5 mThe height of the deep-sea mining robot
l 1.65 mThe length of one track of the deep-sea mining robot on ground
d 0.37 mThe end-to-end distance of two tracks positioned on the same side
b 0.55 mSingle track width
r 0.31 mDriving wheel radius of the track
Table 4. Hyperparameter settings of TD3 algorithm.
Table 4. Hyperparameter settings of TD3 algorithm.
ParameterSymbolValueDescription
Discount Factor γ 0.95Balances immediate rewards and long-term returns
Policy Update Delay N i n t 2Delayed policy update frequency
Actor Network Learning Rate Learning Rate l α 0.0001Learning rate for optimizing the Actor network parameters
Critic Network Learning Rate l β 0.0001Learning rate for optimizing the Critic network parameters
Target Network Soft Update Rate τ 0.0005Updates target networks using Exponential Moving Average (EMA)
Replay Buffer Size N m a x 32,000Maximum capacity of the experience buffer for storing interaction data
Batch Size k b a t c h 512Number of samples drawn from the replay buffer per training step
Target Action Noise Scale σ 1 0.2Standard deviation of Gaussian noise added to the target action for smoothing
Target Action Noise Clipping c 0.5Limits the amplitude of Gaussian noise to ensure stable target action outputs
Action Noise Scale σ 2 0.1Standard deviation of Gaussian noise added to the action for exploration
Exploration Noise Decaying Coefficient λ e x p 0.9Reduces the amplitude of action noise during training to transition from exploration to exploitation
Training Episodes T 8000Defines the total number of training episodes for adequate learning
Random Seed-24Controls the randomness in NumPy (1.26.4), PyTorch (2.4.0), and the environment to ensure reproducibility of experimental results
Table 5. Performance evaluation metrics for path following and obstacle avoidance across training environments.
Table 5. Performance evaluation metrics for path following and obstacle avoidance across training environments.
Training
Environment
Path FollowingObstacle Avoidance
Straight PathS-Shaped Path
Trajectory
Deviation
Average
Speed
Trajectory
Deviation
Average
Speed
Completion
Rate
Strategy
ROE0.1690.8260.4320.8060%Unilateral Obstacle Avoidance
DTE0.4090.8020.6040.79485.4%Bilateral Obstacle Avoidance
OPE0.6850.8100.7340.8036.6%Unilateral Obstacle Avoidance
Table 6. Performance metrics for path following under different multi-step settings.
Table 6. Performance metrics for path following under different multi-step settings.
Step LimitedStraight PathS-Shaped Path
Trajectory DeviationAverage SpeedTrajectory DeviationAverage Speed
n = 1 0.4130.8030.6030.793
n = 2 0.1490.8060.2000.801
n = 3 0.1480.8140.2230.815
n = 4 0.1440.8200.1790.819
n = 5 0.1280.8220.1950.814
n = 6 0.1500.8260.2820.823
Table 7. Performance metrics for multi-directional obstacle avoidance under different multi-step settings.
Table 7. Performance metrics for multi-directional obstacle avoidance under different multi-step settings.
Step LimitedRandom Obstacle Avoidance Test Completion Rate
−135°−90°−45°45°90°135°180°Average
n = 1 72.0%71.4%75.8%71.8%74.6%70.0%74.4%71.0%72.63%
n = 2 77.4%77.2%78.4%74.2%78.2%77.8%80.0%79.6%77.85%
n = 3 94.6%95.0%94.4%93.4%94.4%93.4%92.8%92.8%93.85%
n = 4 100.0%99.2%99.6%99.6%99.6%99.2%98.8%99.6%99.45%
n = 5 100.0%99.8%100.0%100.0%100.0%100.0%100.0%100.0%99.98%
n = 6 96.0%96.0%96.6%97.6%97.4%97.0%97.2%96.6%96.80%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xue, Y.; Yang, J.; Chen, Q.; Mao, J.; Xu, W.; Lu, C. Optimized Deep Reinforcement Learning for Dual-Task Control in Deep-Sea Mining: Path Following and Obstacle Avoidance. J. Mar. Sci. Eng. 2025, 13, 735. https://doi.org/10.3390/jmse13040735

AMA Style

Xue Y, Yang J, Chen Q, Mao J, Xu W, Lu C. Optimized Deep Reinforcement Learning for Dual-Task Control in Deep-Sea Mining: Path Following and Obstacle Avoidance. Journal of Marine Science and Engineering. 2025; 13(4):735. https://doi.org/10.3390/jmse13040735

Chicago/Turabian Style

Xue, Yulong, Jianmin Yang, Qihang Chen, Jinghang Mao, Wenhao Xu, and Changyu Lu. 2025. "Optimized Deep Reinforcement Learning for Dual-Task Control in Deep-Sea Mining: Path Following and Obstacle Avoidance" Journal of Marine Science and Engineering 13, no. 4: 735. https://doi.org/10.3390/jmse13040735

APA Style

Xue, Y., Yang, J., Chen, Q., Mao, J., Xu, W., & Lu, C. (2025). Optimized Deep Reinforcement Learning for Dual-Task Control in Deep-Sea Mining: Path Following and Obstacle Avoidance. Journal of Marine Science and Engineering, 13(4), 735. https://doi.org/10.3390/jmse13040735

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop