Next Article in Journal
Contactless Inductive Sensors Using Glass-Coated Microwires
Previous Article in Journal
Visual-to-Tactile Cross-Modal Generation Using a Class-Conditional GAN with Multi-Scale Discriminator and Hybrid Loss
Previous Article in Special Issue
Remaining Useful Life Prediction for Bearings Across Domains via a Subdomain Adaptation Network Driven by Spectral Clustering
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Policy Switching for Multi-Agent ASVs in Multi-Objective Aquatic Cleaning Environments

by
Dame Seck
1,
Samuel Yanes-Luis
2,
Manuel Perales-Esteve
1,
Sergio Toral Marín
1 and
Daniel Gutiérrez-Reina
1,*
1
Department of Electronic Engineering, University of Sevilla, 41004 Seville, Spain
2
Department of Electronic Technology, University of Sevilla, 41004 Seville, Spain
*
Author to whom correspondence should be addressed.
Sensors 2026, 26(2), 427; https://doi.org/10.3390/s26020427
Submission received: 19 November 2025 / Revised: 3 January 2026 / Accepted: 7 January 2026 / Published: 9 January 2026
(This article belongs to the Special Issue Advances in Wireless Sensor and Mobile Networks)

Abstract

Plastic pollution in aquatic environments is a major ecological problem requiring scalable autonomous solutions for cleanup. This study addresses the coordination of multiple Autonomous Surface Vehicles by formulating the problem as a Partially Observable Markov Game and decoupling the mission into two tasks: exploration to maximize coverage and cleaning to collect trash. These tasks share navigation requirements but present conflicting goals, motivating a multi-objective learning approach. The proposed multi-agent deep reinforcement learning framework involves the utilisation of the same Multitask Deep Q-network shared by all the agents, with a convolutional backbone and two heads, one dedicated to exploration and the other to cleaning. Parameter sharing and egocentric state design leverages agent homogeneity and enable experience aggregation across tasks. An adaptive mechanism governs task switching, combining task-specific rewards with a weighted aggregation and selecting tasks via a reward-greedy strategy. This enables the construction of Pareto fronts capturing non-dominated solutions. The framework demonstrates improvements over fixed-phase approaches, improving hypervolume and uniformity metrics by 14% and 300%, respectively. It also adapts to diverse initial trash distributions, providing decision-makers with a portfolio of effective and adaptive strategies for autonomous plastic cleanup.

1. Introduction

Macro-plastic (plastic > 25 mm) persists in aquatic environments, threatening ecosystems through entanglement, ingestion, and the trophic transfer of toxic additives [1,2]. Despite advances in mitigation and cleanup technologies using Autonomous Surface/Unmanned Surface Vehicles (ASVs/USVs) [3,4,5,6], a primary challenge in plastic cleanup missions remains the uncertainty of trash distribution before deployment [7,8,9]. Plastic contamination is highly dynamic—driven by water currents, wind, and human activity—which undermines static trajectory planning and limits the efficiency of deterministic control or classical optimization approaches. Recent advances in Deep Reinforcement Learning (DRL) have demonstrated the capacity to handle such uncertainty by enabling adaptive, data-driven decision-making. Canonical methods such as Deep Q-Learning (DQL) [10] and its extensions like Rainbow DQN [11] have shown strong performance in high-dimensional control tasks, including pollution monitoring [12,13,14] and trash collection [9]. To extend these capabilities to coordinated multi-robot systems, we model the cleanup mission as a Partially Observable Markov Game (POMG) [15], which naturally captures both the uncertainty of trash locations and the interactions among multiple ASVs operating simultaneously. The POMG generalizes the classical Markov Decision Process (MDP) to multi-agent scenarios where each vehicle senses only partial observations of the environment, reflecting realistic sensing and communication constraints. POMG thereby provides a mathematical foundation for Multi-Agent DRL (MADRL), which leverages parameter sharing to improve scalability and coordination across ASV fleets [16,17].
To manage mission complexity, we decouple the mission into two distinct tasks: exploration, where agents aim to cover the map to localize trash, and cleaning, where they focus on trash collection. Multi-task learning has been used in this context to jointly optimize policies for both objectives using Multi-task DQN (MDQN) architectures [7,8,18]. Unlike prior studies that relied on fixed-phase strategies, conducting exploration first and cleaning afterward [8], we treat task allocation as a dynamic decision variable. Specifically, a binary switch variable ν { 0 , 1 } enables each agent to adaptively select between exploratory and cleaning policies. Therefore, agents can shift toward cleaning when nearby plastic concentrations are high or return to exploration when the local area becomes depleted. This flexible task allocation introduces an inherent trade-off: excessive exploration reduces cleaning time, while insufficient exploration risks leaving undetected debris.
The use of Multi-Objective DRL (MO-DRL) has gained increasing attention for resolving trade-offs between conflicting objectives. Foundational studies such as [19,20,21] developed techniques for learning Pareto-optimal policies through scalarization and preference-conditioned networks. Building on these ideas, by training across varying allocations of ν , we construct Pareto-optimal fronts that capture the trade-offs between exploration coverage and cleaning efficiency. Recent studies have shown that, beyond uncertainty alone, aquatic environments fundamentally violate many of the assumptions underlying standard MO-DRL methods [22]. Even single-objective DRL has been shown to struggle with aquatic navigation due to non-stationary dynamics, degraded sensing and violations of the Markov property [22]. These challenges are further amplified in the multi-objective setting, where sparse, delayed and constrained objectives make reward design and scalarization more difficult, limiting the applicability of existing MO-DRL approaches to realistic aquatic missions. To support adaptive decision-making, we also introduce a reward-greedy policy selection mechanism inspired by submodular optimization principles [23,24]. This mechanism evaluates all possible task assignments across the fleet and selects the one that maximizes a weighted combination of exploration and cleaning performance. By systematically varying the weights, the framework reconstructs a diverse Pareto front, providing decision-makers with a portfolio of strategies tailored to different mission priorities. In summary, we adopt a Multi-Task Multi-Agent DRL (MT-MADRL) framework solved via multi-objective optimization. Our contributions are twofold: (i) a multi-task multi-agent DRL framework for ASV-based plastic cleanup that decouples exploration and cleaning while enabling adaptive task switching; (ii) a multi-objective optimization formulation that leverages reward decomposition and Pareto front construction to balance exploration and cleaning under diverse mission requirements.
The rest of this paper continues as follows: Section 2 introduces the environmental framework, including the scenario characteristics, system dynamics, and the methodology used throughout the study. It also describes the DRL formulation, the policy network architecture and training procedure, the state and reward definitions, and the proposed policy-selection mechanism. Section 3 details the simulation setup and presents the experimental results, including a comparison between the proposed method and existing approaches in the literature. Finally, Section 4 provides concluding remarks and outlines directions for future research.

2. Materials and Methods

2.1. Problem Formulation

We formulate the plastic cleanup mission as a POMG [25] involving N homogeneous ASVs operating in a discretized aquatic environment. The POMG framework naturally captures the agents’ limited perception and the interactions among multiple ASVs. Although agents share observations through centralized communication to construct a collective trash map Y ^ , this map represents only the last known state of the environment and may diverge from the true trash distribution Y due to continuous trash dynamics and sensing delays. Furthermore, regions outside the perception radius of all agents remain unobserved. The overall mission objective is to remove trash as efficiently as possible, which implicitly requires effective coordination, safe navigation, and energy-aware trajectories. In order to locate and remove trash items, it is necessary to explore the environment. Rather than considering this as an implicit part of the cleaning process, we have decided to decouple the learning process into two dedicated policies: an exploration policy, where agents aim to maximize coverage of the environment, and a cleaning policy, where agents collect as much identified trash as possible once locations are known. Both tasks share underlying requirements such as efficient navigation and obstacle avoidance, yet they present conflicting objectives: allocating more effort to exploration improves trash localization but reduces the time available for collection, while prioritizing cleaning without sufficient exploration leads to incomplete recovery. This inherent tension motivates a multi-objective learning framework.
Formally, let π denote a policy and let D be the number of objectives. Each policy produces a vector of performance metrics X π R D across episodes due to stochastic trash distributions. The expected performance is f π = ( f π 1 , , f π D ) , where f π i = E [ X π i ] . In our case, D = 2 , with f π 1 and f π 2 representing expected exploration and cleaning performance, respectively. A policy π * is Pareto-optimal if there is no other policy π such that:
f π i f π * i i { 1 , 2 } , with strict inequality for at least one objective .
The aim of this work is thus twofold: (i) to learn specialized exploration and cleaning policies, and (ii) to construct a Pareto front of solutions that captures the optimal trade-offs between the two objectives.

2.2. Scenario and Vehicle Properties

Let G = ( V , E ) denote a grid-based map of size H × W structured as a connected graph, where
V = { v i , j 1 i H , 1 j W }
is the set of nodes, with each node v i , j V corresponding to a location in the grid. The set of edges connecting adjacent nodes is denoted as E V × V , indicating possible movements between neighboring nodes. Each node connects to its eight immediate neighbors, making the grid eight-connected. Nodes with fewer than eight connections represent the presence of obstacles. The graph G can also be represented as a binary matrix M [ i , j ] { 0 , 1 } of size H × W (see Figure 1b):
M [ i , j ] = 1 if the node v i , j is navigable 0 otherwise
Let N denote the number of ASVs. The fleet position in G at time t is given by P = { p n t n = 1 , 2 , , N } , where p n t = v i , j V is the position of vehicle n. ASVs move synchronously in one of the eight cardinal directions (N, E, S, W, NE, SE, NW, SW), with each simulation step corresponding to one discrete movement. Actions leading to non-navigable nodes ( M [ i , j ] = 0 ) are masked to ensure valid transitions. When multiple ASVs attempt to occupy the same node, conflicts are resolved via a consensus algorithm [14], preventing inter-vehicle collisions. Energy constraints are directly modeled as a finite movement budget corresponding to battery capacity. Each agent’s episode ends after exceeding this distance budget, making policies implicitly energy-aware. Each ASV is also equipped with a camera. The perception area of vehicle n is defined as:
Θ ( p n ) = { v V v p n < ρ }
where ρ is the vision radius and v p n denotes Euclidean distance (see Figure 2).

2.3. Trash Dynamics and Recollection

Let B = { b k = ( x k , y k ) k = 1 , , K } denote the set of trash positions, where K is the total number of items and ( x k , y k ) are continuous coordinates. At the beginning of each episode, trash is generated in 1–10 clusters, with items sampled from a multivariate normal distribution. The dynamics of each item follow:
b k t + 1 = b k t + Δ t w w i n d · ν w i n d + w r a n d · ν r a n d ,
where ν w i n d U ( 1 , 1 ) is sampled once per episode, and ν r a n d U ( 1 , 1 ) is sampled independently at each step. Weights are fixed at w w i n d = 0.1 and w r a n d = 0.01 . Trash items are discretized into a matrix Y N H × W (see Figure 1c and Figure 2), where:
Y [ i , j ] = | { b k B ( x k , y k ) Area ( i , j ) } | ,
with Area ( i , j ) denoting the region corresponding to cell ( i , j ) . Each agent maintains an estimated trash map Y ^ [ i , j ] , initialized to zero and updated whenever v i , j Θ ( p n ) :
Y ^ [ i , j ] Y [ i , j ] .
Agents hare their observations via centralized communication, mitigating some limitations of partial observability but still subject to the uncertainty of trash dynamics. Inter-agent communication delays and robustness to disturbances are not explicitly modeled and are left for future work. However, the learning architecture itself is fleet-size independent, as agents rely on relative state information rather than a fixed team structure. Nevertheless, this property does not imply computational scalability of the task-allocation mechanism. When an ASV occupies p n = v i , j , it collects all trash at that node, with collection capacity assumed unlimited. The number of items collected is:
C ( p n , Y ) = Y [ i , j ] v i , j = p n .

2.4. Deep Reinforcement Learning Framework

Reinforcement Learning (RL) [26] addresses sequential decision-making by modeling the interaction between agents and the environment as a MDP ( S , A , T , R , γ ) . At each step t, an agent observes a state s t , selects an action a t , transitions to s t + 1 T ( s t + 1 | s t , a t ) , and receives a reward r t . The objective is to learn a policy π that maximizes the expected cumulative reward. Q-Learning estimates the action-value function Q π ( s , a ) , the expected return of taking action a in state s under π . Deep Q-Networks (DQNs) [10] approximate Q ( s , a ) with a neural network parameterized by θ , updated via:
Q ( s , a ; θ ) Q ( s , a ; θ ) + α ( r + γ max a Q ( s , a ; θ ) Q ( s , a ; θ ) ) ,
where α is the learning rate. Stability and performance are enhanced through mechanisms such as Experience Replay, Target Networks, Double DQN [27], and the Dueling Architecture [28].
MADRL generalizes DQL to a cooperative setting with N homogeneous agents acting in a shared environment, formalized as a POMG ( N , S , A , O , T , R , γ ) . Each agent n N receives a local observation o n t , selects an action a n t , and receives a reward R n , while the environment evolves according to the joint action a t = ( a 1 t , , a N t ) . Each agent aims to maximize its long-term return:
J π n , π n ( s ) = E a t π t = 0 γ t R n t ( s t , a t , s t + 1 ) s = s t ,
where n denotes all agents except n. In our cooperative ASV cleanup scenario, parameter sharing [29] is applied: all agents use the same policy network, allowing experiences to be aggregated, reducing computational cost, and ensuring scalability. To improve credit assignment, agents are rewarded based on their individual contributions to the team’s performance [30].
To further improve learning, the mission is formulated as a multi-task problem with two tasks: exploration, which aims to maximize coverage and locate trash, and cleaning, which focuses on collecting the identified trash. While the state and action spaces are shared, transition dynamics T ( z ) and rewards R ( z ) differ per task z Z , producing a vectorized reward signal. Following [18], Multi-task Deep Q-Networks (MDQN) employ separate Q-heads per task while sharing the underlying feature representation. This allows knowledge gained in exploration, such as efficient navigation and obstacle avoidance, to transfer to cleaning, improving overall mission performance.

2.5. Policy Network Training

2.5.1. Phase Construction

Following [7], the mission is divided into two phases to handle initially unknown trash locations: Exploration and Cleaning. The transition between phases is controlled by a variable ν { 0 , 1 } , representing the probability of selecting an exploration action:
π ν ( s ) = arg max a Q e ( s , a ) , with probability ν arg max a Q c ( s , a ) , with probability 1 ν
where Q e and Q c are the Q-functions for exploration and cleaning, respectively. In the Exploration phase, ν = 1 at every step, so all actions are chosen according to the exploratory policy Q e , with the primary goal of maximizing map coverage and discovering trash locations. In the Cleaning phase, ν = 0 , so actions are fully governed by the cleaning policy Q c , aiming to recollect the detected trash efficiently.

2.5.2. Action Space

Each autonomous surface vehicle operates in a discrete action space
A = { a 1 , , a 8 } = { N , S , E , W , NE , NW , SE , SW } ,
corresponding to the eight possible movement directions on the discretized grid. Actions that would result in collisions with obstacles or boundary violations are masked and excluded from execution, ensuring safe navigation within the environment.

2.5.3. States

The state representation for agent n is egocentric and composed of three H × W min-max normalized channels:
  • Trash model: Y ^ [ i , j ] , shared among all agents, representing detected trash positions and capturing partial observability.
  • Agent trail position: records the agent’s detection history over the past 10 steps, with current detection cells at 1 and past steps decaying by 0.1 per step, down to 0.1.
  • Position of other agents: the union of detection masks Θ of all other agents, with cells inside the masks set to 1, providing awareness of fleet distribution.
Although this work assumes a homogeneous fleet for clarity, heterogeneity in autonomy, sensors, or speed can be handled through the shared contamination model and agent-specific state channels. Position remains a universal attribute, and the contamination model can be extended to reflect sensor variability.

2.5.4. Rewards

At each step t, each agent n receives a vector of two rewards R t , n = [ R e , n , R c , n ] , corresponding to exploration and cleaning objectives.
The Exploration Reward ( R e ) encourages discovery of new areas while penalizing inactivity and redundancy:
R e , n = r v r , n p i p , n p r p , n
  • Visit reward  r v r , n : proportional to newly discovered cells, shared among overlapping agents:
    r v r , n = U t ( Θ ( p n ) ) U t 1 ( Θ ( p n ) ) η ( v , P )
    where η ( v , P ) = | { p n P v p n < ρ } | counts overlapping coverage.
  • Inactivity penalty  p i p , n : penalizes consecutive steps without discovering new areas:
    p i p , n = 1 δ · ξ , ξ = ξ + 1 , if no new cells discovered 0 , otherwise
    where ξ accumulates the number of consecutive steps without discovering new cells, and δ scales this penalty to control its impact.
  • Redundancy penalty  p r p , n : penalizes overlap with areas explored by self or other agents in the last τ steps:
    p r p , n = 1 | Θ ( p n t ) | · τ t = t τ t 1 m n | Θ ( p n t ) Θ ( p m t ) |
The Cleaning Reward ( R c ) encourages trash collection and efficient movement towards trash:
R c , n = r t c r , n + r m u , n + r d , n p t , n
  • Trash collection reward  r t c r , n = C ( p n , Y ) : rewards collecting trash at current position.
  • Distance reward  r d , n : promotes movement towards nearby trash using inverse Dijkstra distances:
    r d , n = 1 | Θ ( p n t ) | · η ( v , P ) v Θ ( p n ) b B 1 d ( v , b )
  • Model update reward  r m u , n = Y ^ t ( Θ ( p n ) ) Y ^ t 1 ( Θ ( p n ) ) : encourages adaptation to dynamic trash movement.
  • Time penalty  p t , n = 1 : discourages unnecessary delays.

2.5.5. Multi-Task Policy Network

The network containing the two task policies is trained following the Randomized ν -Sampling for Pareto Front Approximation Algorithm as in [8]. All agents share a single MDQN-based policy network composed of a dense Convolutional Neural Network (CNN) with two dueling Q-heads, one for exploration and one for cleaning. Parameter sharing is employed because the agents are homogeneous in both observations and actions, while the egocentric state formulation allows the policy architecture to generalize to arbitrary fleet sizes without structural modification. Each Q-head is trained independently using its respective task-specific rewards through DQL. The utilisation of shared layers facilitates the aggregation of experiences derived from both tasks, thereby reducing computational cost and enhancing sample efficiency. In order to ensure diverse learning across the exploration—cleaning trade-off, the duration of the exploration phase is sampled at the beginning of each episode. This duration determines when ν transitions from 1 (pure exploration) to 0 (pure cleaning) over the course of the mission. By varying the phase length across episodes, the network learns policies covering the full spectrum of task allocations. At the end of training, the network contains two fully trained policies, exploration π e and cleaning π c .

2.6. Reward-Greedy Policy Selection and Pareto Front Construction

To enhance adaptive decision-making, we introduce a Reward-Greedy Policy Selection (RGPS) mechanism. At each decision step, RGPS evaluates all possible task allocations across the fleet, specifically whether each ASV should perform exploration or cleaning. Given two possible tasks per vehicle, this results in 2 N possible task combinations for a fleet of N ASVs. Each combination is simulated to compute its expected scalarized reward, which aggregates exploration and cleaning objectives according to a given weighting of objectives. The combination yielding the highest weighted reward is selected, and each agent executes the corresponding action from its designated policy (exploration or cleaning). This process defines the fleet’s joint behavior at every step. The RGPS mechanism follows the principle of greedy submodular optimization, which is well-established for near-optimal decision-making in adaptive and stochastic settings [23,24]. In submodular systems, greedy maximization provides strong theoretical guarantees for solution quality under uncertainty—properties that align well with our partially observable and dynamically changing aquatic environment. Since the marginal benefit of additional exploratory actions typically diminishes as more trash is detected, the problem exhibits approximately submodular behavior. The Reward-Greedy method is simple to implement and it does not require additional learning. It explicitly considers all feasible task assignments, ensuring near-optimal short-term decisions based on the chosen reward weighting. Moreover, it leverages the separate training of exploration and cleaning policies, allowing agents to specialize while maintaining coordination. Figure 3 shows the diagram that summarizes our proposed algorithm. The RGPS task allocation mechanism is executed on a centralized server (base station) that receives observations from all ASVs, computes the optimal task assignment, and transmits individual action commands back to each vehicle. This centralized architecture leverages existing shore-based infrastructure commonly available in harbor and coastal operations.
Weighting Method: To balance exploration and cleaning, we introduce a weighting coefficient w [ 0 , 1 ] that combines the two objectives into a single scalar optimization function. This methods allow the user to specify preferences, which may be articulated in terms of goals or the relative importance of different objectives. To mitigate bias introduced by any single scalarization, several complementary weighting strategies are employed, including linear, multiplicative, power-based, and exponential formulations. Each scalarization explores different regions of the Pareto front, and their combination significantly improves coverage and diversity. Following the most common general scalarization methods for multi-objective optimization [31], with ( w 1 , w 2 ) = ( w , 1 w ) and ( R 1 , R 2 ) = ( R clean , R explore ) , the combined objective can be expressed using either:
  • Weighted Sum (WS): The most common approach to multi-objective optimization is the weighted sum method, which is essentially a convex combination of the two rewards, it suffers from the disadvantage of not being able to find a diverse set solutions if the Pareto front is non-convex.
    R total = i = 1 2 w i R i
  • Weighted Power (WP): In this method, each reward is raised to a power α before applying the weight. This allows for emphasizing certain objectives non-linearly and can be tuned via the parameter α :
    R total = i = 1 2 w i R i α
  • Weighted Product of Powers (WPOP): This multiplicative approach combines the rewards by taking each reward to the power of its respective weight and then multiplying them. It is useful for emphasizing that all objectives should be high, as a low value in any objective strongly reduces the total scalarized reward:
    R total = i = 1 2 R i w i
  • Exponential Weighted Criterion (EWC): In response to the inability of the weighted sum method to capture points on non-convex portions of the Pareto optimal surface, Ref. [32] propose the exponential weighted criterion, the performance of the method depends on the value of p and usually a large value of p is needed.
    R total = i = 1 k e p w i 1 e p R i
These formulations provide a flexible trade-off between the two competing objectives, enabling systematic exploration of the Pareto front. The RGPS mechanism uses this weighted reward to evaluate all possible policy assignments, selecting the combination that maximizes R total at each step. By repeating this process across different weightings w and multiple episodes, we obtain a set of policies whose expected performance in exploration and cleaning spans the trade-off space. Plotting the expected R explore versus R clean for these policies forms a Pareto front, representing the set of non-dominated strategies. Decision-makers can then select policies from the Pareto front according to mission priorities, ensuring an explicit balance between maximizing trash collection and achieving full coverage of the environment.
Novelty and Relationship to Prior Work: This work builds upon our previous study [8] in which we introduced a multi-task DQN framework for ASV fleet clean-up involving a fixed-phase decomposition between exploration and clean-up. In that approach, all agents transitioned synchronously between phases at a predefined timestep. In order to enable a fair comparison, we have retained the same environment, state representation, reward formulation and multi-task DQN architecture as in [8]. The key novelty of the present work lies in the task coordination mechanism. RGPS is an adaptive, per-agent task selection method that operates at every decision step, replacing the global, time-based phase switch. RGPS enables agents to dynamically specialise based on current environmental conditions, allowing for asynchronous and fine-grained task allocation across the fleet. This results in emergent coordinated behaviour without the need for additional learning modules or explicit communication, representing a significant departure from the fixed-phase strategy employed in [8].

3. Results

The proposed algorithm has been implemented in Python 3 (https://www.python.org/, accessed on 5 January 2025), using the PyTorch (https://pytorch.org/, accessed on 5 January 2025) library for policy optimization. The Gym (https://gymnasium.farama.org/, accessed on 5 January 2025) has been used to simulate the environment, NumPy (https://numpy.org/, accessed on 5 January 2025) and SciPy (https://scipy.org/, accessed on 5 January 2025) libraries for numerical and matrix operations. Pareto Fronts were constructed using Pymoo (https://pymoo.org/, accessed on 5 January 2025). The algorithm’s code and results are accessible in a GitHub repository (https://github.com/dsdiop/MultiProfilePlasticCleaningProblem, accessed on 5 January 2025). All experiments and simulations were run on an Intel Xeon Gold 5220R CPU operating at 2.20 GHz with 187 GB of RAM. Additionally, a NVIDIA GeForce RTX 3090 GPU with 24 GB of VRAM was employed to accelerate training.

3.1. Simulation Settings

The environment consists of four ASVs, each equipped with a detection radius of 2 nodes, a movement budget of 200 units per episode, and a movement step size of 1 node. The ASVs start from four fixed initial positions. Horizontal and vertical movements incur a cost of 1 unit, while diagonal movements cost 2 1.414 units.
Table 1 summarizes the environment and training parameters. These values are intentionally kept consistent with our previous work [8] in order to isolate the impact of the proposed adaptive task allocation mechanism, rather than confounding the comparison with changes in environment or learning hyperparameters.
Greedy Execution: The agents are evaluated over 100 episodes using the Reward-Greedy strategy. For each episode, a weight vector is sampled from a uniform distribution w U ( 0 , 1 ) , where one weight w [ 0 , 1 ] is drawn and the other is set to 1 w , ensuring that the sum of weights equals one. After completing the 100 episodes for a given weight vector, the mean values of the exploration metric (PMV) and the cleaning metric (PTC) are computed. This process is repeated for 100 different weight combinations. Finally, the Pareto Front is constructed from the resulting set of 100 ( PTC , PMV ) metric pairs. The entire procedure is performed separately for each weight-based scalarization method. The path planner that is going to choose the action once the decision to explore or clean is taken by the RGPS agent, is trained as in [8].

3.2. Metrics and Hyperparameter Sensitivity Analysis

The performance is evaluated fof each policy using a set of objective metrics that quantify exploration efficiency and cleaning effectiveness.
  • Percentage of the Map Visited (PMV): It is the exploration objective to be maximized, and it measures the percentage of the map that has been visited during the mission. It is defined as the proportion of the total navigable area that has been visited at least once by any ASV. A higher PMV value indicates a more comprehensive exploration, ensuring that the agents gather sufficient information about the environment’s structure, obstacles, and trash distribution, which is crucial for the subsequent cleaning phase. Mathematically, PMV is given by:
    P M V = i j U T [ i , j ] i j M [ i , j ] · 100
  • Percentage of Trash Cleaned (PTC): It is the cleaning objective to be maximized, and it measures the proportion of trash that has been effectively eliminated from the environment, relative to the total amount of trash present at the beginning of the episode (K). Optimizing PTC requires efficient path planning and coordination among agents to maximize trash collection before exceeding their distance limits. PTC is mathematically defined as:
    P T C = t T p n P C ( p n , Y t ) K · 100
To characterize the trade-offs between these objectives, we compute Pareto front metrics that capture the diversity, extension and quality of the resulting solutions.
  • Consecutive Spacing: The spacing metric measures the variance of distances between consecutive solutions along the Pareto front, sorted by the first objective f 1 . It captures gaps along the front better than nearest-neighbor spacing in 2D.
    For a sorted Pareto front Y N = { y 1 , y 2 , , y N } , the consecutive distances are
    d i = y i + 1 y i , i = 1 , , N 1 ,
    and the spacing metric is defined as the standard deviation of these distances:
    S = 1 N 2 i = 1 N 1 d i d ¯ 2 ,
    where
    d ¯ = 1 N 1 i = 1 N 1 d i .
    A lower S indicates more uniform spacing along the front.
  • Zitzler’s M 3 Metric: The M 3 metric measures the extent of the Pareto front along all objectives. For a Pareto front Y N = { y 1 , y 2 , , y N } in m objectives, let
    Z R i = max 1 j N y j ( i ) min 1 j N y j ( i ) , i = 1 , , m ,
    be the range of the front along the i-th objective. Then the M 3 metric is defined as
    Z M 3 = i = 1 m Z R i .
    A higher M 3 indicates a larger extent of the front across all objectives.
  • Hypervolume (HV): The hypervolume quantifies the size of the objective space dominated by the Pareto Front (PF) with respect to a reference point r = ( r 1 , r 2 ) . For discrete non-dominated points sorted by f 1 (e.g., PTC), the hypervolume can be approximated as the sum of rectangular areas between consecutive PF points:
    H V = i = 1 N ( r 1 f 1 ( i ) ) ( f 2 ( i ) f 2 ( i 1 ) )
    where f 1 ( i ) and f 2 ( i ) denote the objective values of the i-th non-dominated solution in the sorted PF, and f 2 ( 0 ) = r 2 corresponds to the reference point. A higher H V indicates a larger and more dominant Pareto region, reflecting both convergence and diversity.
Hyperparameter Sensitivity Analysis: To assess the robustness of the proposed framework with respect to scalarization hyperparameters, we conducted a sensitivity analysis on the parameters governing the non-linear scalarization methods. Specifically, we evaluated the effect of the exponent a in the Weighted Power (WP) method and the parameter p in the Exponential Weighted Criterion (EWC) method on the resulting Pareto fronts. For WP, we considered a { 1 , 3 , 5 , 7 , 9 } , while for EWC we evaluated p { 1 , 3 , 5 , 7 } . The analysis focuses on the Hypervolume metric as the primary indicator of multiobjective performance, complemented by the M3 and Spacing metrics. The results show that, although absolute metric values vary with the choice of parameters, the qualitative behavior of the framework remains consistent across the tested ranges. Moderate parameter values yield balanced Pareto coverage, whereas more extreme values increasingly bias the scalarization toward specific objectives. Based on this analysis, we select a = 3 for WP and p = 1 for EWC in the experiments reported in this work. Figure 4a shows that increasing the WP exponent a progressively compresses the Pareto front and reduces coverage of balanced trade-offs between cleaning and exploration. We therefore select a = 3 as a moderate setting that preserves a well-spread Pareto front without the strong bias and loss of diversity observed for larger values. Moreover, we select p = 1 for EWC as it yields the highest hypervolume (see Figure 4b) while avoiding the increased bias and Pareto front distortion observed for larger values of p.

3.3. Scalarization Results

As illustrated in Figure 5, the Pareto fronts obtained for each scalarization method under the Reward-Greedy Policy Selection (RGPS) framework are presented, in conjunction with the combined Pareto set which integrates all scalarizations to exploit the complete information available. The corresponding performance metrics are summarised in Table 2. Each Pareto point corresponds to the mean performance over 100 independent evaluation episodes under fixed trained policies, rather than to individual trajectories.
It is evident that the WP method exhibits significantly poorer performance in comparison to the alternative approaches. The Wilcoxon signed-rank test, conducted between Pareto fronts, revealed that the remaining methods do not exhibit statistically significant differences (Table 3). The WS approach achieves the highest hypervolume; however, it is known that it is limited to convex regions of the Pareto front. In contrast, the WPOP and EWC methods, both non-linear scalarizations, achieve slightly lower hypervolume and Pareto extension ( M 3 ) values but yield improved Spacing due to their ability to explore non-convex regions. Furthermore, due to the convex-constrained nature of WS, numerous weight combinations tend to converge towards the boundary of the front, thereby leading to elevated M 3 values.
When all scalarization results are combined, the best overall metrics are achieved, along with an increased number of Pareto-optimal points. Figure 5b illustrates the contribution of each scalarization method to the combined Pareto front. Out of 40 total Pareto points, EWC contributes the majority (47.50%), followed by WS (27.50%) and WPOP (25.00%), while WP does not contribute. This highlights the importance of employing multiple scalarization techniques of distinct natures (linear, non-linear, additive, multiplicative, and exponential) to achieve a well-distributed and comprehensive Pareto front. The combined front represents our proposed contribution and will serve as the basis for comparison against alternative algorithms from the literature.

3.3.1. Comparison

For benchmarking against existing approaches, we adopt the methodology introduced in [8], hereafter referred to as the Fixed-Phase Pareto Set (FP2S). A predefined set of phase transition configurations, denoted as the Phase Duration Set (PDS), is used to generate possible phase durations whose candidate policies are later evaluated for Pareto optimality. A value in the PDS represents the proportion of the mission dedicated to the exploration phase. The PDS ranges from 1.0 to 0.0 in decrements of 0.1 (therefore 11 values), forming the sequence:
P D S q = 1.0 0.1 q , q { 0 , 1 , 2 , , 10 } .
Here, q represents the step index, ensuring a uniform progression from a mission of pure exploration ( P D S 1 = 1.0 ) to pure cleaning ( P D S 0 = 0.0 ). For instance, a P D S q of 0.3 ( q = 7 ) means that 30% of the mission corresponds to exploration (all agents executing exploration behavior), followed by 70% cleaning. During training, the network is periodically evaluated under each PDS configuration. If a newly trained policy surpasses the current best for a specific PDS and objective, it replaces the previous one, ensuring that only the most optimal policies are retained for Pareto evaluation. Then, for each selected PDS value, three distinct policies are stored as Pareto candidates every 500 training episodes: (1) Best Cleaning Policy: The policy achieving the highest cleaning performance within that PDS. (2) Best Exploration Policy: The policy attaining the greatest exploration efficiency within that PDS. (3) Final Trained Policy: The policy obtained at the end of training, which may exhibit improved generalization to unseen scenarios due to prolonged optimization.
Once all candidate policies are trained, Pareto optimality is determined through post-training evaluation. Each policy generated from the Phase Duration Set (PDS) is tested on an independent evaluation set consisting of 200 randomized episodes per environment. During this stage, the average PTC and PMV metrics are measured for each policy. These averaged results are used to determine Pareto optimality and construct the final Fixed-Phase Pareto Set with ours.
Table 4 reports the performance of the FP2S policies across different fixed exploration-to-cleaning phase ratios (PDS). Figure 6 presents a comparative analysis between the Pareto Front obtained with our proposed Combined approach and the FP2S baseline from the literature. The front generated by FP2S exhibits a smaller number of Pareto-optimal solutions, concentrated primarily in regions with high exploration but limited cleaning performance. This scarcity of solutions results in a sparser and less continuous front, which ultimately reduces the achievable hypervolume and diversity. In contrast, the Combined Pareto Front (black crosses) demonstrates a much denser and more evenly distributed set of non-dominated policies, extending further along both objective axes. This indicates that our method achieves a broader and more balanced coverage of the trade-off space between exploration and cleaning.
As detailed in Table 2, our approach achieves a significantly higher hypervolume value (5810.03 compared to 5091.75). The improved hypervolume indicates that the Combined Pareto Front dominates a larger portion of the objective space, offering more favorable trade-offs between exploration efficiency and cleaning effectiveness. Furthermore, the Spacing metric decreases substantially (0.70 versus 2.14), showing that the solutions in our front are more uniformly distributed, providing smoother transitions between neighboring policies. This uniformity is desirable for multi-objective decision-making, as it allows for finer granularity when selecting operating points according to specific mission preferences.
The extension metric M 3 remains comparable between both methods, although our Combined approach achieves similar coverage with a substantially larger number of points (40 versus 9). This richer set of solutions improves the interpretability and applicability of the Pareto front, giving mission planners a broader range of policies to select from based on operational priorities or environmental conditions. It is worth noting that the FP2S front’s lower diversity is primarily due to its reliance on fixed-duration phase transitions, which limits the variability of exploration–cleaning balance during training. By contrast, our approach combines multiple scalarization strategies of different natures—linear, nonlinear, and exponential—allowing the discovery of policies across a wider spectrum of trade-offs.
Overall, the results clearly demonstrate that the proposed Combined Pareto Front not only surpasses the FP2S baseline in terms of quantitative Pareto metrics but also provides greater adaptability, finer policy resolution, and a more comprehensive representation of the underlying exploration–cleaning dynamics. This enhanced front serves as a more informative and flexible decision-making tool for real-world multi-objective environmental cleanup missions.
Figure 7 illustrates the evolution of task allocations across the four agents for three representative scalarization methods, as well as the evolution produced by FP2S for a given P D S . Each line indicates the instantaneous task selected at each time step ( ν = 1 for exploration, ν = 0 for cleaning). As expected, the EWC configuration (Figure 7a), which heavily prioritizes cleaning ( w 1 = 0.991 , w 2 = 0.009 ), produces trajectories dominated by cleaning action selections, reflecting a predominance of cleaning behavior. Conversely, the WS configuration (Figure 7c), which emphasizes exploration ( w 1 = 0.162 , w 2 = 0.838 ), results in substantially fewer cleaning action selections, consistent with exploratory dominance. The WPOP configuration (Figure 7b) lies between these two extremes, producing a balanced alternation between exploration and cleaning phases. Finally, Figure 7d illustrates FP2S, which—as opposed to the adaptive DRL-based policies—applies a predefined task schedule shared synchronously across all agents, and therefore does not constitute an adaptive policy-switching mechanism. However, the temporal evolution of ν within each episode reveals that the agents’ task decisions are not trivially determined by the scalarization weights. Even in strongly biased configurations, agents exhibit frequent task switches and non-monotonic ν trajectories. This behavior can be attributed to the strong interdependence between exploration and cleaning: an agent’s optimal task choice depends not only on its local context but also on the collective state of the environment and the complementary actions of other agents. Consequently, the system demonstrates emergent coordination, where task allocation dynamically adapts to environmental and inter-agent conditions.
Adaptation to Different Scales:
To evaluate the framework’s capacity for generalisation across environments of varying scales and structural complexity, additional experiments were conducted on the Alamillo Lake map (see Figure 8). This map contains 563 navigable nodes, constrained corridors, and internal obstacles. This environment contrasts with Malaga Port, which has 705 navigable nodes (approximately 25% more) and a more open layout. Despite these differences in size and topology, the proposed coordination framework performs consistently (as shown in Figure 9 and Table 5), generating well-formed Pareto fronts in both environments without any environment-specific parameter tuning. Crucially, execution time per decision step remains consistent across environments because task allocation and policy evaluation depend on fleet size rather than map scale. This confirms that the proposed method maintains real-time feasibility when transitioning to structurally different environments. Notably, the Alamillo Lake environment achieves higher overall performance than Malaga Port. This can be attributed to its smaller number of navigable cells, which enables faster spatial coverage under the same step budget. Accelerated exploration increases the likelihood of visiting cells containing trash earlier, which improves the percentage of trash collected directly, since trash is removed immediately upon cell visitation.

3.3.2. Limitations

While the proposed RGPS framework demonstrates significant improvements over fixed-phase approaches in multi-agent ASV cleanup operations, it is important to acknowledge several limitations that constrain its current applicability and suggest directions for future research. The most significant limitation of the RGPS mechanism is its exponential computational complexity with respect to fleet size. At each decision step, RGPS evaluates all possible task allocations across the fleet, requiring 2 N evaluations for N agents. While this exhaustive search guarantees optimal task allocation given the current policies, it becomes computationally intractable for large fleets. Figure 10 illustrates the exponential growth in computation time as fleet size increases. Each evaluation step takes approximately 0.13 s on our hardware for N = 4, resulting in negligible overhead during mission execution. However, extrapolation suggests that for N 10 , real-time execution would become challenging with the current implementation. Our method is designed for and evaluated on small to medium-sized ASV fleets, which are representative of current harbor, coastal, and lake cleanup deployments. For such fleet sizes, the computational overhead remains manageable and compatible with typical control loop frequencies. However, scaling to larger fleets would require algorithmic modifications.

4. Conclusions

This work introduces a multi-agent deep reinforcement learning framework for autonomous plastic cleanup using fleets of ASVs. The proposed system formulates the problem as a POMG, decoupling the mission into two complementary tasks: exploration, to map and locate waste, and cleaning, to collect it efficiently. Both tasks share navigation capabilities but involve conflicting objectives, motivating a multi-objective optimization approach. A shared two-headed Deep Q-Network, with one head dedicated to exploration and the other to cleaning, allows all agents to benefit from parameter sharing and experience aggregation.
To enhance adaptive decision-making, we introduce the Reward-Greedy Policy Selection (RGPS) mechanism, inspired by the principles of adaptive submodularity [23,24]. RGPS evaluates all feasible task allocations across the fleet at each time step and greedily selects the configuration that maximizes a scalarized reward. This explicit optimization accounts for interdependencies between agents’ actions and enables per-agent, per-timestep task allocation without additional learning or training. By leveraging separate policies for exploration and cleaning, RGPS supports coordinated yet specialized behavior, making it well suited to the dynamic and stochastic nature of aquatic environments, particularly in small to medium-sized ASV fleets.
The proposed approach generates a well-distributed Pareto front, revealing diverse strategies across the exploration–cleaning spectrum. Compared to the Fixed-Phase Pareto Set (FP2S) baseline from the literature, our framework achieves a 14% improvement in hypervolume and a 300% increase in uniformity, confirming superior Pareto coverage and policy diversity. Moreover, the system exhibits adaptive behavior across structurally different environments, automatically adjusting phase durations to maximize mission efficiency. The proposed framework is not limited to simulation and is currently being integrated into a real-world mission planning, navigation, and control stack for autonomous surface vehicles. Although real-world deployment poses challenges such as environmental disturbances and communication instability, the methodology remains practically transferable, and ongoing work focuses on experimental validation on physical ASV platforms.
Future work will focus on developing more computationally efficient approaches for constructing Pareto fronts, such as Multi-Objective Evolutionary Algorithms (MOEAs), which can evolve a diverse set of trade-off solutions over successive generations. Another promising direction is to incorporate mission duration as an additional objective, thereby extending the optimization to a three-dimensional Pareto front that captures the interplay between exploration, cleaning, and time efficiency. Moreover, adaptive mechanisms capable of dynamically adjusting the exploration–cleaning balance based on real-time environmental feedback could be explored. This adaptability could be realized through Hierarchical Reinforcement Learning (HRL), where a high-level policy governs the switching between tasks, while low-level controllers handle task-specific decision-making. Additionally, while our current study assumes reliable communication, we recognize that real-world conditions can be more challenging. Future work will evaluate algorithm performance under intermittent or limited-bandwidth communication conditions. Furthermore, future work will prioritize systematic comparisons with state-of-the-art MO-MADRL approaches as reproducibility practices and benchmarking protocols in the field become more standardized.

Author Contributions

Conceptualization, D.S., S.Y.-L. and D.G.-R.; methodology, D.S., S.Y.-L. and D.G.-R.; software, D.S. and S.Y.-L.; validation, D.S., S.Y.-L. and D.G.-R.; formal analysis, D.S., S.Y.-L. and D.G.-R.; investigation, D.S.; resources, D.G.-R. and S.T.M.; writing—original draft preparation, D.S.; writing—review and editing, D.S., S.Y.-L., D.G.-R., S.T.M. and M.P.-E.; visualization, D.S.; supervision, S.Y.-L., D.G.-R., S.T.M. and M.P.-E.; project administration, D.G.-R. and S.T.M.; funding acquisition, D.G.-R. and S.T.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Spanish Ministry of Science, Innovation, and Universities MICIU/AEI/10.13039/501100011033, by FEDER and by the European Union EU under grant PID2024-158365OB-C21.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding authors.

Acknowledgments

The authors would like to acknowledge the support of the Department of Electronic Engineering at the Higher Technical School of Engineering (ETSI), University of Seville, for their technical assistance and resources provided throughout the development of this work.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
POMGPartially Observable Markov Game
ASVAutonomous Surface Vehicle
MLMarine Litter
DRLDeep Reinforcement Learning
DQLDeep Q-Learning
DQNDeep Q-Network
MADRLMulti-Agent Deep Reinforcement Learning
MDQNMulti-Agent Deep Q-Network
MT-MADRLMulti-Task Multi-Agent Deep Reinforcement Learning
CNNConvolutional Neural Network
C-DQLCensoring Deep Q-Learning
MDPMarkov Decision Process
RGPSReward Greedy Policy Selection
WSWeighted Sum
WPWeighted Power
WPOPWeighted Product of Powers
PTCPercentage of Trash Cleaned
PMVPercentage of the Map Visited
HVHypervolume
FP2SFixed-Phase Pareto Set
PDSPhase Duration Set
MOEAsMulti-Objective Evolutionary Algorithms
HRLHierarchical Reinforcement Learning

References

  1. Le, V.G.; Nguyen, H.L.; Nguyen, M.K.; Lin, C.; Hung, N.T.Q.; Khedulkar, A.P.; Hue, N.K.; Trang, P.T.T.; Mungray, A.K.; Nguyen, D.D. Marine macro-litter sources and ecological impact: A review. Environ. Chem. Lett. 2024, 22, 1257–1273. [Google Scholar] [CrossRef]
  2. Egger, M.; Booth, A.M.; Bosker, T.; Everaert, G.; Garrard, S.L.; Havas, V.; Huntley, H.S.; Koelmans, A.A.; Kvale, K.; Lebreton, L.; et al. Evaluating the environmental impact of cleaning the North Pacific Garbage Patch. Sci. Rep. 2025, 15, 16736. [Google Scholar] [CrossRef] [PubMed]
  3. Dunbabin, M.; Grinham, A.; Udy, J. An autonomous surface vehicle for water quality monitoring. In Proceedings of the Australasian Conference on Robotics and Automation (ACRA), Sydney, Australia, 2–4 December 2009; pp. 2–4. [Google Scholar]
  4. Kamarudin, N.; Mohd Nordin, I.N.A.; Misman, D.; Khamis, N.; Razif, M.; Hanim, F. Development of Water Surface Mobile Garbage Collector Robot. Alinteri J. Agric. Sci. 2021, 36, 534–540. [Google Scholar] [CrossRef]
  5. Balestrieri, E.; Daponte, P.; De Vito, L.; Lamonaca, F. Sensors and measurements for unmanned systems: An overview. Sensors 2021, 21, 1518. [Google Scholar] [CrossRef]
  6. Katsouras, G.; Dimitriou, E.; Karavoltsos, S.; Samios, S.; Sakellari, A.; Mentzafou, A.; Tsalas, N.; Scoullos, M. Use of unmanned surface vehicles (USVs) in water chemistry studies. Sensors 2024, 24, 2809. [Google Scholar] [CrossRef]
  7. Diop, D.S.; Luis, S.Y.; Esteve, M.P.; Marín, S.L.T.; Reina, D.G. Decoupling Patrolling Tasks for Water Quality Monitoring: A Multi-Agent Deep Reinforcement Learning Approach. IEEE Access 2024, 12, 75559–75576. [Google Scholar] [CrossRef]
  8. Seck, D.; Yanes, S.; Perales, M.; Gutiérrez, D.; Toral, S. Multiobjective Environmental Cleanup with Autonomous Surface Vehicle Fleets Using Multitask Multiagent Deep Reinforcement Learning. Adv. Intell. Syst. 2025, e202500434. [Google Scholar] [CrossRef]
  9. Barrionuevo, A.M.; Luis, S.Y.; Reina, D.G.; Marín, S.L.T. Optimizing Plastic Waste Collection in Water Bodies Using Heterogeneous Autonomous Surface Vehicles With Deep Reinforcement Learning. IEEE Robot. Autom. Lett. 2025, 10, 4930–4937. [Google Scholar] [CrossRef]
  10. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A.A.; Veness, J.; Bellemare, M.G.; Graves, A.; Riedmiller, M.; Fidjeland, A.K.; Ostrovski, G.; et al. Human-level control through deep reinforcement learning. Nature 2015, 518, 529–533. [Google Scholar] [CrossRef]
  11. Hessel, M.; Modayil, J.; van Hasselt, H.; Schaul, T.; Ostrovski, G.; Dabney, W.; Horgan, D.; Piot, B.; Azar, M.; Silver, D. Rainbow: Combining Improvements in Deep Reinforcement Learning. arXiv 2017, arXiv:1710.02298. [Google Scholar] [CrossRef]
  12. Casado-Pérez, A.; Yanes, S.; Toral, S.L.; Perales-Esteve, M.; Gutiérrez-Reina, D. Variational Autoencoder for the Prediction of Oil Contamination Temporal Evolution in Water Environments. Sensors 2025, 25, 1654. [Google Scholar] [CrossRef]
  13. Luis, S.Y.; Peralta, F.; Córdoba, A.T.; del Nozal, Á.R.; Marín, S.T.; Reina, D.G. An evolutionary multi-objective path planning of a fleet of ASVs for patrolling water resources. Eng. Appl. Artif. Intell. 2022, 112, 104852. [Google Scholar] [CrossRef]
  14. Yanes Luis, S.; Shutin, D.; Marchal Gómez, J.; Gutiérrez Reina, D.; Toral Marín, S. Deep Reinforcement Multiagent Learning Framework for Information Gathering with Local Gaussian Processes for Water Monitoring. Adv. Intell. Syst. 2024, 6, 2300850. [Google Scholar] [CrossRef]
  15. Liu, Q.; Szepesvári, C.; Jin, C. Sample-efficient reinforcement learning of partially observable markov games. Adv. Neural Inf. Process. Syst. 2022, 35, 18296–18308. [Google Scholar]
  16. Zhang, K.; Yang, Z.; Liu, H.; Zhang, T.; Basar, T. Fully Decentralized Multi-Agent Reinforcement Learning with Networked Agents. In Proceedings of the 35th International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; Dy, J., Krause, A., Eds.; Proceedings of Machine Learning Research. Volume 80, pp. 5872–5881. [Google Scholar]
  17. Xia, J.; Luo, Y.; Liu, Z.; Zhang, Y.; Shi, H.; Liu, Z. Cooperative multi-target hunting by unmanned surface vehicles based on multi-agent reinforcement learning. Def. Technol. 2023, 29, 80–94. [Google Scholar] [CrossRef]
  18. Liu, L.T.; Dogan, U.; Hofmann, K. Decoding multitask dqn in the world of minecraft. In Proceedings of the 13th European Workshop on Reinforcement Learning (EWRL), Barcelona, Spain, 3–4 December 2016. [Google Scholar]
  19. Mossalam, H.; Assael, Y.M.; Roijers, D.M.; Whiteson, S. Multi-Objective Deep Reinforcement Learning. arXiv 2016, arXiv:1610.02707. [Google Scholar] [CrossRef]
  20. Li, K.; Zhang, T.; Wang, R. Deep reinforcement learning for multiobjective optimization. IEEE Trans. Cybern. 2020, 51, 3103–3114. [Google Scholar] [CrossRef] [PubMed]
  21. Basaklar, T.; Gumussoy, S.; Ogras, U.Y. Pd-morl: Preference-driven multi-objective reinforcement learning algorithm. arXiv 2022, arXiv:2208.07914. [Google Scholar]
  22. Corsi, D.; Camponogara, D.; Farinelli, A. Aquatic Navigation: A Challenging Benchmark for Deep Reinforcement Learning. Reinf. Learn. J. 2025, 3, 1106–1123. [Google Scholar]
  23. Golovin, D.; Krause, A. Adaptive submodularity: Theory and applications in active learning and stochastic optimization. J. Artif. Intell. Res. 2011, 42, 427–486. [Google Scholar]
  24. Krause, A.; Golovin, D. Submodular function maximization. Tractability 2014, 3, 3. [Google Scholar]
  25. Hansen, E.A.; Bernstein, D.S.; Zilberstein, S. Dynamic programming for partially observable stochastic games. In Proceedings of the AAAI, San Jose, CA, USA, 25–29 July 2004; Volume 4, pp. 709–715. [Google Scholar]
  26. Sutton, R.S.; Barto, A.G. Introduction to Reinforcement Learning; MIT Press: Cambridge, MA, USA, 1998. [Google Scholar]
  27. Van Hasselt, H.; Guez, A.; Silver, D. Deep reinforcement learning with double q-learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; Volume 30. [Google Scholar]
  28. Wang, Z.; Schaul, T.; Hessel, M.; Hasselt, H.; Lanctot, M.; Freitas, N. Dueling Network Architectures for Deep Reinforcement Learning. In Proceedings of the 33rd International Conference on Machine Learning, New York, NY, USA, 20–22 June 2016; Balcan, M.F., Weinberger, K.Q., Eds.; Proceedings of Machine Learning Research. Volume 48, pp. 1995–2003. [Google Scholar]
  29. Gupta, J.K.; Egorov, M.; Kochenderfer, M. Cooperative multi-agent control using deep reinforcement learning. In Proceedings of the Autonomous Agents and Multiagent Systems: AAMAS 2017 Workshops, Best Papers, São Paulo, Brazil, 8–12 May 2017; Revised Selected Papers 16. Springer: New York, NY, USA, 2017; pp. 66–83. [Google Scholar]
  30. Wong, A.; Bäck, T.; Kononova, A.V.; Plaat, A. Deep multiagent reinforcement learning: Challenges and directions. Artif. Intell. Rev. 2023, 56, 5023–5056. [Google Scholar] [CrossRef]
  31. Marler, R.T.; Arora, J.S. Survey of multi-objective optimization methods for engineering. Struct. Multidiscip. Optim. 2004, 26, 369–395. [Google Scholar] [CrossRef]
  32. Athan, T.W.; Papalambros, P.Y. A note on weighted criteria methods for compromise solutions in multi-objective optimization. Eng. Optim. 1996, 27, 155–176. [Google Scholar] [CrossRef]
Figure 1. Examples of the (a) Malaga Port map in Spain (b) transformed into a matrix M [ i , j ] { 0 , 1 } and (c) an example of the matrix Y (trash distribution) at a certain step.
Figure 1. Examples of the (a) Malaga Port map in Spain (b) transformed into a matrix M [ i , j ] { 0 , 1 } and (c) an example of the matrix Y (trash distribution) at a certain step.
Sensors 26 00427 g001
Figure 2. Illustration of the perception areas Θ ( p n ) for two vehicles (A and B) in a discretized environment with a vision radius ρ = 2 . The function η ( v , P ) represents the number of agents covering each cell, while Y [ i , j ] indicates the number of trash items located within that cell after discretization.
Figure 2. Illustration of the perception areas Θ ( p n ) for two vehicles (A and B) in a discretized environment with a vision radius ρ = 2 . The function η ( v , P ) represents the number of agents covering each cell, while Y [ i , j ] indicates the number of trash items located within that cell after discretization.
Sensors 26 00427 g002
Figure 3. Diagram of the proposed algorithm. At each step, the environment provides an observation, and all possible task combinations (corresponding to different ν configurations) are simulated across the fleet. The resulting rewards are computed and scalarized according to user-defined preferences. The Reward-Greedy Policy Selection mechanism then identifies the task combination yielding the highest reward. The selected actions are executed, and the environment advances to produce the next observation.
Figure 3. Diagram of the proposed algorithm. At each step, the environment provides an observation, and all possible task combinations (corresponding to different ν configurations) are simulated across the fleet. The resulting rewards are computed and scalarized according to user-defined preferences. The Reward-Greedy Policy Selection mechanism then identifies the task combination yielding the highest reward. The selected actions are executed, and the environment advances to produce the next observation.
Sensors 26 00427 g003
Figure 4. Hyperparameter sensitivity analysis of scalarization methods. Comparison of 2D Pareto fronts obtained by varying the exponent parameters: (a) a in the Weighted Power method and (b) p in the Exponential Weighted Criterion method.
Figure 4. Hyperparameter sensitivity analysis of scalarization methods. Comparison of 2D Pareto fronts obtained by varying the exponent parameters: (a) a in the Weighted Power method and (b) p in the Exponential Weighted Criterion method.
Sensors 26 00427 g004
Figure 5. Results of the scalarization methods in Málaga Port scenario. (a) Pareto fronts obtained using each scalarization method across 100 different weight configurations, along with the resulting combined Pareto front that aggregates all non-dominated solutions. (b) Contribution of each scalarization method to the combined Pareto front.
Figure 5. Results of the scalarization methods in Málaga Port scenario. (a) Pareto fronts obtained using each scalarization method across 100 different weight configurations, along with the resulting combined Pareto front that aggregates all non-dominated solutions. (b) Contribution of each scalarization method to the combined Pareto front.
Sensors 26 00427 g005
Figure 6. Illustration of the combined pareto front and the FP2S pareto front.
Figure 6. Illustration of the combined pareto front and the FP2S pareto front.
Sensors 26 00427 g006
Figure 7. Evolution of ν (task allocation) for each of the four agents across a single representative episode under different scalarization methods and weighting preferences, including an FP2S example. Each subplot shows instantaneous task selections. (a) EWC ( w 1 = 0.991 , w 2 = 0.009 ) prioritizes cleaning, (b) WPOP ( w 1 = 0.676 , w 2 = 0.324 ) balances both objectives, (c) WS ( w 1 = 0.162 , w 2 = 0.838 ) emphasizes exploration, and (d) example of FP2S with P D S q = 0.5 , illustrating its characteristic switching pattern. All weight values correspond to Pareto-optimal solutions from the combined front.
Figure 7. Evolution of ν (task allocation) for each of the four agents across a single representative episode under different scalarization methods and weighting preferences, including an FP2S example. Each subplot shows instantaneous task selections. (a) EWC ( w 1 = 0.991 , w 2 = 0.009 ) prioritizes cleaning, (b) WPOP ( w 1 = 0.676 , w 2 = 0.324 ) balances both objectives, (c) WS ( w 1 = 0.162 , w 2 = 0.838 ) emphasizes exploration, and (d) example of FP2S with P D S q = 0.5 , illustrating its characteristic switching pattern. All weight values correspond to Pareto-optimal solutions from the combined front.
Sensors 26 00427 g007
Figure 8. Satellite view of the Alamillo Lake.
Figure 8. Satellite view of the Alamillo Lake.
Sensors 26 00427 g008
Figure 9. Results of the scalarization methods in Alamillo Lake scenario. (a) Pareto fronts obtained using each scalarization method across 100 different weight configurations, along with the resulting combined Pareto front that aggregates all non-dominated solutions. (b) Contribution of each scalarization method to the combined Pareto front.
Figure 9. Results of the scalarization methods in Alamillo Lake scenario. (a) Pareto fronts obtained using each scalarization method across 100 different weight configurations, along with the resulting combined Pareto front that aggregates all non-dominated solutions. (b) Contribution of each scalarization method to the combined Pareto front.
Sensors 26 00427 g009
Figure 10. Illustration of the exponential growth in computation time as fleet size increases.
Figure 10. Illustration of the exponential growth in computation time as fleet size increases.
Sensors 26 00427 g010
Table 1. Simulation environment, learning, and scalarization hyperparameters.
Table 1. Simulation environment, learning, and scalarization hyperparameters.
CategoryParameterValue
ScalarizationWeighted Product exponent ( α )3
Exponential Weighted Criterion (p)1
EnvironmentNumber of ASVs (N)4
Detection radius ( ρ )2 nodes
Distance budget200 units
Movement length1 node
TrainingReplay buffer size 1 × 10 6
Batch size128
Learning rate ( L R ) 1 × 10 4
Target network hard update2000 steps
ϵ -greedy interval [ 1 , 0.05 ]
ϵ threshold 50 %
Discount factor ( γ ) 0.99
Table 2. Summary of Pareto front quality metrics for each reward scalarization and the combined front.
Table 2. Summary of Pareto front quality metrics for each reward scalarization and the combined front.
ScalarizationHypervolumeM3SpacingNº of Points
WP5400.87385.04060.919321
WS5779.88755.45111.436819
WPOP5750.23805.06390.687621
EWC5706.16095.25291.087723
Combined5810.02835.49770.703540
Literature5091.75285.79912.14229
Table 3. Wilcoxon signed-rank test results between Pareto fronts for each reward configuration. Significant ( p < 0.05 ) differences are highlighted in bold.
Table 3. Wilcoxon signed-rank test results between Pareto fronts for each reward configuration. Significant ( p < 0.05 ) differences are highlighted in bold.
ComparisonCleaning (p-Value)Exploration (p-Value)
WP vs. WS0.27530.0955
WP vs. WPOP0.02390.5392
WP vs. EWC0.65780.0022
WS vs. WPOP0.29350.1956
WS vs. EWC0.33210.1956
WPOP vs. EWC0.11110.1193
Table 4. FP2S candidate Pareto-optimal policies evaluated over 200 episodes. The policy labels follow the notation Policy (Type @ PDS), where Type { BestExploration , BestCleaning , FinalTrained } policies, and PDS denotes the fixed exploration duration (e.g., 0.7 represents a mission with 70% exploration time followed by cleaning).
Table 4. FP2S candidate Pareto-optimal policies evaluated over 200 episodes. The policy labels follow the notation Policy (Type @ PDS), where Type { BestExploration , BestCleaning , FinalTrained } policies, and PDS denotes the fixed exploration duration (e.g., 0.7 represents a mission with 70% exploration time followed by cleaning).
Policy (Type @ PDS)Cleaning (%)Exploration (%)
Best Exploration @ 1.035.290.8
Best Exploration @ 0.941.389.6
Best Cleaning @ 0.941.887.5
Best Exploration @ 0.848.086.6
Best Exploration @ 0.751.384.1
Best Cleaning @ 0.753.483.4
Final Trained @ 0.753.982.7
Best Exploration @ 0.657.080.4
Best Cleaning @ 0.657.379.3
Table 5. Summary of Pareto front quality metrics for each reward scalarization method and the combined Pareto front on the Alamillo Lake environment.
Table 5. Summary of Pareto front quality metrics for each reward scalarization method and the combined Pareto front on the Alamillo Lake environment.
ScalarizationHypervolumeM3SpacingNº of Points
EWC7483.71895.11271.583512
WPOP6873.03915.58864.43238
WS7507.05334.20342.51246
WP6721.31043.49420.75459
Combined7533.90855.00351.527411
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Seck, D.; Yanes-Luis, S.; Perales-Esteve, M.; Marín, S.T.; Gutiérrez-Reina, D. Adaptive Policy Switching for Multi-Agent ASVs in Multi-Objective Aquatic Cleaning Environments. Sensors 2026, 26, 427. https://doi.org/10.3390/s26020427

AMA Style

Seck D, Yanes-Luis S, Perales-Esteve M, Marín ST, Gutiérrez-Reina D. Adaptive Policy Switching for Multi-Agent ASVs in Multi-Objective Aquatic Cleaning Environments. Sensors. 2026; 26(2):427. https://doi.org/10.3390/s26020427

Chicago/Turabian Style

Seck, Dame, Samuel Yanes-Luis, Manuel Perales-Esteve, Sergio Toral Marín, and Daniel Gutiérrez-Reina. 2026. "Adaptive Policy Switching for Multi-Agent ASVs in Multi-Objective Aquatic Cleaning Environments" Sensors 26, no. 2: 427. https://doi.org/10.3390/s26020427

APA Style

Seck, D., Yanes-Luis, S., Perales-Esteve, M., Marín, S. T., & Gutiérrez-Reina, D. (2026). Adaptive Policy Switching for Multi-Agent ASVs in Multi-Objective Aquatic Cleaning Environments. Sensors, 26(2), 427. https://doi.org/10.3390/s26020427

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop