Next Article in Journal
Multivariable Model Predictive Control of Cleanroom Pressure Cascades
Previous Article in Journal
Online Traffic Obfuscation Experimental Framework for the Smart Home Privacy Protection
Previous Article in Special Issue
The Behavior of an IoT Sensor Monitoring System Using a 5G Network and Its Challenges in 6G Networking
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Movable Wireless Sensor-Enabled Waterway Surveillance with Enhanced Coverage Using Multi-Layer Perceptron and Reinforced Learning

Department of Embedded Systems Engineering, Incheon National University, Incheon 22012, Republic of Korea
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(16), 3295; https://doi.org/10.3390/electronics14163295
Submission received: 30 June 2025 / Revised: 14 August 2025 / Accepted: 17 August 2025 / Published: 19 August 2025
(This article belongs to the Special Issue Wireless Sensor Network: Latest Advances and Prospects)

Abstract

Waterway networking environments present unique challenges due to their dynamic nature, including vessel movement, water flow, and varying water quality. These challenges render traditional static surveillance systems inadequate for effective monitoring. This study proposes a novel wireless sensor-enabled surveillance and monitoring framework tailored to waterway conditions, integrating a two-phase approach with a Movement Phase and a Deployment Phase. In the Movement Phase, a Multi-Layer Perceptron (MLP) guides sensors efficiently toward a designated target area, minimizing travel time and computational complexity. Subsequently, the Deployment Phase utilizes reinforcement learning (RL) to arrange sensors within the target area, optimizing coverage while minimizing overlap between sensing regions. By addressing the unique requirements of waterways, the proposed framework ensures both efficient sensor mobility and resource utilization. Experimental evaluations demonstrate the framework’s effectiveness in achieving high coverage and minimal overlap, with comparable performance to traditional clustering algorithms such as K-Means. The results confirm that the proposed approach achieves flexible, scalable, and computationally efficient monitoring tailored to waterway environments.

1. Introduction

Waterways are spaces that play an important role in modern society and are used for various purposes, such as ship operation, ecosystem maintenance, and water quality management. A waterway is a dynamic environment that is a mixture of physical and ecological diversity, and there are complex elements such as ships going on and off the route, activities of various aquatic organisms, and water quality that changes over time and region. These characteristics mean that waterways have more than just a passageway and are an important axis of natural ecosystems and an essential foundation for human activities. Therefore, understanding and monitoring the environment of waterways is essential in various fields such as logistics management, multipath routing, environmental protection, and water quality improvement [1,2,3,4]. In addition to spatial monitoring, recent research highlights the importance of temporal data validity, particularly through the concept of AoI (Age of Information) [5,6]. In dynamic environments like waterways, delayed or outdated environmental data can significantly reduce the effectiveness of decision-making. Incorporating AoI into surveillance design ensures that the system not only captures wide-area data but also prioritizes data freshness in real-time operations.
However, the surveillance and monitoring work of waterways presents challenges that are inherently different from existing land-based systems. The surveillance system in the land environment is designed based on a relatively fixed sensor network and stable monitoring conditions in combination with AI(Artificial Intelligence), 5G-6G communications, WSN (Wireless Sensor Networks), IoT (Internet of Things), UAVs (Unmanned Aerial Vehicles) [7,8,9,10,11,12,13]. On the other hand, the limitations of the traditional monitoring system are clearly revealed due to the ever-changing fluid environment of the waterways. For example, changes in the moving path or water quality of a ship are difficult to predict, and a fixed sensor network cannot guarantee sufficient coverage and efficient data collection. Figure 1 depicts the waterway surveillance for a specific region detected by a movable sensor. In particular, due to the nature of the waterways, a flexible system design that can efficiently cover a large monitoring area and reflect dynamic changes in real time is required [14,15].
Existing barrier-based sensor systems are designed for surveillance-oriented purposes [16,17,18,19]. However, in a waterway environment, it is important to efficiently collect environmental data over a wide area to detect intruders. To this end, a new approach is needed to monitor a wide range by minimizing overlap when deploying sensors and maximizing the sensing radius. For example, a sensor network with reduced overlap is more suitable for collecting environmental data such as water temperature or pollution level, which enables broader and more even data collection.
This study deals with the problem of using a movable sensor network for monitoring work in a waterway. Unlike the existing static sensor arrangement, in this study, sensors are designed to move to the target area and effectively perform sensing in the area. To this end, this study proposes a system composed of the following two steps. First, the moving step of the sensors to the target area, and second, the arranging step of the sensors being effectively arranged within the target area to maximize sensing efficiency. In the first step, the sensors are learned to move efficiently to the target area using MLP. This reduces the computational complexity of the movement path extraction process and allows a simple movement task to be performed quickly. In the second step, sensors find the optimal arrangement within the target area using RL. In particular, it aims to minimize the overlap of sensing areas between sensors and maximize the sensing radius by reflecting the characteristics of the water channel environment. Unlike conventional security-oriented surveillance, it is based on design goals for more effectively collecting environmental data, such as water temperature and pollution level.
The main contributions of this study are summarized as follows.
  • First, we design a movable sensor-assisted waterway surveillance model with enhanced coverage using a Multi-Layer Perceptron and Reinforced Learning. Also, we present the system overview and settings, as well as make a formal problem definition for a dynamic waterway environment.
  • To resolve the problem, we propose a scheme consisting of Movement Phase and Deployment Phase which provides enhanced coverage and overlapped area minimization. Then, the performances of the proposed scheme are evaluated based on numerical results through extensive simulations with various scenarios and factors, including loss, ratio, train time, traveled distance, and overlapped area.
The remainder of this paper is organized as follows. Section 2 introduces the proposed system with overview, settings and problem definition. Section 3 describes the proposed schemes consisting of Movement Phase and Deployment Phase. Then, in Section 4, the performances of the proposed algorithms are evaluated with detailed discussions through extensive simulations. Finally, in Section 5, we conclude the paper.

2. Proposed System

In this section, we explain the system overview, settings, and problem definitions that are used in the proposed waterway surveillance model.

2.1. System Overview

A waterway is an environment with dynamic characteristics and is used for various purposes such as ship movement, ecosystem activities, and environmental monitoring. These characteristics are essentially different from the environment in which the existing monitoring and monitoring system is designed. While the existing monitoring system is designed to respond to dynamic intruders for a relatively fixed environment, the waterway environment is constantly changing due to dynamic factors such as water flow, ship movement, and water quality changes.
The proposed system consists largely of two stages: the Movement Phase and the Deployment Phase. Those stages work complementarily for efficient operation of the sensor network, with specialized techniques utilized at each stage. In the Movement Phase, the sensor is moved to the target region using an MLP. This step serves to learn and predict the movement direction so that the sensor may reach the target region from the current position. The goal is to effectively move the sensor to the target region while reducing computational complexity. In the Deployment Phase, RL is used to efficiently arrange sensors within the target area. In a waterway environment, the main task is to minimize the overlap of sensing areas between sensors and to maximize the overall coverage. In the Deployment Phase, to achieve this goal, the RL model is trained to optimize the sensor placement. Figure 2 and Figure 3 briefly present an overview of the Movement Phase and Deployment Phase.

2.2. System Setting

  • All simulations were implemented in Python 3.8.
  • The MLP and RL models used in the proposed framework were implemented using PyTorch 2.0.
  • The size of the waterway environment consists of a 2D environment of 100 (m) × 100 (m).
  • The size of the target area, where sensors are intensively conducting sensing operations, is 10 (m) × 10 (m).
  • The sensor used in the proposed system is a movable sensor in the waterway.
  • In a waterway, the movable sensor can only move in four directions: up, down, left, and right, based on the environmental 2D coordinates.

2.3. Problem Definition

This study addresses the problem of achieving efficient surveillance and monitoring in dynamic waterway environments by utilizing a mobile sensor network. Unlike traditional static environments, waterways are characterized by continuous changes due to factors such as water flow, vessel movement, and environmental variability. These dynamic characteristics require sensors to not only respond to moving targets but also monitor static or semi-static phenomena across a constantly shifting landscape. To tackle these challenges, this study proposes a system that ensures efficient sensor movement and deployment within a designated target region.
The problem is divided into two phases: the Movement Phase, which focuses on guiding sensors from random initial positions to the target region, and the Deployment Phase, where sensors are strategically positioned within the target region to optimize sensing coverage while minimizing overlap. The two steps serve complementary roles and aim to propose an effective sensor network that provides computationally efficient movement within the dynamic characteristics of the waterway environment and less overlap of sensing radii in the waterway.

3. Proposed Schemes

The proposed system in this study is divided into two key phases: the Movement Phase and the Deployment Phase. The Movement Phase leverages an MLP to navigate sensors from their initial random positions to a designated target region, while the Deployment Phase employs RL to optimize sensor placement within the target region by maximizing coverage and minimizing overlap.

3.1. Movement Phase with MLP

The Movement Phase serves as the first step in the proposed system, aiming to guide sensors from their initial random positions to a predefined target region. This phase is essential to ensure that sensors are effectively positioned for subsequent deployment while maintaining computational efficiency. Due to the relatively straightforward nature of the movement problem compared with deployment, this phase adopts a Multi-Layer Perceptron (MLP) as a lightweight and fast machine learning model. Figure 4 shows the MLP overview, which is applied to the Movement Phase. The MLP model, a type of feedforward neural network, is trained to map the current sensor position (x,y) to the optimal movement direction (up, down, left, or right) that minimizes the distance to the target region. The input to the model comprises the current coordinates of the sensor, while the output is a probability distribution over the four possible directions. Each probability represents the likelihood that a specific direction will reduce the sensor’s distance to the target region. The model employs a Softmax function to normalize these probabilities and selects the direction with the highest probability, thereby ensuring that the sensor moves closer to the target region with every step.
The proposed architecture with MLP consists of multiple fully connected layers with ReLU activation functions as follows:
h 1 = ReLU ( W 1 · x + b 1 )
h 2 = ReLU ( W 2 · h 1 + b 2 )
h 3 = ReLU ( W 3 · h 2 + b 3 )
output = Softmax ( W 4 · h 3 + b 4 )
where x represents the input coordinates ( x , y ), W 1 , W 2 , W 3 , W 4 are the weights, and b 1 , b 2 , b 3 , b 4 are the biases of each layer. The hidden layers increase and then decrease the number of neurons, enabling efficient feature extraction and decision-making. The output layer produces probabilities for the four movement directions.
L classification = i = 1 N y i log ( y ^ i )
L distance = 0 if d current d next > 0 ( d current d next ) 2 otherwise
L total = α · L classification + β · L distance
The training process of the MLP incorporates two loss functions to achieve optimal movement. The first is the classification loss, where Equation (5), based on cross-entropy, ensures that the predicted movement direction aligns with the optimal direction. Here, y i denotes the ground truth label (one-hot encoded) for the i-th training sample, and y ^ i represents the predicted probability for each possible direction.
The second is the distance-based loss where the Equation (6) encourages the sensor to minimize its Euclidean distance to the target region ( d current ) at each step. The term d next indicates the predicted distance after the next move, and the loss penalizes cases where this value is not smaller than d current .
By combining these two losses as in Equation (7), with α and β representing the weighting factors for classification and distance losses, respectively, the MLP learns to navigate sensors efficiently toward the target region.
This combination leverages the complementary strengths of each loss. The classification loss enhances directional prediction accuracy, and the distance-based loss promotes efficient physical movement toward the target region. Together, they are expected to improve both convergence stability and the model’s ability to generalize across diverse initial positions and dynamic environmental conditions.
The MLP-based Movement Phase provides a computationally efficient and accurate solution for sensor navigation, even in large-scale waterway environments. This phase sets the foundation for the subsequent Deployment Phase, where sensor placement is optimized to maximize coverage and minimize overlap within the target region. Algorithms 1 and 2 describes the procedures in detail.
Algorithm 1 MLP Functions
1:
Function InitializeWeights( i n p u t _ d i m , o u t p u t _ d i m ):
2:
    Return random weights for given dimensions
3:
Function InitializeBias( d i m ):
4:
    Return zeros for given dimension
5:
Function ReLU(x):
6:
    Return max ( 0 , x )
7:
Function Softmax(x):
8:
    e x p _ x exp ( x max ( x ) )
9:
    Return e x p _ x / ( e x p _ x )
10:
Function EncodeInput( p o s i t i o n ):
11:
    Return position as the i n p u t _ v e c t o r
12:
Function SelectMove( p r o b a b i l i t i e s ):
13:
    Return the i n d e x of the sample probabilistically
14:
Function UpdatePosition( c u r r e n t _ p o s i t i o n , m o v e ):
15:
    m o v e _ m a p p i n g
16:
         { 0 : ( 0 , 1 ) , 1 : ( 0 , 1 ) , 2 : ( 1 , 0 ) , 3 : ( 1 , 0 ) }
17:
    ( d x , d y )
18:
         m o v e _ m a p p i n g [ m o v e ]
19:
    Return ( c u r r e n t _ p o s i t i o n . x + d x ,
20:
                                    c u r r e n t _ p o s i t i o n . y + d y )
21:
Function CalculateDistanceToRegion( p o s i t i o n ):
22:
    t a r g e t _ r e g i o n _ x [ x _ m i n , x _ m a x ]
23:
    t a r g e t _ r e g i o n _ y [ y _ m i n , y _ m a x ]
24:
    If  p o s i t i o n . x < x _ m i n
25:
         x _ d i s t a n c e x _ m i n p o s i t i o n . x
26:
    Else if  p o s i t i o n . x > x _ m a x
27:
         x _ d i s t a n c e p o s i t i o n . x x _ m a x
28:
    Else
29:
         x _ d i s t a n c e 0
30:
    If p o s i t i o n . y < y _ m i n
31:
         y _ d i s t a n c e y _ m i n p o s i t i o n . y
32:
    Else if  p o s i t i o n . y > y _ m a x
33:
         y _ d i s t a n c e p o s i t i o n . y y _ m a x
34:
    Else
35:
         y _ d i s t a n c e 0
36:
    Return x _ d i s t a n c e + y _ d i s t a n c e
Algorithm 2 MLP and Movement Phase with MLP
1:
Class MLP:
2:
    Initialize weights ( W 1 , W 2 , W 3 , W 4 )
3:
                     and biases ( b 1 , b 2 , b 3 , b 4 )
4:
Function forward(x):
5:
      h 1 ReLU ( W 1 × x + b 1 )
6:
      h 2 ReLU ( W 2 × h 1 + b 2 )
7:
      h 3 ReLU ( W 3 × h 2 + b 3 )
8:
      o u t p u t Softmax ( W 4 × h 3 + b 4 )
9:
      Return o u t p u t
10:
Function MovementPhase( c u r r e n t _ p o s i t i o n ,
11:
                                         m l p _ m o d e l ):
12:
    i n p u t _ v e c t o r
13:
         EncodeInput ( c u r r e n t _ p o s i t i o n )
14:
    m o v e m e n t _ p r o b a b i l i t i e s
15:
         mlp_model.forward ( i n p u t _ v e c t o r )
16:
    n e x t _ m o v e
17:
         SelectMove ( m o v e m e n t _ p r o b a b i l i t i e s )
18:
    i n i t i a l _ d i s t a n c e
19:
         CalculateDistanceToRegion ( c u r r e n t _ p o s i t i o n )
20:
    n e x t _ p o s i t i o n
21:
         UpdatePosition ( c u r r e n t _ p o s i t i o n , n e x t _ m o v e )
22:
    c u r r e n t _ d i s t a n c e
23:
         CalculateDistanceToRegion ( n e x t _ p o s i t i o n )
24:
    d i f f _ i n i t _ c u r
25:
         ← i n i t i a l _ d i s t a n c e - c u r r e n t _ d i s t a n c e
26:
    t a r g e t _ d i s t a n c e s
27:
         CalculateDistanceToRegion        ( UpdatePosition ( c u r r e n t _ p o s i t i o n , m o v e )
28:
                        for  m o v e  in  [ 0 , 1 , 2 , 3 ] )
29:
    t a r g e t _ d i r e c t i o n
30:
         IndexOfMinimum ( t a r g e t _ d i s t a n c e s )
31:
    l o s s          CalculateLoss ( d i f f _ i n i t _ c u r ,
32:
              m o v e m e n t _ p r o b a b i l i t i e s , t a r g e t _ d i r e c t i o n )
33:
    Backpropagate loss and update model
34:
    Return n e x t _ m o v e

3.2. Deployment Phase with RL

The Deployment Phase is the second step in the proposed system, designed to position sensors optimally within the target region after their arrival. This phase aims to maximize the overall sensing coverage while minimizing overlap between the sensing ranges of different sensors. Given the dynamic and constrained nature of the waterway environment, reinforcement learning (RL) is employed to solve this optimization problem. The RL framework defines three key elements: state, action, and reward. The state represents the current positions of the sensors and the coverage map, reflecting the areas already monitored within the target region. The action corresponds to adjustments in the sensors’ positions, allowing them to explore new configurations within their movement constraints. The reward function is carefully designed to evaluate the effectiveness of sensor placements. It incorporates three components: total coverage, overlap penalty, and boundary penalty. The total reward can be expressed as
Coverage = i = 1 M coverage_map [ i ]
Overlap Penalty = i = 1 N j = i + 1 N max ( 0 , r i + r j d i j )
Boundary Penalty = i = 1 N ( max ( 0 , x i x max ) + max ( 0 , x min x i ) + max ( 0 , y i y max ) + max ( 0 , y min y i ) )
Reward = Coverage ( Overlap Penalty + Boundary Penalty )
Equation (8) represents the total coverage, calculated as the sum of all covered cells coverage _ map [ i ] , where M is the total number of cells in the coverage grid and each element coverage _ map [ i ] is 1 if covered and 0 otherwise.
Equation (9) accounts for the overlap penalty by summing the penalty for overlapping sensing ranges among N sensors, where r i and r j denote the sensing radii of sensors i and j, and d i j is the Euclidean distance between these sensors.
Equation (10) defines the boundary penalty, which penalizes sensors placed outside the defined target boundaries, with x max , x min , y max , y min indicating the limits of the operational area.
These three components are combined in Equation (11), where the coverage area is adjusted by subtracting the two penalty values. The weighting of these penalties is critical to maintaining a balance between maximizing coverage and preventing undesirable configurations such as excessive overlap or boundary violations. Through this formulation, the final reward reflects both the quantity of covered area and the quality of the deployment layout.
The RL model implemented in this phase utilizes the Proximal Policy Optimization (PPO) algorithm [20], which balances exploration and exploitation during training. PPO is a reinforcement learning algorithm that incorporates a policy network, responsible for determining actions, and a value network, which estimates future rewards for given states. In the Deployment Phase, the policy network defined as part of the PPO model takes the state as input and outputs actions, allowing iterative adjustments to sensor positions. Additionally, the PPO algorithm uses the value network to guide the learning process effectively and maintain stability. The environment is initialized with key parameters, including map dimensions, the number of sensors, and the sensing radius. The system updates the coverage map dynamically as sensors are repositioned, ensuring that sensor placements reflect the latest configuration. The step function evaluates the reward for each configuration by calculating the total coverage, overlap penalty, and boundary penalty, optimizing sensor placement iteratively based on these metrics. By continuously improving the placement of sensors, the RL model learns strategies that maximize sensing coverage and minimize redundancy. The DeploymentPhaseWithRL function integrates the trained PPO model with the environment, visualizing the final sensor placements and coverage map for evaluation.
Overall, the RL-based Deployment Phase ensures efficient utilization of sensors within the target region, addressing the dynamic and complex challenges of waterway environments. This phase complements the preceding Movement Phase by optimizing sensor placement for effective surveillance and monitoring. Algorithms 3 and 4 describes the procedures of RL functions and RL-based deployment in detail.
Algorithm 3 RL Functions
1:
Function InitializeEnvironment():
2:
    m a p _ d i m e n s i o n s , n u m _ s e n s o r s , s e n s i n g _ r a d i u s
3:
         ← defined
4:
    a c t i o n _ s p a c e , o b s e r v a t i o n _ s p a c e
5:
         ← set
6:
    Return i n i t i a l i z e d   e n v i r o n m e n t   s t a t e
7:
Function Reset():
8:
    s e n s o r _ c o o r d i n a t e s
9:
         ← r a n d o m l y _ i n i t i a l i z e d _ w i t h i n _ m a p _ b o u n d s
10:
    c o v e r a g e _ m a p
11:
         ← i n i t i a l i z e d a s e m p t y
12:
    c o v e r a g e _ m a p
13:
         ← u p d a t e d w i t h i n i t i a l s e n s o r p l a c e m e n t
14:
    Return i n i t i a l   o b s e r v a t i o n
15:
Function Step( a c t i o n ):
16:
    r e w a r d ← 0
17:
    d o n e ← False
18:
    For each s e n s o r
19:
         c o o r d i n a t e s
20:
              ← updated using a c t i o n
21:
         c o o r d i n a t e s
22:
              ← clipped to remain within map bounds
23:
         c o v e r a g e _ m a p
24:
              ← updated with n e w   s e n s o r   c o o r d i n a t e s
25:
    Compute reward:
26:
         t o t a l _ c o v e r a g e
27:
              ← sum of all covered cells in coverage map
28:
         o v e r l a p _ p e n a l t y
29:
             ← calculated for overlapping sensors
30:
         b o u n d a r y _ p e n a l t y
31:
             ← calculated for sensors out of bounds
32:
         r e w a r d
33:
              ← t o t a l _ c o v e r a g e -
34:
                ( o v e r l a p _ p e n a l t y + b o u n d a r y _ p e n a l t y )
35:
    Return:
36:
         o b s e r v a t i o n
37:
              ← s e n s o r _ c o o r d i n a t e s , c o v e r a g e _ m a p
38:
         r e w a r d t o t a l _ c o v e r a g e - p e n a l t i e s
39:
         d o n e ← True
40:
         i n f o ← auxiliary information
41:
Function _update_coverage( c o o r d ):
42:
    For each cell within the s e n s o r   r a d i u s
43:
         cell ← marked as covered in c o v e r a g e _ m a p
44:
Function _compute_reward( n e w _ c o o r d s ):
45:
    t o t a l _ c o v e r a g e
46:
              ← calculated from coverage map
47:
    o v e r l a p _ p e n a l t y
48:
              ← computed for overlapping sensors
49:
    b o u n d a r y _ p e n a l t y
50:
              ← computed for out-of-bound sensors
51:
    r e w a r d
52:
              ← t o t a l _ c o v e r a g e  −
53:
                ( o v e r l a p _ p e n a l t y + b o u n d a r y _ p e n a l t y )
54:
    Return r e w a r d
Algorithm 4 RL and Deployment Phase with RL
  • Class RL:
  •     Initialize:
  •       Define policy network:
  •          I n p u t   l a y e r   s i z e
  •               ← observation space dimension
  •          H i d d e n   l a y e r s
  •               ← a sequence of fully connected layers
  •                    with R e L U activation
  •          O u t p u t   l a y e r   s i z e ← action space dimension
  •          O u t p u t   a c t i v a t i o n S o f t m a x
  •       Define value network:
  •          I n p u t   l a y e r   s i z e ← observation space dimension
  •          H i d d e n   l a y e r s
  •               ← a sequence of fully connected layers
  •                       with R e L U activation
  •          O u t p u t   l a y e r   s i z e ← estimated state value
  •          O u t p u t   a c t i v a t i o n L i n e a r
  •       P P O   m o d e l ← initialized with
  •                           p o l i c y and v a l u e   n e t w o r k s
  •       l e a r n i n g _ r a t e ← set
  •       g a m m a ← set
  • Function train( e n v i r o n m e n t ):
  •     m o d e l ← trained on the provided e n v i r o n m e n t
  •     Return t r a i n e d _ m o d e l
  • Function DeploymentPhaseWithRL( e n v i r o n m e n t ,
  •                   t r a i n e d _ m o d e l , i n i t i a l _ c o o r d i n a t e s ):
  •     i n i t i a l _ s e n s o r _ c o o r d i n a t e s ← set in e n v i r o n m e n t
  •     p r e d i c t e d _ a c t i o n s t r a i n e d _ m o d e l .predict()
  •     s e n s o r _ c o o r d i n a t e s               ← updated based on p r e d i c t e d _ a c t i o n s
  •     Visualize s e n s o r placements and c o v e r a g e
  •     Return f i n a l   c o v e r a g e   m a p , s e n s o r   p o s i t i o n s

4. Experimental Evaluations

This section evaluates the performance of the proposed movement and deployment stages through detailed numerical experiments. Experimental analysis evaluates the effectiveness of the proposed method by focusing on indicators such as Find Ratio, coverage overlap area size, and learning time, which is the percentage of accurate movement to a specific area. The experimental base environment is configured to be simulated in an environment tailored to the system setting of the Proposed Schemes section, and both stages use parameters that match each algorithm. The Movement Phase evaluates the node’s ability to reach its target location in duplicate with minimal energy consumption, while the Deployment Phase focuses on optimally deploying sensors to maximize coverage and minimize redundancy. To ensure robustness and statistical validity, results are derived based on experiments.

4.1. Movement Phase Evaluation

The Movement Phase is a step in evaluating the process of moving the sensor to the target area in an experimental environment. This step analyzes the sensor’s ability to efficiently reach the target area in its initial position. Therefore, the MLP model is used to determine the direction of movement of each sensor. MLP predicts the optimal direction of movement based on the current coordinate (x, y) of the sensor in the 100 × 100 simulation environment, and selects the direction to maximize the chance of successful arrival in the target area as represented in Figure 5.
In this experiment, we analyzed the performance of the Movement Phase in various initial sensor deployments and target area settings. Key performance indicators are learning time, distance traveled, and Find Ratio, which means the percentage of sensors that reached the target area. We also evaluate whether the proposed loss function was applied properly in the Movement Phase. Identify the efficiency of the proposed migration strategy and MLP-based approach by minimizing losses and optimizing sensor movement. Comprehensively analyze these metrics to verify that the Movement Phase has effectively achieved its intended goals.

4.1.1. Loss

The convergence of the loss function has a direct correlation with the performance of model learning, indicating that the model is gradually learning the optimal pattern. The Movement Phase learning combined Classification Loss with Distance-Based Loss to guide you to learn how to efficiently move to the target area.
For Figure 6, the total loss showed a high value in the early stages of learning, but gradually decreased and showed a tendency to converge stably as learning progressed. This means that the model is gradually learning the best path to the target area. In particular, Classification Loss played an important role in increasing predictive accuracy in the direction of travel, and Distance Loss helped sensors reduce distance from the target area. The combination of these loss functions complements the model to allow simultaneous learning of the optimization of travel paths and access to target areas. Loss convergence results show that the model returns an efficient navigation path by gradually reducing uncertainty during the movement.

4.1.2. Find Ratio and Train Time

Find Ratio shows the percentage of sensors that have successfully accessed the target area within 3000 steps based on a total of 1000 coordinate generations. The performance metrics in Figure 7 are calculated as the average over 100 independent simulation runs for each setting. This is used as an important indicator of how much MLP has generalized learning in the Movement Phase. A higher Find Ratio means that the model has learned to move accurately from initial locations to the target area.
We set the data to 100, 200, 300, 400, and 500 pieces, respectively, and analyzed the performance changes depending on the number of data points. The number of epochs gradually increased from 100 to 500, and the learning rate was set to 0.001. Analysis of the model’s performance in these simulation environments has shown that Find Ratio tends to increase as the number of data points and epochs increases. In particular, with 500 pieces of data, Find Ratio has increased rapidly since the beginning of the epoch, showing high performance. This shows that the more data is used for learning, the better MLP can learn different patterns and generalize the path to the target area. Figure 7 presents the measurements of Find Ratio and train time according to different epochs. For example, when learning with 500 pieces of data, since epoch 200, the Find Ratio remained close to 100%, showing excellent performance. Because the coordinates used in the learning were randomly extracted, the performance might not be constant when the deflected coordinates were extracted. You can see this in a graph, but in some cases, a small number of data (e.g., 400) showed a higher Find Ratio than using more data (e.g., 500). This indicates that random coordinate extraction can overfit a particular pattern or that some significant paths were not learned. Also, even if the initial value of Find Ratio is low, we could see how the Find Ratio gradually improves as the epoch increases. This shows that the MLP model has learned more accurately the direction of gradually moving to the target area through the learning process. However, it should be taken into account that the rate of improvement and final value of the Find Ratio may vary depending on the number of data points and the diversity of initial coordinates. In summary, through Find Ratio analysis, we can check that the performance of the MLP model in the Movement Phase was greatly influenced by the number of data and the learning epoch. Therefore, it is important to maintain a balance between the composition of learning data and the coordinate extraction process, which can further improve the generalization performance of the MLP model.

4.1.3. Traveled Distance

In order to evaluate the performance of the Movement Phase, Manhattan Distance was introduced as a comparison between how to measure the distance of the Movement Phase and how to evaluate it. All reported values represent the average results obtained from 100 independent simulation runs under the same conditions. First, the distance of the Movement Phase is calculated by measuring the total number of movements before the sensor reaches the target area along the direction predicted by the model. This is used as an important indicator of whether the model moves to the target area efficiently. Manhattan Distance, which was used for comparison, is the shortest distance calculation method that reflects the assumption that only four directions (top, bottom, left, right) can be moved. That is, it is a simple and effective criterion for evaluating realistic movement constraints, calculated as the sum of horizontal and vertical travel distances between two points.
As shown in Table 1, Table 2 and Table 3, the distance traveled by the route obtained in the Movement Phase showed a difference of up to 5 percent compared with the shortest path in Manhattan under the same conditions. This indicates that the model has a very efficient path-finding capability.
While the proposed MLP-based approach does not explicitly guarantee the absolute shortest path as calculated by graph-based algorithms, it is designed to prioritize computational efficiency and scalability. Conventional graph-based methods often require exhaustive searches or iterative path recalculations, which can substantially increase computational overhead and hinder real-time applicability in dynamic environments. In contrast, the MLP model generates movement decisions with constant time complexity O(1) after training, enabling rapid inference regardless of the size of the environment. Although the measured travel distances may deviate by up to 5% from the corresponding Manhattan Distances, the proposed model consistently maintains a high Find Ratio. This observation indicates that such minor deviations in path length do not compromise the model’s ability to reach the target area effectively. Consequently, the MLP-based approach achieves a practical balance between path efficiency, decision-making speed, and adaptability, making it suitable for real-world deployment scenarios.

4.2. Deployment Phase Evaluation

The Deployment Phase evaluates the optimal placement of sensors that have successfully reached the target area during the Movement Phase. This phase focuses on maximizing the coverage of the 10 × 10 m target area, where the x and y axes represent the coordinate positions within this area, while minimizing the overlap between the sensor’s sensing ranges as shown in Figure 8. By achieving this balance, the Deployment Phase ensures efficient sensor deployment and optimal resource utilization.
The performance of the Deployment Phase was analyzed under various sensor deployments and target area configurations. The key performance indicators used were coverage ratio and overlap ratio. The coverage ratio measures the proportion of the target area covered by the sensors, while the overlap ratio quantifies unnecessary coverage caused by overlapping sensing ranges between sensors. These metrics provide critical insights into the effectiveness of the proposed deployment strategy. We also evaluate whether the reinforcement learning-based optimization algorithm, specifically Proximal Policy Optimization (PPO), was successfully applied during the Deployment Phase. The PPO algorithm incorporates a loss function that penalizes overlap to minimize redundancy while simultaneously optimizing for uncovered regions to ensure maximum coverage. By analyzing these metrics comprehensively, it can be verified that the Deployment Phase effectively achieved its objectives of maximizing coverage and minimizing overlaps. For example, the sensing redundancy rate was determined by combining the set target area size and the number of sensors. This confirms that the proposed algorithm has effectively achieved the goal of minimizing redundancy for coverage maximization in the Deployment Phase.

4.2.1. Loss

The Deployment Phase focuses on optimizing sensor placement to maximize coverage of the target area while minimizing overlap between sensor sensing ranges. The reward function utilized during this phase incorporates three critical components: coverage, overlap penalty, and boundary penalty. The coverage term measures the proportion of the target area effectively covered by the sensors. The overlap penalty penalizes overlapping regions between sensor sensing ranges, ensuring efficient utilization of sensors and minimizing redundancy. Lastly, the boundary penalty discourages sensor placements outside the allowable area, guiding sensors to remain within the target area boundaries.
As illustrated in Figure 9, the total loss during the Deployment Phase initially starts at a high value due to significant overlap and suboptimal placements. However, as the optimization progresses, the loss decreases substantially, showcasing a stable convergence pattern. This indicates that the proposed PPO-based reinforcement learning algorithm effectively learns to balance maximizing coverage and minimizing overlap. The loss convergence highlights the successful application of the overlap penalty and boundary penalty in the reward function. These penalties ensure that the sensors are optimally distributed across the target area, covering as much area as possible while avoiding unnecessary overlap. This demonstrates that the Deployment Phase achieves its objective of efficient sensor placement. Then, the stable convergence of the loss function underscores the reliability and effectiveness of the Deployment Phase in optimizing sensor placement. By achieving a balance between coverage maximization and overlap minimization, the Deployment Phase demonstrates its capability to support robust and efficient sensor networks in the Target Area.

4.2.2. Overlap Minimization

In the Deployment Phase, minimizing overlap between sensor coverage areas is essential for maximizing total coverage within a limited target area. By reducing overlap, the sensing regions are distributed more efficiently, avoiding redundancy and improving resource utilization. To evaluate the effectiveness of the Deployment Phase, the degree of overlap between sensors is measured. The overlap is calculated using the following Equation:
Overlap Area = π r 2 × 1 d 2 r
where r is the sensing radius, and d is the Euclidean distance between two sensor centers. If d is smaller than 2 r , the sensors overlap, and this formula estimates the overlap area. By summing the overlap areas for all sensor pairs, the total overlap in the deployment is quantified.
To validate the proposed approach, it is compared with a baseline method using K-Means Clustering. K-Means clustering is commonly used to partition data points into evenly distributed groups, ensuring that sensors are positioned uniformly across the area. However, it does not explicitly address overlap minimization, which can result in suboptimal resource utilization. Now, we compare the proposed Deployment Phase with the K-Means clustering scheme to confirm that the region’s overlap rate is reduced, confirming that the maximum coverage region can be achieved.
Table 4 presents a comparison of overlap areas during the Deployment Phase using the proposed reinforcement learning-based approach and the K-Means clustering algorithm. Each value in Table 4 represents the average result from 100 simulation runs for the corresponding sensor configuration. Overall, the results indicate that both methods perform similarly in minimizing overlap. For the case of 3 sensors, both approaches achieved an overlap area of 0, demonstrating their effectiveness in avoiding redundant coverage in low-density scenarios. As the number of sensors increased to 5 and 7, some overlap was observed in both methods, with the reinforcement learning-based approach yielding slightly higher values compared with K-Means clustering. However, the differences remain relatively minor, indicating that the proposed model is capable of achieving performance levels close to K-Means clustering. The results highlight the strength of the proposed Deployment Phase model in adapting to different sensor configurations. Unlike K-Means clustering, which distributes sensors evenly without considering specific constraints, the proposed model incorporates objectives such as maximizing coverage and minimizing penalties for overlapping and boundary violations. This adaptability ensures that the model can handle diverse deployment scenarios while achieving the desired outcomes.

5. Conclusions

This study proposed a novel sensor deployment framework designed to address the challenges of dynamic waterway environments. By integrating a Movement Phase using a Multi-Layer Perceptron (MLP) and a Deployment Phase optimized via reinforcement learning (RL), the framework successfully optimized sensor placement for enhanced coverage and minimized overlap. Through comprehensive evaluations, the effectiveness of the proposed framework was validated across various scenarios. The results demonstrate that the Movement Phase enabled sensors to efficiently navigate toward the designated target area, while the Deployment Phase ensured optimal arrangement to maximize coverage while reducing redundancy. Comparisons with the K-Means clustering algorithm further highlighted the adaptability and performance of the proposed approach, achieving comparable results while incorporating additional constraints for real-world applicability. Furthermore, although the current framework does not explicitly optimize for temporal freshness, future extensions could incorporate AoI-based evaluation to quantify how rapidly and reliably environmental updates are delivered. This would provide an additional performance dimension, particularly relevant in fast-changing aquatic environments where delayed information could lead to suboptimal responses.

Author Contributions

Conceptualization, M.K.; software, M.K.; validation, M.K.; investigation, H.K.; methodology, M.K.; resources, M.K.; data curation, M.K.; writing—draft preparation, M.K. and H.K.; writing—review and editing, H.K.; visualization, M.K. and H.K.; supervision, H.K.; project administration, H.K.; funding acquisition, H.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing not applicable.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
WSNWireless Sensor Networks
IoTInternet of Things
AIArtificial Intelligece
MLPMulti-Layer Perceptron
RLReinforcement Learning
PPOProximal Policy Optimization
AoIAge of Information
UAVUnmanned Aerial Vehicle
2DTwo-Dimensional
FC LayerFully Connected Layer

References

  1. Bingol, E.; Ekici, E. mmWave Beam Selection in Dynamic Multi-Path Environments: A POMDP Approach. In Proceedings of the International Conference on Computing, Networking and Communications (ICNC), Big Island, HI, USA, 19–22 February 2024; pp. 921–926. [Google Scholar]
  2. Mohanti, S.; Roy, D.; Eisen, M.; Cavalcanti, D.; Chowdhury, K.R. L-NORM: Learning and Network Orchestration at the Edge for Robot Connectivity and Mobility in Factory Floor Environments. IEEE Trans. Mob. Comput. 2024, 23, 2898–2914. [Google Scholar] [CrossRef]
  3. Park, S.; Jung, S.; Lee, H.; Kim, J.; Kim, J. Large-Scale Water Quality Prediction Using Federated Sensing and Learning: A Case Study with Real-World Sensing Big-Data. Sensors 2024, 21, 1462. [Google Scholar] [CrossRef] [PubMed]
  4. Choi, M.; Xiang, T.; Kim, J. Intelligent Caching for Seamless High-Quality Streaming in Vehicular Networks: A Multi-Agent Reinforcement Learning Approach. Sensors 2024, 9, 3672–3686. [Google Scholar] [CrossRef]
  5. Ma, Y.; Ma, R.; Lin, Z.; Zhang, R.; Cai, Y.; Wu, W. Improving Age of Information for Covert Communication with Time-Modulated Arrays. IEEE Internet Things J. 2025, 12, 1718–1731. [Google Scholar] [CrossRef]
  6. Zhang, G.; Wei, X.; Tan, X.; Han, Z.; Zhang, G. AoI Minimization Based on Deep Reinforcement Learning and Matching Game for IoT Information Collection in SAGIN. IEEE Trans. Commun. 2025, 73, 5950–5964. [Google Scholar] [CrossRef]
  7. Khanduri, P.; Sharma, P.; Yang, H.; Hong, M.; Liu, J.; Rajawat, K.; Varshney, P.K. STEM: A Stochastic Two-Sided Momentum Algorithm Achieving Near-Optimal Sample and Communication Complexities for Federated Learning. In Proceedings of the 35th International Conference on Neural Information Processing System, Online, 6–14 December 2021; pp. 6050–6061. [Google Scholar]
  8. Feng, S.; Yin, M.; Wang, Y.; Yang, J.; Liang, Y. Towards General Function Approximation in Nonstationary Reinforcement Learning. IEEE J. Sel. Areas Inf. Theory 2024, 5, 190–206. [Google Scholar] [CrossRef]
  9. Yue, S.; Wang, G.; Shao, W.; Zhang, Z.; Lin, S.; Ren, J.; Zhang, J. CLARE: Conservative Model-Based Reward Learning for Offline Inverse Reinforcement Learning. In Proceedings of the International Conference on Learning Representations (ICLR), Kigali, Rwanda, 1–5 May 2023; pp. 1–23. [Google Scholar]
  10. Memos, V.A.; Psannis, K.E. Optimized UAV-based data collection from MWSNs. ICT Express 2023, 9, 29–33. [Google Scholar] [CrossRef]
  11. Yeh, C.; Jo, G.D.; Ko, Y.-J.; Chung, H.K. Perspectives on 6G wireless communications. ICT Express 2023, 9, 82–91. [Google Scholar] [CrossRef]
  12. Kaushik, A.; Singh, R.; Li, M.; Luo, H.; Dayarathna, S.; Senanayake, R.; An, X.; Stirling-Gallacher, R.A.; Shin, W.; Renzo, M.D. Integrated Sensing and Communications for IoT: Synergies with Key 6G Technology Enablers. IEEE Internet Things Mag. 2024, 7, 136–143. [Google Scholar] [CrossRef]
  13. Ko, K.; Byun, I.; Ahn, W.; Shin, W. High-Speed Train Positioning Using Deep Kalman Filter with 5G NR Signals. IEEE Trans. Intell. Transp. Syst. 2022, 23, 15993–16004. [Google Scholar] [CrossRef]
  14. Mao, B.; Tang, F.; Fadlullah, Z.M.; Kato, N. An Intelligent Route Computation Approach Based on Real-Time Deep Learning Strategy for Software Defined Communication Systems. IEEE Trans. Emerg. Top. Comput. 2021, 9, 1554–1565. [Google Scholar] [CrossRef]
  15. Tang, F.; Mao, B.; Fadlullah, Z.M.; Kato, N.; Akashi, O.; Inoue, T.; Mizutani, K. On Removing Routing Protocol from Future Wireless Networks: A Real-time Deep Learning Approach for Intelligent Traffic Control. IEEE Wirel. Commun. 2018, 25, 154–160. [Google Scholar] [CrossRef]
  16. Lee, S.; Lee, S.; Kim, H. Differential Security Barriers for Virtual Emotion Detection in Maritime Transportation Stations with Cooperative Mobile Robots and UAVs. IEEE Trans. Intell. Transp. Syst. 2023, 24, 2461–2471. [Google Scholar] [CrossRef]
  17. Lee, S.; Lee, S.; Choi, Y.; Ben-Othman, J.; Kim, H. GeoSS: Geographic Segmentation Security Barriers for Virtual Emotion Detection with Discriminative Priorities in Intelligent Cooperative Vehicular System. IEEE Trans. Veh. Technol. 2023, 72, 6491–6502. [Google Scholar] [CrossRef]
  18. Lee, S.; Lee, S.; Choi, Y.; Kim, H. Trustworthy Clash-free Surveillance using Virtual Emotion Detection in 6G-assisted Graded Districts. ICT Express 2023, 9, 754–760. [Google Scholar] [CrossRef]
  19. Sedjelmaci, H.; Tourki, K.; Ansari, N. Enabling 6G Security: The Synergy of Zero Trust Architecture and Artificial Intelligence. IEEE Netw. 2024, 38, 171–177. [Google Scholar] [CrossRef]
  20. Schulman, J.; Wolski, F.; Dhariwal, P.; Radford, A.; Klimov, O. Proximal Policy Optimization Algorithms. arXiv 2017, arXiv:1707.06347. [Google Scholar] [CrossRef]
Figure 1. Movable sensor in waterway to specific sensing areas.
Figure 1. Movable sensor in waterway to specific sensing areas.
Electronics 14 03295 g001
Figure 2. Movement phase: trajectory of movable sensor into the target area.
Figure 2. Movement phase: trajectory of movable sensor into the target area.
Electronics 14 03295 g002
Figure 3. Deployment phase: overlapped area minimization by movable sensor.
Figure 3. Deployment phase: overlapped area minimization by movable sensor.
Electronics 14 03295 g003
Figure 4. MLP in Movement Phase: input coordination and direction probability output.
Figure 4. MLP in Movement Phase: input coordination and direction probability output.
Electronics 14 03295 g004
Figure 5. Sensor movement path through Movement Phase to various target areas: (a) Paths of 5 randomly placed sensors navigating toward a 10 × 10 target area centered at (10, 10) during the Movement Phase. (b) Paths of 5 randomly placed sensors navigating toward a 10 × 10 target area centered at (50, 50) during the Movement Phase. (c) Paths of 5 randomly placed sensors navigating toward a 10 × 10 target area centered at (80, 80) during the Movement Phase.
Figure 5. Sensor movement path through Movement Phase to various target areas: (a) Paths of 5 randomly placed sensors navigating toward a 10 × 10 target area centered at (10, 10) during the Movement Phase. (b) Paths of 5 randomly placed sensors navigating toward a 10 × 10 target area centered at (50, 50) during the Movement Phase. (c) Paths of 5 randomly placed sensors navigating toward a 10 × 10 target area centered at (80, 80) during the Movement Phase.
Electronics 14 03295 g005
Figure 6. Loss graph for the training process with 500 data points, 500 epochs, and a learning rate of 0.001.
Figure 6. Loss graph for the training process with 500 data points, 500 epochs, and a learning rate of 0.001.
Electronics 14 03295 g006
Figure 7. Measurements of Find Ratio and train time: (a) Find Ratio by number of data: The graph represents the Find Ratio (%) over different epochs (100 to 500) for varying amounts of training data (#100, #200, #300, #400, and #500). (b) Train time by number of data: The graph shows the training time (in seconds) increasing with the number of epochs and the amount of training data (#100, #200, #300, #400, and #500).
Figure 7. Measurements of Find Ratio and train time: (a) Find Ratio by number of data: The graph represents the Find Ratio (%) over different epochs (100 to 500) for varying amounts of training data (#100, #200, #300, #400, and #500). (b) Train time by number of data: The graph shows the training time (in seconds) increasing with the number of epochs and the amount of training data (#100, #200, #300, #400, and #500).
Electronics 14 03295 g007
Figure 8. Sensor deployment through deployment path to various sensor numbers: (a) The Deployment Phase of 3 sensors in a 10 × 10 target area. Each sensor’s sensing range is shown as a blue circle. (b) The Deployment Phase of 5 sensors in a 10 × 10 target area. As the number of sensors increases, overlapping regions among sensors’ sensing ranges become more noticeable. (c) The Deployment Phase of 7 sensors in a 10 × 10 target area. With a higher sensor density, significant overlap occurs among the sensing regions, enhancing area coverage but also increasing redundancy.
Figure 8. Sensor deployment through deployment path to various sensor numbers: (a) The Deployment Phase of 3 sensors in a 10 × 10 target area. Each sensor’s sensing range is shown as a blue circle. (b) The Deployment Phase of 5 sensors in a 10 × 10 target area. As the number of sensors increases, overlapping regions among sensors’ sensing ranges become more noticeable. (c) The Deployment Phase of 7 sensors in a 10 × 10 target area. With a higher sensor density, significant overlap occurs among the sensing regions, enhancing area coverage but also increasing redundancy.
Electronics 14 03295 g008
Figure 9. Loss graph for the training process with 1024 time steps, 100 epochs, dynamic clip range (0.2 0.1), learning rate of 0.0001, and 100,000 total time steps.
Figure 9. Loss graph for the training process with 1024 time steps, 100 epochs, dynamic clip range (0.2 0.1), learning rate of 0.0001, and 100,000 total time steps.
Electronics 14 03295 g009
Table 1. Comparison of movement distances during the Movement Phase with Manhattan Distances for the 5 paths in the Figure 5a.
Table 1. Comparison of movement distances during the Movement Phase with Manhattan Distances for the 5 paths in the Figure 5a.
Path NumberDistance of Movement PhaseDistance of Manhattan
P18686
P2105103
P33535
P45958
P55350
Table 2. Comparison of movement distances during the Movement Phase with Manhattan Distances for the 5 paths in the Figure 5b.
Table 2. Comparison of movement distances during the Movement Phase with Manhattan Distances for the 5 paths in the Figure 5b.
Path NumberDistance of Movement PhaseDistance of Manhattan
P15050
P23836
P34036
P45049
P56156
Table 3. Comparison of movement distances during the Movement Phase with Manhattan Distances for the 5 paths in the Figure 5c.
Table 3. Comparison of movement distances during the Movement Phase with Manhattan Distances for the 5 paths in the Figure 5c.
Path NumberDistance of Movement PhaseDistance of Manhattan
P17774
P2112106
P34542
P43933
P52525
Table 4. Comparison of overlap area during the Deployment Phase with K-Means clustering in the Figure 8.
Table 4. Comparison of overlap area during the Deployment Phase with K-Means clustering in the Figure 8.
Sensor NumberDeployment Phase Overlap AreaK-Means Clustering Overlap Area
300
53.613.30
710.299.41
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, M.; Kim, H. Movable Wireless Sensor-Enabled Waterway Surveillance with Enhanced Coverage Using Multi-Layer Perceptron and Reinforced Learning. Electronics 2025, 14, 3295. https://doi.org/10.3390/electronics14163295

AMA Style

Kim M, Kim H. Movable Wireless Sensor-Enabled Waterway Surveillance with Enhanced Coverage Using Multi-Layer Perceptron and Reinforced Learning. Electronics. 2025; 14(16):3295. https://doi.org/10.3390/electronics14163295

Chicago/Turabian Style

Kim, Minsoo, and Hyunbum Kim. 2025. "Movable Wireless Sensor-Enabled Waterway Surveillance with Enhanced Coverage Using Multi-Layer Perceptron and Reinforced Learning" Electronics 14, no. 16: 3295. https://doi.org/10.3390/electronics14163295

APA Style

Kim, M., & Kim, H. (2025). Movable Wireless Sensor-Enabled Waterway Surveillance with Enhanced Coverage Using Multi-Layer Perceptron and Reinforced Learning. Electronics, 14(16), 3295. https://doi.org/10.3390/electronics14163295

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop