Next Article in Journal
Confidence-Calibrated Human Activity Recognition
Previous Article in Journal
Fiber Optic Sensors Based on the Faraday Effect
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Motion-Encoded Electric Charged Particles Optimization for Moving Target Search Using Unmanned Aerial Vehicles

by
Mohammed A. Alanezi
1,
Houssem R. E. H. Bouchekara
2,*,
Mohammad S. Shahriar
2,
Yusuf A. Sha’aban
2,
Muhammad S. Javaid
3 and
Mohammed Khodja
4,5
1
Department of Computer Science and Engineering Technology, University of Hafr Al Batin, Hafr Al Batin 31991, Saudi Arabia
2
Department of Electrical Engineering, University of Hafr Al Batin, Hafr Al Batin 31991, Saudi Arabia
3
Department of Electrical and Electronic Engineering, Faculty of Engineering, Imperial College London, London SW7 2AZ, UK
4
Department of Electronics, College of Engineering, Mustaqbal University, Buraidah 51452, Saudi Arabia
5
Department of Electrical Engineering, Faculty of Technology, M’sila University, M’sila 28000, Algeria
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(19), 6568; https://doi.org/10.3390/s21196568
Submission received: 16 August 2021 / Revised: 22 September 2021 / Accepted: 24 September 2021 / Published: 30 September 2021
(This article belongs to the Section Navigation and Positioning)

Abstract

:
In this paper, a new optimization algorithm called motion-encoded electric charged particles optimization (ECPO-ME) is developed to find moving targets using unmanned aerial vehicles (UAV). The algorithm is based on the combination of the ECPO (i.e., the base algorithm) with the ME mechanism. This study is directly applicable to a real-world scenario, for instance the movement of a misplaced animal can be detected and subsequently its location can be transmitted to its caretaker. Using Bayesian theory, finding the location of a moving target is formulated as an optimization problem wherein the objective function is to maximize the probability of detecting the target. In the proposed ECPO-ME algorithm, the search trajectory is encoded as a series of UAV motion paths. These paths evolve in each iteration of the ECPO-ME algorithm. The performance of the algorithm is tested for six different scenarios with different characteristics. A statistical analysis is carried out to compare the results obtained from ECPO-ME with other well-known metaheuristics, widely used for benchmarking studies. The results found show that the ECPO-ME has great potential in finding moving targets, since it outperforms the base algorithm (i.e., ECPO) by as much as 2.16%, 5.26%, 7.17%, 14.72%, 0.79% and 3.38% for the investigated scenarios, respectively.

1. Introduction

Currently, unmanned aerial vehicles (UAVs) are among the most promising research tools of interest due to their potentials for use in numerous practical applications [1,2,3]. UAVs are especially suitable for surveillance and rescue tasks. They are highly capable of working under harsh environmental conditions. Moreover, they can connect with a sensor-rich work environment, ideal for dealing with a variety of tasks [4]. In the case of searching for a lost target using UAVs, several factors are taken into consideration. One of those is the ‘golden time’, a critical period when the probability of finding the target becomes maximal [5]. This probability decreases rapidly with time, due to several factors such as the terrain, weather conditions, attenuation of the initial particulars, the target dynamics, etc.
Therefore, while formulating the problem of searching for a lost target by UAVs, it is required to find a path that can maximize the target detection probability within a given flight time. Usually, the target position information and other search conditions play crucial roles in such formulations [6,7]. Such search problems are found in the literature as probability functions where uncertainties are considered using initial assumptions. Besides, sensor models and search conditions are adequately incorporated in the problem formulation. Articles [6,8] proposed a Bayesian approach to deal with the objective functions of detection probability evaluation of UAV flight paths.
Article [9] proposed a novel path tracking mechanism for high-speed autonomous vehicles based on the combination of model predictive control (MPC) and a PID-speed controller. A lane-level multilayer map model has been proposed in [10] to ease the tracking of a UAV or any other autonomous vehicle, even in a complicated network. A self-assessment based method [11] was used to solve a cooperative search problem, where autonomous vehicles can cooperate. This method was checked on different communication structures and excellently performed in scalability, design complexity, and communication skill. The UAV-based search problem is represented as a control system problem in [12] where the detection and collection of data can work simultaneously with coverage control.
In [7,13], the stochastic Markov process was proposed to represent deterministic target dynamics, meaning that the search scenarios will not affect the process. The initial search map can be modeled as a multivariate normal distribution, where the mean and the variance are computed using the initial information of the target position. While modeling the sensor, continuous Gaussian variable [6] and binary variable [14] are both used with the binary states defined as ‘‘detected’’ and ‘‘non-detected’’.
Because of the involvement of multiple probabilistic variables, such search problems are complex in nature.
The complexity of the search problem varies between NP-hard [15] to NEXP-complete [16]. The former refers to the nondeterministic polynomial-time hardness, while the latter refers to nondeterministic exponential-time completeness. In NEXP-complete type problems, the number of solutions increases exponentially along with the flight time and search dimension. Therefore, solving such a problem by classical methods like differential calculus to find the exact solution is understandably impractical. Therefore, approximated heuristic methods are mostly used. Typical optimization algorithms used in this field include genetic algorithms (GA) [17], ant colony optimization (ACO) [14], one-step look ahead greedy search algorithm [6], k-step look ahead greedy search algorithm [7], Bayesian optimization approaches (BOA) [8], branch and bound approach [18], cross-entropy optimization (CEO) [19], gradient descend methods [20,21], limited depth search algorithms [22], neural networks [23] etc. Among the used algorithms, Refs. [8,14,19,22] presented are the algorithms capable of using more than one UAV for search purposes, thus speeding up the search process.
On the other hand, Refs. [6,7] proposed algorithms to support a specific design of the search problem. However, it is noted that most of the used methods can successfully track moving targets using the binary model of sensor detection. Recently, [24] proposed using motion-encoded particle swarm optimization (MPSO) to solve UAVs’ moving-target search problem. The paper also compared the results with some well-known metaheuristic approaches. It was 24% more efficient in detection and 4.71 times more time-efficient than the original PSO. A review paper [25] nicely gathers all the intelligent optimization algorithms that have been tried on UAV dependent swarm search applications.
The literature identifies different approaches adopted to certify the optimal search divergence, such as target dynamics, constraints, assumptions, and searching mechanisms. The complex nature of optimal search, particularly for fast-moving targets, makes optimal problem formulation and solution strategizing challenging. However, recent advancements in sensors, UAV technologies, and communications have opened up new opportunities for research in the field. The solution must be highly robust in terms of search capacity and possess the properties of a methodical optimization tool like adaptability, computational efficiency, optimality, etc.
Electric charged particle optimization (ECPO) is a new metaheuristic optimization algorithm that was proposed by Bouchekara in 2020 [26]. The algorithm is established based on the interaction phenomenon between electrically charged particles. Different strategies of interacting behavior have made this algorithm a robust one and suitable for application on diverse applications. ECPO is a promising optimization tool that has been compared with several state-of-the-art optimization algorithms. It was tested and verified with a set of 30 functions and one real-life engineering problem. This algorithm resembles an intuitive sense of the motion of an apparently stray and lost animal in the wilderness and thus is expected to give results of analogical significance. However, motion encoding feature has been added with the ECPO in this paper for the first time to improve the searching capabilities of the algorithm for UAV dynamic target search problems.
This paper proposes a motion-encoded electric charged particle optimization (ECPO-ME) algorithm to solve the problem of moving target search. Some of the main features of the paper are highlighted below:
-
The formulation of the optimization problem with a suitable objective function and the required constraints represents the targeted problem accurately.
-
The use of motion-encoding mechanism with ECPO to increase the efficacy of the algorithm. This duo has neither been tried before in the literature nor in solving any optimization problems.
-
Comparing the proposed mechanism with 10 commonly used metaheuristic optimization algorithms strengthens the logic of using it in moving target search applications. It is also compared with MPSO, used in a recently published research paper to solve a similar optimization problem.
-
The presentation of the convergence curves for all the used optimization methods in a single plot to ease the comparison of their performance.
The simulation results demonstrate the superiority of ECPO-ME over all other optimization algorithms in solving the problem of moving target search, which is crucial in ascertaining the presence and position of a misplaced animal in vast landscapes. Additionally, we prove that the motion-encoding strategy improved the efficacy of the ECPO algorithm by a considerable margin.
The remaining paper is structured as follows: Section 2 contains the problem formulation and the objective function development. Section 3 presents the optimization algorithm details to solve the addressed problem. Section 4 exhibits the simulation problem, and, finally, Section 5 draws the concluding remarks of the article.

2. Problem Formulation

The problem addressed in this work is formulated as a search problem that entails modeling of the target, sensor, and belief map, as presented in [24]. These three aspects of the problem formulation are discussed in Section 2.1, Section 2.2 and Section 2.3 respectively and the objective function is developed in Section 2.4.

2.1. Target Model

Consider the variable x Χ , which represents the unknown location of the target in the search problem. To begin the search, the probability distribution function (PDF) which could be any suitable probability distribution based on the target’s most recent information (for example the target’s most recently known position) is used to model the target’s location. For example, reference [24] discussed using a normal distribution centered on the last known location or a uniform distribution if there is no information about the target’s location. The PDF is represented using the belief map, b x 0 , which is a grid map with each cell marked with the probability of the target being in that cell. The belief map is developed by discretizing the search space into a grid of S r × S c cells with their associated probabilities such that if the target is in the search space, then x 0 S b x 0 = 1 .
The target navigation pattern can be modelled as a Markov process. In this work, we assume a conditionally deterministic target whose navigation pattern depends upon the target’s initial position, x 0 . Therefore, probability that a target moves from cell x t 1 to x t , p x t | x t 1 called the transition function is known for all cells x t S . As such, if the target’s initial position is known, then the target’s entire path will be known, which is a standard assumption in target search problems [14].

2.2. Sensor Model

The UAV is equipped with a sensor that carries out independent observations z t at each time step t . An observation is classified based on the results of a detection algorithm; if detected, z t = D t otherwise, z t = D ¯ t . However, the observation likelihood, p ( z t | χ t ) , given sensor model, captures the fact that z t = D t does not guarantee the presence of a target, due to imperfection of the sensor and detection algorithm. Hence it is quite clear that, for a target location χ t , the likelihood of no detection can be obtained as:
p D ¯ t | χ t = 1 p ( D t | χ t )
Here   p ( D t | x t t ) and p D ¯ t | x t represent the likelihood of detection and the likelihood of no detection, respectively.

2.3. Belief Map Update

Given sensor observations, Z 1 : t = z 1 ,   ,   z t and initial distribution b χ 0 , a Bayesian approach can be used to develop the belief map of the target, b χ t . As with all Bayesian approaches, prediction and update operations are required. With the help of Bayesian approaches, both the prediction and update operations can be carried out. For the prediction step, the target motion model is used to propagate the belief map over time according to the equation:
b ^ x t = x t 1 S p χ t | χ t 1 b χ t 1  
where: b χ t 1 is the previous belief map such that b χ t 1 = p χ t 1 | z 1 : t 1 i.e., the conditional proability of target being at χ t 1 given observations up to t 1 ; and the update step involves multiplying the latest conditional observation by b ^ χ t :
b x t = η p z t | χ t b ^ x t
The normalization factor η t , which measures the probability that the target is inside the search area (i.e., χ t S b χ t = 1 ), is defined as:
η t = 1 / χ t S p ( z t | χ _ t ) b ^ x t

2.4. Objective Function

Bayesian theory considerations mean that the probability that the target is not detected at time t during any observation, r t = p ( D ¯ t | z 1 : t 1 ) depends upon two factors: the latest belief map from the prediction phase (2), and the no-detection likelihood (1). This probability r t is defined as:
r t = p D ¯ t | χ t b ^ x χ t ,
which is applicable across the entire search space. From (4) and (5) r t = 1 / η t , and z t = D ¯ t represents a “no detection” event. The combined probability of failing to detect the target over the period of 1 to t R t = p D ¯ 1 : t becomes:
R t = k = 1 t r k = R t 1 r t
And the probability of detecting the target for the first time at time t is given as:
p t = k = 1 t 1 r k 1 r t = R t 1 1 r t
Therefore, the probability of detecting the target in t steps is a summation of p t over t steps known as the cumulative probability P t is given by:
P t = k = 1 t p k = p t 1 + p t
Note that the cumulative probability P t = 1 R t is different from p t . Hence for a finite search time over a time period 1 ,   ,   N ,   the goal of the search strategy is to determine a search path O = o 1 ,   ,   o N that could maximize the cumulative probability P t . Therefore, the search objective function is:
J = t = 1 N p t

3. Motion-Encoded Electric Charged Particles Optimization (ECPO-ME) Algorithm

3.1. Description

The ECPO [26], being inspired by the electric charged particles (ECPs) interactions, is a population-based heuristic algorithm. In the description of the algorithm, the following internal parameters are used:
  • nECP: the total number of ECPs,
  • MaxITER: the maximum number of iterations,
  • nECPI: the number of ECPs which are interacting with themselves in one of the three strategies,
  • naECP: the archive pool size.
The ECP is assumed to search for a better solution by consequent interaction of charged particles in the selected strategy. While interacting, the best particle attracts the comparatively worst one, whereas the worst particle repels the best one. The ECPO is explained in detail in the following sub-section.

3.2. Pseudocode

Algorithm 1 presents the pseudocode for ECPO. All the steps are elaborately described in the following sub-sections.
Algorithm 1 ECPO pseudocode
1 Inputs
ObjFunction (objective function), ProblemSize (dimension of the problem), nECP (number of ECPs), nECPI (number of ECPs in interaction), Strategy, naECP (size of the archive pool) and MaxITER (maximum number of iterations)
2
Output ECP best
3
Initialization()
4
for Iter = 1: MaxITER
5
    Selection()
6
    Interaction()
7
    BoundsCheck()
8
    Diversification()
9
       PopulationUpdate()
10
end for

3.3. Algorithm

3.3.1. Initialization

The working procedure of ECPO starts by generating nECP charged particles within the search space, like all other metaheuristic approaches. The particles are then sorted according to their fitness. To generate the charged particles, a random normal distribution procedure has been adopted.

3.3.2. Archive Pool

Along with the generated population, a separate archive of naECP is created with the best ECPs and is denoted by archiveECP. In this archive, only the best quality ECPs are stored and updated with each iteration.

3.3.3. Selection

The performance of the algorithm much depends on the successful selection operation of ECPs. In this stage, nECPI number of particles are arbitrarily selected from the generated population, sorted according to their fitness, starting from the best to the worst. The particles selected in this phase will go through the next phase- interaction.

3.3.4. Interaction

As aforementioned, not all ECPs interact, only few of them do which are specified by nECPI. In the interaction phase, the selected nECPI particles interact among themselves, in one of the predetermined strategies. To illustrate clearly, let us assume that nECPI = 3 which are sorted as best to worst and denoted as ‘ECP1′, ‘ECP2′, and ‘ECP3′. The best particle among all the particles is denoted as ECPbest. It is worth mentioning that the procedure will remain same for any other value of nECPI. Although any value for nECPI can be used, from the authors experience nECPI = 2 and nECPI = 3 give good results compared to other values.

Strategy 1

In strategy 1, only the best ECP, also known as the ECP best gets interacted with another ECP at the time. Therefore, this strategy generates two new ECPs ( ECP i new 1 and ECP i new 2 )   for the case of three interacting ECPs. The whole procedure is presented in Figure 1.
For ECP 1 :
At first, it gets affected by ECP2 and ECPbest, simultaneously, to move to ECP 1 new 1 . Then, ECP3 and ECPbest affected ECP1 to move to ECP 1 new 2 .
The required force to move ECP1 up to ECP 1 new 1 is represented as:
F = F b e s t 1 + F 21
where: F 21 represents the force of ECP2 on charge particle ECP1 whereas F best 1 is the force on ECP1 from ECPbest.
These two forces can be expressed as follows:
F b e s t 1 = β × E C P b e s t E C P 1
F 21 = β × E C P 1 E C P 2
where β is a randomly generated number through any of the distribution mechanisms, such as normal distribution or gamma distribution.
As discussed earlier, ECPbest attracts ECP1, whereas ECP2 repeals ECP1, because ECP1 is better than ECP2 and worse than ECPbest.
Therefore, the total force pushing ECP1 towards ECP 1 new 1 is given by (as demonstrated in Figure 1a):
E C P 1 n e w 1 = E C P 1 + F = E C P 1 + F b e s t 1 + F 21 = E C P 1 + β × E C P b e s t E C P 1 + β × E C P 1 E C P 2
Similarly, the required force pushing ECP1 towards ECP 1 new 2 is as follows:
E C P 1 n e w 2 = E C P 1 + F = E C P 1 + F b e s t 1 + F 31 = E C P 1 + β × E C P b e s t E C P 1 + β × E C P 1 E C P 3
It is illustrated in Figure 1b.
For ECP 2   :
ECP1 and ECPbest affects simultaneously to move to ECP 2 new 1 . Similarly, ECP2 is affected by ECP3 and ECPbest simultaneously for moving to ECP 2 new 1 . These two scenarios are illustrated in Figure 1c,d, respectively.
For ECP 3   :
This last and third particle, ECP3 is affected by ECP1 and ECPbest, simultaneously, to move to the new particle, ECP 3 new 1 . Later, ECP3 gets affected by ECP2 and ECPbest together to create ECP 3 new 2 (Figure 1e,f).
Equations (13) and (14), used for ECP1, are like the ones used for ECP2 and ECP3. They have not been rewritten here to avoid repetition.

Strategy 2

In the second strategy, the ECPbest does not interact with all of the remaining ECPs; rather, it associates with selected ones. Therefore, in the presented illustration, all the three interacting ECPs will create one new ECP, each, which is called ECP i new (where i represents the indexing of the selected ECP).
For ECP 1 :
ECP1 gets affected by ECP2 and ECP3, simultaneously, which will lead it to move towards ECP 1 new . The resulting force for this movement is given by:
F = F 21 + F 31
where: F 21 refers to the force of ECP2 for ECP1 and F 31 represents the force of ECP3 on the first charged particle, ECP1.
Therefore, the required force of pushing ECP1 towards ECP 1 new is given by (as shown in Figure 2a):
E C P 1 n e w = E C P 1 + F 1 = E C P 1 + F 21 + F 31 = E C P 1 + β × E C P 1 E C P 2 + β × E C P 1 E C P 3
For ECP 2 :
The second particle (ECP2) is affected by ECP1 and ECP3, simultaneously, to create the movement towards ECP 2 new . The resulting force for this incident is denoted by:
F = F 12 + F 32
where: F 12 refers to the force between ECP1 and ECP2 whereas F 32 represents the force between ECP3 and ECP2.
For ECP 3 :
The ECP3 gets affected by first and second particles, together, with the following force:
F = F 13 + F 23
where: F 13 is the force between ECP1 and ECP3 whereas F 23 represents the force between ECP2 and ECP3.
Figure 2b,c represents the forces to move ECP2 to ECP2new and ECP3 to ECP3new, respectively. At the same time, the equation of the force equation will be like the one presented in (16).

Strategy 3

This strategy uses Strategies 1 and 2, both in generating the new ECPs. Therefore, in the case of nECPI = 3, a total of 9 new ECPs will be available, where strategies 1 and 2 will generate 6 and 3 ECPs, respectively.
The product of the interaction phase is a new population of ECPs, which is denoted as newECP. Its size is the same as the original population size, whatever the nECPI or the interaction strategy. This means that the product of the interaction phase will generate, for instance, 35 particles if the initially generated population is 35 particles.

3.3.5. Checking the Bounds

This phase checks whether any new ECPs created in the interaction phase have been generated outside the search space. To verify this, all of the new ECPs are checked. If any particles are found outside the search space, they will be brought back within the assigned boundary.

3.3.6. Diversification

Some part of the new ECP population needs to be diversified with a specific probability, known as the probability of diversification (Pd). In this algorithm, the diversity operator collects information from newECP and archiveECP. The pseudocode of the diversification phase is presented in Algorithm 2.

3.3.7. Diversification

Some part of the new ECP population needs to be diversified with a specific probability, which is known as probability of diversification (Pd). In this algorithm, the diversity operator collects information from newECP and archiveECP. The pseudo code of the diversification phase is presented below in Algorithm 2.
Algorithm 2 Pseudocode for the diversification phase
1 For i = 1 :   newECP
2     For j = 1 :   ProblemSize
3        I f   rand < Pd
4            select a random ECP from the archive pool (k)
5            newECP i , j =   archiveECP k , j
6        End If
7 End For

3.3.8. Population Update

The new population is modified in this phase and stored in the previously created archive pool. The best nECP particles, from rank 1 to nECO, will generate the new population. It will go through an iterative procedure again, as described before.

3.3.9. Criteria for Termination

There could be several ways of terminating for the discussed algorithm. However, in this current ECPO version, the iteration continues for a specified number of times (MaxITER) and then stops. The MaxITER is an important parameter for the ECPO-ME algorithm. It is determined by trying various values and selecting the optimal one. It is obvious that increasing MaxITER will improve the results in most cases, but, at the same time, it will take more time to get the results. Inversely, decreasing MaxITER might lead to less efficient results, since the algorithm might need more iteration to converge to the global minimum. It is therefore suggested to optimally determine the MaxITER parameter.

3.3.10. Constraint Handling

The constraints in ECPO are handled by changing the constrained optimization problems to unconstrained using the penalty factor technique. In this technique, if an ECP fails to fulfill the constraint demand, a penalty factor will be imposed on the objective function. The higher the severity of the violation, the bigger will be the penalty term.

3.3.11. Motion Encoding (ME)

ECPO has yet to undergo any modification or upgrade as a new metaheuristic optimization algorithm. However, this paper has added a motion-encoding feature to the existing approach [26] to improve the search capability of ECPO. It has helped the algorithm cope with the challenging task of searching the dynamic targets within a limited time frame. Therefore, there is the need to encode the particle positions to find the global solution easily. To achieve this, a position needs to be defined as a multi-dimensional vector that can represent the search path [24].
In this paper, UAV motion is used to encode the particle positions. Each searching path is thus seen as a set of UAV motion segments instead of the traditional nodal concept. Thus, the movement of a UAV from one cell to another is tracked accordingly to run the optimization process. The whole mapping procedure of ECPO-ME is such that the particles search the motion space, not the cartesian space.

4. Application, Results, and Discussion

In this section, the performance of the proposed ECPO-ME-based search algorithm is assessed under different scenarios and compared with several other optimization algorithms. All the scenarios and algorithms have been implemented and simulated using the commercial software MATLAB. All calculations were done using a computer equipped with Inter Core i7- 6500U CPU, 2.59 GHz, 8 GB RAM, running the Windows 10 operating system.

4.1. Scenarios

Six scenarios are used, some of which are similar to the ones investigated in [24]. All the scenarios have same grid size w x = w y = 40 but different initial position (or location) of the UAV, target motion model P ( x t | x t 1 )   and belief map b x 0 . The tested scenarios are depicted in Figure 3. The probability map is color-coded (cells with higher probabilities of target presence are colored with warmer colors), the UAV’s initial position is indicated with a white circle, and the dynamics of the moving target are outlined with white arrows.
Scenario A: This scenario has two areas with higher probabilities, located near each other and moving towards the east. However, the likelihood of the upper area is slightly higher than the lower area. The UAV is situated in the middle of the search space. This scenario is depicted in Figure 3a.
Scenario B: In this scenario, there are two areas with high probabilities that are equally spaced from the initial position of the UAV along the y-axis of the search space. However, the probability of the upper area is slightly higher than that of the lower area. Both areas move towards the south-west. The UAV is located at the middle of the search space, and the corresponding scenario is depicted in Figure 3b.
Scenario C: In this scenario, picturized in Figure 3c, only one dense area moves towards the southeast. This scenario tests the adaptability of the searching algorithm. The UAV is located at the middle of the x-axis and under the target along the y-axis of the search space.
Scenario D: In this scenario, there is only one dense area, like in Scenario C. However, the target is moving towards the UAV in this scenario, which is initially located at the south-west region of the search space. This scenario is depicted in Figure 3d.
Scenario E: This scenario, depicted in Figure 3e, consists of two probability areas that are equally spaced from the initial position of the UAV along the x-axis of the search space, and the target is moving towards the north. However, the probability of the right area is slightly higher than that of the left area. The UAV is located in the middle of the search space.
Scenario F: In this scenario, two probability areas are the same distance apart from the initial position of the UAV, and the target is moving towards the north-east. However, the probability of the lower area is slightly higher than the upper area. The UAV, along with the areas, is on the south-west side of the search space. This scenario is depicted in Figure 3f.

4.2. Comparing Algorithms

As mentioned before, the proposed algorithm results are compared to ten well-known algorithms, which are described in the following subsections.
Motion-encoded particle swarm optimization (MPSO): This algorithm is an improved version of the famous PSO algorithm using the ME mechanism. PSO is inspired by the social behavior of bird flocking or fish schooling [27,28]. In the PSO a population of particles (each particle has a position and a velocity) evolves during iterations. During each iteration, the position and velocity of each particle are updated using the best solution achieved by each particle so far, and the overall best solution. The MPSO has been used in [24] to solve a similar problem to the one solved in this paper, i.e., finding a moving target using UAVs. As reported in [24] the obtained results using MPSO were better than some versions of the PSO and many other optimization algorithms. Therefore, it will be interesting to compare the results obtained using the proposed approach with those obtained using MPSO.
Most-valuable-player algorithm (MVPA): This algorithm is an optimization algorithm that optimizes the problem at hand by considering a population of players that compete against each other (individually and in teams) to win the trophy of the most valuable player (individually) and the championship (in teams) [29].
Differential evolution (DE): DE is a population-based optimization algorithm relying on a variety of evolutionary processes. A specific member of the population, representing the solution, iteratively improves itself from one generation to another. This algorithm is famous for its stability and versatility, specifically among population-based search algorithms [30]. The standard DE method does not require gradient information trivializing the condition of differentiability. After the occurrence of a randomly chosen evolutionary event, if a candidate gives a better solution, the adaptation is accepted for that candidate; otherwise, its traits remain unchanged for the next generation. The process is repeated until pre-defined termination criteria are met.
Genetic algorithm (GA): GA was first used by Holland in 1975 based on Darwin’s theory of evolution [31]. Now it is one of the most famous metaheuristics and has been used in many research fields. The objective of the GA is to maximize the payoff of candidate solutions in the population against a cost function (or a fitness function) from the problem domain using three operators: selection, recombination, and mutation.
Electrostatic discharge algorithm (ESDA): This algorithm is an optimization algorithm inspired by electrostatic discharge (ESD). In this algorithm, objects (population of objects) work in a given environment (search space). If these objects are near each other, the ESD may occur (sometimes between two objects and sometimes between three objects). If ESD occurs on a victim object couple of times, the object or some of its components is/are damaged, and it/they must be changed [32].
Biogeography-based optimization (BBO): BBO is also an evolutionary algorithm inspired by nature. As evident from the name, its equations are modeled on biogeography, which covers the distribution and growth of living beings through space and time. Three events decide the fitness of each candidate in every generation: speciation, migration, and extinction. Like DE, BBO maintains the population of candidates solutions after every biogeographic event [33]. Once the termination criterion is met, the candidate with the best objective function is considered the optimal solution.
Artificial bee colony (ABC): ABC is another nature-inspired optimization algorithm based on the foraging pattern of a bee swarm. The bees in a colony belong to either of the three groups; employed, onlookers, and scouts. The employed ones look for the food source around the hive, where the location of the food represents the optimal solution, and the amount of nectar in a food source is equivalent to the fitness value. Once all employed bees bring the food, the onlookers evaluate the quantity of food and determine the next trip of employed bees. Scouts replace a depleted source with a randomly located new food source [34]. All these metaphors are implemented mathematically and iteratively give the optimal solution in a search space.
Gravitational search algorithm (GSA): GSA models the optimization formulation using Newton’s gravitational laws and laws of motion. Many variants of GSA have been obtained by modifying its parameters, including Kbest, velocity, and position. The mass interaction under Newtonian mechanics governs the convergence of the solution. Every mass exerts a force on another mass under a fixed gravity determining the fitness function. Based on the interaction, the velocity and position of each agent are updated [35]. The process continues till the termination criteria are met. Ultimately, the solution converges to the optimal solution.
Teaching–learning-based optimization (TLBO): The TLBO algorithm is inspired by the teaching-learning process. In this algorithm, there are two phases, namely the teacher phase and the learner phase. In the first phase, the population of students improves their knowledge with the teacher who is the best solution among the population. However, in the learner phase, the students interact to improve their knowledge [36].
Black-hole (BH) algorithm: This algorithm is inspired by the black hole theory, a region of spacetime where no object, particle, or even light can escape from it due to its strong gravity. In this algorithm, the population of stars moves in space. If a star approaches the black hole (within a radius called the Schwarzschild radius), it is absorbed by the hole, and a new star is generated to keep the same number of stars over the iterations [37,38].

4.3. Results

The proposed algorithm and the comparing ones were applied to the UAV dynamic target search. Each algorithm has been run 30 times with a population of size 30 and 300 as the maximum number of iterations. The remaining parameters related to each algorithm are selected as recommended in the literature. Table 1 presents the ‘BEST’, the ‘MEAN’, the ‘MEDIAN, the ‘WORST’, the ‘Standard Deviation (SD)’ and the ‘Feasibility Ratio (FR)’ for each algorithm and each scenario. The FR is defined as the percentage of the successful runs over the total number of attempted runs. The following comments can be made from Table 1:
-
For Scenario A, the proposed ECPO-ME algorithm obtained the best results for the ‘BEST’, the ‘MEAN’ and, the ‘MEDIAN values. The TLBO algorithm obtained the highest ‘WORST’ and ‘SD’ Values.
-
For Scenario B, the TLBO obtained a slightly better value than the ECPO-ME in terms of the ‘BEST’ values. However, the ECPO-ME obtained the best results of the ‘MEAN’, the ‘MEDIAN, the ‘WORST’ and the ‘SD’ values compared to the remaining algorithms.
-
For Scenario C and Scenario D, the proposed ECPO-ME outperformed the other algorithms in terms of the statistical performance indicators used in this study. It is worth mentioning that the DE achieved equally good results to ECPO-ME in terms of the ‘BEST’ values.
-
For Scenario E, the proposed ECPO-ME showed better performance than all the remaining algorithms in terms of the ‘BEST’, the ‘MEAN’ and the ‘MEDIAN values while the DE achieved better results in terms of the ‘WORST’ and the ‘SD’ values.
-
For Scenario F, the proposed ECPO-ME achieved a better result than all the remaining algorithms in terms of the ‘BEST’, the ‘MEDIAN’ and the ‘WORST’ values, while the TLBO was better in terms of the ‘MEAN’ and ‘SD’ values.
-
All algorithms gave an FR equal to 100, which reflects that all of them could find a solution (i.e., a path) in all the runs and for all the investigated scenarios except for the ABC algorithm for Scenario D (FR = 93.33%).
-
For scenarios with high probability regions, like Scenario C and Scenario D, the likelihood of finding the target is higher because there is no need to divide or spread the chances of finding the target in other areas.
Figure 4, Figure 5, Figure 6, Figure 7, Figure 8 and Figure 9 depict the optimal search paths obtained for the tested algorithms for Scenario A, Scenario B, Scenario C, Scenario D, Scenario E, and Scenario F, respectively. In these figures, the sketched path represents the searching path. However, the belief map reflects the target at the last step. This map can be compared with the initial map, illustrated in Figure 3, to see the evolution of the target. It can be seen from these figures that, in all scenarios, the proposed ECPO-ME algorithm was capable of following the target paths and, thus, finding out the highest probability region.
Furthermore, the convergence curves of all algorithms are given in Figure 10. It can be seen from these curves that the ECPO-ME algorithm achieves the best results in almost all the investigated scenarios.

4.4. Discussion

The analysis of the obtained results leads to the conclusion that the proposed ECPO-ME algorithm has capability to perform better than many of the well-known and widely used optimization algorithms. For instance, for Scenario A, the ECPO-ME has outperformed the remaining algorithms by 13.30%, 3.74%, 1.75%, 28.20%, 8.82%, 3.02%, 11.02%, 5.00%, 0.32%, 13.54%, respectively. Another example is for Scenario B, the ECPO-ME has outperformed the remaining algorithms (except TLBO) by 5.26%, 21.99%, 6.23%, 0.61%, 22.89%, 15.25%, 5.65%, 13.14%, 46.02%, and 15.53%, respectively.
The success of the ECPO-ME lies first in the search ability of the ECPO itself and secondly, in the motion-encoding mechanism introduced to the ECPO for the first time.
It has been proven earlier that ECPO is superior in terms of its ability to search for the optimal solution, and hence ranked among the top five algorithms in doing so. Yet, the proposed enhancement, ECPO-ME outperforms its predecessor in terms of its search ability, one of the primary reasons for ECPO-ME’s superior performance, as recorded in Table 1.
For the second point, implementing the motion-encoding mechanism with the ECPO has improved its performance for finding a moving target with a UAV. The ME mechanism has prevented the algorithm from generating invalid paths during the search process. Furthermore, the ME mechanism has transformed the search procedure from a cartesian space to a motion space, which has improved the adaptation of ECPO-ME to the dynamics of the target.

5. Conclusions

In this work, a new approach based on the electric charged particles optimization algorithm and the motion-encoding mechanism is presented to search the moving targets using unmanned aerial vehicles. The ME mechanism transforms the cartesian problem into a motion-based problem. The search path is then transformed into a series of motions where the UAV is restricted to the neighbor cells of the current location cell. Extensive simulations (six different scenarios with different complexities) and comprehensive comparisons with other well-known optimization algorithms (ten algorithms) have shown that the proposed approach is highly effective and reliable in searching moving targets using UAVs. Such an approach is highly suitable to use in cattle management, where the animals can be considered the moving targets that are prone to get lost while grazing.
The challenges for future works are multiple. One axis is to investigate the use of multiple UAVs instead of a single UAV, as done in this paper. Many approaches can be adopted, like allocating each UAV to a region, or having all of them search the highest regions. Another interesting point can be the investigation of information sharing between the UAVs, which may significantly improve the search procedure. Another challenging task or research axis can be extending this study to multiple moving targets instead of just one or two targets. The mixture of both axis of research, i.e., using multiple UAVs to search for multiple targets, can lead to a large-scale optimization problem, which is another challenging task to be investigated in the future.
Furthermore, in this paper, only one objective has been considered, with the cumulative probability, P t . Including other objectives, such as fuel consumption by the UAVs or avoidance of any forbidden areas, can define future work extending that reported here. The problem in hand will then become a multi-objective one. From an application point-of-view, this research can be extended to a real-life scenario, wherein the motion of a moving animal can be detected to communicate its location to a nearby control center for further action.

Author Contributions

Conceptualization, H.R.E.H.B. and M.K.; methodology, H.R.E.H.B. and M.S.S.; software, H.R.E.H.B.; validation, H.R.E.H.B., M.S.J. and Y.A.S.; formal analysis, M.S.J. and M.S.S.; investigation, M.K.; resources, H.R.E.H.B.; writing—original draft preparation, M.A.A., H.R.E.H.B., M.S.S., M.S.J., Y.A.S. and M.K.; writing—review and editing, M.A.A., H.R.E.H.B., M.S.S., M.S.J., Y.A.S. and M.K.; visualization, Y.A.S.; supervision, H.R.E.H.B. and M.A.A.; project administration, M.A.A.; funding acquisition, M.A.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Deputyship for Research and Innovation, Ministry of Education in Saudi Arabia, grant number IFP-A-201-2-1”.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Acknowledgments

The authors extend their appreciation to the Deputyship for Research and Innovation, Ministry of Education in Saudi Arabia for funding this research work through the project No IFP-A-201-2-1.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Huo, L.; Zhu, J.; Li, Z.; Ma, M. A Hybrid Differential Symbiotic Organisms Search Algorithm for UAV Path Planning. Sensors 2021, 21, 3037. [Google Scholar] [CrossRef] [PubMed]
  2. Jayaweera, H.M.P.C.; Hanoun, S. UAV Path Planning for Reconnaissance and Look-Ahead Coverage Support for Mobile Ground Vehicles. Sensors 2021, 21, 4595. [Google Scholar] [CrossRef]
  3. Behjati, M.; Mohd Noh, A.B.; Alobaidy, H.A.H.; Zulkifley, M.A.; Nordin, R.; Abdullah, N.F. LoRa Communications as an Enabler for Internet of Drones towards Large-Scale Livestock Monitoring in Rural Farms. Sensors 2021, 21, 5044. [Google Scholar] [CrossRef]
  4. Liu, Y.; Qiu, M.; Hu, J.; Yu, H. Incentive UAV-Enabled Mobile Edge Computing Based on Microwave Power Transmission. IEEE Access 2020, 8, 28584–28593. [Google Scholar] [CrossRef]
  5. Ochoa, S.F.; Santos, R. Human-centric wireless sensor networks to improve information availability during urban search and rescue activities. Inf. Fusion 2015, 22, 71–84. [Google Scholar] [CrossRef]
  6. Bourgault, F.; Furukawa, T.; Durrant-Whyte, H.F. Optimal search for a lost target in a Bayesian world. Springer Tracts Adv. Robot. 2006, 24, 209–222. [Google Scholar] [CrossRef]
  7. Raap, M.; Meyer-Nieberg, S.; Pickl, S.; Zsifkovits, M. Aerial Vehicle Search-Path Optimization: A Novel Method for Emergency Operations. J. Optim. Theory Appl. 2017, 172, 965–983. [Google Scholar] [CrossRef]
  8. Lanillos, P.; Yañez-Zuluaga, J.; Ruz, J.J.; Besada-Portas, E. A Bayesian approach for constrained multi-agent minimum time search in uncertain dynamic domains. In Proceedings of the GECCO 2013 Genetic and Evolutionary Computation Conference, Amsterdam, The Netherlands, 6–10 July 2013; pp. 391–398. [Google Scholar]
  9. Chen, S.; Xiong, G.; Chen, H.; Negrut, D. MPC-based path tracking with PID speed control for high-speed autonomous vehicles considering time-optimal travel. J. Cent. South Univ. 2020, 27, 3702–3720. [Google Scholar] [CrossRef]
  10. Jiang, K.; Yang, D.; Liu, C.; Zhang, T.; Xiao, Z. A Flexible Multi-Layer Map Model Designed for Lane-Level Route Planning in Autonomous Vehicles. Engineering 2019, 5, 305–318. [Google Scholar] [CrossRef]
  11. Sujit, P.B.; Ghose, D. Self assessment-based decision making for multiagent cooperative search. IEEE Trans. Autom. Sci. Eng. 2011, 8, 705–719. [Google Scholar] [CrossRef]
  12. Hu, J.; Xie, L.; Lum, K.Y.; Xu, J. Multiagent information fusion and cooperative control in target search. IEEE Trans. Control Syst. Technol. 2013, 21, 1223–1235. [Google Scholar] [CrossRef]
  13. Furukawa, T.; Bourgault, F.; Lavis, B.; Durrant-Whyte, H.F. Recursive Bayesian search-and-tracking using coordinated UAVs for lost targets. In Proceedings of the IEEE International Conference on Robotics and Automation, Orlando, FL, USA, 15–19 May 2006; Volume 2006, pp. 2521–2526. [Google Scholar]
  14. Perez-Carabaza, S.; Besada-Portas, E.; Lopez-Orozco, J.A.; de la Cruz, J.M. Ant colony optimization for multi-UAV minimum time search in uncertain domains. Appl. Soft Comput. J. 2018, 62, 789–806. [Google Scholar] [CrossRef]
  15. Trummel, K.E.; Weisinger, J.R. Complexity of the optimal searcher path problem. Oper. Res. 1986, 34, 324–327. [Google Scholar] [CrossRef] [Green Version]
  16. Bernstein, D.S.; Givan, R.; Immerman, N.; Zilberstein, S. The complexity of decentralized control of Markov decision processes. Math. Oper. Res. 2002, 27, 819–840. [Google Scholar] [CrossRef]
  17. Goldberg, D.E.; Holland, J.H. Genetic Algorithms and Machine Learning. Mach. Learn. 1988, 3, 95–99. [Google Scholar] [CrossRef]
  18. Eagle, J.N.; Yee, J.R. Optimal branch-and-bound procedure for the constrained path, moving target search problem. Oper. Res. 1990, 38, 110–114. [Google Scholar] [CrossRef] [Green Version]
  19. Lanillos, P.; Besada-Portas, E.; Pajares, G.; Ruz, J.J. Minimum time search for lost targets using cross entropy optimization. In Proceedings of the IEEE International Conference on Intelligent Robots and Systems, Vilamoura, Algarve, Portugal, 7–12 October 2012; pp. 602–609. [Google Scholar]
  20. Gan, S.K.; Sukkarieh, S. Multi-UAV target search using explicit decentralized gradient-based negotiation. In Proceedings of the IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011; pp. 751–756. [Google Scholar]
  21. Mathews, G.; Durrant-Whyte, H.; Prokopenko, M. Asynchronous gradient-based optimisation for team decision making. In Proceedings of the IEEE Conference on Decision and Control, New Orleans, LA, USA, 12–14 December 2007; pp. 3145–3150. [Google Scholar]
  22. Sarmiento, A.; Murrieta-Cid, R.; Hutchinson, S. An efficient motion strategy to compute expected-time locally optimal continuous search paths in known environments. Adv. Robot. 2009, 23, 1533–1560. [Google Scholar] [CrossRef] [Green Version]
  23. Xu, R.; Tang, G.; Xie, D.; Han, L. Three-dimensional neural network tracking control of autonomous underwater vehicles with input saturation. J. Cent. South Univ. 2020, 27, 1754–1769. [Google Scholar] [CrossRef]
  24. Phung, M.D.; Ha, Q.P. Motion-encoded particle swarm optimization for moving target search using UAVs. Appl. Soft Comput. 2020, 97, 106705. [Google Scholar] [CrossRef]
  25. Qiming, Z.; Husheng, W.; Zhaowang, F. A review of intelligent optimization algorithm applied to unmanned aerial vehicle swarm search task. In Proceedings of the 2021 11th International Conference on Information Science and Technology (ICIST), Chengdu, China, 21–23 May 2021; pp. 383–393. [Google Scholar]
  26. Bouchekara, H.R.E.H. Electric Charged Particles Optimization and Its Application to the Optimal Design of a Circular Antenna Array; Springer: Dordrecht, The Netherlands, 2020; ISBN 0123456789. [Google Scholar]
  27. Eberhart, R.; Kennedy, J. A new optimizer using particle swarm theory. In Proceedings of the Sixth International Symposium on Micro Machine and Human Science, Nagoya, Japan, 4–6 October 1995; pp. 39–43. [Google Scholar]
  28. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95-International Conference on Neural Networks, Perth, WA, Australia, 27 Novembre–1 December 1995; Volume 4, pp. 1942–1948. [Google Scholar] [CrossRef]
  29. Bouchekara, H.R.E.H. Most Valuable Player Algorithm: A novel optimization algorithm inspired from sport. Oper. Res. 2017, 20, 139–195. [Google Scholar] [CrossRef]
  30. Georgioudakis, M.; Plevris, V. A Comparative Study of Differential Evolution Variants in Constrained Structural Optimization. Front. Built Environ. 2020, 6, 102. [Google Scholar] [CrossRef]
  31. Holland, J.H. Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence; University of Michigan Press: Massachusetts, MA, USA, 1975. [Google Scholar]
  32. Bouchekara, H.R.E.H. Electrostatic Discharge Algorithm (ESDA): A Novel Nature-Inspired Optimization Algorithm and its Application to Worst-Case Tolerance Analysis of An EMC Filter. IET Sci. Meas. Technol. 2019, 13, 491–499. [Google Scholar] [CrossRef]
  33. Simon, D. Biogeography-based optimization. IEEE Trans. Evol. Comput. 2008, 12, 702–713. [Google Scholar] [CrossRef] [Green Version]
  34. Karaboga, D. Artificial bee colony algorithm. Scholarpedia 2010, 5, 6915. [Google Scholar] [CrossRef]
  35. Mittal, H.; Tripathi, A.; Pandey, A.C.; Pal, R. Gravitational search algorithm: A comprehensive analysis of recent variants. Multimed. Tools Appl. 2020, 80, 7581–7608. [Google Scholar] [CrossRef]
  36. Rao, R.V.; Savsani, V.J.; Vakharia, D.P. Teaching–learning-based optimization: A novel method for constrained mechanical design optimization problems. Comput. Des. 2011, 43, 303–315. [Google Scholar] [CrossRef]
  37. Hatamlou, A. Black hole: A new heuristic optimization approach for data clustering. Inf. Sci. 2013, 222, 175–184. [Google Scholar] [CrossRef]
  38. Bouchekara, H.R.E.H. Optimal power flow using black-hole-based optimization approach. Appl. Soft Comput. J. 2014, 24, 879–888. [Google Scholar] [CrossRef]
Figure 1. Illustration of the interaction between ECPs for strategy 1.
Figure 1. Illustration of the interaction between ECPs for strategy 1.
Sensors 21 06568 g001
Figure 2. Interaction between ECPs—(a) ECP1, (b) ECP2, and (c) ECP3 for Strategy 2.
Figure 2. Interaction between ECPs—(a) ECP1, (b) ECP2, and (c) ECP3 for Strategy 2.
Sensors 21 06568 g002
Figure 3. Search scenarios.
Figure 3. Search scenarios.
Sensors 21 06568 g003aSensors 21 06568 g003b
Figure 4. Search paths for Scenario A, obtained using the tested algorithms.
Figure 4. Search paths for Scenario A, obtained using the tested algorithms.
Sensors 21 06568 g004
Figure 5. Search paths for Scenario B, obtained using the tested algorithms.
Figure 5. Search paths for Scenario B, obtained using the tested algorithms.
Sensors 21 06568 g005
Figure 6. Search paths for Scenario C, obtained using the tested algorithms.
Figure 6. Search paths for Scenario C, obtained using the tested algorithms.
Sensors 21 06568 g006
Figure 7. Search paths for Scenario D, obtained using the tested algorithms.
Figure 7. Search paths for Scenario D, obtained using the tested algorithms.
Sensors 21 06568 g007
Figure 8. Search paths for Scenario E, obtained using the tested algorithms.
Figure 8. Search paths for Scenario E, obtained using the tested algorithms.
Sensors 21 06568 g008
Figure 9. Search paths for Scenario F, obtained using the tested algorithms.
Figure 9. Search paths for Scenario F, obtained using the tested algorithms.
Sensors 21 06568 g009
Figure 10. Convergence curves of the ECPO-ME for all scenarios.
Figure 10. Convergence curves of the ECPO-ME for all scenarios.
Sensors 21 06568 g010aSensors 21 06568 g010b
Table 1. Statistical analysis of the obtained results using each algorithm. The bold in table shows the best results.
Table 1. Statistical analysis of the obtained results using each algorithm. The bold in table shows the best results.
ECPO-MEECPOMPSOMVPADEGAESDABBOABCGSATLBOBH
Scenario ABEST0.205510.201160.181380.198100.201970.160300.188850.199480.185110.195720.204850.18101
MEAN0.192390.165550.060210.110850.189600.133680.147300.178510.149660.042580.192070.13570
MEDIAN0.191180.180910.032250.122250.189420.133150.151590.181320.160030.000950.191150.13742
WORST0.182090.031350.000000.000020.172500.105510.099420.117880.001310.000000.183030.06931
SD0.005930.041740.063050.061000.006300.014660.026090.016960.037070.062030.005870.02547
FR100100100100100100100100100100100100
Scenario BBEST0.276340.262520.226520.260140.274660.224860.239780.261560.244240.189250.276890.23919
MEAN0.257240.231260.106700.172720.245160.185630.183280.215220.194520.057170.252160.19299
MEDIAN0.258200.241530.112740.196130.249090.189320.183790.231960.201960.041390.254620.19735
WORST0.235300.146980.011050.036880.149280.151750.118020.110550.014710.000090.228760.13430
SD0.009640.025390.056590.060690.023390.018600.037860.044660.048920.052810.012930.02803
FR100100100100100100100100100100100100
Scenario CBEST0.686620.640700.584420.643610.686620.548350.642210.668110.611420.511140.672210.60402
MEAN0.641580.526140.301430.370150.621700.467970.475610.551090.494440.228600.632690.49182
MEDIAN0.649970.559790.314200.394190.635380.472100.486700.593580.524230.236930.636310.50913
WORST0.571620.261470.002400.014340.354320.362360.275490.231890.269330.000000.559670.36322
SD0.028760.106430.172890.154260.061930.053060.088450.106330.094060.136200.031110.07498
FR100100100100100100100100100100100100
Scenario DBEST0.554310.493090.470900.460950.554310.402200.514440.502740.464970.334580.537420.39930
MEAN0.488490.387660.243900.316730.454520.296750.312330.35144-0.204180.465750.28810
MEDIAN0.498870.389310.239980.326750.443470.296980.304510.33692-0.204860.462630.27994
WORST0.401480.266340.049820.092410.324580.218070.188260.22614-0.000020.373010.18647
SD0.031200.056200.081770.092980.063390.044110.078790.07564-0.088910.040440.05650
FR10010010010010010010010093.33100100100
Scenario EBEST0.205180.203580.189010.187260.204770.178680.186860.201300.195590.171880.205180.16870
MEAN0.190080.173670.119790.139860.186150.137530.144000.178090.162000.070770.188930.13598
MEDIAN0.188700.181540.129040.146250.183950.140290.148650.177250.164680.076340.188300.13544
WORST0.171870.110000.000210.022990.172770.093050.045210.154010.044780.000120.169690.07844
SD0.008010.023620.050920.035050.007870.018390.034340.011070.028100.055530.008670.02228
FR100100100100100100100100100100100100
Scenario FBEST0.227280.219850.160050.176610.221950.164160.202170.221950.205720.177210.221750.18615
MEAN0.200070.180200.073340.108510.196500.106380.130380.191420.150220.063850.200080.13070
MEDIAN0.210240.185730.072960.134800.201600.113100.144400.204650.158920.036870.209300.13941
WORST0.166500.081880.001290.005640.136400.041280.031710.081110.000410.000700.166480.02830
SD0.019780.032260.050250.054270.022340.033780.048130.033340.052070.061280.018200.04135
FR100100100100100100100100100100100100
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Alanezi, M.A.; Bouchekara, H.R.E.H.; Shahriar, M.S.; Sha’aban, Y.A.; Javaid, M.S.; Khodja, M. Motion-Encoded Electric Charged Particles Optimization for Moving Target Search Using Unmanned Aerial Vehicles. Sensors 2021, 21, 6568. https://doi.org/10.3390/s21196568

AMA Style

Alanezi MA, Bouchekara HREH, Shahriar MS, Sha’aban YA, Javaid MS, Khodja M. Motion-Encoded Electric Charged Particles Optimization for Moving Target Search Using Unmanned Aerial Vehicles. Sensors. 2021; 21(19):6568. https://doi.org/10.3390/s21196568

Chicago/Turabian Style

Alanezi, Mohammed A., Houssem R. E. H. Bouchekara, Mohammad S. Shahriar, Yusuf A. Sha’aban, Muhammad S. Javaid, and Mohammed Khodja. 2021. "Motion-Encoded Electric Charged Particles Optimization for Moving Target Search Using Unmanned Aerial Vehicles" Sensors 21, no. 19: 6568. https://doi.org/10.3390/s21196568

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop