Escolar Documentos
Profissional Documentos
Cultura Documentos
Pelotas, 2019
Patrick Guimarães de Vargas
Pelotas, 2019
Insira AQUI a ficha catalográfica
(solicite em http://sisbi.ufpel.edu.br/?p=reqFicha)
Patrick Guimarães de Vargas
Banca Examinadora:
Prof. Dr. Paulo Roberto Ferreira Jr (orientador)
Doutor em Computação pela Universidade Federal do Rio Grande do Sul.
Agradeço. . .
Só sei que nada sei.
— S ÓCRATES
RESUMO
1 INTRODUCTION . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
2 RELATED WORK . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
3 GEOMETRIC FLIGHT PATTERNS . . . . . . . . . . . . . . . . . . . . . . . 19
4 REAL-TIME SEARCH METHODS . . . . . . . . . . . . . . . . . . . . . . . . 20
5 PATROLLING STRATEGIES FOR NC-DRONE . . . . . . . . . . . . . . . . 23
5.1 Evaporation Strategy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24
5.2 Time-based Strategy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25
5.3 Quadrant Strategy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25
6 GENETIC ALGORITHMS . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27
6.1 Initialize Population . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27
6.2 Fitness Function . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27
6.3 Selection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
6.4 Crossover . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
6.5 Mutation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
7 OUR GENETIC OPTIMIZATION . . . . . . . . . . . . . . . . . . . . . . . . . 31
8 SIMULATION ENVIRONMENT . . . . . . . . . . . . . . . . . . . . . . . . . . 33
9 EXPERIMENTAL RESULTS . . . . . . . . . . . . . . . . . . . . . . . . . . . 35
9.1 Single Strategies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36
9.2 Combination of Strategies . . . . . . . . . . . . . . . . . . . . . . . . . . 37
9.3 Scenarios with No-Fly Zones . . . . . . . . . . . . . . . . . . . . . . . . 38
10 CONCLUSION . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42
REFERENCES . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
1 INTRODUCTION
Unmanned Aerial Vehicles (UAVs) are autonomous aerial platforms with no pilot
aboard. These platforms are usually controlled by a human navigator on the ground,
a ground control station (GCS), or even an intelligent system. A large number of tasks
are solved with the help of these vehicles, such as search and rescue, inspection,
surveillance, collecting data, and payload delivery. Many of these tasks are related
to the Patrolling Problem, which consists of inspecting an area of interest regularly
while gathering information using the onboard sensors of the UAV (HEFFERAN; CLIFF;
FITCH, 2016).
In such a dynamic environment, the UAV must visit and revisit every point of the
area to keep the collected data updated. Thus, patrolling performance is measured
concerning the number of visits in each region, the uniform distribution of visits, and the
interval evenness between those visits (SAMPAIO; SOUSA; ROCHA, 2016). The UAVs
have different constraints compared to terrestrial vehicles, such as payload, endurance,
and maneuverability. Simple geometric flight patterns, such as back-and-forth and
spiral (ANDERSEN, 2014), are often employed in patrolling missions considering these
constraints. However, these movements have pre-defined and repetitive behavior that
can be easily interpreted by an external observer and avoided by an intelligent target,
for instance.
Real-Time Search Methods (RTSM) may be explored to suppress this flaw, per-
forming the patrolling over an area and searching for a target, while keeping motion
unpredictable. These methods were introduced to solve land vehicle problems (KOE-
NIG; LIU, 2001) and expanded to aerial missions (CABREIRA et al., 2018).
In the work of CABREIRA et al. (2018), it is proposed the NC-Drone algorithm, an
RTSM, which is an extension of the well-known Node Counting (NC)(PIRZADEH; SNY-
DER, 1990). This algorithm allows UAVs to explore a grid-discretized scenario through
a pheromone-scheme. Additional strategies were also implemented to improve the
NC-Drone performance, such as Time-based (TS), Evaporation (ES), and Quadrant
(QUAD). The TS considers not only the amount of pheromone to make a decision but
also the time when the last visit occurred. The ES can reduce pheromone in places not
15
visited for a while to attract vehicles. Despite promising results regarding the perfor-
mance improvement of the algorithm, the parameters of the proposed strategies were
obtained empirically (KAPPEL et al., 2019).
This work proposes a parameter optimization for patrolling strategies based on the
NC-Drone algorithm using the Genetic Algorithm (GA). The optimization model em-
ploys a fitness function that can tune the parameters according to the selected perfor-
mance metric. We aim to obtain the set of parameters for the strategies by which its
best performance comes out considering a uniform distribution of visits. We compare
single and combined approaches with empirical values against methods with optimized
parameters by the GA. Experimental results show that the QUAD-ES GA strategy over-
comes all strategies, improving the NC-Drone up to 18.3% depending on the scenario.
This work is summarized as follows: The related work is discussed in Chapter
2. Geometric flight patterns and Real-time search methods are discussed in Chapter
3 and Chapter 4, respectively. Chapter 5 revises the NC-Drone algorithm and the
patrolling strategies. Chapter 6 revises the Genetic Algorithm. Chapter 7 presents
the genetic approach and describes the parameter optimization. The test platform is
presented in Chapter 8. Chapter 9 presents and discusses achieved results. Finally,
Chapter 10 comes up with the conclusion and mentions future works.
2 RELATED WORK
this deterministic model is that the next steps are traceable for whoever is executing
the mission while they are unpredictable for external observers.
An algorithm exploring the process of evaporation called EVAP is proposed to solve
the patrolling problem in multi-agent systems of an unknown environment (CHU et al.,
2007). In this approach, the pheromones lose intensity as time goes. This process
is employed to represent the arising uncertainty about an unvisited place. Later, Glad
et al. (GLAD et al., 2008) compared the EVAP algorithm against the VAW (WAGNER;
LINDENBAUM; BRUCKSTEIN, 1999), which is a similar strategy that also explores the
evaporation. In the EVAP method, agents move and then drop pheromone in the last
visited cell while the VAW algorithm does the opposite, dropping pheromone and then
moving. However, this particular behavior of the VAW leads to the encounter of agents
in the same cell more often. VAW has a more straightforward evaporation function than
EVAP. The authors also created a combination of both approaches called EVAW, which
has the evaporation function of the VAW and follows the ordered steps of moving and
dropping pheromones of the EVAP.
The NC-Drone (CABREIRA et al., 2018), which aims to minimize the UAV turning
maneuvers, already marks the visited cells with pheromone and move to places with
the least amount. Furthermore, another approach also explores a variation of VAW as
the evaporation function (KAPPEL et al., 2019), keeping the simplicity and dropping
pheromone values linearly as the time goes. The optimization using GA intends to im-
prove this function finding the parameter setup by which obtains the best performance.
As for genetic algorithms (GA), a survey is presented by THENGADE; DONDAL (2012)
showing the basic functionalities of a GA, such as selection, crossover, and muta-
tion and their advantages and disadvantages. Also, the authors offer a comparison
between GAs and other algorithms and heuristics for problem-solving
The work of PARADZIK et al. (2016) uses evolutionary strategies to terrain cove-
rage using multi-UAVs. A Genetic Algorithm (GA) is applied to decompose the area in
rectangles to be covered by the UAVs. This method explores two types of pheromo-
nes on these areas for the agent covering the map. Search pheromones describe the
uncertainty of a location. The higher the concentration of this type of pheromone, the
higher the need to visit that area. Path pheromones mark the specific regions visited
by the UAVs. Therefore, other agents should search for places with the least amount
of path pheromone to prevent covering the same locations.
A Genetic Algorithm was applied to optimize the coverage of vacuum cleaner robots
in (YAKOUBI; LASKRI, 2016), where a scenario is split into a grid containing obstacles
marked as unreachable cells. The agent starts with an unknown map and discovers
the area along the way. This approach ends up with a static coverage path planning
strategy to further optimize the path. However, due to the problem of having an in-
telligent observer who can memorize the route and evade the agent, the randomness
18
of the RTSM is a desirable feature. A terrain field coverage on 3D scenarios for agri-
cultural robots is presented in (HAMEED, 2014). The authors were able to reduce by
10% to 15% the operational time using a Genetic Algorithm approach. Furthermore,
they cut the energy consumption in 6.5% by optimizing the driving angle in the 3D
configurations.
GA optimization is usually combined with other methods to generate coverage tra-
jectories (LI; WANG; SUN, 2016; TRUJILLO et al., 2016; HAYAT et al., 2017; BOUZID;
BESTAOUI; SIGUERDIDJANE, 2017). Despite the optimization, all the mentioned
works create predefined coverage paths that can be easily recognized by intelligent
targets during the mission execution. Our goal is to find the ideal configuration of pa-
rameters for different patrolling strategies using the GA. These strategies influence the
behavior of the UAV during the coverage while intending to improve the performance
metrics related to uniform distribution and interval of visits. Our unforeseeable trajecto-
ries are indirectly affected by the parameters’ optimization provoked by the GA but still
are generated online using the NC-Drone algorithm.
3 GEOMETRIC FLIGHT PATTERNS
Despite being the most common flight patterns, for more specific problems, like
military surveillance, these types of geometric patterns can not be used, due to its
trajectory being easily predictable. Once the pattern is known, an intelligent target can
avoid the UAV by moving in the opposite direction or hiding in a place for a while.
4 REAL-TIME SEARCH METHODS
2 4 3 3 4 3
3 2 2 2
4 1 3
1 3 1
the pheromone marks left by the robots, a value, named u-value, is associated with
each cell of the environment. The u-value indicates the number of visits performed
by all robots on each location. A robot senses the pheromones in the neighbor cells.
Since the goal is to explore the scenario, the vehicles use the u-value to select its next
location, always choosing the cell with the lowest amount. When visiting a cell, the
robot also marks the spot by writing pheromone. Updating the pheromone value of the
current cell before moving to the next location. When two or more cells have the same
minimum-value, it randomly decides the cell be visited.
There are four main RTSM: Node Counting (NC)(PIRZADEH; SNYDER,
1990), Learning Real-Time A* (LRTA*)(KORF, 1990), Thrun’s Value Update Rule
(TVUR)(THRUN, 1992) e Wagner’s Value Update Rule (WVUR) (WAGNER; LINDEN-
BAUM; BRUCKSTEIN, 1997).
The NC algorithm (Algorithm 1) checks the neighborhood and selects the cell with
the lowest u-value. In the case of two or more cells with the same amount of u-value,
the next position is chosen randomly. Then the robot updates the u-value of the current
cell and moves to the next selected location. This process is repeated until a stop
criterion is reached, such as the number of simulation cycles.
The other methods, LRTA*, TVUR, WVUR, uses the same algorithm as the NC.
What differentiates each RTSM is the adopted update rule. Node Counting approach
adds one unit to the current value. LRTA* changes the value of the current cell u(s) by
the value of the next cell to be visited by the vehicle u(next(s)) added by one unit. TVUR
compares the value of the current cell u(s) with the value of the next cell to be visited
u(next(s)). The highest value is increased by one unit and replaces the original value
of the current cell u(s). WVUR presents a conditional update rule where the value of
the current cell u(s) is increased by one unit if it is not larger than the value of the next
cell u(next(s)). If the condition is not met, the value of the current cell is not updated.
Table 1 presents the update rules for the four main RTSM.
22
The unpredictability of the Real-Time Search Methods comes from the fact that
when two or more cells have the same pheromone value, the robot randomly chooses
its next location, as shown in Figure 3 . This behavior leads to a massive number
3 0 3
?
? ?
0 0
1 1 1
Algorithm 2: NC-Drone
Input: A set of cells (c1, ...cp)
Output: A set of metric values (m1, ...mp)
10 repeat
11 Read the u-value of the neighbor cells;
12 Select the neighbor cell with lowest u-value;
13 if there is a tie and one of the cells is aligned with the sweep direction then
14 Choose the aligned cell as the next place to be visited;
15 else
16 Randomly choose one of the cells;
17 end if
18 Update with +1 the u-value of the current cell;
19 Move to the selected neighbor cell;
20 until Number of cycles is reached;
The reading and writing pheromone scheme depends on a ground control station
(GCS) to operate in UAVs. The GCS monitors the scenario, and the robot swarm.
Besides, it tracks the least visited places and sends the corresponding motion coordi-
nates to the UAVs. However, this system always requires communication between the
base and the UAVs to perform every action. Thus, a novel matrix-based communica-
tion protocol is also proposed in (CABREIRA et al., 2018). In this multi-robot system,
each aerial robot keeps an internal matrix. This matrix is a representation of the sce-
24
nario, where the vehicle computes the visits performed by itself. Instead of reading and
writing in the scenario, the robot follows its internal matrix. The swarm can exchange
information about the visited places and synchronize their matrices. The system is de-
signed to work in case of a lack of communication, a critical issue that may compromise
the mission. However, in our work, we focus on the centralized NC-Drone as we are
aiming to obtain the best performance from the algorithm.
3 3 3 4 3 3 4 3
2 3 2 2 2 2 4 2
3 2 3 3 2 3 3 3
One of the main concerns of the patrolling problem is to maintain a uniform distribu-
tion of visits, both spatial and temporal. In this way, additional patrolling strategies were
also proposed to improve the performance of the NC-Drone (KAPPEL et al., 2019). All
the parameters for the patrolling strategies were obtained empirically. In this regard,
our work aims to tune these parameters through GA. The strategies consists of Evapo-
ration (Chapter 5.1) and Time-Base (Chapter 5.2). We also proposed a new strategy
called Quad (Chapter 5.3).
Genetic algorithms are a class of optimization procedures that aims to find the
optimum solution to a problem. John Holland first introduced them in 1975 (HOLLAND,
1975). They follow the Darwinian principle of reproduction and survival of the fittest,
where the most adapted individuals spread their offspring to the next generations. GA
is excellent at exploring a large and complex space to find the value close to the global
optimum. The only thing to know is the search space, and the GA will generate a
high-quality solution. Unlike traditional search heuristics, the GA does not evaluate
and improve a single solution but a set of possible solutions, a so-called population.
GAs have been proposed to solve a variety of problems, as they can work in any
search space. They can be used to design neural networks (MILLER; TODD; HEGDE,
1989), solve the travelling salesman problem (RAZALI; GERAGHTY et al., 2011), path
planning (YAKOUBI; LASKRI, 2016), and many more.
The GA process can be divided into five phases: Initialize population, Fitness func-
tion, Selection, Crossover, and Mutation.
A1 0 0 0 0 0 0 Gene
A2 1 1 1 1 1 1 Chromosome
A3 0 1 1 0 0 1
A4 1 0 0 0 1 1 Population
and generate the offspring is determined by the fitness score. The higher the score,
the higher is the probability of being selected.
6.3 Selection
Selection aims to determine which individuals will be selected and combined to
produce the next generation. Based on their fitness score, a pair of individuals is
selected to generate two new individuals. The population size is fix, the least adapted
chromosomes are eliminated, providing space for the new offspring.
6.4 Crossover
Once the individual has been selected, they need to be combined to generate a
new offspring. The solution for this is called crossover. A crossover point, Figure 8a,
is randomly chosen within the parents. The children are created by exchanging the
genes of parents among themselves until the crossover point is reached, as shown in
Figure 8b. The new offspring, Figure 8c, is then added to the population. Crossover
does not always occur. Sometimes, based on a set probability, the children can be
direct copies of the parent.
6.5 Mutation
After selection and crossover, a new population is produced. As some of the in-
dividuals can be direct copies of their parents, we need to allow a small chance of
mutation. In this process, a gene of the individual is chosen at random, and its value is
29
A1 0 0 0 0 0 0 A1 0 0 0 0 0 0 A5 1 1 1 0 0 0
A2 1 1 1 1 1 1 A2 1 1 1 1 1 1 A6 0 0 0 1 1 1
change. Figure 9, shows the mutation selecting the third gene and changing its value
to zero. The mutation is relatively simple; however, it is vital to ensure genetic diversity
within the population.
Before Mutation
A5 1 1 1 0 0 0
After Mutation
A5 1 1 0 0 0 0
The genetic algorithm process, Figure 10, repeats until the population has conver-
ged, which means that the new offspring is not significantly different from the previ-
ous generation, or the process repeats for a number of generations. The number of
generations varies from problem to problem. More complex problems require more
generations until the GA can adapt and provide a suitable response.
30
Initialize Population
Fitness Calculation
Terminate? Results
Selection
Crossover
Mutation
The patrolling strategies improved the performance of the NC-Drone algorithm ac-
cording to the experimental results obtained in the work of KAPPEL et al. (2019).
However, the values of the parameters of such strategies were empirically set up. Thus,
a Genetic Algorithm (GA) optimization is presented to define the values of the set of
parameters for the patrolling strategies. Initially, the GA randomly generates a set of
candidate solutions called population. Each solution represents an individual of the
population, and it is encoded as a chromosome. In our GA optimization approach, th-
ree parameters encoded in a binary array compose the chromosomes, the Evaporation
Time (ET), Evaporation Factor (EF), and Time Threshold (TT). The values for the ET
and TT vary from 1 to 1000, while the values for the EF vary from 0 to 1.
A fitness function evaluates all the individuals, measuring the quality of the candi-
date solutions. Our fitness function is based on the Quadratic Mean of Intervals (QMI)
(SAMPAIO; SOUSA; ROCHA, 2016). The QMI corresponds to frequency-regularity
equilibrium, a patrolling requirement that combines the frequency of visits and even-
ness of intervals between these visits. This performance metric balances values of
average, maximum, and standard deviation of intervals. Equation (1) presents the GA
fitness function based on the QMI performance metric:
1/2
visits(x)+1
1 X X
QM I = (ixj )2 (1)
Nintervals x∈cells j=1
where ixj represents the intervals between the visits j to the cells x and Nintervals cor-
responds to the total number of intervals. The lower the QMI, the more uniform is the
patrolling path.
The next step is to select the solutions to reproduce and spread their offspring for
the next generation. The proposed GA adopts the roulette-wheel selection, where the
fittest candidates have a higher chance of being picked for reproduction. This probabi-
lity is proportional to the amount by which its fitness is higher than its competitor’s fit-
ness. Then, the selected individuals, also known as parents, are submitted to a uniform
crossover to reproduce and generate a new set of possible solutions. In this process,
32
the parents can contribute evenly with their chromosomes for the children, or children
could be direct copies of their parents, depending on the probability of crossover. Our
GA applies a uniform crossover with a chance of 80%.
PARAMETER INFORMATION
Population 80 individuals
Selection Roulette-wheel
Crossover Uniform with 80%
Mutation Rate of 10%
Generation 200
Finally, the new generation goes through a mutation process.The individuals have
a 10% chance to suffer mutation. A process called “big mutation” is also employed
to overcome premature or local-best solution. In this process, the GA multiplies the
fitness of the best individual by α - in our case, α = 0.6. If the result is lower than the
mean of the population fitness, the probability of mutation is five times greater.
All new individuals replace the original population, and the GA restarts a new cycle
of execution. The algorithm runs for 200 generations and selects the best individual as
the ideal configuration of parameters for the patrolling strategies. Table 2 presents the
summary of GA. This particular configuration was obtained empirically. Different rates
of crossover and mutation were tested, as well as a variable number of individuals and
generations.
8 SIMULATION ENVIRONMENT
The test platform, presented in Figure 11, was built using the PyGame library
(SHINNERS, 2019). We based our simulation environment on the work of CABREIRA
et al. (2018). Our platform has two types of agents: Patches and Drones. We have a
grid-discretized scenario, where each one of the grid cells is modeled as patches. The
patches have a fixed position on the scenario. They keep information related to the cell,
like the number of visits, the interval between visits, the time of the last visit, and the
amount of pheromones. The Drones, representing the UAVs, move over the patches
on the scenario using the NC-Drone algorithm. They can read and write information
on the patches.
34
The platform allows us to combine different strategies for NC-Drone easily. We can
also experiment with different parameter values for said strategies. The number of
UAVs can be modified, as well as the number of simulation cycles. The simulation can
be performed step by step on-screen to observe and analyze the agents’ behavior or
off-screen to improve simulation performance.
All the patches start with the number of visits equals to zero. The UAVs are released
on the scenario with an interval of one cycle between each agent. We run m rounds of
n cycles for each experiment. We then compute the Mean and Standard deviation of
the evaluation metrics.
For the GA, we used a python framework for genetic algorithm called Gaft 1 . This
framework provides built-in genetic operators for target optimization and plugin inter-
faces for users to define their own genetic operators and on-the-fly analysis for al-
gorithm testing. First, we need to define the population. Gaft uses tuples to repre-
sent the chromosomes. We select the range and the population size. Then based
on these parameters, Gaft will generate the population. For selection, we have four
built-in operators: roulette wheel selection, exponential ranking selection, linear ran-
king selection, and tournament selection. After the selection operator is defined, we
need to select the crossover and mutation operators. Gaft supports uniform crosso-
ver and flip bit mutation. As we select these operators, we also need to define their
probabilities. We then create an engine that will run the GA. The most important part
is to defined our fitness function. The GA selects the best individual based on the
fitness score. If we want to minimize the fitness function, we can add a minimiza-
tion. In this way, the best candidates are the ones with the lowest score. At last,
we defined the number of generations. After the GA is completed, Gaft will gene-
rate a text file containing the best individual and its fitness value for each generation.
More details about our platform and the GA framework can be found on GitHub in
https://https://github.com/patrickguima/NC-Drone-Simulator.
1
Gaft Documentation, https://gaft.readthedocs.io/en/latest/
9 EXPERIMENTAL RESULTS
PARAMETER INFORMATION
Grid Size 50 x 50
UAVs 4
Cycles 10.000
Simulations 30
The QMI metric was described in Chapter 7. The SDF computes the distribution
equality of visits over the cells. The lower the value, the more uniform the frequency is.
Equation (2) illustrates the SDF performance metric:
" #1/2
1 X
SDF = ( f req(x) − Favg )2 (2)
|cells| x∈cells
where f req(x) corresponds to the frequency of visits on each cell x, Favg is the average
frequency of all cells and cells is the total number of cells present in the scenario.
36
NTM is usually correlated with the energy spent to perform a coverage path. Every
time a UAV needs to execute a turning maneuver, it must decelerate, rotate, and ac-
celerate again. These maneuvers increase covering time and, consequently, the ne-
cessary energy to carry out a mission. Thus, the lower the NTM, the lower the energy
consumption is. Equation (3) shows the NTM performance metric:
turns(u)
X X
NT M = tuk (3)
u∈uavs k=1
where tuk denotes the number of turning maneuvers performed by each UAV u.
PARAMETERS
APPROACHES
ET EF TT
ES 100 0.1 x
ES GA 1 0.32 x
TS x x x
TS GA x x 4
QUAD x x x
The results for the Time-based Strategy (TS) are also presented in Table 5. The
original TS has not a Time Threshold (TT) parameter. The TS-GA introduces this
parameter with T T = 4 and presents a QMI improvement of around 3% compared to
the original NC-Drone. In the TS-GA, the UAV selects the less recently visited cell
if there is a difference of at least four cycles of simulation between the last visit in
the two cells with the same value. Otherwise, the UAV randomly chooses one of the
cells. Despite the parameter optimization, the Student’s T-Test states that the difference
between the TS and TS-GA is considered to be not statistically significant.
Finally, the Quadrant Strategy (QUAD) presents a QMI improvement of around 2%
and reduces in 59% the standard deviation compared to the original NC-Drone. Consi-
37
dering all strategies, the ES-GA obtains the best results in the QMI performance metric,
followed by the TS-GA. Despite the smaller improvement in the QMI, the TS-GA stra-
tegy maintains the SDF and NTM with similar results compared to the NC-Drone, while
ES-GA increases the SDF by 62% and NTM by 121%. The experimental results of the
patrolling strategies in a scenario free of obstacles are presented in Table 5.
PARAMETERS
APPROACHES
ET EF TT
ES-TS 100 0.1 x
ES-TS GA 1 0.32 4
QUAD-TS x x x
QUAD-TS GA x x 4
QUAD-ES 100 0.1 x
QUAD-ES GA 1 0.83 x
QUAD-ES-TS 100 0.1 x
QUAD-ES-TS GA 5 0.32 3
sets T T = 4 and presents a QMI improvement of around 4.3% compared to the NC-
38
Drone. According to the Student’s T-test, the difference between the QUAD-TS and
QUAD-TS GA is considered to be not statistically significant.
QUAD-ES mixes Quadrant and Evaporation strategies with the following empirical
parameters ET = 100 and EF = 0.1. The GA parameter optimization sets ET = 1
and EF = 0.83 for the QUAD-ES GA. This strategy presents the best results among all
combinations — the percentage of QMI improvement is around 10.7% over the QUAD-
ES and 18.3% over the NC-Drone.
The final combination includes all previously presented strategies, Quadrant, Time-
based, and Evaporation. The empirical QUAD-ES-TS adopts the same configuration
with ET = 100 and EF = 0.1, while the GA version sets ET = 5, EF = 0.32, and
T T = 3. The QUAD-ES-TS GA shows an improvement of 7.7% and 15.8% compared
to the QUAD-ES-TS and NC-Drone without the parameter optimization.
Three strategies are marked with * in Table 7, QUAD-TS, QUAD-ES GA, and QUAD-
ES-TS. These strategies have zero standard deviation in all performance metrics,
which means that the UAV moves through the scenario using the same trajectory in
all simulations. Figure 12 illustrates the patrolling pattern generated by the QUAD-ES
GA strategy, the best solution among the combination of strategies.
20
18
16
14
12
10
0
0 2 4 6 8 10 12 14 16 18 22
The complete results for the series of combination of patrolling strategies are pre-
sented in Table 7.
QMI
SDF NTM
APPROACHES
M SD M SD M SD
NC-Drone 750.13 11.58 0.75 0.09 5.6k 366.51
ES-TS 703.89 0.40 0.70 0.05 10.7k 3.55
ES-TS GA 678.74 1.72 1.23 0.13 12.4k 117.66
QUAD-TS* 718.23 0.00 0.89 0.00 6.8k 0.00
QUAD-TS GA 717.60 5.71 0.89 0.04 6.2k 202.04
QUAD-ES 685.97 3.23 0.93 0.04 11.6k 256.49
QUAD-ES GA* 612.64 0.00 0.81 0.00 4.5K 0.00
QUAD-ES-TS* 684.00 0.00 0.92 0.00 11.7k 0.00
QUAD-ES-TS GA 631.32 8.09 0.90 0.05 7.0k 1.0k
8 presents the resulting parameter optimization in the three scenarios containing NFZ.
These values are the ideal configuration that optimizes the QMI results of each patrol-
ling strategy presented in Table 9. The ET parameter remains almost constant, and
the interval between each evaporation varies only between 1 and 2 cycles of simulation
depending on the scenario. The ES parameter varies between 0.37 and 0.90 depending
on the scenario and the adopted strategy. Finally, the TT parameter changes from 3 in
the TS to 898 in the ES-TS.
The QMI results for all strategies, and the respective combinations are presented
in Table 9. The SDF and NTM results are presented in Table 10 and Table 11, Res-
pectively. As one can see, they get worse as the QMI improves. Once again, the
approaches whose parameters were optimized through the GA present better results
than the strategies with the empirical values, with the exception of the TS GA and
QUAD-TS GA. The ES GA shows a statistical difference compared to the empirical ES
40
and improves the QMI of the NC-Drone in around 8% in the three scenarios. Consi-
dering the TS and the QUAD-TS, there is no statistically significant difference between
the empirical approaches and the optimized ones by the GA. The ES-TS GA presents
a slight improvement compared to the original ES-TS.
The QUAD-ES GA and QUAD-ES-TS GA strategies obtain the best results among
all combinations. Both approaches present no statistical difference in the first two
scenarios. However, the QUAD-ES GA overcomes QUAD-ES-TS GA in scenario 3.
The percentage of improvement of the QUAD-ES GA compared to the NC-Drone is
around 10% in scenarios 1 and 2 and almost 12% in scenario 3.
Table 9 – QMI results for the combination of patrolling strategies using GA in
scenarios with NFZ.
This work proposes a parameter optimization for patrolling strategies based on the
NC-Drone algorithm using GA. These strategies consist of Evaporation, Time-based,
and Quadrant, and the parameters are the Evaporation Time (ET), Evaporation Factor
(EF), and Time Threshold (TT). The optimization model explores the QMI metric as
a fitness function to tune the parameters. We aim to minimize this metric to obtain a
uniform distribution of visits.
We present experimental results of single and combined strategies over scenarios
free of obstacles and containing no-fly zones. We compare approaches with empiri-
cal parameters against solutions with values optimized by the GA. The QUAD-ES GA
emerges as the most promising solution for the patrolling problem. This strategy over-
comes all combinations in all different scenarios. As future work, we intend to further
explore the patrolling strategies in specific applications, such as marine traffic, and
implement the solutions in UAVs to perform real flights.
43
REFERENCES
ANDERSEN, H. L. Path Planning for Search and Rescue Mission using Multicop-
ters. 2014. Dissertação (Mestrado em Ciência da Computação) — Institutt for teknisk
kybernetikk, Norway.
CHU, H. N. et al. Swarm approaches for the patrolling problem, information propa-
gation vs. pheromone evaporation. In: TOOLS WITH ARTIFICIAL INTELLIGENCE,
2007. ICTAI 2007. 19TH IEEE INTERNATIONAL CONFERENCE ON, 2007. Anais. . .
[S.l.: s.n.], 2007. v.1, p.442–449.
GLAD, A.; SIMONIN, O.; BUFFET, O.; CHARPILLET, F. Theoretical study of ant-
based algorithms for multi-agent patrolling. In: EUROPEAN CONFERENCE ON AR-
44
HAMEED, I. A. Intelligent coverage path planning for agricultural robots and autono-
mous machines on three-dimensional terrain. Journal of Intelligent & Robotic Sys-
tems, [S.l.], v.74, n.3-4, p.965–983, 2014.
HAYAT, S.; YANMAZ, E.; BROWN, T. X.; BETTSTETTER, C. Multi-objective UAV path
planning for search and rescue. In: ROBOTICS AND AUTOMATION (ICRA), 2017
IEEE INTERNATIONAL CONFERENCE ON, 2017. Anais. . . [S.l.: s.n.], 2017. p.5569–
5574.
HEFFERAN, B.; CLIFF, O. M.; FITCH, R. Adversarial Patrolling with Reactive Point
Processes. In: ARAA AUSTRALASIAN CONFERENCE ON ROBOTICS AND AUTO-
MATION (ARAA, 2016), 2016. Proceedings. . . [S.l.: s.n.], 2016. p.39–46.
KAPPEL, K. S. et al. Strategies for Patrolling Missions with Multiple UAVs. Journal of
Intelligent & Robotic Systems, [S.l.], p.1–17, 2019.
KOENIG, S.; LIU, Y. Terrain Coverage with Ant Robots: A Simulation Study. In:
FIFTH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS, 2001. Pro-
ceedings. . . [S.l.: s.n.], 2001. p.600–607.
KUIPER, E.; NADJM-TEHRANI, S. Mobility models for UAV group reconnaissance ap-
plications. In: INTERNATIONAL CONFERENCE ON WIRELESS AND MOBILE COM-
MUNICATIONS (ICWMC’06), 2006., 2006. Anais. . . [S.l.: s.n.], 2006. p.33–33.
LI, D.; WANG, X.; SUN, T. Energy-optimal coverage path planning on topographic map
for environment survey with unmanned aerial vehicles. Electronics Letters, [S.l.], v.52,
n.9, p.699–701, 2016.
MILLER, G. F.; TODD, P. M.; HEGDE, S. U. Designing Neural Networks using Genetic
Algorithms. In: ICGA, 1989. Anais. . . [S.l.: s.n.], 1989. v.89, p.379–384.
ÖST, G. Search path generation with UAV applications using approximate convex
decomposition.
PIRZADEH, A.; SNYDER, W. A unified solution to coverage and search in explored and
unexplored terrains using indirect control. In: IEEE INTERNATIONAL CONFERENCE
ON ROBOTICS AND AUTOMATION, 1990. Proceedings. . . [S.l.: s.n.], 1990. p.2113–
2119.
ROSALIE, M.; DANOY, G.; CHAUMETTE, S.; BOUVRY, P. From random process to
chaotic behavior in swarms of UAVs. In: ACM SYMPOSIUM ON DEVELOPMENT
AND ANALYSIS OF INTELLIGENT VEHICULAR NETWORKS AND APPLICATIONS,
6., 2016. Proceedings. . . [S.l.: s.n.], 2016. p.9–15.
SAMPAIO, P.; SOUSA, R.; ROCHA, A. New Patrolling Strategies with Short-Range
Perception. In: XIII LATIN AMERICAN ROBOTICS SYMPOSIUM AND IV BRAZILIAN
ROBOTICS SYMPOSIUM, 2016., 2016. Proceedings. . . [S.l.: s.n.], 2016.
TRUJILLO, M. M. et al. Optimized flight path for 3D mapping of an area with structures
using a multirotor. In: UNMANNED AIRCRAFT SYSTEMS (ICUAS), 2016 INTERNA-
TIONAL CONFERENCE ON, 2016. Anais. . . [S.l.: s.n.], 2016. p.905–910.
YAKOUBI, M. A.; LASKRI, M. T. The path planning of cleaner robot for coverage region
using genetic algorithms. Journal of innovation in digital ecosystems, [S.l.], v.3, n.1,
p.37–43, 2016.