Supporting Search and Rescue Operations With UAVs
Supporting Search and Rescue Operations With UAVs
Abstract— Search and rescue operations can greatly benefit aftermath of Hurricane Katrina, two UAVs were used to survey
from the use of autonomous UAVs to survey the environment the damaged area in the search for trapped survivors [1]. The
and collect evidence about the position of a missing person. growing recognition of the potential of using UAVs for search
To minimize the time to find the victim, some fundamental
parameters need to be accounted for in the design of the search and rescue applications is supported by an increasing number
algorithms: 1) quality of sensory data collected by the UAVs; 2) of works in the areas of image recognition for victim detection,
UAVs energy limitations; 3) environmental hazards (e.g. winds, path planning and task allocation [2]–[4].
trees); 4) level of information exchange/coordination between In this paper, our focus is the analysis of the performance of
UAVs. different search techniques when the time to find the victim
In this paper, we discuss how these parameters can affect
the search task and present some of the research avenues we (also referred to as target) is the optimization criterion. In
have been exploring. We then study the performance of different particular, we compare the benefit of sharing data between
search algorithms when the time to find the victim is the UAVs with different search techniques based on greedy heuris-
optimization criterion. tics, potential fields, and partially observable Markov decision
process.
I. I NTRODUCTION
The remainder of the paper is organized as follows. We first
Search and rescue operations are often characterized by a describe our hardware platforms in Sec. II. We emphasize
similar set of constraints: time is critical and any delay can some research challenges of search and rescue applications
result in dramatic consequences – potentially human losses; with UAVs in Sec. III. We then present the performance
operational environments are unfriendly, e.g. disaster scenes, evaluation of different search strategies in Sec. IV and Sec. V.
forests, etc. Using Unmanned Aerial Vehicles (UAVs) can We conclude this paper in Sec. VI.
provide a critical support to search and rescue operations.
UAVs are agile, fast, can exhibit autonomous behaviours and II. UAV QUADROTOR
hence perform operations hard to execute by human operators, The hardware platforms we are using are off-the-shelf,
at low operating costs. In a typical scenario, UAVs will be electrically-powered quadrotors. We are currently using several
deployed in an area of interest, perform sensory operations to models from Ascending Technologies (Fig. 2). These vehicles
collect evidence of the presence of a victim, and report their have the advantage of being easy to prepare, control and fly.
collected information to a remote ground station or rescue The payload they can carry is in the order of a few hundred
team (Fig. 1). UAVs have already demonstrated their benefit grammes. With the current battery capacity available, the flight
time of the UAVs is restricted to a few tens of minutes. Each
of these platforms has been redesigned to accommodate a 1.6
GHz Intel Atom board with 1GB of RAM and 16GB SSD
that can provide the computation abilities.
III. R ESEARCH C HALLENGES
To optimize the tasks of the UAVs during the search
operation, several factors need to be accounted for in the
design of the search strategy:
• Quality of sensory data: In search and rescue operations,
to properly assess the quality/trustworthiness of the
Fig. 1. Rescue scenario with UAVs: the UAVs fly over an area the victim
is believed to be located, gather information on its potential location, and information reported to the rescue team is of paramount
transmit it back to a remotely located rescue team. importance. If a UAV is surveying an area, a victim
should not be missed, i.e. the probability of false negative
in search and rescue operations by helping responders to focus should remain low; at the same time, the probability
their search efforts while avoiding hazards. In 2006, in the of false positive, i.e. the probability that a victim is
Fig. 2. UAVs in use in the SUAAVE (Sensing Unmanned Autonomous Aerial VEhicles) project.
considered as detected when it is actually not there, and cost of connection (necessity to travel to a given
should remain low to avoid sending a ground rescue rendezvous point). Since it is also critical to regularly
team to a place of no interest. The challenge is then to inform the ground station of the evolution of the
establish accurate models of the quality of the sensory mission, it is important to evaluate the benefit to
data obtained from the UAVs. maintain the network fully or partially connected.
Other questions that require further investigation are:
• Energy limitations: One of the critical aspects of the How frequently should the connections be estab-
UAV platforms we are using is their limited flight time. lished between the UAVs? How frequently should
In preliminary experiments, we have established that the the UAVs report to the ground station?
current consumption when the UAV is flying and when
IV. S EARCH A LGORITHMS
it is hovering is similar. One way to save energy is then
to have the UAV remain idle (rotors off). Depending on In this work, we consider that UAVs are equipped with
the tasks to accomplish and on the capabilities of the downward-pointing cameras to detect victims on the ground.
UAVs, several search strategies can be envisionned such By changing altitudes, UAVs can change the size of their
as alternating periods of flights and periods of rest, or observation areas. The higher they fly, the bigger their obser-
sending only a subset of the UAVs to accomplish certain vation area. But as the UAV flies higher up, the level of detail
tasks whilst the rest remains idle (recharging batteries decreases. Using these observations, we have shown that with
for instance). a single UAV, changing altitudes is a valid control strategy that
can speed up the search process [5], [6]. A question we are
• Environmental hazards. During its flight, a UAV should investigating in this work is the impact of altitude on the search
be capable of avoiding environmental hazards (trees, strategy. When several UAVs are deployed, the complexity of
building, etc.) as well as avoiding collisions with other the problem increases. In an initial work [7], we investigated a
UAVs. It is also necessary to account for the fact that simple fusion algorithm in which all observations made by all
depending on the position of the UAV, some areas on UAVs are exchanged when the UAVs come into communica-
the ground might be occluded. tion range and are locally maintained. We use this mechanism
in this work when communication between UAVs occur. We
• Information sharing: Two aspects of the information also consider differences in quality of sensory data, as well
sharing process need to be considered: data fusion and as we account for the presence of obstacles during the search
network connectivity. operations. We implicitly account for energy constraints by
looking at search strategies that minimize the time to find the
– Data fusion: When multiple UAVs are deployed, the victim.
sensory data they collect can be shared and fused Search algorithms for search and rescue operations should
to generate a complete picture of the environment – be able to cope with the uncertainties of real-world deploy-
which can in turn guide the search process. This task ment. Hence real-time approaches are more appropriate and
is all the more challenging as any solution that will can be divided into three main categories:
be proposed needs to account for limitations in terms 1) Greedy heuristics
of processing, memory storage, energy consumption, 2) Potential-based heuristics
network availability and so on. 3) Partially Observable Markov Decision Process
– Network Connectivity: During deployment, commu- (POMDP) based heuristics
nications can occur between UAVs and between
UAV-Ground Station. The connections can happen To maintain the information on the probability of the victim
opportunistically or can be scheduled. A trade-off location, each UAV maintains a grid-based probabilistic map
needs to be made between reward of establishing a (belief map) composed of cells that represent the discretization
connection (resulting in exchanges of information) of the search space. Each cell contains the probability that the
143
Authorized licensed use limited to: UNIVERSITY OF BRISTOL. Downloaded on November 05,2024 at 00:16:48 UTC from IEEE Xplore. Restrictions apply.
target is present in it. These maps are commonly used for ap- • Rand: Pick s� at random (the version of Rand where
plications such as surface mapping, exploration or navigation UAVs can communicate is referred to as M S Rand).
(e.g. [8]). • Cov: Pick s that provides the highest coverage.
�
144
Authorized licensed use limited to: UNIVERSITY OF BRISTOL. Downloaded on November 05,2024 at 00:16:48 UTC from IEEE Xplore. Restrictions apply.
• Attractive potentials should increase as the UAV moves 48;:
away from the goal. When the UAV reaches the goal, no 4 89:
23!1'-14#5#&1=
forces from the goal should act upon it.
• Repulsive potentials should increase as the UAV comes <
48<:
9 ;
affect the UAV motion when it is far from the obstacle.
Practically, in our search and rescue scenario, we assigned
to each cell a weight that depends on its nature: obstacles
23!1'-14#5#&16
and already-visited locations. The weight of each cell
7
that is visited is increased by Kv after each visit. A
potential is computed for each grid cell according to Alg. 2.
Initialization
For each cell that is an obstacle, assign maxWeight
s = start location .)*/(01
%&'(# !"#$%&'(#
%)*+#,-"*(
while true do
assign weight w(s) = min(w(s) + Kv , maxW eight)
for each cell do Fig. 4. Scenario for the implementation of Hierarchical POMDP with 2
compute distance dv to visited cells UAVs flying at 2 different altitudes.
U (cell) = w(cell)/d2v
end
s’=neighboring cell with minimum potential • γ: discount factor.
move UAV to location s’ The goal is to find a set of actions that maximizes
� the expected
if target in s’ then sum of rewards over an infinite horizon E[ t γ t R(st , at )].
Report Base Station
end In our implementation, we considered that each UAV is able
s = s’ to compute its policy based on information locally maintained.
end
Time constraints have been enforced due to the size of the
Algorithm 2: Potential-based search algorithm problem and the number of states that could potentially lead to
hours of computation Since UAVs can fly at different altitudes,
C. Partially Observable Markov Decision Process we can differentiate the tasks of the UAVs as a function of their
Partially Observable Markov Decision Processes (POMDPs) altitude (Fig. 4). This can be modelled using a POMDP with
are a generalisation of a Markov Decision Process (MDP) to different observations models and set of actions for each set
situations where the system can be modelled by a MDP, but of UAVs (referred to as Hierarchical POMDP).
the underlying states are unobservable. Instead, the state dis- For simplicity, we consider that the UAVs can fly at only two
tribution must be inferred based on a model of the world and distinct altitudes.
local observations where, typically, sensors provide partial and We evaluate three strategies:
noisy information about the world. The POMDP framework 1) POMDP: UAVs fly at the same altitude but do not
is general enough to model a variety of real-world sequential exchange any information.
decision processes. An exact solution to a POMDP yields the 2) MS POMDP: UAVs fly at the same altitude and ex-
optimal action for each possible belief over the world states. change information when in communication range.
The optimal action maximizes (or minimizes) the expected 3) Hierarchical POMDP (HPOMDP): UAVs fly at different
reward (or cost) of the agent over a time horizon. The sequence altitudes and exchange information when in communi-
of optimal actions is known as the optimal policy and is chosen cation range.
so as to optimise the expected reward (or cost).
V. S IMULATIONS
Formally, a POMDP is represented by the following n-tuple:
{S, A, O, b0 , T.Ω, R, γ}, with: We consider 3D terrains discretized into cells of unit length.
In our first evaluations, we only consider trees as obstacles that
• S: finite set of discrete states,
we model in 3D space by cylinders (Fig. 5). Each cylinder is
• A: set of actions,
further discretized into an integer number of grid cells. In our
• O: set of observations,
model, we account for the fact that, depending on its position
• b0 : initial belief states distribution,
relative to the obstacles, a UAV may not be able to observe a
• T (s, a, s ): probability of transition from s to s when
� �
certain number of cells on the ground (shadowing effect).
taking action a,
We consider a sensing model accounting for false positive
• Ω(o, s , a): probability of observing o from state s after
� �
and false negative:
taking action a,
• Probability(sensing target at height h—target)=1 − βh
• R(s, a): reward when executing action a in state s,
145
Authorized licensed use limited to: UNIVERSITY OF BRISTOL. Downloaded on November 05,2024 at 00:16:48 UTC from IEEE Xplore. Restrictions apply.
!"#$%#&"'()*+%,%&'&()*+&,&-&./00+&1&#&.2/30.&
"
$ #
!
Fig. 6. Test configuration: top view of the test area with 2 UAVs at the
Fig. 5. Example of the effect of the presence of obstacles on the sensing bottom left corner, 5 trees and 1 goal (victim) centrally located.
range: a tree (cylinder) blocks a UAV field of view, and makes some areas
on the ground not observable.
146
Authorized licensed use limited to: UNIVERSITY OF BRISTOL. Downloaded on November 05,2024 at 00:16:48 UTC from IEEE Xplore. Restrictions apply.
Time to find the victim [in s] 70
60
50
40
30
20
10
0
Rand MS_Rand Cov MaxSum MaxMin MaxMax MinNeigh LA_MaxSum LA_MaxMin LA_MaxMax LA_MinNeigh Pot MS_Pot POMDP MS_POMDP HPOMDP
Search Strategy
Fig. 7. Simulation results showing the time to find the victim (completion time) with different search techniques.
147
Authorized licensed use limited to: UNIVERSITY OF BRISTOL. Downloaded on November 05,2024 at 00:16:48 UTC from IEEE Xplore. Restrictions apply.