TY - JOUR
T1 - Distributed Reinforcement Learning for Flexible and Efficient UAV Swarm Control
AU - Venturini, Federico
AU - Mason, Federico
AU - Pase, Francesco
AU - Chiariotti, Federico
AU - Testolin, Alberto
AU - Zanella, Andrea
AU - Zorzi, Michele
PY - 2021/9
Y1 - 2021/9
N2 - Over the past few years, the use of swarms of Unmanned Aerial Vehicles (UAVs) in monitoring and remote area surveillance applications has become widespread thanks to the price reduction and the increased capabilities of drones. The drones in the swarm need to cooperatively explore an unknown area, in order to identify and monitor interesting targets, while minimizing their movements. In this work, we propose a distributed Reinforcement Learning (RL) approach that scales to larger swarms without modifications. The proposed framework relies on the possibility for the UAVs to exchange some information through a communication channel, in order to achieve context-awareness and implicitly coordinate the swarm's actions. Our experiments show that the proposed method can yield effective strategies, which are robust to communication channel impairments, and that can easily deal with non-uniform distributions of targets and obstacles. Moreover, when agents are trained in a specific scenario, they can adapt to a new one with minimal additional training. We also show that our approach achieves better performance compared to a computationally intensive look-ahead heuristic.
AB - Over the past few years, the use of swarms of Unmanned Aerial Vehicles (UAVs) in monitoring and remote area surveillance applications has become widespread thanks to the price reduction and the increased capabilities of drones. The drones in the swarm need to cooperatively explore an unknown area, in order to identify and monitor interesting targets, while minimizing their movements. In this work, we propose a distributed Reinforcement Learning (RL) approach that scales to larger swarms without modifications. The proposed framework relies on the possibility for the UAVs to exchange some information through a communication channel, in order to achieve context-awareness and implicitly coordinate the swarm's actions. Our experiments show that the proposed method can yield effective strategies, which are robust to communication channel impairments, and that can easily deal with non-uniform distributions of targets and obstacles. Moreover, when agents are trained in a specific scenario, they can adapt to a new one with minimal additional training. We also show that our approach achieves better performance compared to a computationally intensive look-ahead heuristic.
KW - Artificial intelligence
KW - Drones
KW - Reinforcement learning
KW - Sensors
KW - Surveillance
KW - Training
KW - Transfer learning
KW - Wireless communication
KW - distributed decision making
KW - mobile robots
KW - neural network applications.
UR - http://www.scopus.com/inward/record.url?scp=85102309125&partnerID=8YFLogxK
U2 - 10.1109/TCCN.2021.3063170
DO - 10.1109/TCCN.2021.3063170
M3 - Journal article
SN - 2332-7731
VL - 7
SP - 955
EP - 969
JO - IEEE Transactions on Cognitive Communications and Networking
JF - IEEE Transactions on Cognitive Communications and Networking
IS - 3
ER -