Multi-Agent Dynamic Area Coverage Based on Reinforcement Learning with Connected Agents


Creative Commons License

Aydemir F., ÇETİN A.

COMPUTER SYSTEMS SCIENCE AND ENGINEERING, cilt.45, sa.1, ss.215-230, 2023 (SCI-Expanded) identifier identifier

  • Yayın Türü: Makale / Tam Makale
  • Cilt numarası: 45 Sayı: 1
  • Basım Tarihi: 2023
  • Doi Numarası: 10.32604/csse.2023.031116
  • Dergi Adı: COMPUTER SYSTEMS SCIENCE AND ENGINEERING
  • Derginin Tarandığı İndeksler: Science Citation Index Expanded (SCI-EXPANDED), Scopus, PASCAL, Aerospace Database, Applied Science & Technology Source, Communication Abstracts, Computer & Applied Sciences, Metadex, zbMATH, Civil Engineering Abstracts
  • Sayfa Sayıları: ss.215-230
  • Anahtar Kelimeler: Dynamic environments, multi-agent reinforcement learning, dynamic area coverage, ALGORITHM, BEHAVIOR
  • Gazi Üniversitesi Adresli: Evet

Özet

Dynamic area coverage with small unmanned aerial vehicle (UAV) systems is one of the major research topics due to limited payloads and the difficulty of decentralized decision-making process. Collaborative behavior of a group of UAVs in an unknown environment is another hard problem to be solved. In this paper, we propose a method for decentralized execution of multi-UAVs for dynamic area coverage problems. The proposed decentralized decision-making dynamic area coverage (DDMDAC) method utilizes reinforcement learning (RL) where each UAV is represented by an intelligent agent that learns policies to create collaborative behaviors in partially observable environment. Intelligent agents increase their global observations by gathering information about the environment by connecting with other agents. The connectivity provides a consensus for the decision-making process, while each agent takes decisions. At each step, agents acquire all reachable agents' states, determine the optimum location for maximal area coverage and receive reward using the covered rate on the target area, respectively. The method was tested in a multi-agent actor-critic simulation platform. In the study, it has been considered that each UAV has a certain communication distance as in real applications. The results show that UAVs with limited communication distance can act jointly in the target area and can successfully cover the area without guidance from the central command unit.