skip to main content
10.1145/3477314.3507052acmconferencesArticle/Chapter ViewAbstractPublication PagessacConference Proceedingsconference-collections
research-article

A reinforcement learning-based path planning for collaborative UAVs

Published:06 May 2022Publication History

ABSTRACT

Unmanned Aerial Vehicles (UAVs) are widely used in search and rescue missions for unknown environments, where maximized coverage for unknown devices is required. This paper considers using collaborative UAVs (Col-UAV) to execute such tasks. It proposes to plan efficient trajectories for multiple UAVs to collaboratively maximize the number of devices to cover within minimized flying time. The proposed reinforcement learning (RL)-based Col-UAV scheme lets all UAVs share their traveling information by maintaining a common Q-table, which reduces the overall time and the memory complexities. We simulate the proposed RL Col-UAV scheme under various simulation environments with different grid sizes and compare the performance with other baselines. The simulation results show that the RL Col-UAVs scheme can find the optimal number of UAVs required to deploy for the diverse simulation environment and outperforms its counterparts in finding a maximum number of devices in a minimum time.

References

  1. J. Cui and et al., "Adaptive uav-trajectory optimization under quality of service constraints: A model-free solution," IEEE Access, 2020.Google ScholarGoogle Scholar
  2. K. Nguyen and et al., "3d uav trajectory and data collection optimisation via deep reinforcement learning," arXiv preprint, 2021.Google ScholarGoogle Scholar
  3. H. La and et al., "Multirobot cooperative learning for predator avoidance," IEEE Transactions on Control Systems Technology, 2014.Google ScholarGoogle Scholar
  4. H. Pham and et al., "Reinforcement learning for autonomous uav navigation using function approximation," in 2018 IEEE International Symposium on Safety, Security, and Rescue Robotics (SSRR), 2018.Google ScholarGoogle Scholar
  5. D. Ebrahimi and et al., "Autonomous uav trajectory for localizing ground objects: A reinforcement learning approach," IEEE Transactions on Mobile Computing, 2020.Google ScholarGoogle Scholar
  6. R. Hadidi and et al., "Distributed perception by collaborative robots," IEEE Robotics and Automation Letters, 2018.Google ScholarGoogle Scholar
  7. L. Kong and et al., "Adasharing: Adaptive data sharing in collaborative robots," IEEE Transactions on Industrial Electronics, 2017.Google ScholarGoogle Scholar
  8. D. Popescu and et al., "A collaborative uav-wsn network for monitoring large areas," Sensors, 2018.Google ScholarGoogle Scholar
  9. X. Zhu and et al., "Model of collaborative uav swarm toward coordination and control mechanisms study," Procedia Computer Science, 2015.Google ScholarGoogle Scholar
  10. Watkins and et al., "Q-learning," Machine learning, 1992.Google ScholarGoogle Scholar
  11. R. Sutton and et al., "Reinforcement learning: an introduction mit press," Cambridge, MA, 1998.Google ScholarGoogle Scholar
  12. M. Yi and et al., "Deep reinforcement learning for fresh data collection in uav-assisted iot networks," in INFOCOM WKSHPS, IEEE, 2020.Google ScholarGoogle Scholar
  13. Battocletti, Reinforcement Learning approach for cooperative UAVs exploration of critical environments. PhD thesis, Politecnico di Torino, 2021.Google ScholarGoogle Scholar

Index Terms

  1. A reinforcement learning-based path planning for collaborative UAVs
      Index terms have been assigned to the content through auto-classification.

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Conferences
        SAC '22: Proceedings of the 37th ACM/SIGAPP Symposium on Applied Computing
        April 2022
        2099 pages
        ISBN:9781450387132
        DOI:10.1145/3477314

        Copyright © 2022 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 6 May 2022

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article

        Acceptance Rates

        Overall Acceptance Rate1,650of6,669submissions,25%
      • Article Metrics

        • Downloads (Last 12 months)99
        • Downloads (Last 6 weeks)10

        Other Metrics

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader