ABSTRACT
Unmanned Aerial Vehicles (UAVs) are widely used in search and rescue missions for unknown environments, where maximized coverage for unknown devices is required. This paper considers using collaborative UAVs (Col-UAV) to execute such tasks. It proposes to plan efficient trajectories for multiple UAVs to collaboratively maximize the number of devices to cover within minimized flying time. The proposed reinforcement learning (RL)-based Col-UAV scheme lets all UAVs share their traveling information by maintaining a common Q-table, which reduces the overall time and the memory complexities. We simulate the proposed RL Col-UAV scheme under various simulation environments with different grid sizes and compare the performance with other baselines. The simulation results show that the RL Col-UAVs scheme can find the optimal number of UAVs required to deploy for the diverse simulation environment and outperforms its counterparts in finding a maximum number of devices in a minimum time.
- J. Cui and et al., "Adaptive uav-trajectory optimization under quality of service constraints: A model-free solution," IEEE Access, 2020.Google Scholar
- K. Nguyen and et al., "3d uav trajectory and data collection optimisation via deep reinforcement learning," arXiv preprint, 2021.Google Scholar
- H. La and et al., "Multirobot cooperative learning for predator avoidance," IEEE Transactions on Control Systems Technology, 2014.Google Scholar
- H. Pham and et al., "Reinforcement learning for autonomous uav navigation using function approximation," in 2018 IEEE International Symposium on Safety, Security, and Rescue Robotics (SSRR), 2018.Google Scholar
- D. Ebrahimi and et al., "Autonomous uav trajectory for localizing ground objects: A reinforcement learning approach," IEEE Transactions on Mobile Computing, 2020.Google Scholar
- R. Hadidi and et al., "Distributed perception by collaborative robots," IEEE Robotics and Automation Letters, 2018.Google Scholar
- L. Kong and et al., "Adasharing: Adaptive data sharing in collaborative robots," IEEE Transactions on Industrial Electronics, 2017.Google Scholar
- D. Popescu and et al., "A collaborative uav-wsn network for monitoring large areas," Sensors, 2018.Google Scholar
- X. Zhu and et al., "Model of collaborative uav swarm toward coordination and control mechanisms study," Procedia Computer Science, 2015.Google Scholar
- Watkins and et al., "Q-learning," Machine learning, 1992.Google Scholar
- R. Sutton and et al., "Reinforcement learning: an introduction mit press," Cambridge, MA, 1998.Google Scholar
- M. Yi and et al., "Deep reinforcement learning for fresh data collection in uav-assisted iot networks," in INFOCOM WKSHPS, IEEE, 2020.Google Scholar
- Battocletti, Reinforcement Learning approach for cooperative UAVs exploration of critical environments. PhD thesis, Politecnico di Torino, 2021.Google Scholar
Index Terms
- A reinforcement learning-based path planning for collaborative UAVs
Recommendations
Reinforcement Learning for UAV Attitude Control
Autopilot systems are typically composed of an “inner loop” providing stability and control, whereas an “outer loop” is responsible for mission-level objectives, such as way-point navigation. Autopilot systems for unmanned aerial vehicles are ...
Real-time path planning of controllable UAV by subgoals using goal-conditioned reinforcement learning
AbstractThe conventional path planning problem for an unmanned aerial vehicle (UAV) typically involves a pre-defined environment and mission, with the objective of reaching a single target point. However, in order to perform different missions,...
Highlights- We propose a real-time autonomous path planning of UAVs in unknown environments.
A Service-Oriented Middleware for Building Collaborative UAVs
For a while, Unmanned Arial Vehicles (UAVs) use was limited to military applications, however recently UAVs are also used for a wide range of civilian applications. Some of these UAV applications may involve multiple UAVs that must cooperate to achieve ...
Comments