
doi: 10.2139/ssrn.4430575
Unmanned aerial vehicle decision-making issues are increasingly being addressed using reinforcement learning (RL) (UAVs). The current advances in RL-based algorithms for UAV applications, encompassing both single-agent and swarm scenarios, are thoroughly reviewed in this work. First, the basic concepts of RL and its variants are introduced, followed by an overview of the state-of-the[1]art RL algorithms that have been applied to UAV navigation, path planning, and obstacle avoidance. The study then examines real-time learning concerns, model selection, and exploration-exploitation trade-offs, as well as challenges and potential for employing RL in UAV systems. In order to further the use of RL in UAVs, future research initiatives are also suggested. They include creating hybrid methods that integrate RL with other methodologies and incorporating human feedback and domain expertise into the learning process. Overall, this work demonstrates the potential of this approach to improve the autonomy, adaptability, and resilience of UAV systems and serves as a significant resource for researchers and those interested in applying RL to UAVs.
UAV, Aircraft Vehicle, Autonomy, Reinforcement
UAV, Aircraft Vehicle, Autonomy, Reinforcement
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 0 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
