
A methodology is presented in this paper for stochastic optimal control of unmanned aerial vehicle performing the task of perimeter patrol. The optimal control problem is modeled as a Markov decision processes, and an approximate policy iteration algorithm is used for the cost-to-go function (value function) by introducing Gaussian process regression, resulting in improved quality of the decisions made while retaining computationally feasibility. The approximate dynamic programming (ADP) framework is developed to tackle the issues, in which situations standard dynamic programming algorithms become computationally too demanding. As a nonparametric ADP algorithm, the Gaussian processes that provide the combination of the prior and noise models presents a sub-solution in a lower dimensional space by exploiting kernel-based method. The numerical results that corroborate the effectiveness of the proposed methodology are also provided.
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 0 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
