
This study proposes an autonomous aircraft taxi-agent that can be used to recommend the pilot the optimal speed profile to achieve optimal fuel burn and to arrive on time at the target position on the taxiway while considering potential interactions with surrounding traffic. The problem is modeled as a control decision problem which is solved by training the agent under a Deep Reinforcement Learning (DRL) mechanism, using Proximal Policy Optimization (PPO) algorithm. The reward function is designed to consider the fuel burn, taxi-time, and delay-time. Thus, the trained agent will learn to taxi the aircraft between any pair of locations on the airport surface timely while maintaining safety and efficiency. As the result, in more than 97.8% of the evaluated sessions, the controlled aircraft can reach the target position with the time difference within the range of [-20,5] seconds. Moreover, compared with actual fuel burn, the proposed autonomous taxi-agent demonstrated a reduction of 29.5%, equivalent to the reduction of 13.9 kg of fuel per aircraft. This benefit in fuel burn reduction can complement the emission reductions achieved by solving other sub-problems, such as pushback control and taxi-route assignments to achieve much higher performance. Civil Aviation Authority of Singapore (CAAS) National Research Foundation (NRF) Submitted/Accepted version This research is supported by the National Research Foundation, Singapore, and the Civil Aviation Authority of Singapore, under the Aviation Transformation Programme.
:Civil engineering::Transportation [Engineering], :Computer science and engineering::Computing methodologies::Artificial intelligence [Engineering], Fuel Burn, Autonomous Taxi, :Computer science and engineering::Computing methodologies::Simulation and modeling [Engineering], Reinforcement Learning, Optimal Speed
:Civil engineering::Transportation [Engineering], :Computer science and engineering::Computing methodologies::Artificial intelligence [Engineering], Fuel Burn, Autonomous Taxi, :Computer science and engineering::Computing methodologies::Simulation and modeling [Engineering], Reinforcement Learning, Optimal Speed
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 0 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
