
arXiv: 2308.09733
handle: 1959.4/unsworks_61809
Multi-objective Markov decision processes are sequential decision-making problems that involve multiple conflicting reward functions that cannot be optimized simultaneously without a compromise. This type of problems cannot be solved by a single optimal policy as in the conventional case. Alternatively, multi-objective reinforcement learning methods evolve a coverage set of optimal policies that can satisfy all possible preferences in solving the problem. However, many of these methods cannot generalize their coverage sets to work in non-stationary environments. In these environments, the parameters of the state transition and reward distribution vary over time. This limitation results in significant performance degradation for the evolved policy sets. In order to overcome this limitation, there is a need to learn a generic skill set that can bootstrap the evolution of the policy coverage set for each shift in the environment dynamics therefore, it can facilitate a continuous learning process. In this work, intrinsically motivated reinforcement learning has been successfully deployed to evolve generic skill sets for learning hierarchical policies to solve multi-objective Markov decision processes. We propose a novel dual-phase intrinsically motivated reinforcement learning method to address this limitation. In the first phase, a generic set of skills is learned. While in the second phase, this set is used to bootstrap policy coverage sets for each shift in the environment dynamics. We show experimentally that the proposed method significantly outperforms state-of-the-art multi-objective reinforcement methods in a dynamic robotics environment.
FOS: Computer and information sciences, anzsrc-for: 4611 Machine Learning, Computer Science - Machine Learning, 4 Quality Education, Computer Science - Artificial Intelligence, anzsrc-for: 46 Information and Computing Sciences, anzsrc-for: 4602 Artificial Intelligence, Basic Behavioral and Social Science, 004, Machine Learning (cs.LG), Computer Science - Robotics, Artificial Intelligence (cs.AI), 46 Information and Computing Sciences, 4602 Artificial Intelligence, mechatronics and robotics, 4611 Machine Learning, Behavioral and Social Science, Robotics (cs.RO), anzsrc-for: 4007 Control engineering
FOS: Computer and information sciences, anzsrc-for: 4611 Machine Learning, Computer Science - Machine Learning, 4 Quality Education, Computer Science - Artificial Intelligence, anzsrc-for: 46 Information and Computing Sciences, anzsrc-for: 4602 Artificial Intelligence, Basic Behavioral and Social Science, 004, Machine Learning (cs.LG), Computer Science - Robotics, Artificial Intelligence (cs.AI), 46 Information and Computing Sciences, 4602 Artificial Intelligence, mechatronics and robotics, 4611 Machine Learning, Behavioral and Social Science, Robotics (cs.RO), anzsrc-for: 4007 Control engineering
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 2 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
