
doi: 10.7151/dmps.1034
In stochastic dynamic programming problems, the transition function usually depends on the current state, current action, and random factors, referred in this paper as disturbances. For Markov decision processes, the disturbances form a sequence of independent random values. This paper studies stochastic dynamic programming problems when the decision-maker knows the value of a disturbance at the current state before he or she selects a decision. This paper is motivated by applications to transportation problems.
Markov and semi-Markov decision processes, certainty equivalence principle, Stochastic programming, distance property, stochastic dynamic programming, Dynamic programming, dominant policy, Markov decision processes
Markov and semi-Markov decision processes, certainty equivalence principle, Stochastic programming, distance property, stochastic dynamic programming, Dynamic programming, dominant policy, Markov decision processes
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 0 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
