Abstract
Markov Decision Processes (MDPs) are extensively
used to encode sequences of decisions with probabilistic
effects. Markov Decision Processes with Imprecise
Probabilities (MDPIPs) encode sequences of decisions
whose effects are modeled using sets of probability distributions.
In this paper we examine the computation
of Γ-maximin policies for MDPIPs using multilinear
and integer programming. We discuss the application
of our algorithms to “factored” models and to a recent
proposal, Markov Decision Processes with Set-valued
Transitions (MDPSTs), that unifies the fields of probabilistic
and “nondeterministic” planning in artificial
intelligence research.
Original language | English |
---|---|
Title of host publication | Proceedings of the 5th International Symposium on Imprecise Probability: Theories and Applications |
Pages | 395-404 |
Number of pages | 10 |
Publication status | Published - 2007 |
Event | The 5th International Symposium on Imprecise Probability: Theories and Applications - Prague, Czech Republic Duration: 16 Jul 2007 → 19 Jul 2007 |
Conference
Conference | The 5th International Symposium on Imprecise Probability: Theories and Applications |
---|---|
Country/Territory | Czech Republic |
City | Prague |
Period | 16/07/2007 → 19/07/2007 |