Book contents
- Frontmatter
- Contents
- List of contributors
- Preface
- 1 Inference and estimation in probabilistic time series models
- I Monte Carlo
- II Deterministic approximations
- III Switching models
- IV Multi-object models
- V Nonparametric models
- VI Agent-based models
- 17 Optimal control theory and the linear Bellman equation
- 18 Expectation maximisation methods for solving (PO)MDPs and optimal control problems
- Index
- Plate section
- References
18 - Expectation maximisation methods for solving (PO)MDPs and optimal control problems
from VI - Agent-based models
Published online by Cambridge University Press: 07 September 2011
- Frontmatter
- Contents
- List of contributors
- Preface
- 1 Inference and estimation in probabilistic time series models
- I Monte Carlo
- II Deterministic approximations
- III Switching models
- IV Multi-object models
- V Nonparametric models
- VI Agent-based models
- 17 Optimal control theory and the linear Bellman equation
- 18 Expectation maximisation methods for solving (PO)MDPs and optimal control problems
- Index
- Plate section
- References
Summary
Introduction
As this book demonstrates, the development of efficient probabilistic inference techniques has made considerable progress in recent years, in particular with respect to exploiting the structure (e.g., factored, hierarchical or relational) of discrete and continuous problem domains. In this chapter we show that these techniques can be used also for solving Markov decision processes (MDPs) or partially observable MDPs (POMDPs) when formulated in terms of a structured dynamic Bayesian network (DBN).
The problems of planning in stochastic environments and inference in state space models are closely related, in particular in view of the challenges both of them face: scaling to large state spaces spanned by multiple state variables, or realising planning (or inference) in continuous or mixed continuous-discrete state spaces. Both fields developed techniques to address these problems. For instance, in the field of planning, they include work on factored Markov decision processes [5, 17, 9, 18], abstractions [10], and relational models of the environment [37]. On the other hand, recent advances in inference techniques show how structure can be exploited both for exact inference as well as for making efficient approximations. Examples are message-passing algorithms (loopy belief propagation, expectation propagation), variational approaches, approximate belief representations (particles, assumed density filtering, Boyen–Koller) and arithmetic compilation (see, e.g., [22, 23, 7]).
In view of these similarities one may ask whether existing techniques for probabilistic inference can directly be translated to solving stochastic planning problems.
- Type
- Chapter
- Information
- Bayesian Time Series Models , pp. 388 - 413Publisher: Cambridge University PressPrint publication year: 2011
References
- 2
- Cited by