Abstract
Morton and Wecker (1977) stated that the value iteration algorithm solves a dynamic program's policy function faster than its value function when the limiting Markov chain is ergodic. I show that their proof is incomplete, and provide a new proof of this classic result. I use this result to accelerate the estimation of Markov decision processes and the solution of Markov perfect equilibria.
Original language | English (US) |
---|---|
Pages (from-to) | 43-65 |
Number of pages | 23 |
Journal | Quantitative Economics |
Volume | 10 |
Issue number | 1 |
DOIs | |
State | Published - Jan 2019 |
Keywords
- C01
- C13
- C15
- C61
- C63
- C65
- Markov decision process
- Markov perfect equilibrium
- dynamic discrete choice
- nested fixed point
- nested pseudo-likelihood
- relative value iteration
- strong convergence
ASJC Scopus subject areas
- Economics and Econometrics