Markov Decision Processes with Exogenous Variables

28 Pages Posted: 21 Feb 2017 Last revised: 6 Feb 2018

Date Written: February 5, 2018

Abstract

I present two algorithms for solving dynamic programs with exogenous variables: endogenous value iteration and endogenous policy iteration. These algorithms are like relative value iteration and relative policy iteration, except they discard the variation in the value function due solely to the exogenous variables (this variation doesn't affect the policy function). My algorithms are always at least as fast as relative value iteration and relative policy iteration, and are faster when the endogenous variables converge to their stationary distributions faster than the exogenous variables.

Keywords: Markov Decision Process; Dynamic Programming; Relative Value Iteration; Strong Convergence; Exogenous Variables

Suggested Citation

Bray, Robert, Markov Decision Processes with Exogenous Variables (February 5, 2018). Available at SSRN: https://ssrn.com/abstract=2919946 or http://dx.doi.org/10.2139/ssrn.2919946

Robert Bray (Contact Author)

Kellogg School of Management ( email )

2001 Sheridan Road
Evanston, IL 60208
United States

Do you have negative results from your research you’d like to share?

Paper statistics

Downloads
364
Abstract Views
1,821
Rank
142,699
PlumX Metrics