A General Theory of Markovian Time Inconsistent Stochastic Control Problems

Tomas Björk, Agatha Murgochi

Research output: Working paperResearch

Abstract

We develop a theory for stochastic control problems which, in various ways, are time inconsistent in the sense that they do not admit a Bellman optimality principle. We attach these problems by viewing them within a game theoretic framework, and we look for Nash subgame perfect equilibrium points. For a general controlled Markov process and a fairly general objective functional we derive an extension of the standard Hamilton-Jacobi-Bellman equation, in the form of a system of on-linear equations, for the determination for the equilibrium strategy as well as the equilibrium value function. All known examples of time inconsistency in the literature are easily seen to be special cases of the present theory. We also prove that for every time inconsistent problem, there exists an associated time consistent problem such that the optimal control and the optimal value function for the consistent problem coincides with the equilibrium control and value function respectively for the time inconsistent problem. We also study some concrete examples.
Original languageEnglish
Place of Publicationwww
PublisherSSRN: Social Science Research Network
Edition2
Number of pages55
Publication statusPublished - 17 Sep 2010

Keywords

  • Time Consistency
  • Time Inconsistent Control
  • Dynamic Programming
  • Time Inconsistency
  • Stochastic Control
  • Hyperbolic Discounting
  • Meanvariance
  • Bellman Equation
  • Hamilton-Jacobi-Bellman

Cite this