Deep reinforcement learning in a monetary model

Staff working papers set out research in progress by our staff, with the aim of encouraging comments and debate.
Published on 26 September 2025

Staff Working Paper No. 1,142

By Mingli Chen, Rama Cont, Andreas Joseph, Michael Kumhof, Xinlei Pan, Wei Xiong and Xuan Zhou

We propose deep reinforcement learning (DRL) as a general approach to bounded rationality in dynamic stochastic general equilibrium (DSGE) models. Agents are represented by deep artificial neural networks and learn to maximise their intertemporal objective function by interacting with an a priori unknown environment. Applying this approach to a model from the adaptive learning literature, DRL agents can learn all equilibria irrespective of local stability properties. However, learning is slow and may be unstable without the imposition of early stopping criteria. These findings can have implications for the use and interpretation of DRL agents and of DSGE models more generally.

Deep reinforcement learning in a monetary model