On Passivity, Reinforcement Learning and Higher-Order Learning in Multi-Agent Finite Games

08/13/2018
by   Bolin Gao, et al.
0

In this paper, we propose a passivity-based methodology for analysis and design of reinforcement learning in multi-agent finite games. Starting from a known exponentially-discounted reinforcement learning scheme, we show that convergence to a Nash distribution can be shown in the class of games characterized by the monotonicity property of their (negative) payoff. We further exploit passivity to propose a class of higher-order schemes that preserve convergence properties, can improve the speed of convergence and can even converge in cases whereby their first-order counterpart fail to converge. We demonstrate these properties through numerical simulations for several representative games.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset