Title
A multiagent reinforcement learning algorithm using extended optimal response
Abstract
Stochastic games provides a theoretical framework to multiagent reinforcement learning. Based on the framework, a multiagent reinforcement learning algorithm for zero-sum stochastic games was proposed by Littman and it was extended to general-sum games by Hu and Wellman. Given a stochastic game, if all agents learn with their algorithm, we can expect that the policies of the agents converge to a Nash equilibrium. However, agents with their algorithm always try to converge to a Nash equilibrium independent of the policies used by the other agents. In addition, in case there are multiple Nash equilibria, agents must agree on the equilibrium where they want to reach. Thus, their algorithm lacks adaptability in a sense. In this paper, we propose a multiagent reinforcement learning algorithm. The algorithm uses the extended optimal response which we introduce in this paper. It will converge to a Nash equilibrium when other agents are adaptable, otherwise it will make an optimal response. We also provide some empirical results in three simple stochastic games, which show that the algorithm can realize what we intend.
Year
DOI
Venue
2002
10.1145/544741.544831
AAMAS
Keywords
DocType
ISBN
stochastic game,extended optimal response,optimal response,nash equilibrium,theoretical framework,multiagent reinforcement,zero-sum stochastic game,simple stochastic game,reinforcement learning,multiple nash equilibrium,nash equilibria,q learning
Conference
1-58113-480-0
Citations 
PageRank 
References 
22
1.22
8
Authors
2
Name
Order
Citations
PageRank
Nobuo Suematsu1548.99
Akira Hayashi2519.08