Abstract | ||
---|---|---|
The principle of service composition based on multi-agent is that multi-agent can coordinate to reach Pareto-optimal Nash equilibrium. Reinforcement learning algorithms can be used to deal with the coordination problem in cooperative games. In this paper, the multi-agent coordination problems in cooperative games for different user preference is investigated. In our case, each agent can represent a user’s preference, and it finally learns a policy that is best fit for that user. Most previous works study the deterministic gain of a state. However, in practical service environments, the gain may be nondeterministic due to unstable Quality of Service (QoS). In addition, user preference should be considered. To avoid local optimal solution, we let each agent randomly change interacting partners in each iteration. Thus, an agent can learn its optimal strategy by interacting repeatedly with the rest of agents representing different user preference. The experimental results show that our reinforcement learning algorithm outperforms other learning methods. |
Year | DOI | Venue |
---|---|---|
2017 | 10.1016/j.future.2016.06.039 | Future Generation Computer Systems |
Keywords | Field | DocType |
Distributed artificial intelligence,Cooperative games,Multi-agent coordination,QoS,Reinforcement learning | Coordination game,Nondeterministic algorithm,Computer science,Quality of service,Service composition,Artificial intelligence,Nash equilibrium,Reinforcement learning algorithm,Distributed computing,Reinforcement learning | Journal |
Volume | ISSN | Citations |
68 | 0167-739X | 2 |
PageRank | References | Authors |
0.36 | 10 | 2 |
Name | Order | Citations | PageRank |
---|---|---|---|
Lei Yu | 1 | 21 | 12.87 |
Junxing Zhang | 2 | 137 | 13.64 |