Abstract | ||
---|---|---|
Distributed agents concurrently learning to coordinate in a multiagent system can suffer from considerable amounts of agent noise. This is the noise that arises from the non-stationarity of the learning environment for each individual agent since other agents in the system are also constantly updating their policies, thereby continually shifting the goal posts for successful coordination. In this work, we propose a method to reduce agent noise by allowing individual agents to probabilistically determine whether or not to undergo policy updates. We show that using this method to adapt the number of actively learning agents over time provides improvements in convergence speed of the team as a whole without affecting the final converged learning performance. |
Year | DOI | Venue |
---|---|---|
2018 | 10.5555/3237383.3238017 | PROCEEDINGS OF THE 17TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS (AAMAS' 18) |
Keywords | Field | DocType |
Multiagent Learning, Agent Noise, Reasoning about Action | Convergence (routing),Computer science,Multiagent learning,Learning environment,Distributed computing | Conference |
Citations | PageRank | References |
0 | 0.34 | 0 |
Authors | ||
3 |
Name | Order | Citations | PageRank |
---|---|---|---|
Jen Jen Chung | 1 | 21 | 9.92 |
Scott Chow | 2 | 3 | 1.21 |
kagan tumer | 3 | 1632 | 168.61 |