Title
Learning to Cooperate: Application of Deep Reinforcement Learning for Online AGV Path Finding
Abstract
Multi-agent path finding (MAPF), naturally exists in applications like picking-up and dropping-off parcels by automated guided vehicles (AGVs) in the warehouse. Existing algorithms, like conflict-based search (CBS), windowed hierarchical cooperative A* (WHCA), and other A* variants, are widely used to find the shortest paths in different manners. However, in real-world environments, MAPF cases are dynamically generated and need to be solved in real time. In this work, a decentralized multi-agent reinforcement learning (MARL) framework with multi-step ahead tree search (MATS) strategy is proposed to make efficient decisions. Through performing experiments on a 30*30 grid world and a real-world warehouse case, our proposed MARL policy is proved to be capable of: 1) scaling to a large number of agents in real-world environment with online response time within acceptable levels; 2) outperforming existing algorithms with shorter path length and solution time, as the number of agents increases.
Year
DOI
Venue
2020
10.5555/3398761.3399080
AAMAS '19: International Conference on Autonomous Agents and Multiagent Systems Auckland New Zealand May, 2020
DocType
ISBN
Citations 
Conference
978-1-4503-7518-4
0
PageRank 
References 
Authors
0.34
0
5
Name
Order
Citations
PageRank
Yi Zhang100.34
Qian Yu277.22
Yichen Yao300.68
Haoyuan Hu401.35
Yinghui Xu517220.23