Title
Massively Parallel Methods for Deep Reinforcement Learning
Abstract
We present the first massively distributed architecture for deep reinforcement learning. This architecture uses four main components: parallel actors that generate new behaviour; parallel learners that are trained from stored experience; a distributed neural network to represent the value function or behaviour policy; and a distributed store of experience. We used our architecture to implement the Deep Q-Network algorithm (DQN). Our distributed algorithm was applied to 49 games from Atari 2600 games from the Arcade Learning Environment, using identical hyperparameters. Our performance surpassed non-distributed DQN in 41 of the 49 games and also reduced the wall-time required to achieve these results by an order of magnitude on most games.
Year
Venue
Field
2015
CoRR
Architecture,Hyperparameter,Massively parallel,Computer science,Bellman equation,Distributed algorithm,Artificial intelligence,Learning environment,Artificial neural network,Machine learning,Reinforcement learning
DocType
Volume
Citations 
Journal
abs/1507.04296
83
PageRank 
References 
Authors
5.85
9
14
Name
Order
Citations
PageRank
arun nair1835.85
praveen srinivasan2835.85
sam blackwell3835.85
cagdas alcicek4835.85
rory fearon5835.85
alessandro de maria6835.85
vedavyas panneershelvam7835.85
mustafa suleyman862824.43
Charles Beattie9232995.49
Stig Petersen10232995.83
shane legg11835.85
volodymyr mnih12835.85
Koray Kavukcuoglu1310189504.11
David Silver148252363.86