Title
Exploration by Random Network Distillation.
Abstract
We introduce an exploration bonus for deep reinforcement learning methods that is easy to implement and adds minimal overhead to the computation performed. The bonus is the error of a neural network predicting features of the observations given by a fixed randomly initialized neural network. We also introduce a method to flexibly combine intrinsic and extrinsic rewards. We find that the random network distillation (RND) bonus combined with this increased flexibility enables significant progress on several hard exploration Atari games. In particular we establish state of the art performance on Montezumau0027s Revenge, a game famously difficult for deep reinforcement learning methods. To the best of our knowledge, this is the first method that achieves better than average human performance on this game without using demonstrations or having access to the underlying state of the game, and occasionally completes the first level.
Year
Venue
Field
2018
international conference on learning representations
Random graph,Distillation,Artificial intelligence,Artificial neural network,Machine learning,Mathematics,Reinforcement learning,Computation
DocType
Volume
Citations 
Journal
abs/1810.12894
24
PageRank 
References 
Authors
0.62
31
4
Name
Order
Citations
PageRank
Yuri Burda1934.47
Harrison A Edwards21517.12
Amos J. Storkey395594.20
Oleg Klimov41153.60