Title
Historical Best Q-Networks for Deep Reinforcement Learning
Abstract
The popular DQN algorithm is known to have some instability and variability which make its performance poor sometimes. In prior work, there is only one target network, the network that is updated by the latest learned Q-value estimate. In this paper, we present multiple target networks which are the extension to the Deep Q-Networks (DQN). Based on the previously learned Q-value estimate networks, we choose several networks that perform best in all previous networks as our auxiliary networks. We show that in order to solve the problem of determining which network is better, we use the score of each episode as a measure of the quality of the network. The key behind our method is that each auxiliary network has some states that it is good at handling and guides the agent to make the right choices. We apply our method to the Atari 2600 games from the OpenAI Gym. We find that DQN with auxiliary networks significantly improves the performance and the stability of games.
Year
DOI
Venue
2018
10.1109/ICTAI.2018.00012
2018 IEEE 30th International Conference on Tools with Artificial Intelligence (ICTAI)
Keywords
Field
DocType
Deep Reinforcement Learning,Multiple Target Networks,Deep Q-Networks (DQN),Deep Learning
Computer science,Feature extraction,Artificial intelligence,Artificial neural network,Machine learning,Reinforcement learning
Conference
ISSN
ISBN
Citations 
1082-3409
978-1-5386-7450-5
2
PageRank 
References 
Authors
0.41
4
5
Name
Order
Citations
PageRank
Wenwu Yu14340185.95
Rui Wang213953.65
Ruiying Li320.41
Jing Gao4216.58
Xiaohui Hu5178.10