Title
ACtuAL: Actor-Critic Under Adversarial Learning.
Abstract
Generative Adversarial Networks (GANs) are a powerful framework for deep generative modeling. Posed as a two-player minimax problem, GANs are typically trained end-to-end on real-valued data and can be used to train a generator of high-dimensional and realistic images. However, a major limitation of GANs is that training relies on passing gradients from the discriminator through the generator via back-propagation. This makes it fundamentally difficult to train GANs with discrete data, as generation in this case typically involves a non-differentiable function. These difficulties extend to the reinforcement learning setting when the action space is composed of discrete decisions. We address these issues by reframing the GAN framework so that the generator is no longer trained using gradients through the discriminator, but is instead trained using a learned critic in the actor-critic framework with a Temporal Difference (TD) objective. This is a natural fit for sequence modeling and we use it to achieve improvements on language modeling tasks over the standard Teacher-Forcing methods.
Year
Venue
Field
2017
arXiv: Machine Learning
Temporal difference learning,Cognitive reframing,Discriminator,Computer science,Generative modeling,Artificial intelligence,Generative grammar,Machine learning,Adversarial system,Minimax problem,Reinforcement learning
DocType
Volume
Citations 
Journal
abs/1711.04755
1
PageRank 
References 
Authors
0.34
25
7
Name
Order
Citations
PageRank
Anirudh Goyal126420.97
Nan Rosemary Ke214013.74
Alex Lamb326818.84
R Devon Hjelm413513.28
Chris Pal52140106.53
Joelle Pineau62857184.18
Yoshua Bengio7426773039.83