Title
InfoBot: Transfer and Exploration via the Information Bottleneck.
Abstract
A central challenge in reinforcement learning is discovering effective policies for tasks where rewards are sparsely distributed. We postulate that in the absence of useful reward signals, an effective exploration strategy should seek out {it decision states}. These states lie at critical junctions in the state space from where the agent can transition to new, potentially unexplored regions. We propose to learn about decision states from prior experience. By training a goal-conditioned policy with an information bottleneck, we can identify decision states by examining where the model actually leverages the goal state. We find that this simple mechanism effectively identifies decision states, even in partially observed settings. In effect, the model learns the sensory cues that correlate with potential subgoals. In new environments, this model can then identify novel subgoals for further exploration, guiding the agent through a sequence of potential decision states and through new regions of the state space.
Year
Venue
Field
2019
International Conference on Learning Representations
Sensory cue,Artificial intelligence,Information bottleneck method,State space,Mathematics,Machine learning,Reinforcement learning
DocType
Volume
Citations 
Journal
abs/1901.10902
4
PageRank 
References 
Authors
0.38
26
8
Name
Order
Citations
PageRank
Anirudh Goyal126420.97
Riashat Islam21628.27
Daniel Strouse340.38
Zafarali Ahmed450.73
Matthew M Botvinick549425.34
Hugo Larochelle67692488.99
Sergey Levine73377182.21
Yoshua Bengio8426773039.83