Title
NerveNet: Learning Structured Policy with Graph Neural Networks
Abstract
We address the problem of learning structured policies for continuous control. In traditional reinforcement learning, policies of agents are learned by MLPs which take the concatenation of all observations from the environment as input for predicting actions. In this work, we propose NerveNet to explicitly model the structure of an agent, which naturally takes the form of a graph. Specifically, serving as the agentu0027s policy network, NerveNet first propagates information over the structure of the agent and then predict actions for different parts of the agent. In the experiments, we first show that our NerveNet is comparable to state-of-the-art methods on standard MuJoCo environments. We further propose our customized reinforcement learning environments for benchmarking two types of structure transfer learning tasks, i.e., size and disability transfer. We demonstrate that policies learned by NerveNet are significantly better than policies learned by other models and are able to transfer even in a zero-shot setting.
Year
Venue
Field
2018
international conference on learning representations
Graph,Computer science,Transfer of learning,Graph neural networks,Artificial intelligence,Concatenation,Benchmarking,Machine learning,Reinforcement learning
DocType
Citations 
PageRank 
Conference
13
0.50
References 
Authors
21
4
Name
Order
Citations
PageRank
Tingwu Wang1142.20
Renjie Liao2765.59
Lei Jimmy Ba38887296.55
Sanja Fidler42087116.71