Title
Learning to Learn without Gradient Descent by Gradient Descent.
Abstract
We learn recurrent neural network optimizers trained on simple synthetic functions by gradient descent. We show that these learned optimizers exhibit a remarkable degree of transfer in that they can be used to efficiently optimize a broad range of derivative-free black-box functions, including Gaussian process bandits, simple control objectives, global optimization benchmarks and hyper-parameter tuning tasks. Up to the training horizon, the learned optimizers learn to tradeoff exploration and exploitation, and compare favourably with heavily engineered Bayesian optimization packages for hyper-parameter tuning.
Year
Venue
Field
2017
ICML
Gradient descent,Global optimization,Computer science,Bayesian optimization,Recurrent neural network,Descent direction,Gaussian process,Artificial intelligence,Machine learning,Learning to learn
DocType
Citations 
PageRank 
Conference
26
0.95
References 
Authors
25
7
Name
Order
Citations
PageRank
Yutian Chen168036.28
Matt Hoffman222714.27
Sergio Gomez Colmenarejo3422.43
Misha Denil439726.18
Timothy P. Lillicrap54377170.65
Matthew M Botvinick649425.34
Nando De Freitas73284273.68