Abstract | ||
---|---|---|
We introduce a hierarchical model for efficient placement of computational graphs onto hardware devices, especially in heterogeneous environments with a mixture of CPUs, GPUs, and other computational devices. The algorithm learns to assign graph operations to groups and to allocate those groups to available devices. The grouping and device allocations are learned jointly. The proposed algorithm is trained by a policy gradient method and requires no human intervention. Experiments with widely-used computer vision and natural language models show that our algorithm can find optimized, non-trivial placements for TensorFlow (TF) computational graphs with over 80,000 operations. In addition, our approach outperforms placements by human experts as well as a previous state-of-the-art placement method based on deep reinforcement learning. Our method achieves reductions in runtime of up to 60.6% per training step when applied to models such as Neural Machine Translation. |
Year | Venue | Field |
---|---|---|
2018 | international conference on learning representations | Gradient method,Graph operations,Computer science,Machine translation,Device placement,Natural language,Artificial intelligence,Deep learning,Hierarchical database model,Machine learning,Reinforcement learning |
DocType | Citations | PageRank |
Conference | 8 | 0.45 |
References | Authors | |
9 | 6 |
Name | Order | Citations | PageRank |
---|---|---|---|
Azalia Mirhoseini | 1 | 238 | 18.68 |
Anna Goldie | 2 | 75 | 5.17 |
Hieu D. Pham | 3 | 1008 | 32.80 |
Benoit Steiner | 4 | 1486 | 49.38 |
Quoc V. Le | 5 | 8501 | 366.59 |
Jeffrey Dean | 6 | 11804 | 457.69 |