Title
Towards Self-Tuning Parameter Servers
Abstract
Recent years, many applications have been driven advances by the use of Machine Learning (ML). Nowadays, it is common to see industrial-strength machine learning jobs that involve millions of model parameters, terabytes of training data, and weeks of training. Good efficiency, i.e., fast completion time of running a specific ML training job, therefore, is a key feature of a successful ML system. While the completion time of a long-running ML job is determined by the time required to reach model convergence, that is also largely influenced by the values of various system settings. In this paper, we contribute techniques towards building self-tuning parameter servers. Parameter Server (PS) is a popular system architecture for large-scale machine learning systems; and by self-tuning we mean while a long-running ML job is iteratively training the expert-suggested model, the system is also iteratively learning which system setting is more efficient for that job and applies it online. Our techniques are general enough to various PS-style ML systems. Experiments on TensorFlow show that our techniques can reduce the completion times of a variety of long-running TensorFlow jobs from 1.4× to 18×.
Year
DOI
Venue
2020
10.1109/BigData50022.2020.9378141
2020 IEEE International Conference on Big Data (Big Data)
Keywords
DocType
ISSN
ML job,model convergence,system setting,self-tuning parameter servers,popular system architecture,large-scale machine,expert-suggested model,PS-style ML systems,TensorFlow jobs,industrial-strength machine learning jobs,model parameters,training data,fast completion time,ML training job
Conference
2639-1589
ISBN
Citations 
PageRank 
978-1-7281-6252-2
0
0.34
References 
Authors
0
7
Name
Order
Citations
PageRank
Chris Liu112.04
Pengfei Zhang212.03
Bo Tang344.80
Hang Shen400.34
Ziliang Lai512.72
Eric Lo681351.50
Fu-lai Chung724434.50