Title
Scaling distributed machine learning with the parameter server
Abstract
We propose a parameter server framework for distributed machine learning problems. Both data and workloads are distributed over worker nodes, while the server nodes maintain globally shared parameters, represented as dense or sparse vectors and matrices. The framework manages asynchronous data communication between nodes, and supports flexible consistency models, elastic scalability, and continuous fault tolerance. To demonstrate the scalability of the proposed framework, we show experimental results on petabytes of real data with billions of examples and parameters on problems ranging from Sparse Logistic Regression to Latent Dirichlet Allocation and Distributed Sketching.
Year
DOI
Venue
2014
10.1145/2640087.2644155
OSDI
Field
DocType
Citations 
Latent Dirichlet allocation,Petabyte,Computer science,Matrix (mathematics),Real-time computing,Ranging,Artificial intelligence,Consistency model,Distributed computing,Asynchronous communication,Fault tolerance,Machine learning,Scalability
Conference
119
PageRank 
References 
Authors
4.22
24
9
Search Limit
100119
Name
Order
Citations
PageRank
Mu Li191342.35
andersen david g24823345.31
Jun Woo Park31696.47
Alexander J. Smola4196271967.09
Amr Ahmed5174392.13
Vanja Josifovski62265148.84
James Long71236.92
Eugene J. Shekita83630574.21
Bor-Yiing Su932418.28