Abstract | ||
---|---|---|
We categorize parallel machine learning applications into four types of computation models and propose a new set of model-centric computation abstractions. This work sets up parallel machine learning as a combination of training data-centric and model parameter-centric processing. The analysis uses Latent Dirichlet Allocation (LDA) as an example, and experimental results show that an efficient parallel model update pipeline can achieve similar or higher model convergence speed compared with other work. |
Year | DOI | Venue |
---|---|---|
2016 | 10.1145/2926534.2926539 | BeyondMR@SIGMOD |
Field | DocType | Citations |
Convergence (routing),Data mining,Latent Dirichlet allocation,Active learning (machine learning),Computer science,Theoretical computer science,Model of computation,Artificial intelligence,Computational learning theory,Computation,Categorization,Online machine learning,Machine learning | Conference | 2 |
PageRank | References | Authors |
0.38 | 13 | 3 |
Name | Order | Citations | PageRank |
---|---|---|---|
Bingjing Zhang | 1 | 521 | 25.17 |
Bo Peng | 2 | 9 | 2.91 |
Judy Qiu | 3 | 743 | 43.25 |