Back to Search Start Over

Primitives for Dynamic Big Model Parallelism

Authors :
Lee, Seunghak
Kim, Jin Kyu
Zheng, Xun
Ho, Qirong
Gibson, Garth A.
Xing, Eric P.
Lee, Seunghak
Kim, Jin Kyu
Zheng, Xun
Ho, Qirong
Gibson, Garth A.
Xing, Eric P.
Publication Year :
2014

Abstract

When training large machine learning models with many variables or parameters, a single machine is often inadequate since the model may be too large to fit in memory, while training can take a long time even with stochastic updates. A natural recourse is to turn to distributed cluster computing, in order to harness additional memory and processors. However, naive, unstructured parallelization of ML algorithms can make inefficient use of distributed memory, while failing to obtain proportional convergence speedups - or can even result in divergence. We develop a framework of primitives for dynamic model-parallelism, STRADS, in order to explore partitioning and update scheduling of model variables in distributed ML algorithms - thus improving their memory efficiency while presenting new opportunities to speed up convergence without compromising inference correctness. We demonstrate the efficacy of model-parallel algorithms implemented in STRADS versus popular implementations for Topic Modeling, Matrix Factorization and Lasso.

Details

Database :
OAIster
Publication Type :
Electronic Resource
Accession number :
edsoai.on1106203475
Document Type :
Electronic Resource