1. Experience Paper: Towards enhancing cost efficiency in serverless machine learning training
- Author
-
Pablo Gimeno Sarroca and Marc Sanchez-Artigas
- Subjects
Cost efficiency ,Computer science ,business.industry ,media_common.quotation_subject ,020206 networking & telecommunications ,Cloud computing ,02 engineering and technology ,Filter (higher-order function) ,Machine learning ,computer.software_genre ,Nagging ,Matrix decomposition ,Task (computing) ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,Key (cryptography) ,Use case ,Artificial intelligence ,business ,computer ,media_common - Abstract
Function-as-a-Service (FaaS) has raised a growing interest in how to "tame" serverless to enable domain-specific use cases such as data-intensive applications and machine learning (ML), to name a few. Recently, several systems have been implemented for training ML models. Certainly, these research articles are significant steps in the correct direction. However, they do not completely answer the nagging question of when serverless ML training can be more cost-effective compared to traditional "serverful" computing. To help in this task, we propose MLLess, a FaaS-based ML training prototype built atop IBM Cloud Functions. To boost cost-efficiency, MLLess implements two key optimizations: a significance filter and a scale-in auto-tuner, and leverages them to specialize model training to the FaaS model. Our results certify that MLLess can be 15X faster than serverful ML systems [24] at a lower cost for ML models (such as sparse logistic regression and matrix factorization) that exhibit fast convergence.
- Published
- 2021