1. Data Efficiency for Large Recommendation Models
- Author
-
Jain, Kshitij, Xie, Jingru, Regan, Kevin, Chen, Cheng, Han, Jie, Li, Steve, Li, Zhuoshu, Phillips, Todd, Sussman, Myles, Troup, Matt, Yu, Angel, and Zhuo, Jia
- Subjects
Computer Science - Information Retrieval ,Computer Science - Machine Learning - Abstract
Large recommendation models (LRMs) are fundamental to the multi-billion dollar online advertising industry, processing massive datasets of hundreds of billions of examples before transitioning to continuous online training to adapt to rapidly changing user behavior. The massive scale of data directly impacts both computational costs and the speed at which new methods can be evaluated (R&D velocity). This paper presents actionable principles and high-level frameworks to guide practitioners in optimizing training data requirements. These strategies have been successfully deployed in Google's largest Ads CTR prediction models and are broadly applicable beyond LRMs. We outline the concept of data convergence, describe methods to accelerate this convergence, and finally, detail how to optimally balance training data volume with model size.
- Published
- 2024