1. Improving Hyperparameter Optimization with Checkpointed Model Weights
- Author
-
Mehta, Nikhil, Lorraine, Jonathan, Masson, Steve, Arunachalam, Ramanathan, Bhat, Zaid Pervaiz, Lucas, James, and Zachariah, Arun George
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence ,Statistics - Machine Learning ,68T05 ,I.2.6 ,G.1.6 ,D.2.8 - Abstract
When training deep learning models, the performance depends largely on the selected hyperparameters. However, hyperparameter optimization (HPO) is often one of the most expensive parts of model design. Classical HPO methods treat this as a black-box optimization problem. However, gray-box HPO methods, which incorporate more information about the setup, have emerged as a promising direction for more efficient optimization. For example, using intermediate loss evaluations to terminate bad selections. In this work, we propose an HPO method for neural networks using logged checkpoints of the trained weights to guide future hyperparameter selections. Our method, Forecasting Model Search (FMS), embeds weights into a Gaussian process deep kernel surrogate model, using a permutation-invariant graph metanetwork to be data-efficient with the logged network weights. To facilitate reproducibility and further research, we open-source our code at https://github.com/NVlabs/forecasting-model-search., Comment: See the project website at https://research.nvidia.com/labs/toronto-ai/FMS/
- Published
- 2024