Back to Search
Start Over
A context-based meta-reinforcement learning approach to efficient hyperparameter optimization.
- Source :
-
Neurocomputing . Mar2022, Vol. 478, p89-103. 15p. - Publication Year :
- 2022
-
Abstract
- [Display omitted] • A context helps the agent to learn knowledge from previous tasks. • A simple multi-task objective meta-training procedure is proposed to run faster. • A quadratic penalty technique is used to preserve the historical experience. • A predictive model is applied to further accelerate the adaptation phase. In this paper, we present a context-based meta-reinforcement learning approach to tackle the challenging data-inefficiency problem of Hyperparameter Optimization (HPO). Specifically, we design an agent which sequentially selects hyperparameters to maximize the expected accuracy of the machine learning algorithm on the validation set. First, we design a context variable that learns the latent embedding of prior experience, and the agent can solve the new tasks efficiently conditioned on it. Second, we employ a multi-task objective method that aims to maximize the average reward across all the meta-training tasks to meta-train the agent. Third, in the adaptation phase, we introduce a quadratic penalty technique to achieve better performance of the agent. Finally, to further improve the efficiency in the adaptation phase, we use a predictive model to evaluate the accuracy of machine learning algorithm instead of training it. We evaluate our approach on 18 real-world datasets and the results demonstrate that our approach outperforms other state-of-the-art optimization methods in terms of test set accuracy and runtime performance. [ABSTRACT FROM AUTHOR]
Details
- Language :
- English
- ISSN :
- 09252312
- Volume :
- 478
- Database :
- Academic Search Index
- Journal :
- Neurocomputing
- Publication Type :
- Academic Journal
- Accession number :
- 155090546
- Full Text :
- https://doi.org/10.1016/j.neucom.2021.12.086