Back to Search Start Over

Episodic Training for Domain Generalization

Authors :
Li, Da
Zhang, Jianshu
Yang, Yongxin
Liu, Cong
Song, Yi-Zhe
Hospedales, Timothy M.
Publication Year :
2019

Abstract

Domain generalization (DG) is the challenging and topical problem of learning models that generalize to novel testing domains with different statistics than a set of known training domains. The simple approach of aggregating data from all source domains and training a single deep neural network end-to-end on all the data provides a surprisingly strong baseline that surpasses many prior published methods. In this paper, we build on this strong baseline by designing an episodic training procedure that trains a single deep network in a way that exposes it to the domain shift that characterises a novel domain at runtime. Specifically, we decompose a deep network into feature extractor and classifier components, and then train each component by simulating it interacting with a partner who is badly tuned for the current domain. This makes both components more robust, ultimately leading to our networks producing state-of-the-art performance on three DG benchmarks. Furthermore, we consider the pervasive workflow of using an ImageNet trained CNN as a fixed feature extractor for downstream recognition tasks. Using the Visual Decathlon benchmark, we demonstrate that our episodic-DG training improves the performance of such a general-purpose feature extractor by explicitly training a feature for robustness to novel problems. This shows that DG training can benefit standard practice in computer vision.<br />Comment: ICCV'19 CR version and fix Table 5. Code is now available at https://github.com/HAHA-DL/Episodic-DG

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.1902.00113
Document Type :
Working Paper