Back to Search Start Over

Dense Network Expansion for Class Incremental Learning

Authors :
Hu, Zhiyuan
Li, Yunsheng
Lyu, Jiancheng
Gao, Dashan
Vasconcelos, Nuno
Publication Year :
2023

Abstract

The problem of class incremental learning (CIL) is considered. State-of-the-art approaches use a dynamic architecture based on network expansion (NE), in which a task expert is added per task. While effective from a computational standpoint, these methods lead to models that grow quickly with the number of tasks. A new NE method, dense network expansion (DNE), is proposed to achieve a better trade-off between accuracy and model complexity. This is accomplished by the introduction of dense connections between the intermediate layers of the task expert networks, that enable the transfer of knowledge from old to new tasks via feature sharing and reusing. This sharing is implemented with a cross-task attention mechanism, based on a new task attention block (TAB), that fuses information across tasks. Unlike traditional attention mechanisms, TAB operates at the level of the feature mixing and is decoupled with spatial attentions. This is shown more effective than a joint spatial-and-task attention for CIL. The proposed DNE approach can strictly maintain the feature space of old classes while growing the network and feature scale at a much slower rate than previous methods. In result, it outperforms the previous SOTA methods by a margin of 4\% in terms of accuracy, with similar or even smaller model scale.<br />Accepted by CVPR2023

Details

Language :
English
Database :
OpenAIRE
Accession number :
edsair.doi.dedup.....21e64f4c8413f214626638e92141e368