Abstract
Existing works study the Class Incremental learning (CIL) problem with the assumption that the data for previous classes are absent, or only a small subset of samples (known as exemplars) are accessible. Differently, we propose a new and practical setting called retrospective CIL, where all the previous data are accessible, but with bounded training budgets for old data replay. Since only a small subset of old samples can be replayed, it brings a new research problem, i.e., dynamically sampling old data along the incremental training process. As incremental learning particularly suffers from catastrophic forgetting, we propose to use the forgettability of the old samples as the sampling priorities to favour the forgotten samples during the dynamic sampling process. To achieve this, we introduce a forgetting rate metric with graph-based propagation to estimate the sample forgettability. The proposed method brings improvements on two benchmark datasets.
Original language | English |
---|---|
Title of host publication | 2021 IEEE International Conference on Multimedia and Expo, ICME 2021 |
Publisher | IEEE Computer Society |
ISBN (Electronic) | 9781665438643 |
DOIs | |
Publication status | Published - 2021 |
Externally published | Yes |
Event | 2021 IEEE International Conference on Multimedia and Expo, ICME 2021 - Shenzhen, China Duration: Jul 5 2021 → Jul 9 2021 |
Publication series
Name | Proceedings - IEEE International Conference on Multimedia and Expo |
---|---|
ISSN (Print) | 1945-7871 |
ISSN (Electronic) | 1945-788X |
Conference
Conference | 2021 IEEE International Conference on Multimedia and Expo, ICME 2021 |
---|---|
Country/Territory | China |
City | Shenzhen |
Period | 7/5/21 → 7/9/21 |
Bibliographical note
Publisher Copyright:© 2021 IEEE
ASJC Scopus Subject Areas
- Computer Networks and Communications
- Computer Science Applications
Keywords
- catastrophic forgetting
- continual learning
- lifelong learning