Revisiting Class-Incremental Learning with Pre-Trained Models: Generalizability and Adaptivity are All You Need
2S-Lab, Nanyang Technological University
The code repository for "Revisiting Class-Incremental Learning with Pre-Trained Models: Generalizability and Adaptivity are All You Need" (IJCV 2024) in PyTorch. If you use any content of this repo for your work, please cite the following bib entry:
@article{zhou2024revisiting,
author = {Zhou, Da-Wei and Cai, Zi-Wen and Ye, Han-Jia and Zhan, De-Chuan and Liu, Ziwei},
title = {Revisiting Class-Incremental Learning with Pre-Trained Models: Generalizability and Adaptivity are All You Need},
journal = {International Journal of Computer Vision},
year = {2024}
}
[08/2024] Update all training jsons and logs.
[08/2024] Accepted to IJCV.
[03/2023] arXiv paper has been released.
[03/2023] Code has been released.
Class-incremental learning (CIL) aims to adapt to emerging new classes without forgetting old ones. Traditional CIL models are trained from scratch to continually acquire knowledge as data evolves. Recently, pre-training has achieved substantial progress, making vast pre-trained models (PTMs) accessible for CIL. Contrary to traditional methods, PTMs possess generalizable embeddings, which can be easily transferred for CIL. In this work, we revisit CIL with PTMs and argue that the core factors in CIL are adaptivity for model updating and generalizability for knowledge transferring. (1) We first reveal that frozen PTM can already provide generalizable embeddings for CIL. Surprisingly, a simple baseline (SimpleCIL) which continually sets 1 the classifiers of PTM to prototype features can beat state-of-the-art even without training on the downstream task. (2) Due to the distribution gap between pre-trained and downstream datasets, PTM can be further cultivated with adaptivity via model adaptation. We propose AdaPt and mERge (Aper), which aggregates the embeddings of PTM and adapted models for classifier construction. Aper is a general framework that can be orthogonally combined with any parameter-efficient tuning method, which holds the advantages of PTM’s generalizability and adapted model’s adaptivity. (3) Additionally, considering previous ImageNet-based benchmarks are unsuitable in the era of PTM due to data overlapping, we propose four new benchmarks for assessment, namely ImageNet-A, ObjectNet, OmniBenchmark, and VTAB. Extensive experiments validate the effectiveness of Aper with a unified and concise framework.
A simple baseline (SimpleCIL) beats SOTA even without training on the downstream task. AdaPt and mERge (Aper) extends SimpleCIL with better adaptivity and generalizability. Four new benchmarks are proposed for assessment.
We provide the processed datasets as follows:
- CIFAR100: will be automatically downloaded by the code.
- CUB200: Google Drive: link or Onedrive: link
- ImageNet-R: Google Drive: link or Onedrive: link
- ImageNet-A:Google Drive: link or Onedrive: link
- OmniBenchmark: Google Drive: link or Onedrive: link
- VTAB: Google Drive: link or Onedrive: link
- ObjectNet: Onedrive: link You can also refer to the filelist and processing code if the file is too large to download.
These subsets are sampled from the original datasets. Please note that I do not have the right to distribute these datasets. If the distribution violates the license, I shall provide the filenames instead.
The md5sum information can be found in this issue.
You need to modify the path of the datasets in ./utils/data.py
according to your own path.
Ensure that the "shuffle" entry in the JSON file is set to false for the VTAB benchmark.
To prepare your JSON files, refer to the settings in the exps
folder and run the following command.
All main experiments from the paper (see Table 1) are already provided in the exps
folder, you can simply execute them to reproduce the results found in the logs
folder.
python main.py --config ./exps/[configname].json
This repo is based on CIL_Survey and PyCIL.
The implementations of parameter-efficient tuning methods are based on VPT, AdaptFormer, and SSF.
If you have any questions, please contact me via email or open an issue.