Mammoth is built to streamline the development and benchmark of continual learning research. With more than 60 methods and 20 datasets, it includes the most complete list competitors and benchmarks for research purposes.
The core idea of Mammoth is that it is designed to be modular, easy to extend, and - most importantly - easy to debug.
With Mammoth, nothing is set in stone. You can easily add new models, datasets, training strategies, or functionalities.
Check out our guides on using Mammoth for continual learning research
- π₯ Install with
pip install -r requirements.txt
or run it directly withuv run python main.py ...
Note: PyTorch version >= 2.1.0 is required for scaled_dot_product_attention. If you cannot support this requirement, uncomment the lines 136-139 under
scaled_dot_product_attention
inbackbone/vit.py
. - π Use
main.py
or./utils/main.py
to run experiments. - 𧩠New models can be added to the
models/
folder. - π New datasets can be added to the
datasets/
folder.
The following command will run the model derpp
on the dataset seq-cifar100
with a buffer of 500 samples the some random hyperparameters for lr, alpha, and beta:
python main.py --model derpp --dataset seq-cifar100 --alpha 0.5 --beta 0.5 --lr 0.001 --buffer_size 500
To run the model with the best hyperparameters, use the --model_config=best
argument:
python main.py --model derpp --dataset seq-cifar100 --model_config best
NOTE: the
--model_config
argument will look for a file<model_name>.yaml
in themodels/configs/
folder. This file should contain the hyperparameters for the best configuration of the model. You can find more information in the documentation.
See the documentation for a detailed guide on how to create a new model.
See the documentation for a detailed guide on how to create a new dataset.
All the code is under active development. Here are some of the features we are working on:
- π§ New models: We are continuously working on adding new models to the repository.
- π New training modalities: New training regimes, such a regression, segmentation, detection, etc.
- π Openly accessible result dashboard: The ideal would be a dashboard to visualize the results of all the models in both their respective settings (to prove their reproducibility) and in a general setting (to compare them). This may take some time, since compute is not free.
All the new additions will try to preserve the current structure of the repository, making it easy to add new functionalities with a simple merge.
Mammoth currently supports more than 60 models, with new releases covering the main competitors in literature.
Click to expand model list
- AttriCLIP: A Non-Incremental Learner for Incremental Knowledge Learning (AttriCLIP):
attriclip
. - Bias Correction (BiC):
bic
. - CaSpeR-IL (on DER++, X-DER with RPC, iCaRL, and ER-ACE):
derpp_casper
,xder_rpc_casper
,icarl_casper
,er_ace_casper
. - CODA-Prompt: COntinual Decomposed Attention-based Prompting for Rehearsal-Free Continual Learning (CODA-Prompt) - Requires
pip install timm==0.9.8
:coda-prompt
. - Continual Contrastive Interpolation Consistency (CCIC) - Requires
pip install kornia
:ccic
. - Continual Generative training for Incremental prompt-Learning (CGIL):
cgil
- Contrastive Language-Image Pre-Training (CLIP):
clip
(static method with no learning). - CSCCT (on DER++, X-DER with RPC, iCaRL, and ER-ACE):
derpp_cscct
,xder_rpc_cscct
,icarl_cscct
,er_ace_cscct
. - Dark Experience for General Continual Learning: a Strong, Simple Baseline (DER & DER++):
der
andderpp
. - DualPrompt: Complementary Prompting for Rehearsal-free Continual Learning (DualPrompt) - Requires
pip install timm==0.9.8
:dualprompt
. - Efficient Lifelong Learning with A-GEM (A-GEM, A-GEM-R - A-GEM with reservoir buffer):
agem
,agem_r
. - Experience Replay (ER):
er
. - Experience Replay with Asymmetric Cross-Entropy (ER-ACE):
er_ace
. - eXtended-DER (X-DER):
xder
(full version),xder_ce
(X-DER with CE),xder_rpc
(X-DER with RPC). - Function Distance Regularization (FDR):
fdr
. - Generating Instance-level Prompts for Rehearsal-free Continual Learning (DAP):
dap
. - Gradient Episodic Memory (GEM) - Unavailable on windows:
gem
. - Greedy gradient-based Sample Selection (GSS):
gss
. - Greedy Sampler and Dumb Learner (GDumb):
gdumb
. - Hindsight Anchor Learning (HAL):
hal
. - Image-aware Decoder Enhanced Γ la Flamingo with Interleaved Cross-attentionS (IDEFICS):
idefics
(static method with no learning). - Incremental Classifier and Representation Learning (iCaRL):
icarl
. - Joint training for the General Continual setting:
joint_gcl
(only for General Continual). - Large Language and Vision Assistant (LLAVA):
llava
(static method with no learning). - Learning a Unified Classifier Incrementally via Rebalancing (LUCIR):
lucir
. - Learning to Prompt (L2P) - Requires
pip install timm==0.9.8
:l2p
. - Learning without Forgetting (LwF):
lwf
. - Learning without Forgetting adapted for Multi-Class classification (LwF.MC):
lwf_mc
(from the iCaRL paper). - Learning without Shortcuts (LwS):
lws
. - LiDER (on DER++, iCaRL, GDumb, and ER-ACE):
derpp_lider
,icarl_lider
,gdumb_lider
,er_ace_lider
. - May the Forgetting Be with You: Alternate Replay for Learning with Noisy Labels (AER & ABS):
er_ace_aer_abs
. - Meta-Experience Replay (MER):
mer
. - Mixture-of-Experts Adapters (MoE Adapters):
moe_adapters
. - Online Continual Learning on a Contaminated Data Stream with Blurry Task Boundaries (PuriDivER):
puridiver
. - online Elastic Weight Consolidation (oEWC):
ewc_on
. - Progressive Neural Networks (PNN):
pnn
. - Random Projections and Pre-trained Models for Continual Learning (RanPAC):
ranpac
. - Regular Polytope Classifier (RPC):
rpc
. - Rethinking Experience Replay: a Bag of Tricks for Continual Learning (ER-ACE with tricks):
er_ace_tricks
. - Semantic Two-level Additive Residual Prompt (STAR-Prompt):
starprompt
. Also includes the first-stage only (first_stage_starprompt
) and second-stage only (second_stage_starprompt
) versions. - SLCA: Slow Learner with Classifier Alignment for Continual Learning on a Pre-trained Model (SLCA) - Requires
pip install timm==0.9.8
:slca
. - Slow Learner with Classifier Alignment (SLCA):
slca
. - Synaptic Intelligence (SI):
si
. - Transfer without Forgetting (TwF):
twf
. - ZSCL: Zero-Shot Continual Learning:
zscl
.
NOTE: Datasets are automatically downloaded in data/
.
- This can be changed by changing the
base_path
function inutils/conf.py
or using the--base_path
argument. - The
data/
folder should not be tracked by git and is created automatically if missing.
Click to expand dataset list
Mammoth currently includes 23 datasets, covering toy classification problems (different versions of MNIST), standard natural-image domains (CIFAR, Imagenet-R, TinyImagenet, MIT-67), fine-grained classification domains (Cars-196, CUB-200), aerial domains (EuroSAT-RGB, Resisc45), medical domains (CropDisease, ISIC, ChestX).
- Sequential MNIST (Class-Il / Task-IL):
seq-mnist
. - Permuted MNIST (Domain-IL):
perm-mnist
. - Rotated MNIST (Domain-IL):
rot-mnist
. - MNIST-360 (General Continual Learning):
mnist-360
. - Sequential CIFAR-10 (Class-Il / Task-IL):
seq-cifar10
. - Sequential CIFAR-10 resized 224x224 (ViT version) (Class-Il / Task-IL):
seq-cifar10-224
. - Sequential CIFAR-10 resized 224x224 (ResNet50 version) (Class-Il / Task-IL):
seq-cifar10-224-rs
. - Sequential Tiny ImageNet (Class-Il / Task-IL):
seq-tinyimg
. - Sequential Tiny ImageNet resized 32x32 (Class-Il / Task-IL):
seq-tinyimg-r
. - Sequential CIFAR-100 (Class-Il / Task-IL):
seq-cifar100
. - Sequential CIFAR-100 resized 224x224 (ViT version) (Class-Il / Task-IL):
seq-cifar100-224
. - Sequential CIFAR-100 resized 224x224 (ResNet50 version) (Class-Il / Task-IL):
seq-cifar100-224-rs
. - Sequential CUB-200 (Class-Il / Task-IL):
seq-cub200
. - Sequential ImageNet-R (Class-Il / Task-IL):
seq-imagenet-r
. - Sequential Cars-196 (Class-Il / Task-IL):
seq-cars196
. - Sequential RESISC45 (Class-Il / Task-IL):
seq-resisc45
. - Sequential EuroSAT-RGB (Class-Il / Task-IL):
seq-eurosat-rgb
. - Sequential ISIC (Class-Il / Task-IL):
seq-isic
. - Sequential ChestX (Class-Il / Task-IL):
seq-chestx
. - Sequential MIT-67 (Class-Il / Task-IL):
seq-mit67
. - Sequential CropDisease (Class-Il / Task-IL):
seq-cropdisease
. - Sequential CelebA (Biased-Class-Il):
seq-celeba
. This dataset is multi-label (i.e., trains with binary cross-entropy)
@article{boschini2022class,
title={Class-Incremental Continual Learning into the eXtended DER-verse},
author={Boschini, Matteo and Bonicelli, Lorenzo and Buzzega, Pietro and Porrello, Angelo and Calderara, Simone},
journal={IEEE Transactions on Pattern Analysis and Machine Intelligence},
year={2022},
publisher={IEEE}
}
@inproceedings{buzzega2020dark,
author = {Buzzega, Pietro and Boschini, Matteo and Porrello, Angelo and Abati, Davide and Calderara, Simone},
booktitle = {Advances in Neural Information Processing Systems},
editor = {H. Larochelle and M. Ranzato and R. Hadsell and M. F. Balcan and H. Lin},
pages = {15920--15930},
publisher = {Curran Associates, Inc.},
title = {Dark Experience for General Continual Learning: a Strong, Simple Baseline},
volume = {33},
year = {2020}
}
We take great pride and care in the reproducibility of the models in Mammoth and we are commited to provide the community with the most accurate results possible. To this end, we provide a REPRODUCIBILITY.md file in the repository that contains the results of the models in Mammoth.
The performance of each model is evaluated on the same dataset used in the paper and we report in REPRODUCIBILITY.md the list of models that have been verified. We also provide the exact command used to train the model (most times, it follows python main.py --model <model-name> --dataset <dataset-name> --model_config best
).
We encourage the community to report any issues with the reproducibility of the models in Mammoth. If you find any issues, please open an issue in the GitHub repository or contact us directly.
Disclaimer: Since there are many models in Mammoth (and some of them predate PyTorch), the process of filling the REPRODUCIBILITY.md file is ongoing. We are working hard to fill the file with the results of all models in Mammoth. If you need the results of a specific model, please open an issue in the GitHub repository or contact us directly.
Does this mean that the models that are not in the REPRODUCIBILITY.md file do not reproduce?
No! It means that we have not yet found the appropriate dataset and hyperparameters to fill the file with the results of that model. We are working hard to fill the file with the results of all models in Mammoth. If you need the results of a specific model, please open an issue in the GitHub repository or contact us directly.
Pull requests are welcome!
Please use autopep8 with parameters:
--aggressive
--max-line-length=200
--ignore=E402
If you're interested in a version of this repo that only includes the original code for "Dark Experience for General Continual Learning: a Strong, Simple Baseline" or "Class-Incremental Continual Learning into the eXtended DER-verse", please use the following tags:
neurips2020
for DER (NeurIPS 2020).
tpami2023
for X-DER (TPAMI 2022).