Exemplar-free Continual Learning of Vision Transformers via Gated Class-Attention and Cascaded Feature Drift Compensation
Marco Cotogni, Fei Yang, Claudio Cusano, Andrew D. Bagdanov and Joost van de Weijer
ArXiv version of our Paper is available.
In order to use our code, it is possible to create a conda environment using the requirements.txt file and Python 3.9.
For training the model please run:
python3 main.py --options options/data/DATASET_SCENARIO options/data/DATASET_ORDER
options/model/DATASET_GCAB_FDC --name GCAB --data-path PATH
--output-basedir OUTPUTDIR --compress COMPRESS_FILE.txt --report REPORT
For evaluating the trained model:
python3 main.py --options options/data/DATASET_SCENARIO options/data/DATASET_ORDER
options/model/DATASET_GCAB_FDC --name GCAB --data-path PATH
--output-basedir OUTPUTDIR --compress COMPRESS_FILE.txt --report REPORT
--resume PATH_TO_THE_MODEL --eval
If you are considering using our code or you want to cite our paper please use:
@article{cotogni2022gated,
author = {Cotogni, Marco and Yang, Fei and Cusano, Claudio and Bagdanov, Andrew D. and van de Weijer, Joost},
title = {Gated Class-Attention with Cascaded Feature Drift Compensation for Exemplar-free Continual Learning of Vision Transformers},
journal={arXiv preprint arxiv:2211.12292},
year={2022}
}
Our code is based on: