-
Install the lightningdit environment first.
-
Install additional packages:
pip install -r vavae_requirements.txt
-
Taming-Transformers is also needed for training.
Get it by running:
git clone https://github.com/CompVis/taming-transformers.git cd taming-transformers pip install -e .
Then modify
./taming-transformers/taming/data/utils.py
to meet torch 2.x:export FILE_PATH=./taming-transformers/taming/data/utils.py sed -i 's/from torch._six import string_classes/from six import string_types as string_classes/' "$FILE_PATH"
-
Modify training config as you need.
-
Run training by:
bash run_train.sh vavae/configs/f16d32_vfdinov2.yaml
Your training logs and checkpoints will be saved in the
logs
folder. We train VA-VAE with 4x8 H800 GPUs.
-
We provide a training log here for reference. All of our experimental variants are provided here.
-
Put your checkpoint path into
lightningdit/tokenizer/configs/vavae_f16d32.yaml
and uselightningdit/evaluate_tokenizer.py
to evaluate the model.
VA-VAE's training is mainly built upon LDM. Thanks for the great work!