This repo is based on fairseq
repository (lates version, 1.0.0a0) https://github.com/pytorch/fairseq
An implementation of AdaBelief optimizer compatible with fairseq
is in fairseq/optim
.
The original AdaBelief implementation is in https://github.com/juntang-zhuang/Adabelief-Optimizer
Code for transformer to work with PyTorch 1.1 and CUDA9.0 is at: https://github.com/juntang-zhuang/transformer-adabelief
Link to AdaBelief repository https://github.com/juntang-zhuang/Adabelief-Optimizer
PyTorch==1.6.0
pip install --editable .
cd examples/translation/
bash prepare-iwslt14.sh
cd ../..
TEXT=examples/translation/iwslt14.tokenized.de-en
fairseq-preprocess --source-lang de --target-lang en \
--trainpref $TEXT/train --validpref $TEXT/valid --testpref $TEXT/test \
--destdir data-bin/iwslt14.tokenized.de-en \
--workers 20
sh run_adam.sh
Results saved in folder adam
sh run_adabelief.sh
Results saved in folder adabelief
Adam | Adabelief |
---|---|
35.02 | 35.17 |