Use -d device
argument.
It seems so, but it is out of the control of us.
Since this is an re-implementation upon a pytorch template for clearity (the original codebase has too many unrelated functionalities and has some "hacks" in coding), although we tried our best in reimplementing what is described, we still could not guarantee 100% that the implementation is the same. However, we ran experiments on all three datasets with all common settings and found the final output to be at least as high as reported most of the time. If you get higher accuracy, then you are good since we also observe this sometimes and the reported ones are average ones, not the best ones. If you get lower accuracy and you think it's the issue with implementation, please contact us (see README and the contact method).
The checkpoint needs to be placed with config.json.
Make sure that n_gpus
equals to number of your devices you specified.
python train.py -r path_to_checkpoint
The config will be automatically saved with checkpoint and loaded. If you encountered problems with resuming, please contact the code author.
If you set fp16
in utils, it will enable fp16 training. However, this is susceptible to change (and may not work on all settings or models) and please double check if you are using it since we don't plan to focus on this part if you request help. Only some models work (see autograd
in the code). We do not plan to provide support on this because it is not within our focus (just for faster training and less memory requirement).
See train.py
for available options to tune.