Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Showing equivalency b/w two branches. #1

Closed
wants to merge 60 commits into from

Conversation

taylanbil
Copy link
Owner

No description provided.

taylanbil and others added 30 commits June 13, 2019 23:23
optimizer fix
progress bar comment out temporarily
some changes to train_tpu
int mask instead of float
* Loss function replaced with an equivalent logic that doesn't resize
tensors.
* cli args changed to guarantee consistency
* collate_tokens function in fairseq/data/data_utils.py overwritten to
guarantee consistency
some irrelevant changes to train_tpu.py
taylanbil and others added 24 commits July 26, 2019 16:30
+ Tried to include more explanation why skip optimizer step this time
Delete optimizer step in Fairseq's trainer
* Adding tpu capabilities to train.py

* flush when printing for better user experience

* separated cli_main into parse_args, maingpu and maintpu
deleted unused line in datautils.py
* Adding tpu capabilities to train.py

* flush when printing for better user experience

* separated cli_main into parse_args, maingpu and maintpu
deleted unused line in datautils.py

* Enumerate the loader

* enumerate the loader
…arch#10)

* Add option to assert on training and/or validation loss

* applied suggestion
* initial commit for multiprocess api

* indentation fixes and import fix

* no need to softlink, fix save/load

* Remove the hacks to only save from master ordinal as xm.save takes care of that

* fix indentation; 3 -> 4 spaces

* Moved xu.eprints after spawn and dropping last batches better
@taylanbil taylanbil closed this Dec 27, 2019
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants