You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
input shape temp update
clean up
dataset updates
clean up
move tensor idx to matrix op inside apply_mask
use tensor operators to replace tensor indexing, passed consistency test verification
Minor improvements
Fix bucketpadlendataset
Moved mask matrices creation to dataset prep.
Remove dynamism, apply mask correctly, add some guardrails, some cleanups.
Send device data to cpu b4 logging.
Fix data bucketing for RawAudioDataset, refactor bucketing functions, fix filling w/ -inf in wav2vec2, minor cleanups
Sample size computeation during data prep to reduce atens, dont call item in log_scalar, minor cleanups
Remove extra validation atens, clean up marking step and sending to cpu.
Correct loss computation for w2v2 criterion + refactor index_put
Fix bug in index_put + fix integer division
Dont call float on extra logs, clean up comment.
Correct accuracy computation, refactor xla tensor check.
Adjust loss computation so it works w/ binary cross entropy.
Remove sending log outputs back to cpu after allreduce.
Dont sample padded states when sampling negatives + correct mi in loss computation.
Fixing config issues after rebase
Fix bug in negatives from everywhere
Fixing config issue for TPU after rebase
Taylans changes on top of rebase
Use float on cpu if fp16 when filling w/ -inf in w2v2 (#5)
* Use float on cpu if fp16 when filling w/ -inf in w2v2
* xla -> self.xla
* make logging_output_can_be_summed a regular method instead of staticmethod.
Make tpu codepath work w/ hydra. (#6)
* Make tpu codepath work w/ hydra.
* Share and pass down model related args to rawaudiodataset correctly.
* fp16 bug fix on non-xla devices (self._inftensor change)
* use index_put to avoid dynamicity in model's fwd.
* Get rid of some unnecessary warnings for tpus to clean up stderr.
* Send logging outputs to cpu b4 logging to reduce atens.
* Util function to move cpu tensors to tpu.
* Use the util function to handle dummy batches to avoid crash at the
end of epoch in distributed training.
* fixing configs for precompute mask indices
Co-authored-by: Bilge Acun <acun@fb.com>
0 commit comments