-
Notifications
You must be signed in to change notification settings - Fork 505
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support fp8e5m2 dtype #7740
Support fp8e5m2 dtype #7740
Conversation
|
@lsy323 we should include an accuracy correctness analysis in our benchmarking and documentation. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
approve to unblock, mostly lgtm we can have a follow up pr to fix review comments
Hi @JackCaoG, @miladm, thank you for reviewing! Sorry for the delayed respond for the comments.
In this PR the unit test is included to verify the correctness of the fp8 dtype and matmul op. Let's have documentation in a separate PR (After e4m3 variants are ready) |
Add support for
torch.fp8e5m2
dtype.Right now
torch_xla/csrc/tensor_util.cpp
contains many duplicated codes, in the next PR I'll try to refactor it to make it cleaner.The other fp8 variants and will be added in the following PRs
Test:
Added fp8 unit test