-
Notifications
You must be signed in to change notification settings - Fork 5.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[hybrid performance] softmax mask fuse op #33841
Conversation
Thanks for your contribution! |
845deef
to
907037c
Compare
907037c
to
04c3eff
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM for the modification of atol and LGTM for op benchmark ci.
Sorry to inform you that 170cde8's CIs have passed for more than 7 days. To prevent PR conflicts, you need to re-run all CIs manually. |
170cde8
to
3b18c23
Compare
74b218e
to
00c0a44
Compare
00c0a44
to
7151fcc
Compare
b15c513
to
c959fcf
Compare
c959fcf
to
cb9f173
Compare
cb9f173
to
6e6ff16
Compare
6e6ff16
to
f01ffe0
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LG API
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
PR types
New features
PR changes
OPs
Describe
fuse mask elementwise add and softmax together, for transformer used
general pass:
fused pass:
performance, based on PaddleNLP/GPT under AMP:
loss curve, for PaddleNLP/GPT under AMP:


average loss diff for 20,000 steps: 0.0077
loss diff between fused pass and no fuse pass:
Currently, this OP only supports fp16 dtype.
To use this op from python side, follow these codes for static mode:
Follow these codes for dynamic mode: