-
Notifications
You must be signed in to change notification settings - Fork 5.9k
[hybrid performance] softmax mask fuse op #33841
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
Thanks for your contribution! |
845deef to
907037c
Compare
907037c to
04c3eff
Compare
ForFishes
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
Xreki
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM for the modification of atol and LGTM for op benchmark ci.
|
Sorry to inform you that 170cde8's CIs have passed for more than 7 days. To prevent PR conflicts, you need to re-run all CIs manually. |
170cde8 to
3b18c23
Compare
74b218e to
00c0a44
Compare
00c0a44 to
7151fcc
Compare
b15c513 to
c959fcf
Compare
c959fcf to
cb9f173
Compare
cb9f173 to
6e6ff16
Compare
6e6ff16 to
f01ffe0
Compare
ForFishes
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
XiaoguangHu01
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LG API
PangHua
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
PR types
New features
PR changes
OPs
Describe
fuse mask elementwise add and softmax together, for transformer used
general pass:
fused pass:
performance, based on PaddleNLP/GPT under AMP:
loss curve, for PaddleNLP/GPT under AMP:


average loss diff for 20,000 steps: 0.0077
loss diff between fused pass and no fuse pass:
Currently, this OP only supports fp16 dtype.
To use this op from python side, follow these codes for static mode:
Follow these codes for dynamic mode: