-
Notifications
You must be signed in to change notification settings - Fork 5.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Preln_Layernorm_Shift_Partition #47099
Conversation
你的PR提交成功,感谢你对开源项目的贡献! |
@@ -112,6 +112,7 @@ const std::vector<std::string> kTRTSubgraphPasses({ | |||
"layernorm_shift_partition_fuse_pass", // | |||
"merge_layernorm_fuse_pass", // | |||
"preln_residual_bias_fuse_pass", // | |||
"preln_layernorm_x_fuse_pass", // |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
这个pass是不是也能放在原生gpu里面,如果把新增的plugin放到phi算子里,通过通用plugin应该也能接进来
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
后面 @weishengying 完善后,可以按照新方案接入plugin
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
因为目前只有trt算子 所以没有放到原生pass里
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM for preln_layernorm pass
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
PR types
Performance optimization
PR changes
OPs
Describe
layernorm_shift_partition + element_add融合
在swin模型上有3%的加速效果