Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[oneDNN] Optimize fused elementwise kernel #59663

Merged
merged 1 commit into from
Dec 6, 2023

Conversation

LLee233
Copy link
Contributor

@LLee233 LLee233 commented Dec 4, 2023

PR types

Bug fixes

PR changes

Others

Description

This PR aims for int8 case in #59252 when config.enable_mkldnn_int8() is activated. For float & int8, paddle will go through different passes and hence different kernels. So for int8 case, the refreshed kernel can not utilizes what former PR #59421 optimizes.

Copy link

paddle-bot bot commented Dec 4, 2023

你的PR提交成功,感谢你对开源项目的贡献!
请关注后续CI自动化测试结果,详情请参考Paddle-CI手册
Your PR has been submitted. Thanks for your contribution!
Please wait for the result of CI firstly. See Paddle CI Manual for details.

@paddle-bot paddle-bot bot added the contributor External developers label Dec 4, 2023
@LLee233 LLee233 changed the title Optimize fused elementwise kernel [oneDNN] Optimize fused elementwise kernel Dec 5, 2023
@LLee233
Copy link
Contributor Author

LLee233 commented Dec 5, 2023

Hi @yuanlehome, would you mind helping check this PR? Since the CI coverage seems not work... The coverage result didn't show up and I restarted the check but the result was as still. Thanks~

@yuanlehome
Copy link
Contributor

Hi @yuanlehome, would you mind helping check this PR? Since the CI coverage seems not work... The coverage result didn't show up and I restarted the check but the result was as still. Thanks~

Okay, please wait for the latest CI results.

@LLee233
Copy link
Contributor Author

LLee233 commented Dec 5, 2023

Hi @yuanlehome, would you mind helping check this PR? Since the CI coverage seems not work... The coverage result didn't show up and I restarted the check but the result was as still. Thanks~

Okay, please wait for the latest CI results.

Hi @yuanlehome, seems it's still no result...

@LLee233
Copy link
Contributor Author

LLee233 commented Dec 6, 2023

@xinyu-intel, @vivienfanghuagood, @yuanlehome, hi, would you mind helping review this PR? Thanks~

const auto src_y_memory = handler.AcquireSecondSrcMemory(non_const_y);
const auto src_x_memory =
handler.swin_case ? (x.numel() == y.numel()
? handler.AcquireExtendSrcMemory(non_const_x, 0)
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

what is extendsrc for?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This aligns with former PR #59421. Since we need to manually broadcast src1/src2, I hereby name such operations as "extend".

@xinyu-intel xinyu-intel merged commit 0d4bbd6 into PaddlePaddle:develop Dec 6, 2023
29 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
contributor External developers Intel
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants