Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[CPU] Enabled FP16 Compressed FC on models with PagedAttention #26279

Conversation

dmitry-gorokhov
Copy link
Contributor

No description provided.

Copy link
Contributor

@usstq usstq left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM, vcvtph2ps used for decompressing FP16 into FP32 introduced overhead according to tests shown in https://jira.devtools.intel.com/browse/CVS-133453, but given the fact:

  • most serving workload with big-batch-size (compute-bounded) would be running on Xeon with (src-bf16, wei-bf16) or (src-f16, wei-f16) kernel
  • most likely workload running on AI PC, would use small batch-size and (src-f32, wei-f16) can help no matter it's SDPA or PA.

so we can safely enable it

@dmitry-gorokhov dmitry-gorokhov added this pull request to the merge queue Oct 14, 2024
Merged via the queue into openvinotoolkit:master with commit 7250c1e Oct 14, 2024
153 checks passed
@dmitry-gorokhov dmitry-gorokhov deleted the feature/fp16_compressed_fc branch October 14, 2024 07:19
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
category: CPU OpenVINO CPU plugin
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants