-
Notifications
You must be signed in to change notification settings - Fork 10.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add AVX_VNNI support for intel x86 processors #4301
Comments
Same question. |
I build windows rocm build of llamacpp, it does support VNNI. |
why rocm support AVX_VNNI ?? |
This should be helpful I guess : https://www.intel.com/content/www/us/en/content-details/791610/optimizing-and-running-llama2-on-intel-cpu.html |
^ Based on this comment i have implemented this PR |
This issue was closed because it has been inactive for 14 days since being marked as stale. |
Prerequisites
Please answer the following questions for yourself before submitting an issue.
Feature Description
Enable AVX_VNNI ( we already have AVX512 ). Enable support for AVX_VNNI will make intel processor able to use int8 computation.
Motivation
Intel x86 processors are still very popular among consumer and data centers hardwares. Enable AVX_VNNI will make the performance of llama cpp on these processors better.
The text was updated successfully, but these errors were encountered: