Replies: 1 comment 3 replies
-
For llamas+ instruct loras, only 8bit is functional currently. Maybe you are just comparing them inaccurately? If you're using a converted 4bit model of https://huggingface.co/chavinlo/alpaca-native/discussions |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I've read that the 4-bit version shouldn't be noticeably different than the original 16-bit version. However, it seems significantly worse, atleast for the 7B version which I tested.
8-bit (bitsandbytes):
4-bit (GPTQ):
Beta Was this translation helpful? Give feedback.
All reactions