-
Notifications
You must be signed in to change notification settings - Fork 5.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
No module named 'llama_inference_offload' on Arch Linux #879
Comments
Follow the steps here https://github.com/oobabooga/text-generation-webui/wiki/LLaMA-model#installation |
Specifically, the |
Also, I had better luck with vicuna-13b-4bit-128g on cuda. You will probably need to specify |
This failed with following error: :( |
Ah I'm guessing arch has a newer one available than Ubuntu, for example. I would install g++ manually, staying on 11.x, then try |
#850 looks relevant to that |
I had the same issue with fedora 37. To fix this I did the following. conda install -c conda-forge gxx If that doesn't work try conda install gcc_linux-64==11.2.0 |
same error on windows 11 |
Got this working on Arch. Here are the steps:
That should do it. Just did this with a fresh install so it should not be missing any steps. |
EDIT: Seem like this might be a problem with having mismatched cuda and nvcc versions. (You probably don't need to reinstall linux, i just did it out of frustration and found out the script afterward . Running that script should work as it will delete all previously installed drivers for you.) Full output:
|
When facing the original problem, I somehow missed the need for the Make sure the hierarchy of directories goes like this: Hope this helps! |
This issue has been closed due to inactivity for 6 weeks. If you believe it is still relevant, please leave a comment below. You can tag a developer in your comment. |
Describe the bug
Try to run server.py in following: python server.py --wbits 4 --groupsize 128 and get error No module named 'llama_inference_offload' I did this fix: #400 (comment) did not help.
Is there an existing issue for this?
Reproduction
Run following command: python server.py --wbits 4 --groupsize 128
Screenshot
No response
Logs
System Info
The text was updated successfully, but these errors were encountered: