Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ModuleNotFoundError: No module named 'llama_inference_offload' #869

Closed
1 task done
xiaojunhh opened this issue Apr 7, 2023 · 6 comments
Closed
1 task done

ModuleNotFoundError: No module named 'llama_inference_offload' #869

xiaojunhh opened this issue Apr 7, 2023 · 6 comments
Labels
bug Something isn't working stale

Comments

@xiaojunhh
Copy link

Describe the bug

ModuleNotFoundError: No module named 'llama_inference_offload'I searched, and someone had the same problem as me before, but he used an AMD graphics card, while mine was an Nvidia 3080

Welcome to bitsandbytes. For bug reports, please submit your error trace to: https://github.com/TimDettmers/bitsandbytes/issues

CUDA SETUP: CUDA runtime path found: D:\GPT-fast\oobabooga-windows\oobabooga-windows\installer_files\env\bin\cudart64_110.dll
CUDA SETUP: Highest compute capability among GPUs detected: 8.6
CUDA SETUP: Detected CUDA version 117
CUDA SETUP: Loading binary D:\GPT-fast\oobabooga-windows\oobabooga-windows\installer_files\env\lib\site-packages\bitsandbytes\libbitsandbytes_cuda117.dll...
The following models are available:

  1. opt-6.7b
  2. vicuna-13b-GPTQ-4bit-128g

Which one do you want to load? 1-2

2

Loading vicuna-13b-GPTQ-4bit-128g...
Traceback (most recent call last):
File "D:\GPT-fast\oobabooga-windows\oobabooga-windows\text-generation-webui\server.py", line 293, in
shared.model, shared.tokenizer = load_model(shared.model_name)
File "D:\GPT-fast\oobabooga-windows\oobabooga-windows\text-generation-webui\modules\models.py", line 100, in load_model
from modules.GPTQ_loader import load_quantized
File "D:\GPT-fast\oobabooga-windows\oobabooga-windows\text-generation-webui\modules\GPTQ_loader.py", line 14, in
import llama_inference_offload
ModuleNotFoundError: No module named 'llama_inference_offload'

Is there an existing issue for this?

  • I have searched the existing issues

Reproduction

My graphics card is Nvidia 3080, in a conda environment with pytorch/cuda, run:

pip install -r requirements.txt
Then in this repository:

pip install -e . The result is not good

Screenshot

No response

Logs

Welcome to bitsandbytes. For bug reports, please submit your error trace to: https://github.com/TimDettmers/bitsandbytes/issues
================================================================================
CUDA SETUP: CUDA runtime path found: D:\GPT-fast\oobabooga-windows\oobabooga-windows\installer_files\env\bin\cudart64_110.dll
CUDA SETUP: Highest compute capability among GPUs detected: 8.6
CUDA SETUP: Detected CUDA version 117
CUDA SETUP: Loading binary D:\GPT-fast\oobabooga-windows\oobabooga-windows\installer_files\env\lib\site-packages\bitsandbytes\libbitsandbytes_cuda117.dll...
The following models are available:

1. opt-6.7b
2. vicuna-13b-GPTQ-4bit-128g

Which one do you want to load? 1-2

2

Loading vicuna-13b-GPTQ-4bit-128g...
Traceback (most recent call last):
  File "D:\GPT-fast\oobabooga-windows\oobabooga-windows\text-generation-webui\server.py", line 293, in <module>
    shared.model, shared.tokenizer = load_model(shared.model_name)
  File "D:\GPT-fast\oobabooga-windows\oobabooga-windows\text-generation-webui\modules\models.py", line 100, in load_model
    from modules.GPTQ_loader import load_quantized
  File "D:\GPT-fast\oobabooga-windows\oobabooga-windows\text-generation-webui\modules\GPTQ_loader.py", line 14, in <module>
    import llama_inference_offload
ModuleNotFoundError: No module named 'llama_inference_offload'

System Info

Nvidia 3080
@xiaojunhh xiaojunhh added the bug Something isn't working label Apr 7, 2023
@xiaojunhh
Copy link
Author

windows11

@da3dsoul
Copy link
Contributor

da3dsoul commented Apr 7, 2023

#879

@vague-score
Copy link

i resolved this problem in my installation by adding llama-cpp-python==0.1.23 to the requirements.txt and then running the intall.bat one click installer

@github-actions github-actions bot added the stale label May 7, 2023
@github-actions
Copy link

github-actions bot commented May 7, 2023

This issue has been closed due to inactivity for 30 days. If you believe it is still relevant, please leave a comment below.

@github-actions github-actions bot closed this as completed May 7, 2023
@Snimm
Copy link

Snimm commented May 9, 2023

I am having the same issue

@TalhaUusuf
Copy link

same issue

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working stale
Projects
None yet
Development

No branches or pull requests

5 participants