-
-
Notifications
You must be signed in to change notification settings - Fork 5.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
AttributeError: 'NoneType' object has no attribute 'Llama' #4817
Comments
Same problem here, Xeon E5-2660 V1, 40gb of ram, Windows Server 2022 Datacenter 21H2, in a Proxmox virtual machine. Using dolphin-2.2.1-mistral-7b.Q4_K_M.gguf model. |
I had this issue as well with an ARM-based VPS (Hetzner Cloud, 8c 16gb). The machines are using KVM for virtualization, both installed with Debian 12. I deleted the ARM VPS again, but I'm open to do some debugging if anyone wants me to. |
I managed to work around the issue by explicitly specifying the version of Using the nowheels or cpu_only_noavx2 requirements.txt is required for me as I run into issue #4887 otherwise on aarch64. Maybe one of you could try simply adding See diff of my changes:
|
This issue has been closed due to inactivity for 6 weeks. If you believe it is still relevant, please leave a comment below. You can tag a developer in your comment. |
Bump, this issue is still unresolved. This is another way to fix it for me within the diff --git a/requirements_noavx2.txt b/requirements_noavx2.txt
index fc2795cb..73a64ede 100644
--- a/requirements_noavx2.txt
+++ b/requirements_noavx2.txt
@@ -45,6 +45,7 @@ https://github.com/oobabooga/llama-cpp-python-cuBLAS-wheels/releases/download/te
https://github.com/oobabooga/llama-cpp-python-cuBLAS-wheels/releases/download/textgen-webui/llama_cpp_python_cuda_tensorcores-0.2.38+cu121avx-cp310-cp310-win_amd64.whl; platform_system == "Windows" and python_version == "3.10"
https://github.com/oobabooga/llama-cpp-python-cuBLAS-wheels/releases/download/textgen-webui/llama_cpp_python_cuda_tensorcores-0.2.38+cu121avx-cp311-cp311-manylinux_2_31_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" and python_version == "3.11"
https://github.com/oobabooga/llama-cpp-python-cuBLAS-wheels/releases/download/textgen-webui/llama_cpp_python_cuda_tensorcores-0.2.38+cu121avx-cp310-cp310-manylinux_2_31_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" and python_version == "3.10"
+llama-cpp-python==0.2.38; platform_machine == "aarch64"
# CUDA wheels
https://github.com/jllllll/AutoGPTQ/releases/download/v0.6.0/auto_gptq-0.6.0+cu121-cp311-cp311-win_amd64.whl; platform_system == "Windows" and python_version == "3.11" |
The error probably means that the llama_cpp module could not be imported by the script modules/llamacpp_model.py
and try to load a model again. The new error message should be more specific which should make it easier to get help. Same would work if the llama_cpp_cuda or llama_cpp_cuda_tensorcores should be loaded. Don't forget to reverse that changes after troubleshooting. Perhaps in the future an error output can be integrated into text-generation-webui if neither llama_cpp, llama_cpp_cuda, nor llama_cpp_cuda_tensorcores could be loaded. Good luck! |
Hi, |
Hello, Maybe also interesting: tensorflow - amahendrakar |
Hi @burrizza, just reinstalled with noavx2 but got same error. |
Hello @gianlucasullazzo, Good luck! |
At least on ARM64, My diff previously posted simply adds an install option for ARM64 machines to solve this. @gianlucasullazzo Can you share what hardware and OS you're using? |
Hi @RealStickman, @burrizza. About the file used, I closed cmd so I have no more access to that log. Is there a log file inside install folders? |
What Python version are you using? Prebuilt wheels are only specified for Python 3.10 and 3.11 on Windows. You can try |
I have 3.10.11. |
Same issue on macOS 15. |
Describe the bug
Whenever I try to load a model in cpu only mode the model doesn't load and I see this error message.
I noticed there is some discussion about this at #4098.
Is there an existing issue for this?
Reproduction
Try to load a model?
Screenshot
No response
Logs
The text was updated successfully, but these errors were encountered: