Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Running the Alpaca LoRA does not work on GPU #412

Closed
1 task done
Arargd opened this issue Mar 18, 2023 · 3 comments
Closed
1 task done

Running the Alpaca LoRA does not work on GPU #412

Arargd opened this issue Mar 18, 2023 · 3 comments
Labels
bug Something isn't working

Comments

@Arargd
Copy link

Arargd commented Mar 18, 2023

Describe the bug

When attempting to run the LoRA on GPU results in nothing happening.

This occurs for my non-wsl, wsl, and seperate Linux boot on my machine.

It does however generate and work properly when using the --cpu option.

Also to note I had to replace the bitsandbytes_cpu.so with bitsandbytes_cuda117.so to function, if its in any way related.

Is there an existing issue for this?

  • I have searched the existing issues

Reproduction

run the command-line: python server.py --listen --load-in-8bit, select 7b weights, go to parameters, select alpaca-lora-7b, using default prompt and parameters select generate, nothing happens other than a log with 0 tokens and a transformers warning.

Screenshot

No response

Logs

Output generated in 0.23 seconds (0.00 tokens/s, 0 tokens)

C:\Users\Arargd\miniconda3\envs\textgen\lib\site-packages\transformers\generation\utils.py:1374: UserWarning: You are calling .generate() with the `input_ids` being on a device type different than your model's device. `input_ids` is on cuda, whereas the model is on cpu. 
You may experience unexpected behaviors or slower generation. 
Please make sure that you have put `input_ids` to the correct device by calling for example input_ids = input_ids.to('cpu') before running `.generate()`.

System Info

Windows 10 (both WSL and non WSL)
Linux Ubuntu 22.04.2
RTX 2070 8GB
32 GB 3600 MHz DDR4 RAM
@Arargd Arargd added the bug Something isn't working label Mar 18, 2023
@bartman081523
Copy link

You can try this
#400 (comment)

@Arargd
Copy link
Author

Arargd commented Mar 19, 2023

You can try this #400 (comment)

I have commented in that issue, but yes that works for getting me to generate normally without LoRA. With LoRA attached it refuses to generate anything on GPU, however.

@bartman081523
Copy link

bartman081523 commented Mar 19, 2023

it refuses to generate anything on GPU, however.

https://www.reddit.com/r/LocalLLaMA/comments/11o6o3f/how_to_install_llama_8bit_and_4bit/
#332

also please post a log, and have you updated the repo with git pull?

@Arargd Arargd closed this as completed Mar 19, 2023
@Arargd Arargd closed this as not planned Won't fix, can't repro, duplicate, stale Mar 19, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants