Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

llama_cpp_server::supervisor: crates/llama-cpp-server/src/supervisor.rs:94: llama-server <chat> exited with status code 1 #2544

Closed
danny-su opened this issue Jun 28, 2024 · 15 comments

Comments

@danny-su
Copy link

danny-su commented Jun 28, 2024

Describe the bug
/opt/homebrew/bin/tabby serve --device metal --port 8088 --model TabbyML/CodeGemma-2B --chat-model Deepseek-V2-Lite-Chat --parallelism 1

Warp 2024-06-28 18 24 33

Information about your version
Please provide output of tabby --version
tabby 0.13.0

Information about your GPU
Please provide output of nvidia-smi
mps

Additional context
Add any other context about the problem here.

@wsxiaoys
Copy link
Member

Seems the prompt template is broken for Deepseek lite chat - looking

@wsxiaoys
Copy link
Member

wsxiaoys commented Jun 28, 2024

After investigation - i noted the llama.cpp version we pinned hasn't supported deepseek v2 style chat template. As a result it's not usable with latest tabby distribution. I've removed it from the registry.

As workaround, please follow discussion of #2451 to see how to connect tabby to an external http endpoint (that has support for deepseek v2)

@kibabyte
Copy link

any ETA when this will be usable out of the box?

@wsxiaoys
Copy link
Member

Tabby does bi-weekly patch release - as long as the fix is integrated in upstream llama.cpp, we shall be able to integrated them.

@wsxiaoys
Copy link
Member

wsxiaoys commented Jul 4, 2024

Please take a look at 0.13.1-rc.3, where Deepseek-Lite-V2 is now ready.

@kibabyte
Copy link

kibabyte commented Jul 4, 2024

Thank you very much, will look into it!

@kibabyte
Copy link

kibabyte commented Jul 4, 2024

I have been unable to find "0.13.1-rc.3" in this repo. Mind pointing me in the right direction please?

@kibabyte
Copy link

kibabyte commented Jul 5, 2024

What would be the model_ids I would need for the command on Windows? I am currently trying to test but unable to find the proper ID to put in.

@wsxiaoys
Copy link
Member

wsxiaoys commented Jul 6, 2024

Hi - it's not in official registry, but you might try creating one by yourself, or you can also use my forked registry at https://github.com/wsxiaoys/registry-tabby (wsxiaoys/Deepseek-V2-Lite-Chat)

@moqimoqidea
Copy link
Contributor

@wsxiaoys

Hi, In my tests llama.cpp b3267 outputs nonsensical GGGGG content, similar to the community issue 8254 behavior, and looks like it needs to be upgraded to version llma.cpp again.

llama.cpp issue 8254: Bug: Failed to load quantizied DeepSeek-V2-Lite-Chat model

@wsxiaoys
Copy link
Member

wsxiaoys commented Jul 8, 2024

I also noticed the issue and it seems really the problem with system message - I've send out a patch #2596 to fix it for 0.13.1 (it's removed in main branch anyway). Will tag a new rc soon.

@moqimoqidea
Copy link
Contributor

The patch you are referencing appears to be the chat feature, I tested it with llama.cpp b3267 version using tabby server v0.13.1-rc6 and noticed that the templates in the chat feature seem to be behaving with this instead of the user's content. I noticed that the file you are referencing is openai_chat.rs, not sure if it will affect the output of llama.cpp as the engine. I report this issue to share information, and I will look into this problem later.

Let me clarify one thing: in my previous test, I tested code completion using DeepSeek-Coder-V2-Lite-Base, which outputs nonsensical GGGGG content. Share my progress with you: I just used the newer llama.cpp b3334 version and my case is working fine. FYI.

@wsxiaoys
Copy link
Member

wsxiaoys commented Jul 8, 2024

The release https://github.com/TabbyML/tabby/releases/tag/v0.13.1-rc.8 is ready for testing for the DeepSeek-V2-Lite-Chat. Please give it a try.

Hi @moqimoqidea - if you still encounter errors with other models, please file a new issue for tracking. Thank you!

@wsxiaoys
Copy link
Member

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

4 participants