Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Question] PandaGPT with llama.cpp #12

Open
ningshanwutuobang opened this issue Jul 1, 2023 · 0 comments
Open

[Question] PandaGPT with llama.cpp #12

ningshanwutuobang opened this issue Jul 1, 2023 · 0 comments

Comments

@ningshanwutuobang
Copy link

I have tried to use llama.cpp for PandaGPT in panda_gpt_llama_cpp.
The script get poor performance. Is there any thing wrong for the procedure? Or is it just the limit of the model or q4_1 precision?

The following are my steps.

  1. Obtain vicuna v0. Use FastChat@v0.1.10 to merge llama-13b-hf and vicuna-13b-delta-v0.
  2. Merge lora weights to vicuna v0.
  3. Convert it to ggml format and quantize it to q4_1. The result is ggml-pandagpt-vicuna-merge.
  4. The script is located in panda_gpt_llama_cpp.

The model seems to recognize <Img>...</Img> labels.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant