-
Notifications
You must be signed in to change notification settings - Fork 221
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Revert "fix: use /infill for llama.cpp code-completions (#513)" #533
Conversation
…-llama-infill # Conflicts: # src/main/kotlin/ee/carlrobert/codegpt/codecompletions/CodeCompletionRequestFactory.kt
Hey, just curious why ggerganov/llama.cpp#7102 (comment) prompted this? |
We use GGUF models from HF (e.g https://huggingface.co/Qwen/CodeQwen1.5-7B-Chat-GGUF) which (afaik) do not have the special token modification from your PR (ggerganov/llama.cpp#7166), therefore we can't use them with |
You can use mine which does. :) |
Thats nice, but we would need a working GGUF for all models that we support, the list is quite long (see HuggingFaceModel) with different FIM prompt templates (see InfillPromptTemplate) and its continously growing with new models coming up. We want the same solution for every existing and new model, so |
In that case I think you will have to maintain your own copies, which shouldn't be that hard though using |
Thanks for your suggestion but its easier for us to just maintain the FIM templates in our project and using the |
This reverts commit 8de72b3.
As discussed in #510 this reverts the switch from
/completion
to/infill