-
Notifications
You must be signed in to change notification settings - Fork 165
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Phi-3 mini stop token not recognized #74
Comments
Per the tokenizer config, the end of text is:
It looks like people are manually modifying the config or manipulating the runtime. Ideally the shipping tokenizer config would indicate the correct eos tokens. That said, the swift-transformers code only allows a single eos token. We have a couple options here:
|
In Python, we do a combination of 1 (change the config) and 3. The argument is nice because it adds a lot of flexibility. Regarding 2, it's not clear to me if the original model authors intended the model to have two EOS tokens or if it was just an oversight not to update it in the tokenizer... |
OK, then let's plan on adding #3 -- that will make it flexible at least. |
OK, I am working on the model configuration for #53 so it will be included with that. |
#76 should fix this |
This Phi 3 model used in the LLMEval app doesn't behave as expected. It looks like the stop token is not being recognized.
Prompt:
Output:
Related issues:
ggerganov/llama.cpp#6903
nomic-ai/gpt4all#2271
huggingface/swift-transformers#98
The text was updated successfully, but these errors were encountered: