-
Notifications
You must be signed in to change notification settings - Fork 4.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Using train_with_template on mistral end up in a model with a loop #3055
Comments
Hi, I never seen such wired issue before, will test this case when I find myself available. I did not familiar with mistral template, if it is the same as Llama, which use |
Hi, I think this is a model specific behavior. Some models need to specifically learn to generate the stop tokens.
If you would like to have a try on your Mistral model, which is to change the line
|
Thanks @congchan for looking into this. I have the same kind of issue with vicuna-7b, but it is a bit more random:
Then after several runs (the last one is a loop):
I will try with the solution you suggest! 🙏 |
I use
train_with_template.py
withmistralai/Mistral-7B-Instruct-v0.2
Then I run the model:
I get:
There might be a problem with the
stop
token or something?The text was updated successfully, but these errors were encountered: