-
Notifications
You must be signed in to change notification settings - Fork 464
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Can we get an ollama.stop? #287
Comments
I second this as well. And to keep things clear. I'm talking about making available in the ollama python lib the same behavior you get when you run |
+1 |
The generate and chat methods both accept a float for keep-alive. When hot-switching models, I set my keep-alive=0 at the call and it should unload immediately (at least that's been my observation). |
Great - thanks for bringing - will scope in :) |
Also, as I understand, there can be multiple generations at once through |
ollama stop was a great addition. I was hoping we could get stop added to the API library.
Thanks
The text was updated successfully, but these errors were encountered: