-
Notifications
You must be signed in to change notification settings - Fork 5.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Context Shifting #4588
Comments
This issue has been closed due to inactivity for 6 weeks. If you believe it is still relevant, please leave a comment below. You can tag a developer in your comment. |
Is there any way to get this for Exl2? |
This was closed as stale. Did it ever get implemented @oobabooga? Literally this is driving me to use KoboldCpp. As soon as you hit context limits in Oobabooga it becomes obnoxious in comparison. :/ |
Wish too know too |
Thanks @oobabooga you rock |
I remember when I started to use this project (when it just created) with pygmalion 6b a long time ago, it worked exactly like that - AI were shifted context and forgot old info without any slow down. Then at some moment, it stopped worked like that (maybe when exllama appeared). I wonder why that happened... |
Description
About 10 days ago, KoboldCpp added a feature called Context Shifting which is supposed to greatly reduce reprocessing. Here is their official description of the feature:
Any chance this gets added to Ooba as well?
Additional Context
Reddit thread: https://www.reddit.com/r/LocalLLaMA/comments/17ni4hm/koboldcpp_v148_context_shifting_massively_reduced/
llama.cpp pull: ggerganov/llama.cpp#3228
kobold.cpp 1.48.1 release: https://github.com/LostRuins/koboldcpp/releases/tag/v1.48.1
The text was updated successfully, but these errors were encountered: