Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Running --local load my local model over and over again everytime a message is sent #666

Closed
xinranli0809 opened this issue Oct 20, 2023 · 5 comments
Labels
Bug Something isn't working Local Model Running a local model

Comments

@xinranli0809
Copy link

Describe the bug

I managed to load a local mistral 7b model and was able to chat back and forth with the interpreter, but the time it takes to begin streaming the tokens is extremely long after I send a message, considering I get decent token/s once it starts streaming.

Upon a closer look at the memory usage in the task manager, I see that every time I press enter and send a message, the program loads a brand new Ooba server and a GGUF model! This is confirmed by adding a print("Ooba starting and model loading!") before Line 84 in Ooba's llm.py and that an additional python process will pop up in the task manager and occupy 8GB RAM after sending every message.

Please see the screenshots for details.

Reproduce

  1. interpreter --local
  2. chat with it and see how long it takes before begin streaming the response
  3. Monitor your RAM usage, the model is loaded into RAM every time you send a message

Expected behavior

  1. Model load only once per session

Screenshots

image
image

Open Interpreter version

0.1.10

Python version

3.11.6

Operating System name and version

Windows 10

Additional context

I am also getting an interpreter.procedure being a NoneType error, for some reason... But I don't think these are related issues. I tried putting the procedures_db.json in the same directory as get_relevant_procedures_string.py and loaded it manually to make sure it was presenting. The NoneType error went away but a new model is still being loaded after every message.

@xinranli0809 xinranli0809 added the Bug Something isn't working label Oct 20, 2023
@Notnaton
Copy link
Collaborator

Notnaton commented Oct 23, 2023

There are two bugs here

@xinranli0809
Copy link
Author

There are two bugs here

It works, thanks!

@ericrallen ericrallen added the Local Model Running a local model label Oct 27, 2023
@ericrallen
Copy link
Collaborator

It works, thanks!

Hey there, @xinranli0809!

Does that mean this issue is resolved?

@Notnaton
Copy link
Collaborator

It works, thanks!

Hey there, @xinranli0809!

Does that mean this issue is resolved?

@ericrallen
This is an ooba issue, once ooba is merged with KillianLucas/ooba#8 it should be fixed.

@ericrallen
Copy link
Collaborator

Thanks for all of the awesome work, @Notnaton!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Bug Something isn't working Local Model Running a local model
Projects
None yet
Development

No branches or pull requests

3 participants