Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

"Request timed out" #1

Closed
maxbengtzen opened this issue Sep 20, 2024 · 1 comment · Fixed by #2
Closed

"Request timed out" #1

maxbengtzen opened this issue Sep 20, 2024 · 1 comment · Fixed by #2

Comments

@maxbengtzen
Copy link

maxbengtzen commented Sep 20, 2024

First of all - magical tool! Just what I've been searching for.

I got it set up pretty easily. I first tried with a local Ollama setup on my Intel NUC (11th gen) and small models (gemma:2b and phi3:latest 3.8b), but without success. I then tried setting it up with a OpenAI API key with success.

I would like to get the local Ollama setup to work.

The log in Portainer tells me that "generated an exception: Request timed out.". I suspect the answers take too long to report as successful? Is there any way to extend the timeout period?

If I summarize an article via Open WebUI and gemma:2b I see the following information:

image

That's OK for me. 1m 15s. I don't need real time speed. And since it's fetching the unread count each minute, I guess I would be OK with it buffering the workload. I may also have to work with white-/blacklisting to not run at 100% all of the time.

While we're at it; might want to rename white-/blacklist to deny-/allow-list?

@Qetesh
Copy link
Owner

Qetesh commented Sep 20, 2024

Thank you for your interest in the project!

As you noted, this error typically results from an OpenAI API timeout. I also use Ollama. If you run Ollama on the host where the container is located, the correct address is http://host.docker.internal:11434/v1, or you can set ollama's environment variables like OLLAMA_HOST=“0.0.0.0” and OLLAMA_ORIGINS=”*”.

Considering similar use cases, I’ll soon add request timeout and thread count parameters to the configuration file. Please note that multithreading is enabled by default, which may result in higher request times and memory usage than expected.

The suggestion about the deny-/allow-list change is a good one, and I’ll implement it as well.

Thanks again for your feedback.

@Qetesh Qetesh linked a pull request Sep 21, 2024 that will close this issue
@Qetesh Qetesh closed this as completed in #2 Sep 21, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants