-
Notifications
You must be signed in to change notification settings - Fork 2.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Suggestion: See what the reasoning models are thinking before they give their output. #3086
Comments
I tried adding: "openrouter/deepseek/deepseek-r1": {
"remove_reasoning": false
}, in |
Hey, I also tried a few times looking for how to enable this as I assumed it already exists surely... |
There were config options that spoke about tweaking reasoning effort. |
see #3073 |
@zwilch Sorry but adding this in my
I also tried How did you get the thinking tokens to be printed in the other thread? |
Remove reasoning goes in See here: https://aider.chat/docs/config/adv-model-settings.html#model-settings |
Here's an example of a fireworks model setting using that: aider/aider/resources/model-settings.yml Line 614 in b0d58d1
|
OOOOHHHHH I was wondering why this page began with json, but the rest of it was in yaml. I never realized that one was is But sadly, I tried what you said, and it still didn't seem to work.
My
I am on master branch btw. But still unable to see the "thinking"/reasoning tokens. |
Oh, sorry |
I see, is there a similar Or does this require new feature development? |
From what I understand, we want to send the following two params in the HTTP POST Request body: Only mentioning See the streaming example in Postman below: Since these two params
But that didn't work either. I couldn't find any Log for Aider's HTTP requests. |
With this change:
I verified that the reasoning tokens are sent by the API. But when I started playing around with The main issue is: I added:
and got:
See how the first delta object we receive is already after it's done thinking? Yet I can see that
"I'll go through" is what the reasoning strings say if you put them together. Any pointers? @paul-gauthier ? |
@V4G4X can you share the raw response from openrouter // what it looks like when you call the api directly? I suspect the value isn't retuned in chunks without a specific parameter |
Yes, that's right, you have to send both This is the command I used to call the OpenRouter API directly:
This is the dump file: r1_raw.txt These are the first couple of lines:
Hope this helps. |
Issue
Since I was facing a lot of wait time using reasoning models like DeepSeek R1 and Perplexity: Sonar Reasoning in Aider,
i.e an average wait time in minutes even for simple prompts like:
I tried testing them out in Openrouter's chatroom.
I noticed that the models/APIs were not lagging, but that they took a lot of time to think before they responded.
And I could see what they were thinking as they did.
It would help my user experience A LOT, if I could see this thought process when using Aider.
Do I have to wait because it's thinking, is it thinking in the right direction? (If it's not, I can cancel the request and direct it better)
Or is it the API just stuck?
Since the OpenRouter chatroom can get the reasoning tokens, I assume that we can too?
Version and model info
aider 0.72.3
model: openrouter/deepseek/deepseek-r1
The text was updated successfully, but these errors were encountered: