Auto mode has become exceptionally effective!

I would like to thank Cursor for the exceptional Auto mode performance.
The quality is now indistinguishable from Claude-4, making me forget other models exist.

I exclusively use Auto mode. Working with Flutter, despite its complexity, has become seamless.

Thank you, Cursor. Your efforts are outstanding.

5 Likes

I have also found it to be exceptional, until today. For the entire day, every response has taken a minimum of 10 seconds and upwards of 30-50 seconds. And the quality of the responses has been garbage! I sure would like it to perform how it has performed every day until today!

Hi drubbling,

I wanted to offer a different perspective regarding the waiting times you’ve been experiencing with auto mode. From my observations, I’m starting to believe that the issue might be connected to exceeding a certain command limit. Specifically, I suspect it could be related to hitting a 500-command threshold. It seems plausible that once this particular limit is reached, there’s a noticeable and consistent decrease in performance, leading to the longer response times you’ve described.

It might be worth investigating if your usage patterns align with this theory, as it could help pinpoint the cause of the recent performance degradation.

Thanks,
Bashar

So you’ve also experienced this?

I’m still pretty new to Cursor. What do you consider a ā€˜command’? 500 over what time frame? Today I’ve only been trying to chat with Ask/Auto to set up a few rules for a brand new project. Not even a single code file yet.

With GPT-5 now available at a great price, I think they moved from GPT-4.1 to GPT-5. The output costs a little more, but the lower input cost makes a big difference for them, which is why it’s so much better.

Does anyone know how I can enable auto mode and test it?

In your screenshots to the right of Agent^ is another selector set to o3^. If you click on that second selector it should give you a toggle to turn on Auto.

1 Like

Might be related to API key Auto-Mode is not available - #2 by deanrie

There was also a bug where previously used api keys caused the issue, deanrie mentions that lower in that thread

It was great yesturday because it was thinking, TODAY THEY TURNED IT OFF??? Bring back thinking for auto!!!

1 Like

TRUE, please return the ā€œthinkingā€œ @cursor

1 Like

When i ask Auto which model it is, I always get Sonnet 4 (non-thinking), so yeah, that’s why it’s good!

1 Like

I am actually quite a fan of NON-Thinking models! I used Claude 4 Sonnet thinking for the first few months. The whole ā€œthought processā€ stuff seemed really cool.

Then I learned that thinking models cost twice as many requests. Further, they seemed to go off doing more of their own thing, than non-thinking models as well. (ESPECIALLY Gemini at the time…man that thing is a hyper opinionated bulldozer, and it doesn’t like Sonnet code!)

After working with Cursor for a while I learned the value of planning, so I would first ask the agent to generate a plan and report to me then wait for further instruction. I’d refine the plan a bit before finally telling the agent to enact the plan (or more often, enact it one phase at a time, I now even have rules about how plans should be created in a multi-phasic fashion, etc.)

I realized that by planning, AND using a thinking model, sometimes the two were at odds, AND I was RIPPING through requests at an insane rate. So I switched from primarily using Sonnet thinking, to Sonnet non-thinking. That seemed to give me a lot more requests, and I honestly found the results more acceptable, not to mention faster (a non-thinking model just does, it doesn’t wast time pretending to cogitate while burning tokens needlessly.)

I’ve been using gpt-5-fast today, and I am sitting and waiting a LOT more as the thinking does its thing, and wastes my time (while hoping it really truly is FREE right now!) when I figure Sonnet would have been done already. This model is pretty fast, so it seems to move more quickly than Sonnet when it actually acts, but sometimes the ā€œthinkingā€ process can roll on for a good while before it actually acts, and when I already have a plan of action I want to follow, the thinking is really quite annoying and seems quite useless…

1 Like

How? I did not able to ask this question ! how did you ask the LLM what is the model?

1 Like

for the love of god NO….the thinking was aweful, it belongs and should stay on manual mode only

1 Like

I agree. Auto is great and appears to use Claude Sonnet 4 most of the time.

If they can offer a thinking mode for free this would be HUGELY HELPFUL. Have option for ā€œthinking autoā€ and ā€œnon-thinking autoā€. Shouldn’t have to be forced to pay to have a thinking model given GPT-OSS and QWEN!

1 Like

I also found Auto model way more useful comparing to even recent experience. I’d like to stick to it as much as possible, since it’s include into plan and with new pricing I very quickly reach the limits.

My biggest concern is performance. Sometimes 100 line of code is generated 5 minutes. I also experience hang-out request, which force me to stop the current one and prompt again;

Kudos to the team and I hope they address the performance issues soon!

1 Like

is so good that enters infinite loop every 10 seconds