Hey, thanks for the report. Max Mode is mainly intended for models with a context window larger than 200k tokens - like Gemini 3 Pro, GPT 5.1, and Grok 4.
For Claude Opus 4.5, the standard context window is already 200k tokens, so Max Mode might not provide a noticeable increase.
A few clarification questions:
What behavior did you expect from Max Mode? For example, the ability to include more files in context?
Did you test the same requests with and without Max to compare?
I’ve got a repeatable issue with Anthropic models in Cursor during the “planning next moves” phase.
It will sit on “planning next moves” for ~5 minutes at a time.
The actual code execution / implementations are fast; the slowdown is only between steps when it goes back to “planning next moves”.
For some prompts this happens 3–4 times in a single run, so one prompt can take 15–20 minutes end to end.
This only started in the last 24–48 hours.
Scope of the problem:
Only in Cursor. The same Anthropic models behave normally in AntiGravity and Claude Code.
Only Anthropic models. Other models in Cursor (e.g. GPT-5.1) are “as expected” and very fast.
It’s not a networking issue on my side:
DNS ~3–4 ms
Connect/TLS to api2.cursor.sh ~80 ms each, total ~250 ms
Pings ~80–260 ms
Health checks to your API all return 200 and complete quickly.
So this looks like something specific to how Cursor is calling Anthropic for the planning phase, not my connection or machine. Please check whatever internal calls you’re making during “planning next moves” for Anthropic models over the last 24–48 hours.