Sonnet 4 / GPT-5 usage period is too short, Auto model is highly problematic

hi @matrix_code thank you for the detailed post. Here are a few remarks from my experience:

  • Please check if you can optimize your Sonnet 4 and GPT-5 usage by reducing context size and optimizing Agent runs as this will give you more usage. Check out Understanding LLM Token Usage if this helps you with best practices.
  • We are improving Auto mode by adding better models and improving handling. Feel free to file bug reports with issues you find with Auto.

As to why there is a different usage to apps like ChatGPT:

  • Context matters a lot and increasing context means that more unrelated code, comments etc is presented to AI for processing. This confuses models and increases mistakes by AI.
  • AI providers do have usage optimizations in their own apps like Claude chat and ChatGPT but they also have hourly limits after which they downgrade to smaller models which is in Cursor the users choice.

Additionally, note that the cost is essentially set by the AI providers and since powerful LLMs are very large and must run on large machines with huge graphic cards the cost for those is much higher than smaller models. See the price difference between Sonnet and Opus where gap is 5 times in cost.