We are currently seeing some ongoing issues with our database infrastructure, which may be causing issues both in-app and on cursor.com around usage reporting.
Please keep an eye on status.cursor.com as that may be having an effect here!
We are aware of the bad reporting issue for Ultra plans on cursor.com - this is being worked on by the team too, and should be fixed soon.
If the rate limiting metrics were not a black box to the user, there would be less frustration on our end. Does the governing algorithm details have to be secret ? Can there be a rate limit display showing how many percent of the rate limit you have used in real time ? My codebase employs rate limiting with my api providers I’m interested in how Cursor tracks commpute units
The models page documentation no longer says what the request pricing per model is only the context size. How about puting the request pricing back, and including a rate limit quantifier / factor ? Also auto mode is a no go zone for me. Some models I will never use again, deleting code blocks is a blacklist offense, and it happens frequently - I wont mention names.
If there is new executive members the founders have brought in to the mix to raise revnues, that are not programmers, this could be the real problem here.
Taking a break from coding is fundamental to development - you gotta pull back and think about design objectively, on a regular basis. Less is more.
This is a summary of the information on Cursor-Rate Limits, generated by AI.
Cursor rate limits are based on compute usage, not a fixed number of requests.
Here’s a concise summary of how Cursor’s rate limits work.
TL;DR
Cursor no longer uses a hard monthly request quota (like 500 requests/month).
Instead, it applies rate limits based on your compute usage, which depends on the model, message length, attached files, and conversation context.
There are two types of limits: burst (for short, intense usage) and local (refills every few hours).
When you hit a limit, you can switch models, upgrade your plan, or enable usage-based pricing.
Key Points
1. How rate limits work
Burst rate limit: Lets you use a lot of compute in a short time, but refills slowly.
Local rate limit: Refills fully every few hours (the exact time is not specified).
Limits are based on compute, not request count. Using more powerful models or longer messages consumes more compute.
2. What happens if you hit a limit?
You’ll get a notification and three options:
Switch to a model with higher limits (e.g., Sonnet instead of Opus).
Upgrade to a higher plan (e.g., Ultra).
Enable usage-based pricing to pay for extra requests.
3. Legacy Pro Plan
If you prefer a fixed request quota, you can switch back to the old Pro Plan (500 requests/month) in Dashboard > Settings > Advanced.
4. Transparency issues
Many users complain that Cursor does not disclose the exact compute units, refill times, or per-model limits.
This lack of transparency makes it hard to plan your workflow, especially for long or bursty coding sessions.
5. Practical advice
If you need predictable, uninterrupted usage (e.g., for long coding sessions or AI agent workflows), consider the legacy plan or prepare to use usage-based pricing.
Monitor your usage and be ready to switch models or plans if you hit a limit.
Example
If you’re coding for 6 hours straight and using a powerful model, you might hit a rate limit unexpectedly. You’ll have to wait a few hours, pay extra, or switch to a less powerful model.