"Hi” Message Used 13K+ Tokens – Why Is Token Usage So High?

Hi everyone,

I’m seeing unexpectedly high token usage in Cursor, and I’m trying to understand why.

For example, at 12:32 PM, I sent a simple message:

“Hi”

But it used 13.3K tokens (and another similar one used 13.4K tokens).

This seems extremely high for such a small message. My usage is getting consumed very quickly because even basic prompts are using thousands of tokens.

A few questions:

  • Why would a one-word message like “Hi” consume 13K+ tokens?

  • Is Cursor automatically sending the full project context or multiple files?

  • Does Auto mode include large amounts of background context by default?

  • How can I reduce token usage for simple interactions?

I’ve attached a screenshot of the usage dashboard showing the 12:32 PM entry.

Would really appreciate clarification on how token calculation works here.

Thanks!

I think this is caused by additional Rules and Context Stuff. But its just a guess from me.

but how can i fix this?
because there is not context and even no mcp server is connected..

I think this are internal rules from the cursor team to make the models work better in generel and/or in conjunction with cursor.

in 2 days i hit 84% of usage limit in pro plan, How is this possible?

Hey, the ~13K tokens for a “Hi” message is actually expected baseline behavior, not a bug.

Every request to the AI includes built-in overhead that can’t be removed:

  • Cursor’s internal system prompt, which tells the AI how to behave
  • Built-in tool definitions like file editing, terminal, search, and more, even if you don’t use MCP
  • Any active rules, even if they’re minimal

For reference, other AI coding assistants have similar baselines. Claude Code CLI uses about 20K tokens just for “hello,” so 13K is actually on the lower end.

A couple of important things about token costs:

  • Most of this is cache read tokens, which are 10x cheaper than regular input tokens
  • In the same chat session, later messages reuse the cached context, so the overhead doesn’t grow linearly

To reduce overall usage:

  1. Keep chats short and focused on a single task
  2. Start new chats often instead of extending long conversations
  3. Remove any unnecessary rules or MCP servers
  4. Keep an eye on which model you’re using since some use more tokens than others

About hitting 84% in 2 days, if you’re using Agent mode a lot, especially with more expensive models, usage can add up fast. You can check the breakdown on your usage dashboard at https://cursor.com/settings to see which models and features are using the most.

There’s a longer discussion here: Saying ‘hello’ uses 122,000 tokens – the cache usage seems inefficient

Let me know if you have more questions.

About hitting 84% in 2 days, if you’re using Agent mode a lot, especially with more expensive models, usage can add up fast. You can check the breakdown on your usage dashboard at https://cursor.com/settings to see which models and features are using the most.

and tell me why
this message was sent by cursor team

Here’s your usage breakdown for this billing period:

Tell me solution how can i fix this

Auto mode is expensive Pricing | Cursor Docs model pricing is here Models | Cursor Docs you can choose cheap model for some routine tasks and smart model for something complicated.

Still sad that before middle of February I used GPT-5.2 for 60$/month at Pro+ and now it’s only enough for 2-3 days.

same, somehow i spent 3k last month which is unbelievable, contacted support, ai response told me i was double charged, then once redirected to a human they said, nah thats real.
Im a single user using on not even a daily basis, no idea how it scaled up to 3k but it did.

Then I talk to friends who use claudecode to see how much they spend and its sub $500 and they run it daily.
But yes usage numbers seem extremely inflated as of the last 2 months.