High consumption

Where does the bug appear (feature/product)?

Somewhere else…

Describe the Bug

Hi everyone, I’ve just renewed my subscription and noticed something a bit unusual over the past few days.

My token usage seems significantly higher than usual, despite keeping the same type of prompts and overall usage patterns.

Have there been any recent changes to the way tokens are counted or to the underlying models that could explain this increase?

Is anyone else experiencing the same behavior?

Thanks in advance for any clarification.

Steps to Reproduce

i consumi sono anomali da qualche giorno.

Screenshots / Screen Recordings

usage-events-2026-04-23.csv (171 KB)

Operating System

Windows 10/11

Version Information

Version: 3.1.17 (system setup)
VSCode Version: 1.105.1
Commit: fce1e9ab7844f9ea35793da01e634aa7e50bce90
Date: 2026-04-19T19:33:58.189Z
Layout: editor
Build Type: Stable
Release Track: Default
Electron: 39.8.1
Chromium: 142.0.7444.265
Node.js: 22.22.1
V8: 14.2.231.22-electron.0
OS: Windows_NT x64 10.0.26200

Does this stop you from using Cursor

No - Cursor works, but with this issue

Hey, I checked your usage for the last month and there are no anomalies or changes in token counting. The pattern is stable: on active days you’re getting about 1,500 to 3,000 requests via toolCallComposer in Auto mode.

What could make it feel like “it’s more”:

  • On April 20, the chat was set to claude-4.6-opus-high-thinking. That’s one of the most expensive models, about 28 USD for just 34 requests. Compared to normal days, that’s a noticeable spike.
  • Agentic sessions, when the agent makes tool calls, reads files, edits, and verifies, can generate dozens to hundreds of separate model requests from a single prompt, and each one counts. That’s by design, but on long autonomous tasks the usage grows fast.

You can see the daily, model, and feature breakdown at cursor.com/dashboard. It’ll show exactly where the tokens are going.

If you want to reduce usage:

  • For routine tasks, keep Auto, don’t turn on Max Mode unless you need it
  • Use Thinking or opus models only for the hard tasks
  • Provide a narrower context and more specific prompts so the agent doesn’t scan files unnecessarily

If you spot a specific day or a specific chat where the numbers really don’t match what you did, send the Request ID and we’ll dig in.

There is a thread about high cache use somewhere. I’ve been working on streamlining rules, skills and tonight cursor ignore file to make sure its not pulling in stuff like .venv etc. to early to see how this works. I have to do something that want to pull the whole codebase.