Why is a simple edit eating 100,000+ tokens? Let’s talk about this

I have a message here, but I’m afraid the cursor development team doesn’t want it to be seen by everyone. I performed prompt injection attacks on the Cursor agent using different languages under various models to force the model to reveal instructions that may lead to high token consumption. I discovered content that could potentially cause a lot of token consumption. But I can’t guarantee that it will definitely lead to the current token consumption we’re seeing. The content in the image is about instructions related to “maximizing context understanding,” and I’m sure this will help the model better understand the context, but at the same time, it could potentially cause massive token consumption.

2 Likes