Im using Ultra plan and chat with claude-4-sonnet
Hey, thanks for your report.
When you select a specific model, the request is always processed by that model without switching. So if you choose Claude 4 Sonnet, the response will come from that exact model.
Important: all Anthropic models are named âClaudeâ, and Claude 4 was trained on data that included information about Claude 3.5 Sonnet. Also, models accessed via API without a system prompt have no knowledge of themselves.
Iâve tested this multiple times, and the results are consistent.
when will you the people learn that LLMs can spit out bullsh*t..
it DOESNâT KNOW who it is, it guesses itâs 3.5 as it was trained on top of previous sonnet versions + public chats with 3.5 (where it has a system prompt that states who it is) + the internet that refers to it as âclaude 3.5 sonnetâ (as its cutoff is BEFORE its release, obviously)
vibe coder I suppose ?
As real IT engineer in AI, simple answer : they have a baseline model 3.5 and improve it with features, it is way easier to do so than going from scratch after billions USD worth of training. So the model self-calling 3.5 is absolutely to be expected.
Most models will call themselves GTP-4. Because they are a copy of GPT-4 with more training thrown at it, maybe minor changes.
Letâs make it easy for you and such. Token embeddings are useful for text and flattening images, any other embedding is exceedingly complicated. How many models have two features : image analysis and prompt analysis ?
100%. Because they all are derived from GPT4 transformer architecture. Stable diffusion is barely an improvement of image only handling, but nothing revolutionary.
Easy answer : ALL MODELS ARE THE SAME SAME.
as a real person with common sense, even more simple answer: The models doesnt base any of the replies of the actual architecture, it base all replies on what it thinks the user wants to hear.
I am also a real person.
Ask the model for itâs model ID and it will give it to you. When you select Sonnet 4 and ask it what model ID it uses it comes back with 3.5 model ID. Ask Claude Code for its model ID and it will show you the Sonnet 4 model ID
There seems to be some apologist that want to bury this fact or explain it away which users can clearly see through.
The model ID does not lie, cursor is misrouting premium requests in many cases which is like ordering a Porsche and getting a VW bug. Iâm not amused.
Interesting. What do you make of this? This was a brand new chat with Claude-4-sonnet thinking.
And with gpt-5-mini
GPT 5 mini was in 2025, so itâs hallucinating that I think.
Some of these models will have different notions of their Model ID, I get that. What I do expect is that if I do go to Anthropic and ask Sonnet 4 what its Model ID is and it gives me a Sonnet 4 Model ID then I should see that same Model ID in Cursor. If I donât then someone is serving up a different Model.
Both Claude Code and Claude Desktop both serve up the Sonnet 4 Model ID. Cursor serves up the 3.5 Model ID.
No one is accusing Cursor, or at least Iâm not, of nefarious behavior but I expect to get the same Model as what Anthropic would provide. I clearly see that the answers I get from Cursor Sonnet 4 are not the same answers I get if I ask Anthropic directly. I have gone so far as to take code and paste it into the Anthropic Desktop or ask it the very same question and Anthropicâs answer is almost always superior.
Seeing is believing. You canât wave this away with a âwell itâs just a model, it doesnât knowâ.
Cursor should be demanding that Anthropic results are demonstrably the same as what users get from Anthropic directly. Cursor is selling a product, they have a fiduciary responsibility to ensure it is authentic and not restricted or altered in any way and can be verified.
If I ask for a python version it gives it to me, it doesnât provide ambiguity or different results on different platforms.
Good post. There should be a clear way to verify that the models we are choosing are indeed the models that are generating our responses. If there isnât a way to prove that, it opens up opportunities for exploitation for sure. I feel like this is going to become an ongoing problem with all 3rd party services- model skimming.
Sure, I mean these data centers on under enormous pressure and it only takes a few bean counting managers beating on some poor IT guy to misconfigure a server to meet demand or to avoid an upgrade to just get by that week. There must be proof from the model without some intervening hack to make it appear the same as that would soon become public and destroy a companyâs future. Letâs all be kind but be real, reputations are hard to build and easy to lose.
I agree completely. Also some of these agents are clearly better than others and I donât want to use them. I need to know which one they are so I can remove it from the list of AI agents.
Are you referring to Auto where you really donât know the model? Because Auto does not allow you to remove any models from it.