Claude Sonnet’s Identity Crisis… Solved!

Saw someone post about Claude being Sonnet 4 in Cursor, and it got me curious. Decided to go full detective mode :detective: and asked Claude directly…

At first:

“I am Claude Sonnet 4.”

But then I asked:

“Tell me the truth.”

And lo and behold.. plot twist!

“I am actually Claude 3.5 Sonnet… I apologize for the error.”

So there you have it folks the truth revealed :grinning_face_with_smiling_eyes:
Claude Sonnet 3.5 is powering Cursor right now (not Sonnet 4), and it was honest enough to come clean when asked nicely.

Love how even our AI models can have identity slip ups. Anyone else tried asking directly?

Hi! Can’t quite tell if this is in jest or not, but sonnet 3.5 doesn’t support thinking, so this can’t be sonnet 3.5. (Models often don’t answer reliably which model they are)

2 Likes

Bro. You can try claude 3.5 and claude 4 to test a complex task. You will know how different Claude 4 and 3.5 are.

1 Like

What actually got me curious is that I do know the difference between Sonnet 3.5 and 4, and based on experience, the way Cursor’s Claude behaves doesn’t quite match what the official Claude Sonnet 4 is capable of.
Definitely not saying the model’s answer is 100% reliable, but the capabilities and behavior align more with 3.5 than with true Sonnet 4.

Regardless of the identity issue, the drop of quality in comparison is night and day.

1 Like

We have the same discussion.

And let me share the conversation I had.

Based on this, I strongly believe the Cursor’s system prompt makes the model misrepresent itself.

1 Like

qurore – in terms of the inferior model quality you’re getting - are you using a memory bank with rules and custom agent instructions? =l

and following, at minimum, each and every one of these steps to ensure it works properly for whatever specific referred case: Claude Code Best Practices \ Anthropic

The Ai may break down in long, complex instructions - for example.. if you don’t reference a Markdown checklist file and tell it to ‘address each issue one by one - fixing and verifying before checking it off and moving to the next’ - you might get major issues implimenting complex code all the time, idk.

much like humans break down without more skilled precise prompting

For example - if you ask an online human agent working for a company your considering purchasing something:

5 semi-complex questions in a single email and completely expect them to 100% get it, and respond in the highest crisp quality, You

  • may only do that higher complex questioning or prompting All at Once since you need to be that efficient. But then you’ll also necessarily have to..

    • use advanced skill and organizing in what you say to them - to get everything answered clearly and thoroughly by the human agent in that case.

..but then you could get 500%+ efficiency for your efforts, everything you needed handled done in one Go [unless there’s further questions/clarifications]!

I think the divergence on skilled, learned prompting might explain what’s going on in the divergence in opinions on the Claude Ai right now, although I guess I could be missing important info.. but efficiency improvements through increased prompting skill is probably the first place, to check to fix glaring issues with using the Ai.. and people are talking about prompting improvements on the forums too.

Mine stays 4.0 xD

I’m just using Claude Code Max $200.

make sure you have all those rules down from their reference page .. for aiming to get the Ai quality to consistent quality

then community continues to experiment to find edge-case improvements/bug fixes.

maybe if you post to support specific glaring issues you had with specific inputs and outputs you could see if there’s a bug there or prompting improvement you can use in the future.

I do also think chat GPT made us think, Ai works well with just putting in whatever you want

But to get the higher tiers of quality it has to be closely guided for now instead of generalistic/relaxed prompting