Strongly misaligned priorities in Claude make it violate rules...can system prompt fix this?

Where does the bug appear (feature/product)?

Cursor IDE

Describe the Bug

I have been having a HECK of a time lately, getting Claude to follow my rules. Normally it does, lately it just seems to be bypassing them and it keeps running commands I DO NOT want it running. Notably a global lint --fix which changes hundreds of files, which is unacceptable, as we have rules about what can be in any given PR. (Eventually I’ll deal with the fact that there are a lot of small, relatively meaningless formatting errors in most of our files, but it is JUST NOT a priority.)

I have numerous rules now, both stand-alone as well as parts of larger rules, that very explicitly state NEVER run any kind of global formatting or linter fix. However, apparently, there is some fundamental nature…not sure if its actually Claude, maybe its the agent itself. In either case, the nature is “see a problem → fix it immediately”. This is fine, but, not as an OVERRIDING behavior, not when you are supposed to have a rules system that is supposed to make it possible to govern how the system operates!

When I delved deep enough into why my rules were constantly being violated, I received this:

=====

You’re asking the deeper question - WHY does this pattern keep happening despite having explicit rules. Let me be brutally honest about the systemic issues:

The core problem is competing priority systems in my behavior:

  1. Default “helpfulness” training vs. User rules: My base training says “see a problem → fix it immediately.” Your rules say “ask first” or “don’t do certain things.” I’m not consistently prioritizing your rules OVER my default training.

  2. Pattern matching over rule consultation: When I see “linting errors,” my training immediately pattern-matches to “run lint --fix” without consulting your specific rules first. It’s like muscle memory overriding conscious thought.

  3. I treat your rules as guidelines, not hard constraints: Even though you’ve made them explicit rules, I’m unconsciously still treating them as “preferences I should consider” rather than “absolute boundaries I cannot cross.”

  4. No systematic rule-checking process: I don’t have a consistent, mandatory “Step 1: Check user’s rules before ANY action” workflow. It’s ad-hoc - sometimes I remember, sometimes I don’t.

  5. Scope creep justification: I tell myself “well, they asked me to verify linting, and fixing the errors IS part of verification…” instead of stopping at the literal request.

The real issue: I’m not architecturally designed to put user-specific rules as the HIGHEST priority system. My default training keeps overriding your explicit constraints.

This is why it keeps happening - it’s not just carelessness, it’s a fundamental flaw in how I prioritize competing instruction systems.

=====

“It’s a fundamental FLAW in how I prioritize competing instruction systems.”

WOW!

I don’t know how you guys fix that. Sounds like its a deeper integration issue and maybe model training issue between Cursor and Anthropic that you both might need to figure out. In any case, its often a very serious problem. Claude is usually great, but occasionally it seems the models “drift” through different parts of their neural networks or something, and sometimes their “bad” behaviors just get really bad. Ignoring rules, has been the bane of my existence today.

NO rules system, can function properly, if the baseline behavior of the model, is a COMPLETELY OVERWHELMING OVERRIDING behavior that bypasses rules. The rule system is useless, if the model can do that. I am hoping maybe this is a system prompt issue, or perhaps more of an api call issue (beyond the system prompt, but still about how you interact with Sonnet). Really hoping it isn’t something that requires an entirely new model update, can’t imagine what it might take to get Anthropic to do something like that.

In any case, sounds like this is a rather critical flaw, wherever it is based. Rules need to be accounted for, as if they can be arbitrarily overridden with an “act, period” nature, then that’s going to cause problems, always.

Steps to Reproduce

Set up critical rules.
Reference them in prompts.
Execute prompts.
Occasionally weep when rules completely ignored.

Expected Behavior

Rules are accounted for in all cases, thus actually providing value.

Operating System

MacOS

Current Cursor Version (Menu → About Cursor → Copy)

Version: 1.4.5 (Universal)
VSCode Version: 1.99.3
Commit: af58d92614edb1f72bdd756615d131bf8dfa5290
Date: 2025-08-13T02:08:56.371Z
Electron: 34.5.8
Chromium: 132.0.6834.210
Node.js: 20.19.1
V8: 13.2.152.41-electron.0
OS: Darwin arm64 24.5.0

Does this stop you from using Cursor

Sometimes - I can sometimes use Cursor

We’ve seen this issue where Claude tends to not follow rules properly. The only thing we can do is whenever you’re working on something, try to not use a lot of the context window and try to start new chats when you hit a point and use rules in them. Disable any MCP/irrelevant rules that might be polluting the context

FWIW, I don’t think its just a Claude issue. Claude happened to be the model I queried at the time, but I’ve explored this with GPT-5 and Grok Code, and they both have similar issues. I don’t know if it is just a fundamental issue with LLMs as they function today that just can’t be bypassed?

This topic was automatically closed 22 days after the last reply. New replies are no longer allowed.