Cursor implemented something a while back where it started acting like how ChatGPT does when it's in its auto mode.
Essentially, choosing when it was going to use what model/reasoning effort on its own regardless of my preferences. Basically moved to dumber models while writing code in between things, producing some really bad results for me.
Anecdotal, but the reason I will never talk about Cursor is because I will never use it again. I have barred the use of Cursor at my company, It just does some random stuff at times, which is more egregious than I see from Codex or Claude.
ps. I know many other people who feel the same way about Cursor and other who love it. I'm just speaking for myself, though.
ps2. I hope they've fixed this behavior, but they lost my trust. And they're likely never winning it back.
Same here. Auto mode is NOT ok. Sadly, smaller models cannot be trusted with access to Bash.
Don’t use the “auto” model and you will be fine.
You just described their “auto” behavior, which I’m guessing uses grok.
Using it with specific models is great, though you can tell that Anthropic is subsidizing Claude Code as you watch your API costs more directly. Some day the subsidy will end. Enjoy it now!
And cursor debugging is 10x better, oh my god.
I have switched to 70% Claude Code, 10% Copilot code reviews (non anthropic model), and 20% Cursor and switch the models a bit (sometimes have them compete — get four to implement the same thing at the same time, then review their choices, maybe choose one, or just get a better idea of what to ask for and try again).