Essentially, choosing when it was going to use what model/reasoning effort on its own regardless of my preferences. Basically moved to dumber models while writing code in between things, producing some really bad results for me.
Anecdotal, but the reason I will never talk about Cursor is because I will never use it again. I have barred the use of Cursor at my company, It just does some random stuff at times, which is more egregious than I see from Codex or Claude.
ps. I know many other people who feel the same way about Cursor and other who love it. I'm just speaking for myself, though.
ps2. I hope they've fixed this behavior, but they lost my trust. And they're likely never winning it back.
You just described their “auto” behavior, which I’m guessing uses grok.
Using it with specific models is great, though you can tell that Anthropic is subsidizing Claude Code as you watch your API costs more directly. Some day the subsidy will end. Enjoy it now!
And cursor debugging is 10x better, oh my god.
I have switched to 70% Claude Code, 10% Copilot code reviews (non anthropic model), and 20% Cursor and switch the models a bit (sometimes have them compete — get four to implement the same thing at the same time, then review their choices, maybe choose one, or just get a better idea of what to ask for and try again).
Why would you do that to yourself? Reviewing 4 different solutions instead of 1 is 4 times the amount of work.
I ended up spending time just clicking "Accept file" 20x now and then, accepting changes from past 5 chats...
PR reviews and tying review to git make more sense at this point for me than the diff tracking Cursor has on the side.
Cancelling my cursor before next card charge solely due to the review stuff.
I’m on claude code $100 plan and never worry about any of that stuff and I think I am using it much more than they use cursor.
Also, I prefer CC since I am terminal native.