(Similar guidance goes for writing tools & whatnot - give the LLM exactly and only what it needs back from a tool, don’t try to make it act like a deterministic program. Whether or not they’re capital-I intelligent, they’re pretty fucking stupid.)
I also use Github Copilot which is just $10/mo. I have to use the official copilot though, if I try to 'hack it' to work in Claude Code it burns thru all the credits too fast.
I am having a LOT of great luck using Minimax M2 in Claude Code, its very cheap, and it works so good.. its close to Sonnet in Claude Code. I use this tool called cc-switch to swap out different models for Claude Code.
(I just learned ChatGPT 5.2 Pro is $168/1mtok. Insanity.)
If Claude makes a yawn or similar, I know it’s parsed the files. It’s not been doing so the last week or so, except for once out of five times last night.
“You’re absolutely right! I see here you don’t want me to break every coding convention you have specified for me!”
I've used it pretty extensively over the year and never had issues with this.
If you hit autocompact during a chat, it's already too long. You should've exported the relevant bits to a markdown file and reset context already.
I think you may be observing context rot? How many back and forths are you into when you notice this?
Real semi-productive workflow is really a "write plans in markdowns -> new chat -> implement few things -> update plans -> new chat, etc".
I'm sure there are workarounds such as resetting the context, but the point is that god UX would mean such tricks are not needed.