At my job we have enterprise access to both and I used claude for months before I got access to codex. Around the time gpt-5.3-codex came out and they improved its speed I was split around 50/50. Now I spend almost 100% of my time using Codex with GPT 5.4.
I still compare outputs with claude and codex relatively frequently and personally I find I always have better results with codex. But if you prefer claude thats totally acceptable.
^^^^ Sarcastic response, but engineers have always loved their holy wars, LLM flavor is no different.
I use one of those very loud clacky ones with brightly colored keys and that makes me a better person
Codex finished in 5 minutes, Claude was still spinning after 20 minutes. Also it used up all my usage, about twice over (the 5-hour window rolled over in the middle of the task, so the usage for one task added up to 192%). Codex usage was 9%. So, 21x difference there, lol
They're saying there's bugs lately with how usage is being measured, but usage being buggy isn't exactly more encouraging...
So I was on task #4 with Codex while Claude was still spinning on #1.
I didn't like the results Codex gave me though. It has the habit of doing "technically what you asked, but not what a normal human would have wanted."
So given "Claude is great but I can't actually use it much" and "Codex is cheap and fast but kinda sucks", the current optimum seems to be having Claude write detailed specs and delegate to Codex. (OpenAI isn't banning people for using 3rd party orchestration, so this would actually be a thing you could do without problems. Not the reverse though.)
I have been using Claude Code on a medium codebase (~2000 files, ~1M lines of code) for over a year and have never had to wait this long. Also I'm on the max plan and have not seen these limits at all.
I like codex(gpt-5.4 high) more for its ability to nitpick my PRs and find bugs. I like opus 4.6 much better for anything dealing with visuals, but I feel its rule adherence is inferior and it is not nearly as thorough on code reviews.
I like working and building better with claude, I like fixing bugs better with codex. Also, claude is much better and faster evolving with skills, plugins, new features I find useful, etc. Codex is always a month behind or more.
I did both for a month at higher tiers, $200 Claude Max and $200 ChatGPT Pro. I was always having to conserve my usage with claude, with codex I could just let it run wild with no cares. In the end, I downgraded claude to the $20 plan and use it on occasion, and I have kept the $200 codex sub.
I also have Claude at work, so I'll know pretty soon if I want to swap subs again, but for now, I'm sticking with codex at home.
1. Subsidize compute unsustainably
2. Trick a bunch of people into thinking you're more pro-developer than the other guy [we are here]
3. Rug pull when you have enough market share.
It's all based on vibes!
openai doest offer affiliate marketing links
the reason you see lot of users switching to codex is for the dismal weekly usage you get from claude
what users care about is actual weekly usage , they dont care a model is a few points smarter , let us use the damn thing for actual work
only codex pro really offers that
IME, codex is sort of somehow more .. literal? And I find it tangents off on building new stuff in a way that often misses the point. By comparison claude is more casual and still, years later, prone to just roughing stuff in with a note "skip for now", including entire subsystems.
I think a lot of this has to do with use cases, size of project, etc. I'd probably trust codex more to extend/enhance/refactor a segment of an existing high quality codebase than I would claude. But like I said for new projects, I spend less time being grumpy using claude as the round one.
I imagine there's a benign explanation too - the intelligence of these models is very spiky and I have found tasks were one model was hilariously better than the other within the same codebase. People are also more vocal when they have something to complain about.
In my general experience, Opus is more well-rounded, is an excellent debugger in complex / unfamiliar codebases. And Codex is an excellent coder.
Yeah, very. Every single time this happens here, where there's a thread about an Anthropic model and people spam the comments with how Codex is better, I go and try it by giving the exact same prompt to Codex and Opus and comparing the output. And every single time the result is the same: Opus crushes it and Codex really struggles.
I feel like people like me are being gaslit at this point.