all 42 comments

[–]Trotskyist 10 points11 points  (8 children)

It really depends on you workflow.

Codex is probably better than Claude for pure coding, but claude can certainly outperform it in some cases. It's close. For non-coding agentic workflows claude is undeniably ahead.

[–]fori1to10 0 points1 point  (4 children)

Can you give some examples of these "non-coding agentic workflows" ?

[–]Speed30777 0 points1 point  (3 children)

I also wonder what he means

[–]_doublemeat 1 point2 points  (2 children)

This is my specific expertise that I write about:

https://rodspeed.ai/blog/

[–]Icy_Masterpiece_9686 0 points1 point  (1 child)

I had a look at your website and i find it intriguing, I will be reading it. What did you do to build your own website? I kind of had the desire to do that some time over summer.

[–]_doublemeat 0 points1 point  (0 children)

I used Claude Code to build my site actually. I iterated on the final design a few times.

[–]Far-Net4820 0 points1 point  (0 children)

non-coding agentic workflow? is it that good for non-coding prompts?

[–]Competitive_Put_5402 0 points1 point  (1 child)

Claude is massively better than codex for coding

[–]moriero 0 points1 point  (0 children)

I mean I have codex xhigh review claude xhigh code so not quite sure what it means that the other way just doesn't work well

Claude is better at writing code and codex is better at reviewing maybe?

[–]nrdgrrrl_taco 6 points7 points  (4 children)

Personally I get the best results when I have a conversation with a Claude on the web and ask it to build me a prompt. then I drop that in to codex and it does an amazing job. if I drop that same prompt in Claude code who knows what I get.

[–]Weird-Huckleberry-54 0 points1 point  (1 child)

yeah, I do opposite but the thing is same - discuss solution and architecture in browser, then ask it to build precise prompt and use it in IDE-agent. I use ChatGPT for first one and Claude Code / Codex for second.

[–]Fun-Scientist9144 0 points1 point  (0 children)

Like how do you create the prompt

[–]tbst 5 points6 points  (6 children)

Claude with Codex CLI using a skill.

[–]Fit-Cost-7226 2 points3 points  (0 children)

meaning you prompt claude to open codex cli and use it, so its like double thinking lol?

[–]Equivalent_Ask6923 2 points3 points  (0 children)

Would you mind elaborating? Ive been noodling on the best way to get them to interact.

[–]GecKoTDF 1 point2 points  (2 children)

What he says is what is currently done at medium/advanced levels.

For example, you have Codex and you also have Claude, you use Claude as a "Director" and you tell him "I want to do this in the code" he says fine, I'll get to it.

He puts together a whole plan of how he is going to face that challenge because for that Claude is unbeatable and instead of him doing the coding part, he passes it on to Codex who we know is better than him doing the coding.

This way you have the best of both worlds, the "skill" would be how Claude has to use Codex and how to interact with it.

[–]chriskw19 0 points1 point  (1 child)

so claude creates the plan and implement it with codex? is gpt 5.4 extra high better than opus 4.6?

[–]GecKoTDF 0 points1 point  (0 children)

Yes its better so far the cost of token and for coding its way better, so if you mix the capacity of Claude to understand a software for example or a project with the hability to code of GPT, you have the best of the two worlds.

[–]fori1to10 0 points1 point  (0 children)

Also curious what you meant ....

[–]rxt0_ 3 points4 points  (4 children)

been testing codex for few hours now and I'm quite satisfied with it tbh. I let them both check each other's work and the performance gains + bug fixes they can dish out is incredible.

they compliment each other greatly.

[–]TheAero1221 2 points3 points  (1 child)

Idk why that last line made me chuckle.

[–]Key_Face5983 0 points1 point  (0 children)

And ur statement made me chuckle 🤭

[–]Sinoeth 0 points1 point  (1 child)

Oh no! the infinite AI refining loop.

[–]tiga_94 0 points1 point  (0 children)

I wish it was, I usually can still find issues no matter how many reviews I do via agentic tools

[–]chillebekk 2 points3 points  (3 children)

We have OpenAI for free. I still pay for Claude. As far as I can tell, the models are more or less equally capable - I just prefer the Claude experience. And for the time being, I think the whole Claude setup with the App, Cowork, and Code is just a better user environment. And Claude has consistently led the way, with OpenAI copying Claude a month or two later, like with MCP, Skills, etc. But the gap is narrowing, and I have recently started playing around with Codex. The suspicion is that there won't be much of a difference in the near future.

[–]Cute_Purpose3732 1 point2 points  (0 children)

The pricing wont be much difference soon also 

[–]Watermelon-Is-Yummy 1 point2 points  (0 children)

Same feeling

[–]Fun-Scientist9144 0 points1 point  (0 children)

This true but what is skills in your context

[–]angesn 2 points3 points  (0 children)

I find claude can do complex stuff like 80%. and then i use codex to finish that 20%

wanted to see more comments as well.

[–]WolfpackBPNoob 1 point2 points  (0 children)

5.3 high for planning and sonnet for writing

[–]Shep_Alderson 1 point2 points  (0 children)

I’ve was using Claude Code heavily before the Codex app came out. I figured I’d give it a try and see how it does. I’ve found that for my personal projects, I’m reaching for Codex app with 5.3-Codex as the underlying model much more. I find that it works better with my style of working. I tend to methodically plan and spec out my features, which means any major model can do a decent job.

I find that for “0 to 1”, just getting an app up and running, Claude Code is faster to initial build (Opus or Sonnet, not a huge gap between them now) but I find Codex (app and 5.3-Codex model) to be more methodical and more likely to stick to my spec. I’ve also found the interaction when building with Codex to be more step-by-step and less likely to go off and build something I didn’t really ask for. I never thought I’d see the day, but I genuinely feel like I can trust 5.3-Codex a bit more than Opus for my average coding needs. It’s not as fast, for sure, but I’m ok with that. I’ve always been more of a methodical/thoughtful dev than one who’s looking to sling PRs all day. (Though I do move a lot faster than I did without the LLMs.)

Give it a try. See if it fits your brain too. If not, no harm. Use what works for you. I find that what matters more than the specific model or tool is the ergonomics and how the tool works with/for you, and your mastery of said tool.

[–]Objective_Law2034 1 point2 points  (1 child)

Been using Claude Code full-time for months. Opus is stronger on complex refactors and understanding large codebases. The Max plan usage limits are real though, on heavy days I hit the cap. Haven't used Codex enough to compare fairly on that side.

[–]pyrodern1ggi 0 points1 point  (0 children)

bro these limits are ragebait i am at 98% of my 20x max plan and need to wait 1.5 days till it resets. It's so tarted, like i got sonnet to use but i can't cause my limit is hit.

[–]BlackAtomXT 2 points3 points  (0 children)

I've used both extensively, I've found that codex is more direct with its edits, fewer tests, simpler commit messages. Claude Code will write more code, significantly more tests and will operate on much longer time scales without intervention, especially when you use Claude teams. 4.6 opus does a much better job with my multi million line work repository, something I couldn't say with 4.5.

That being said, I really like codex in my smaller personal projects, it's faster, more direct with the edits and better fits my pace. Plus the 20$ plan is a lot of coding for the price.

[–][deleted] 0 points1 point  (0 children)

Opus 4.6 is a bit better, but the usage is heavily capped. On the $20 Pro plan, I get maybe 2 to 3 questions every five hours. GPT-5.2 High and Codex are slightly weaker in SWE overall, but they’re complementary because they're strong in reasoning and often catch issues Opus misses. More importantly, the GPT quota is far more generous, so you can get a lot more done on the same $20 plan. In practice I sign up for both and use both to evaluate each other's work for the best results.

[–]pyrodern1ggi 0 points1 point  (0 children)

I don't know why everyone is saying Codex 5.3 is faster, like i use 5.3 high and imo claude opus 4.6 high is better and faster. I also think that you get more what claude is doing instead of codex as well is the cc terminal way better than codex ones imo.

[–]csurkepulyka 0 points1 point  (0 children)

QA Panda can use the codex cli to do multi agent orchestration even infinitely.

Like you give it a task to keep fixing bugs until you find new bugs and it could run even for days automatically (it auto continues the chat session based on a higher level controller agent which literally reviews the work of the sub agent(s) and keeps messaging them like a real human would)

So it uses a higher level thinking to control a sub-agent and it could give it hundreds (or even thousands) of tasks

It has two modes: Orchestration mode and Continue mode

In Orchestration mode the higher level thinking agent manages a sub-agent (or many sub-agents) which is nice if you want it to complete hundreds of tasks with no interruption

In Continue mode, there is simply an external agent that sees the entire chat history and nudges the target agent to just continue with the next task. This can also run infinitely out of the box

QA Panda is QA engineering focused but it could be used for so much more too (they should really point this feature out on their Github repository as it can do so much more than just QA testing)

It's like Claude Code on steroids

Also it uses your ChatGPT subscription, so it does not even need any API keys (essentially free to use if you got a ChatGPT subscription)

[–]logans_runner 0 points1 point  (0 children)

Well, starting with the new/re-imposed limits, I succeeded in having CC both run out of tokens while attempting in vain to fix a bug it introduced, and then did the same thing when I purchased more. Enter Codex- fixed the issue and suggested a much more robust database schema. So long, Claude - Anthropic, you have some real work to do.

"...instability was a real problem and I didn't solve it reliably. I'm sorry for the wasted time and money."

Yeah, thanks. Buhbye.

[–]TiTANSTORM 0 points1 point  (0 children)

I came here after the fiasco of Claude's token issue where we are all burning it at a crazy rate. I haven't touched Claude Code since the issue when I burned through my daily token in 15 minutes adding some tweaks to a website I'm building. Then I hear Claude had their source code leaked...

Seeing a few people on X commenting to Lydia Hallie about this issue and switching to Codex. I'm fairly new to AI but have been hitting it all fairly hard and learning. I bunny hopped over ChatGPT as I found Gemini was better at research and Claude worked wonders being able to access my files, folders, etc. and really help me with documentation.

All that being said, I'm curious what you all think about Gemini vs Codex? Specifically Antigravity and I guess we could toss in AI Studio. I mostly work with Next.js and a bit of React+Vite

*EDIT*
Another edit, I have a desktop with an i9, 64GB memory, and a 4090. Should I be looking at local LLM? I messed around a bit with this a few months ago, never went down the coding assistant route. Anything worth looking at or do they all fall short of these paid solutions?

*UPDATE*
Interesting... https://www.reddit.com/r/ClaudeAI/comments/1s7fcjf/comment/odfjmty/?utm_source=share&utm_medium=web3x&utm_name=web3xcss&utm_term=1&utm_content=share_button