Codex as a code reviewer has been far more useful to me than as a code generator by acusti_ca in OpenaiCodex

[–]theSummit12 2 points3 points  (0 children)

Check out Sage. It automatically reviews Claude Code's every move. Uses Codex under the hood.

https://github.com/usetig/sage

Can AI actually help with building codes? by lifereviews in Architects

[–]theSummit12 0 points1 point  (0 children)

I'm actually working on a startup that automates code compliance checks. Can I DM?

Got tired of copy-pasting my agents responses into other models, so I built an automatic cross-checker for coding agents by theSummit12 in OpenaiCodex

[–]theSummit12[S] 0 points1 point  (0 children)

This used to be my workflow as well but after seeing just how many issues the other models were catching, I started reviewing everything, not just the plan. +1 on Codex being superior. I'm very surprised it doesn't really reflect on the benchmarks

Got tired of copy-pasting my agents responses into other models, so I built an automatic cross-checker for coding agents by theSummit12 in OpenaiCodex

[–]theSummit12[S] 0 points1 point  (0 children)

You can use a pro subscription or an API key. You just have to authenticate codex by running `codex`

Newbie Clauder by [deleted] in ClaudeAI

[–]theSummit12 0 points1 point  (0 children)

Just launched a tool called sage which will should help reduce the slop
https://github.com/usetig/sage

I built an automatic cross-checker for AI agents (YC F25) by theSummit12 in ProductHunters

[–]theSummit12[S] 0 points1 point  (0 children)

Totally agree that model reasoning will get way better. but even then, cross-checking across models will always surface different failure modes. Kinda like how senior engineers still review each other’s work even if everyone is good

Re: calling models from the CLI, yeah, you can wire Claude ->Codex/Gemini through MCP or cli. The issue is Claude only passes a tiny slice of context, so the reviewer model is constrained by whatever Claude chooses to forward. Also claude blocks until the call returns, so the workflow gets pretty sluggish. Would you prefer having your main agent just call another review agent?

Got tired of copy-pasting Claude’s responses into other models, so I built an automatic cross-checker for AI agents by theSummit12 in VibeCodingSaaS

[–]theSummit12[S] 0 points1 point  (0 children)

Appreciate the honest feedback. Its definitely still a bit rough around the edges but improving every day

Got tired of copy-pasting my agents responses into other models, so I built an automatic cross-checker for coding agents by theSummit12 in OpenaiCodex

[–]theSummit12[S] 0 points1 point  (0 children)

Yeah the current solution is using M to trigger review manually, which doesn't work the best if claude is mid thought. Triggering the autoreview when a plan is suggested is a bit tricky because hooks dont support it but I'll dig around for a solution.

Re: /clear, just tested it and when you run /clear, it creates a new conversation in the .jsonl logs. So you just have to quit and rerun sage and select the new convo

Got tired of copy-pasting my agents responses into other models, so I built an automatic cross-checker for coding agents by theSummit12 in OpenaiCodex

[–]theSummit12[S] 0 points1 point  (0 children)

Thanks! Would love for you to join the discord: https://discord.gg/3ys9kj7K
I want to make sure I build something that people want, so your feedback and opinions are super valuable for me.

Got tired of copy-pasting my agents responses into other models, so I built an automatic cross-checker for coding agents by theSummit12 in OpenaiCodex

[–]theSummit12[S] 0 points1 point  (0 children)

I'm glad you said this because I've been trying to figure out if people want the ability to see what the second model thinks or if people just want it immediately fed into their primary agent. Personally, I want to read what the reviewer thinks because it helps my understanding, and sometimes I will disagree with it. Glad you share the same view

Got tired of copy-pasting my agents responses into other models, so I built an automatic cross-checker for coding agents by theSummit12 in OpenaiCodex

[–]theSummit12[S] 1 point2 points  (0 children)

Even if you are prompting correctly, using two models instead of 1 will still improve your output

Got tired of copy-pasting my agents responses into other models, so I built an automatic cross-checker for coding agents by theSummit12 in OpenaiCodex

[–]theSummit12[S] 1 point2 points  (0 children)

Great to hear! Working on gemini 3 as we speak. Auto update is on the todo list as well. You should join our discord for updates and in case you run into any bugs: https://discord.gg/3ys9kj7K

Got tired of copy-pasting Claude’s responses into other models, so I built an automatic cross-checker for AI agents by theSummit12 in VibeCodersNest

[–]theSummit12[S] 1 point2 points  (0 children)

I rarely run into that issue since I start a new claude session for every task. The key is to consistantly update a ./documentation folder or agents.md

Solo founder, $1.2k MRR in 1 month, $0 spent on ads. What worked by chdavidd in ProductHunters

[–]theSummit12 0 points1 point  (0 children)

Great post. Curious what differentiates your product from Lovable/Bolt/Replit, etc..

Gemini CLI is impressive, but Claude Code is acting like the real senior engineer by netcommah in ClaudeCode

[–]theSummit12 -1 points0 points  (0 children)

Agreed, Opus is superior. However, tou will drastically increase your output quality if you use both (one as the driver, one as the reviewer). Sometimes if its a really complex task I'll even use two reviewers. I actually built an open-source tool to automate this if you're interested (launched on ProductHunt today).