Recently, I’ve been running Codex alongside Claude Code and pasting every response into Codex to get a second opinion. It worked great… I experienced FAR fewer bugs, caught bad plans early, and was able to benefit from the strengths of each model.
But obviously, copy-pasting every response is slow and tedious.
So, I looked for ways to automate it. Tools like just-every/code replace Claude Code entirely, which wasn’t what I wanted.
I also experimented with having Claude call the Codex MCP after every response, but ran into a few issues:
- Codex only sees whatever limited context Claude sends it.
- Each call starts a new thread, so Codex has no memory of the repo or previous turns (can’t have a multi-turn discussion).
- Claude becomes blocked until Codex completes the review.
Other third-party MCP solutions seemed to have the same problems or were just LLM wrappers with no agentic capabilities.
Additionally, none of these tools allowed me to choose to apply or ignore the feedback, so it wouldn’t confuse the agent if unnecessary or incorrect.
I wanted a tool that was automatic, persistent, and separate from my main agent. That’s why I built Sage, which runs in a separate terminal and watches your coding agent in real time, automatically cross-checking every response with other models (currently just OpenAI models, Gemini & Grok coming soon).
Unlike MCP tools, Sage is a full-fledged coding agent. It reads your codebase, makes tool calls, searches the web, and remembers the entire conversation. Each review is part of the same thread, so it builds context over time.
Would love your honest feedback. Feel free to join our Discord to leave feedback and get updates on new projects/features https://discord.com/invite/bsFz2ppgMU
Top comments (0)