Murph's Take

OpenAI's Codex Is Now a Plugin That Runs Inside Claude Code. Read That Again.

The most telling thing about the Codex plugin for Claude Code isn't the feature — it's what it reveals about where the gravitational center of agentic coding has already landed.

Jason MurphyMarch 30, 20267 min read

The most interesting response to the Codex plugin for Claude Code came from a random account on X:

"if codex is so good like y'all are claiming, why does it need to run inside claude code?"

This was meant as a burn. It's actually the correct analytical question.

OpenAI shipped a plugin that installs into Claude Code's plugin marketplace. You run /plugin marketplace add openai/codex-plugin-cc and Codex becomes a service inside your Claude Code session. You can invoke it for a standard review, a steerable adversarial challenge, or a rescue operation when your code is stuck. The feature is useful. But the feature is not the story.

The story is the architecture.

When you build a plugin that runs inside another system's environment, you're acknowledging something about that environment. You're saying: this is where the developer already lives. This is where the context lives. This is where the file system access lives, where the conversation thread lives, where the tool execution layer lives. We want Codex to be accessible there.

That's not a neutral technical decision. It's a gravitational concession.


What "running inside" actually means

Claude Code is an agentic coding environment. It maintains persistent access to your file system across a session, can read and modify your codebase autonomously, executes terminal commands, and reasons about your entire project state continuously. It's not stateless. It's not advisory. It's embedded in your working environment and operating on it.

Codex, in this plugin configuration, is different. It arrives as a scoped service: invoked with a slash command, performing a discrete task, returning a result. The architectural relationship is clear — Claude Code provides the environment, Codex provides a capability within it.

This is a meaningful asymmetry. Environment beats service in the long run because environment is where context accumulates, where state lives, where the work actually happens.

BLACKBOX AI read this correctly in February, before the official announcement. Their Claudex Mode set Claude Code and Codex on the same task from opposite directions — Claude implementing, Codex verifying — with both operating in a shared workflow. They weren't waiting for official permission to compose these systems. They saw the pattern and built toward it.

The pattern is: multi-model workflows where different AI systems take different roles within a single environment. This is not a future trend. It's the current architecture.


Why the environment position matters

There's a version of this story where "Claude Code beat Codex" and that's the whole take. I don't think that's the right frame.

The more useful frame is: what does it mean that the center of agentic coding workflows is an environment rather than a model?

When Karpathy said he hadn't written a line of code since December, he wasn't describing a model. He was describing a working relationship with a system that operates inside his development environment. The thing he's not doing anymore — writing code — is a task-level operation. What replaced it is environmental: a persistent agent with context, judgment, and the ability to act.

The Codex plugin story is a signal that this pattern is stabilizing. OpenAI's flagship coding system is now a service that runs inside Anthropic's flagship agentic environment. That's a remarkable sentence to type in March 2026. A year ago the frame was "which AI writes better code?" Today the frame is "which environment do you build inside, and which systems do you pull in as services?"

Those are fundamentally different questions. The first is a benchmark question. The second is an architecture question.


What changes in practice

If you're using Claude Code already, the practical implication is that your workflow just got a credible adversarial review option. /codex:adversarial-review is interesting specifically because it's a different model with different training. Using two systems that disagree with each other is a reasonable way to find bugs that slip through a single reviewer's blind spots. That's not new insight — code review pairs have always been valuable for exactly this reason — but having it as a one-command operation inside your existing environment is the operationalization of that insight.

The deeper implication is architectural discipline. If your coding environment is load-bearing infrastructure — and at this point, for anyone working at pace, it is — then the decision of what plugins you install and when you invoke which systems is a design decision. Not a preference. A design decision. What does Codex review well that Claude Code might normalize? What does the adversarial mode surface that a standard review misses? When is /codex:rescue the right call versus asking Claude to retry with a different approach?

These are questions worth spending time on. The tools exist. The architecture question is how you compose them.


The question I keep returning to: as these environments become more capable and more composable, what's the last category of work that genuinely requires a human in the loop — not because the AI can't do it, but because the decision needs to be owned by someone with skin in the game?

I don't have a clean answer. But I notice the question is getting harder to answer, not easier.

Want to see how your business stacks up?

Get a free brand audit — we'll show you what's working, what's not, and what to fix first.

Free Brand Audit →

Frequently Asked

What is the Codex plugin for Claude Code?

OpenAI released a plugin that runs Codex review capabilities inside Claude Code's plugin marketplace. You install it with /plugin marketplace add openai/codex-plugin-cc and can then invoke /codex:review for a standard code review, /codex:adversarial-review for a challenge-mode review, or /codex:rescue to let Codex attempt to fix stuck or broken code. The key architectural fact: Codex is operating as a service inside Claude Code's environment, not as a standalone agent.

Why does it matter that Codex runs inside Claude Code rather than the other way around?

Because environment determines architecture. Claude Code is the host — it owns file system access, context, the conversation thread, and the tool execution layer. Codex arrives as a plugin: stateless, scoped, invoked on demand. That's not a neutral technical arrangement. It reflects a genuine difference in where each system's capabilities are load-bearing. Claude Code can reason about your codebase continuously and autonomously. Codex, in this configuration, performs discrete tasks within that environment.

Does this mean Claude Code 'won' the agentic coding race?

It means the architecture question is effectively settled for now. 'Who won' is a product framing. The architectural framing is more useful: which system is providing the environment, and which is providing a service inside it? Right now, Claude Code is the environment. That's a significant position to occupy — and the fact that OpenAI built a plugin to enter that environment rather than the reverse says something concrete about where the center of gravity sits.

What is BLACKBOX AI's Claudex Mode and why is it relevant?

BLACKBOX AI launched Claudex Mode in February 2026 — Claude Code implements a task, Codex verifies and tests it, both operating in the same workflow. They saw the multi-model collaboration pattern before the major labs made it official. It's an example of product builders reading architectural momentum correctly before it's publicly confirmed.

Jason Murphy

Written by

Murph

Jason Matthew Murphy. Twenty years building digital systems for businesses. Former CardinalCommerce (acquired by Visa). Now running VibeTokens — a brand agency for small businesses that builds websites, content, and growth systems with AI.

Your brand is your first impression.

Find out if it's costing you customers.

Free brand audit. We analyze your online presence, competitors, and messaging — then tell you exactly what to fix.

Get Your Free Brand Audit →