
TL;DR
Opus 4.7 vs GPT-5.5, the new Codex CLI vs the Claude skills ecosystem. An opinionated April 2026 verdict on which terminal agent to reach for, by job.
In April 2026, the terminal-agent question is no longer "which CLI is more capable." Both Claude Code and Codex are competent enough to ship real production work in real repos. The question now is which one fits which job - because the two products have visibly diverged.
Claude Code optimizes for extensibility on top of a planning model. Opus 4.7 is the thinking head; skills, sub-agents, hooks, MCP servers, and plugins are the body. The bet is that you will want to bend the agent to your repo and your team.
Codex optimizes for a tightly integrated agent loop with strong defaults. GPT-5.5, the rebuilt Codex CLI, the new app-server, the in-app browser, and the automatic reviewer are designed to behave well out of the box without much customization.
Both bets are reasonable. They lead to different daily ergonomics.
A quick state-of-the-world before the verdict, because anything older than April is already stale.
Anthropic released Claude Opus 4.7 on April 16. Roughly 13% better than Opus 4.6 on a 93-task internal coding benchmark, with stronger vision and noticeably more taste on UI and document tasks. Pricing held at $15 / $75 per million tokens. Sonnet 4.6 still scores 79.6% on SWE-bench Verified at $3 / $15. Haiku 4.5 sits at $1 / $5 with roughly Sonnet 4-tier coding.
OpenAI released GPT-5.5 on April 24. Inside Codex, OpenAI explicitly says it produces better results with fewer tokens than GPT-5.4. The Codex changelog over the last month also added Unix socket transport for the app-server, sticky environments, remote plugin install, automatic reviewer agents that gate risky approvals, in-app browser hand-off for local dev servers, and codex exec --json reasoning-token output.
Google shipped Gemini 3 Pro and Antigravity on April 22. Relevant context, but it does not change the head-to-head between the two terminal agents.
This is closer than the marketing suggests. On hard, multi-file refactors in real repos, both Opus 4.7 and GPT-5.5 produce working diffs most of the time. The differences:
Net: if you measure SWE-bench-style numbers, they look similar. If you measure your own happiness on a Tuesday, the personalities diverge.
# Same task, two agents
claude -p "add a /healthz endpoint with 200 OK and a tiny test"
codex exec "add a /healthz endpoint with 200 OK and a tiny test"
For tasks at that altitude, Codex usually finishes first.
Get the weekly deep dive
Tutorials on Claude Code, AI agents, and dev tools - delivered free every week.
This is where Claude Code is currently in a different league.
The skills ecosystem became real this month. The community-curated claudemarketplaces.com directory crossed 150 skills in March and the open-source claude-code-plugins-plus-skills marketplace lists 423 plugins, 2,849 skills, and 177 agents. A skill is a Markdown file:
~/.claude/skills/deploy-vercel/SKILL.md
A plugin bundles skills, MCP servers, slash commands, and sub-agents into one installable unit. Hooks let you run shell commands at lifecycle events. Sub-agents let you fan work out cleanly. None of this requires SDK code.
Codex's plugin model exists - the recent changelog added remote plugin install and marketplace upgrades - but it is younger, smaller, and less culturally embedded. If you want a community library to copy from on day one, Claude Code wins.
If your team already has an AGENTS.md or DESIGN.md and a folder of skills, that investment compounds in Claude Code. Move to Codex and most of it does not transfer.
Codex catches up here, and arguably surpasses Claude Code.
The new automatic reviewer agent in Codex CLI gates risky approvals through a separate agent before they execute. Permission profiles round-trip across TUI sessions, user turns, MCP sandbox state, and shell escalation. The in-app browser lets Codex click through a real local app to verify a fix. codex exec --json reports reasoning-token usage so you can budget cost programmatically.
Claude Code's hook system is more flexible (you can run any shell command on PreToolUse, PostToolUse, Stop), but Codex's defaults out of the box are tighter. If you want a junior teammate to run an agent and not break prod, Codex is the safer first install.
Real numbers for a typical 4-hour coding session, all approximate:
For pricing tiers, see our Q2 2026 AI coding tools pricing breakdown.
Pick Claude Code when:
AGENTS.md / CLAUDE.md / DESIGN.md and want the agent to actually read themPick Codex when:
Use both when:
Here is the configuration most heavy users I trust are running this week.
~/.claude/settings.json:
{
"model": "claude-opus-4-7",
"subagent_model": "claude-haiku-4-5"
}
~/.codex/config.toml:
model = "gpt-5.5"
auto_review = true
Then alias them so your fingers pick the right tool:
alias plan="claude" # ambiguous, big-picture
alias do="codex" # tight, well-scoped
It sounds silly. It works.
Both products are converging on "agent that reads your repo, plans, edits, runs, verifies." They will keep getting closer on raw ability. The differentiation is going to be:
If I had to bet, the team that wins is the team whose users build things on top of it without permission. That favors Claude Code in the long run. But Codex's April 2026 release is the closest the gap has been, and on a strict cost-per-task basis it is currently the better default for "small, scoped" coding work.
For a deeper field comparison including Cursor and OpenCode, see our four-way matchup.
Technical content at the intersection of AI and development. Building with AI agents, Claude Code, and modern dev tools - then showing you exactly how it works.
Anthropic's agentic coding CLI. Runs in your terminal, edits files autonomously, spawns sub-agents, and maintains memory...
View ToolOpenAI's cloud coding agent. Runs in a sandboxed container, reads your repo, executes tasks, and submits PRs. Uses GPT-5...
View ToolOpenAI's open-source terminal coding agent built in Rust. Runs locally, reads your repo, edits files, and executes comma...
View ToolOpenAI's flagship. GPT-4o for general use, o3 for reasoning, Codex for coding. 300M+ weekly users. Tasks, agents, web br...
View ToolLimit which tools a subagent can access.
Claude CodeConfigure Claude Code for maximum productivity -- CLAUDE.md, sub-agents, MCP servers, and autonomous workflows.
AI AgentsDeep comparison of the top AI agent frameworks - architecture, code examples, strengths, weaknesses, and when to use each one.
AI Agents
Composio: Connect AI Agents to 1,000+ Apps via CLI (Gmail, Google Docs/Sheets, Hacker News Workflows) Check out Composio here: http://dashboard.composio.dev/?utm_source=Youtube&utm_channel=0426&utm_...

Anthropic has released Channels for Claude Code, enabling external events (CI alerts, production errors, PR comments, Discord/Telegram messages, webhooks, cron jobs, logs, and monitoring signals) to b...

Claude Code “Loop” Scheduling: Recurring AI Tasks in Your Session The script explains Claude Code’s new “Loop” feature (an evolution of the Ralph Wiggins technique) for running recurring prompts that...

From Claude Opus 4.7 and GPT-5.5 to Andrej-karpathy-skills and EvoMap - the AI dev tools actually shipping the last 30 d...

A new study from nrehiew quantifies a problem every Claude Code, Cursor, and Codex user has felt: models making huge dif...
Hacker News keeps arguing about Claude Code, Codex, skills, MCP, and orchestration. Under the noise, the same four truth...

New tutorials, open-source projects, and deep dives on coding agents - delivered weekly.