oh-my-claudecode: Building Multi-Agent Teams That Actually Ship Code
Hook
What if you never had to learn Claude Code’s API? What if you could spawn three Codex workers in separate tmux panes, have them independently review different modules, then synthesize their findings—all from a single natural language command?
Context
AI coding assistants have evolved from autocomplete tools to conversational pair programmers, but orchestrating multiple agents remains painfully manual. If you want Claude to handle authentication while Codex reviews security and Gemini redesigns the UI, you’re stuck context-switching between three different chat windows, copy-pasting context, and manually merging their outputs. The cognitive overhead defeats the purpose of automation.
oh-my-claudecode (OMC) treats multi-agent collaboration as a first-class primitive. Instead of learning Claude Code’s experimental teams API or writing orchestration glue, you describe what you want built and let the framework handle task decomposition, parallel execution, and verification loops. It’s positioned as a zero-learning-curve wrapper: the tagline is literally ‘Don’t learn Claude Code. Just use OMC.’ With 16,238 stars and topics spanning agentic-coding, multi-agent-systems, and parallel-execution, it’s clearly resonating with developers tired of babysitting AI tools.
Technical Insight
OMC’s architecture centers on a five-stage pipeline: team-plan → team-prd → team-exec → team-verify → team-fix. When you run a team command, the planner breaks your request into discrete tasks, the PRD stage formalizes requirements, executors work in parallel, verifiers check outputs, and fixers loop until tests pass. This isn’t theoretical—the framework literally won’t stop until verification succeeds, creating a persistent execution model rare in AI tooling.
The real innovation is hybrid execution. OMC supports three worker types: native Claude teams (via CLAUDE_CODE_EXPERIMENTAL_AGENT_TEAMS=1 flag), tmux-spawned CLI workers for Codex/Gemini, and a tri-model synthesis mode. Here’s what tmux workers look like in practice:
# Spawn 2 Codex workers in separate tmux panes
omc team 2:codex "review auth module for security issues"
# Simultaneously run 2 Gemini workers on UI
omc team 2:gemini "redesign UI components for accessibility"
# Claude handles the business logic
omc team 1:claude "implement the payment flow"
# Check status of the auth review
omc team status auth-review
Each worker is a real CLI process in its own tmux pane—not simulated concurrency. Workers spawn on-demand and terminate when their task completes, avoiding idle resource consumption. This is a fundamentally different model from most AI orchestration tools, which multiplex agents within a single process.
The /ccg skill showcases the tri-model synthesis approach:
/ccg Review this PR — architecture (Codex) and UI components (Gemini)
This routes the architecture question to Codex via /ask codex, the UI question to Gemini via /ask gemini, then has Claude synthesize both responses into a unified recommendation. It’s a practical answer to the “which model is best?” question: use all of them, weighted by their strengths.
Installation follows two paths. The marketplace plugin route integrates directly into Claude Code:
/plugin marketplace add https://github.com/Yeachan-Heo/oh-my-claudecode
/plugin install oh-my-claudecode
/setup
/omc-setup
The npm CLI path gives you standalone orchestration:
npm i -g oh-my-claude-sisyphus@latest
Note the naming confusion: the repo is oh-my-claudecode, but the npm package is oh-my-claude-sisyphus. This trips up first-time users expecting consistency.
The deep interview mode deserves special attention. When you run /deep-interview "I want to build a task management app", the system uses Socratic questioning to expose hidden assumptions before writing any code. It measures requirement clarity across weighted dimensions—a recognition that vague prompts produce vague code. This upfront investment in clarity is the opposite of the “just start coding” mentality most AI tools encourage, and it’s a philosophically different approach to agentic coding.
Cost optimization happens through smart model routing. OMC claims to save 30-50% on tokens by matching task complexity to model capability—simple refactors go to cheaper models, architectural decisions route to premium ones. The framework also extracts problem-solving patterns from past executions and reuses them, creating a learning loop without manual configuration.
Gotcha
The entire native teams feature depends on CLAUDE_CODE_EXPERIMENTAL_AGENT_TEAMS=1, an unstable flag that could break or disappear in any Claude Code update. You’re betting on experimental features stabilizing, which is risky for production workflows.
Package naming creates real friction. The repo, commands, and documentation say oh-my-claudecode, but you install oh-my-claude-sisyphus. This isn’t just aesthetic—it breaks muscle memory and complicates CI/CD scripts. Version 4.4.0 removed the Codex/Gemini MCP servers entirely, forcing a migration to CLI workers. If you built automation around the old x and g providers, you’re rewriting.
The tmux dependency is a hard requirement for CLI workers. No tmux, no Codex/Gemini parallelization. This kills Windows compatibility unless you’re running WSL2, and it assumes your environment supports background sessions. The framework also assumes the codex and gemini CLIs are installed and authenticated—there’s no graceful fallback if they’re missing.
Verdict
Use oh-my-claudecode if you’re orchestrating complex coding tasks that genuinely benefit from parallel agent execution—refactoring a monolith, auditing security across multiple modules, or coordinating UI and backend work simultaneously. It shines when you want multi-model collaboration (Claude + Codex + Gemini) without writing orchestration code, and when you trust experimental Claude Code features enough to build workflows around them. The deep interview mode alone justifies adoption if you work with non-technical stakeholders who provide vague requirements. Skip it if you need stable, production-hardened tooling without experimental dependencies, if you’re on Windows without WSL2, or if your tasks are simple enough for a single-agent conversation. The learning system and cost optimization are compelling, but the reliance on unstable flags and tmux makes this a power-user tool, not a beginner-friendly assistant. If you want control over every AI interaction rather than autonomous orchestration, stick with Cursor or Aider.