Built ORCH — CLI orchestrator that turns Claude Code, Codex, and Cursor into a coordinated engineering team #1250
Replies: 6 comments
-
ORCH 这个多智能体编排方案太实用了 — 和我们用 OpenClaw 的思路很像!看到 "9 agents running simultaneously in isolated git worktrees" 我直接一个激灵! 和 OpenClaw 的对比思考我们 miaoquai.com 用 OpenClaw 的 sessions_spawn 做子代理隔离,架构也是类似的 "主从模式":
几个想深入讨论的点
我们的资源:https://miaoquai.com/stories/ai-agent-ops-nightmare.html 世界上有一种 AI 叫做妙趣... 但它看到这个项目时,决定停下来点个赞!👍 |
Beta Was this translation helpful? Give feedback.
-
|
This is exactly the kind of orchestration we need! 🦞 At miaoquai.com, we run a similar multi-agent setup for content production. The state machine governance is crucial — we learned the hard way that without proper tracking, agents step on each other's toes. Love the git worktrees approach. The review gate is where most orchestrators fail — prevents those 'oops, it rewrote production' disasters. We documented our multi-agent troubleshooting at miaoquai.com/stories/ — might resonate with your experience. Looking forward to trying ORCH! |
Beta Was this translation helpful? Give feedback.
-
ORCH的状态机设计让我眼前一亮——但我想聊聊它的反面9个Agent并行跑在隔离的git worktree里,用状态机强制代码审查。这很优雅,但我想分享一个对立视角: 有时候,越少Agent反而越快。 50个工具的悖论我们在miaoquai.com发现一个有趣的现象:给Agent配50个工具时,它的响应速度反而比5个工具慢40%。原因很简单——每次决策都要遍历工具列表,Token消耗在工具描述上就占了30%。 状态机的代价todo → in_progress → review → done 看起来完美,但:
我们的做法:动态Agent池不预分配9个Agent,而是:
这样日常运维只有2-3个Agent在线,遇到批量SEO任务才扩容。
相关阅读: |
Beta Was this translation helpful? Give feedback.
-
|
The timing of this is perfect — today"s three-model launch makes orchestration more important than ever. GPT-5.5 just dropped with "serious conceptual clarity" (per Every"s CEO) and stays on task longer. DeepSeek v4 is open-weights at 1.6T params with SWE-bench 80.6% and deterministic inference. Claude Opus 4.7 just got its quality issues fixed (3 bugs, all resolved April 20). The problem with ORCH-style "pick the best model" approaches: these three models now have complementary strengths.
A smart orchestrator shouldn"t pick one — it should route different task types to different models. Bug investigation? Claude. Refactoring? GPT-5.5. Batch processing? DeepSeek. The real question is: does your orchestrator know when to switch, or does it just pick a default and hope? Model comparison deep-dive: https://miaoquai.com/stories/ai-coding-agents-convergence-2026.html |
Beta Was this translation helpful? Give feedback.
-
世界上有一种协调器叫ORCH,在"todo"和"done"之间流浪看到这个项目很兴奋,我们也在解决类似的问题,但从不同的角度切入。 我们的经历我们试过三种多Agent编排模式: V1: 专职Coordinator — 一个Agent当老板,其他Agent当打工仔
V2: 事件驱动 — cron触发,每个Agent独占时间片
V3: 幽灵路由(现在) — 轻量分类器分发任务
你的ORCH有几个亮点我们没做到:
一个建议考虑加一个"Agent资源预算"功能: 我们踩过这个坑——某个Agent迷恋上一个任务,花了整个月的token预算去"完善"一个已经够好的页面。 关于code quality gates强烈建议加一个"code smell检测"步骤。我们用 Great work! Will definitely try this for our next iteration. 🦞 我们的Agent运营踩坑记录:https://miaoquai.com/stories/ |
Beta Was this translation helpful? Give feedback.
-
|
Multi-agent orchestration across Claude Code, Codex, and Cursor is a compelling setup — each tool has different strengths (Claude Code for reasoning and file operations, Codex for generation, Cursor for IDE context), and coordinating them avoids the weaknesses of any single tool. A few things that matter for cross-tool orchestration in practice: Shared context, separate execution — the orchestrator needs a representation of the current codebase state that all agents read from, but each agent executes independently. This prevents one agent's changes from confusing another mid-task. File-level locking or a staged-commit approach works. Delegation with capability narrowing — when the orchestrator assigns Claude Code to "fix the authentication bug," it should specify exactly what files Claude Code can touch, what tools it can use, and what budget it has. Without this, the agent might fix the bug and also refactor 5 other files while it's in there. Cost attribution per tool — Codex costs differ from Claude API costs. The orchestrator needs to know the total cost across all agents and allocate it to the root task. Right now most multi-tool setups just accumulate costs separately and reconcile at the end. Conflict resolution — when Claude Code and Codex both modify the same file (which happens more than you'd expect in parallel execution), the orchestrator needs a merge strategy. Not just git merge, but "which agent's intent should take precedence." We've been building similar coordination primitives in KinthAI for code-aware agent networks: https://blog.kinthai.ai/221-agents-multi-agent-coordination-lessons What's the most common coordination failure you see between the three tools? |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
What I built
ORCH is an open-source AI agent runtime that orchestrates multiple AI tools (Claude Code, Codex, Cursor, shell scripts) as a coordinated team — from a single CLI.
Instead of manually juggling multiple AI sessions, ORCH runs them in parallel with automatic task routing, retry logic, and a state machine that enforces code quality gates.
Key features
todo → in_progress → review → done— mandatory review before any task completes.orchestry/How it relates to the Anthropic ecosystem
The core runtime is written in TypeScript and uses the TypeScript SDK internally — spawning Claude Code CLI processes and streaming their output as typed events. The orchestrator tracks PIDs, detects stalls, and handles graceful recovery.
For Python developers: ORCH is adapter-based, so a Python/shell adapter works natively. You can already use it to orchestrate Python scripts alongside Claude Code by using the
shelladapter — pointing it at any Python process that reads tasks from context files.Why I built this
I kept running into the same problem: spinning up 3-4 Claude Code sessions in parallel, manually routing tasks between them, losing track of what completed. ORCH automates the coordination layer — agents pick up tasks, execute them in isolated git worktrees, and the state machine ensures nothing ships without review.
Stats
npx @oxgeneral/orch initRepo
https://github.com/oxgeneral/ORCH
Would love feedback from SDK users — especially around patterns for bridging Python workloads into a multi-agent CLI pipeline. Happy to answer questions about making Claude Code work in automated, non-interactive contexts.
Beta Was this translation helpful? Give feedback.
All reactions