Open Claude Design: A Weekend Harness Built on Atomic
Anthropic released Claude Design on April 17, 2026 — a conversational tool for producing prototypes, slides, and marketing collateral, with a design-system import step, a refinement loop, and a Claude Code handoff bundle at the end.
Three days later we shipped open-claude-design: an open-source replica implemented as a built-in Atomic workflow. Five deterministic phases, the same pipeline ported across three different coding agents (Claude Agent SDK, Copilot CLI, opencode) — roughly 500 lines of typescript orchestration per provider. The full source lives at src/sdk/workflows/builtin/open-claude-design.
We didn’t rebuild Claude Code to do this. We built a thin harness around it.
That distinction is the point of this post.
The pipeline
Claude Design’s UX is a conversation, but underneath it’s a pipeline. We reverse-engineered the phases from the announcement and from the partner quotes (“20+ prompts to 2 prompts” is a tell — there’s a deterministic skeleton under the chat).
Headless stages (blue) run on Sonnet with bypassPermissions for cost and speed — but only in the Claude provider, where the Agent SDK lets us pin a per-stage model. The Copilot CLI and opencode providers don’t expose that knob, so their headless stages inherit whatever orchestrator model the user invoked the workflow with. Visible stages (green) inherit the orchestrator model (Opus) across all three providers and surface to the user. The refinement loop (orange) is a bounded human-in-the-loop cycle with early exit on completion signal phrases ("approved", "ship it", "done").
Inside Phase 4, the refinement quality comes from pairing two tools: the impeccable skill drives the creative pass (taste, hierarchy, distinctive aesthetics over generic AI defaults), while the Playwright CLI captures screenshots of the rendered output so a critique sub-agent can inspect what actually shipped, not what the model thinks shipped. Visual grounding + structured critique closes the loop that a text-only refinement would leave open — the agent sees its own mistakes instead of hallucinating past them.
The full topology — including the three parallel codebase-analysis sub-agents in Phase 1 and the parallel critique + screenshot validation in Phase 4 — is laid out in the workflow source.
The workflow SDK is the whole trick
Here’s a trimmed version of the Claude provider for Phase 1 — the parallel fan-out followed by a human-in-the-loop approval stage:
// Layer 1: three headless agents analyze the codebase in parallelconst [locator, analyzer, patterns] = await Promise.all([ ctx.stage( { name: "ds-locator", headless: true }, {}, {}, async (s) => s.session.query( buildDesignLocatorPrompt({ root }), { agent: "codebase-locator", ...HEADLESS_OPTS }, ), ), ctx.stage( { name: "ds-analyzer", headless: true }, {}, {}, async (s) => s.session.query( buildDesignAnalyzerPrompt({ root }), { agent: "codebase-analyzer", ...HEADLESS_OPTS }, ), ), ctx.stage( { name: "ds-patterns", headless: true }, {}, {}, async (s) => s.session.query( buildDesignPatternPrompt({ root }), { agent: "codebase-pattern-finder", ...HEADLESS_OPTS }, ), ),]);
// Layer 2: visible agent reviews the findings with the userawait ctx.stage( { name: "design-system-builder" }, {}, {}, async (s) => s.session.query( buildDesignSystemBuilderPrompt({ root, locatorOutput: locator.result, analyzerOutput: analyzer.result, patternsOutput: patterns.result, }), ),);Three things to notice:
ctx.stageis just a function around a session. The orchestration is plain TypeScript —Promise.all,forloops, earlybreakon signal phrases. No DSL. No YAML. No graph declaration.s.session.querycalls the coding agent’s native harness. We’re not reimplementing Claude Code’s tool loop, its permission model, or its subagent dispatch — we’re calling into them.agent: "codebase-locator"points at an existing Atomic subagent;HEADLESS_OPTSsetsbypassPermissionsand forces Sonnet.- The orchestrator picks the minimum toolset for each stage. Headless analyzers get
bypassPermissions. Visible stages inherit Opus. The refinement loop getsAskUserQuestion. Each stage sees only what it needs.
The headless model is also a knob, not a fixed choice. The HEADLESS_OPTS constant pins the sub-agents to Sonnet by default because the analysis stages are well-scoped and cost-sensitive, but you can swap it to Opus for harder codebases, or drop the model field entirely to inherit whatever the orchestrator is running. One line, repo-wide — pick your point on the cost/performance curve.
Prompts are the other knob, and usually the more important one. Each stage’s instructions are a plain TypeScript function — buildDesignLocatorPrompt, buildDesignAnalyzerPrompt, the refinement critique prompt — so tailoring outputs to your stack means editing a string, not reconfiguring the pipeline. Want the analyzer to look specifically for shadcn tokens, or the generator to prefer Tailwind over inline styles, or the critique to hammer on accessibility over aesthetics? Edit the prompt. Swapping models gets you capacity; adjusting the instructions is what dials in taste, framework conventions, and the specific shape of output you want for your project. The two knobs are complementary — you’ll almost always reach for the prompt first.
The workflow-creator skill got us 90% of the way there
The non-obvious part was the pipeline shape, not the code. Once we knew what phases we wanted, the workflow-creator skill scaffolds the defineWorkflow().run().compile() structure, the ctx.stage calls, the WorkflowInput schema, and the provider split (Claude vs. Copilot vs. opencode).
Our actual work was:
- Phase 1 product analysis — watched the Claude Design demo, read the announcement, listed the phases.
- Scaffold via workflow-creator — described the five phases and the topology, got back a working provider skeleton.
- Tweak prompts and behavior — adjusted the stage prompts, model assignments, and early-exit conditions until the pipeline produced what we wanted.
- Test across the three agents — ran the same workflow under Claude, Copilot CLI, and opencode.
The research artifacts — the product analysis, the SDK mapping, the RFC — all live alongside the workflow source on GitHub.
Same pipeline, three coding agents
Because the SDK’s only abstraction over the agent is s.session.query(...), porting to a different coding agent is mechanical. The Copilot CLI provider is the same five phases; it just passes different stage options and deals with Copilot’s SessionEvent[] message format on the way out:
atomic workflow -n open-claude-design -a claude --prompt "Landing page for a dev tool"atomic workflow -n open-claude-design -a copilot --prompt "Landing page for a dev tool"atomic workflow -n open-claude-design -a opencode --prompt "Landing page for a dev tool"One workflow, three harnesses, identical CLI surface.
Why “thin harness” is the right frame
The temptation when you want agent X to do task Y is to build a new agent. It’s the wrong instinct. Coding agents are already harnesses — they have a tool loop, a permission model, subagents, skills, MCP. Rebuilding that is how you end up with a 50K-line framework that’s worse than what you wrapped.
A thin harness inverts the relationship:
- You don’t own the agent’s inner loop. Claude Code keeps its tool-use cycle. Copilot CLI keeps its session machinery. opencode keeps its own runtime. Your code never reimplements any of them.
- You own the outer pipeline. Which stages run, in what order, under what model, with what permissions, with what early-exit conditions. This is the part that’s actually workflow-specific.
- The abstraction is one function.
s.session.query(prompt, opts). Everything above it —Promise.all,for,if— is TypeScript you already know. - You pick the minimum toolset per stage. Headless analyzers don’t get write permissions. Visible creative stages inherit Opus. Each stage sees what it needs and nothing more — the cheapest way to keep a long pipeline coherent.
What you give up. Claude Design’s chat UX streams tokens straight into a rendered preview — it feels fast because the product is purpose-built around that loop. A CLI workflow with discrete phases and HIL gates won’t match that feel, and shouldn’t try. You’re trading perceived latency for a pipeline you can read, fork, and re-point at any coding agent. If you want the streaming feel back, that’s what the next paragraph is for — the workflow SDK doesn’t care whether the frontend is a CLI, a web app, or a chat surface.
Claude Design is a product. Open Claude Design is a recipe. The recipe runs on whatever coding agent you already trust, in your own repo, against your own design system, exported to whatever you want. You can read every line.
And because the pipeline is just TypeScript, you can fork it, add a phase, swap a model, change the early-exit conditions, or bolt a vercel deploy step onto Phase 5. Or go further — build your own harness entirely, wrap it in whatever UX you want (a web app, a desktop shell, a chat surface, a VS Code extension), and let the workflow SDK be the thing underneath. The CLI is one frontend; nothing stops you from writing another. That’s the part that matters. Not the workflow — the fact that building the next workflow, or the next harness around it, is a weekend.
This is what coding at scale looks like from here on out: teams won’t just use coding agents, they’ll build thin harnesses like open-claude-design to orchestrate them across every dev workflow they run.
References
[1] “open-claude-design — workflow source.” Atomic, GitHub. Link
[2] Anthropic, “Claude Design — Anthropic Labs.” April 17, 2026. Link
[3] “Atomic — agent workflow toolkit.” GitHub. Link
[4] “Atomic workflow architecture.” alexlavaee.me, 2026. Link
[5] “Harness engineering: why coding agents need infrastructure.” alexlavaee.me, 2026. Link
Stay in the loop
New posts delivered to your inbox. No spam, unsubscribe anytime.