Module 8 · Lesson 8.1

The shape of orchestration.

Module 8 is the first module where you stop holding one steering wheel and start composing two or three at once. The agent that does research does not also draft the brief. The agent that drafts the brief does not also check citations. The thing you own and build is the pipeline between them. This lesson names the three shapes orchestration actually takes, gives you a two-question test for picking the right one, installs the one-agent-too-many test that tells you when not to orchestrate, and names runaway as the failure mode the rest of the module is designed to prevent.

Stage 1 of 3

Read & Understand

5 concept blocks

From directing one agent to composing several CORE

Every module through Module 7 has been about one student talking to one agent at a time. You held the steering wheel in Module 3 against a repository. You held it in Module 4 against the open web. You held it in Module 5 against a mailbox and a calendar. In Module 6 you let a schedule hold the wheel for you, but the wheel was still steering one agent through one task. In Module 7 you elevated the prompt itself into named, reusable skills and plugins — still one agent, now carrying better instructions.

Module 8 is the first module where you stop holding one wheel and start composing two or three or four of them. The agent that does research does not also draft the brief. The agent that drafts the brief does not also check the citations. The agent that checks citations does not also package the draft for your email. Each of those jobs becomes its own small agent, with its own scope, its own instructions, and its own narrow piece of the work — and the thing you build and own is the pipeline between them.

That move sounds small. It is not. A system made of three agents handing off to one another is a different kind of object from one agent doing a larger job. It has behaviors a single agent cannot have — a stage can run in parallel with a sibling, a supervisor can retry a failed worker, a downstream agent can reject upstream output as insufficient. It also has failure modes a single agent cannot have — a vague handoff silently corrupts everything downstream, a hidden retry loop quietly spends fifty times what you budgeted, a supervisor misjudges which of four workers succeeded and promotes the wrong output.

The central claim of this module is that the quality of an orchestrated system is set by the quality of its interfaces, not the quality of its agents. A brilliant drafter that is handed bad notes produces a bad draft. A merely competent drafter that is handed tightly scoped input produces an acceptable draft. The place to invest your attention — from the very first sketch of the pipeline — is the spaces between the agents, not the agents themselves. Lesson 8.2 makes that into a concrete discipline called the handoff contract. This lesson is where you learn to see the shape that contract is going to sit inside.

A useful way to name the shift: through Module 7 you learned to direct an agent. In Module 8 you learn to compose agents. A director tells one person what to do; a composer writes a score that three or four people can follow together without the composer standing behind each of them. The mental model for the rest of the module is the composer, not the director.

Three shapes orchestration actually takes CORE

Nearly every multi-agent system you will build, install, or read about lands in one of three shapes. You can combine them — a branch of a hierarchical system can be a sequential pipeline; a stage of a sequential pipeline can fan out into parallel workers — but the three primitives are the vocabulary. Name the shape before you build.

1. Sequential pipeline. Agent A produces output; agent B reads that output and produces its own; optionally agent C reads B’s output and produces the final artifact. The work has an order dependency — you cannot draft the brief before the research is done, you cannot check citations before the draft exists. Sequential is the shape most students reach for first, and most of the time it is the right shape. A three-agent research → draft → review pipeline is sequential. A two-agent triage → reply-draft pipeline (Module 5 extended) is sequential. A scheduled gather → summarize pipeline (Module 6 extended) is sequential.

The sequential pipeline is easy to reason about because the state at each step is visible — the artifact agent A wrote is the artifact agent B reads, and you can open both. Cost is additive, not multiplicative. The weakness of the sequential pipeline is that a bad early stage contaminates everything after it. Debugging means reading the trace from the top and finding the first stage whose output drifted.

2. Parallel workers. Several agents run at once on independent subtasks, and the results are combined. The work has no order dependency among the parallel siblings — the draft-social-post agent does not need the draft-blog-post agent to finish first; both start from the same source material and produce different artifacts. Parallel is the right shape when the same input feeds several independent outputs, or when a single input can be sliced and each slice processed without reference to the others (summarize each of ten source articles in parallel, then combine the summaries).

Parallel workers are cheaper in wall-clock time than running the same work sequentially — a student whose pipeline would take twelve minutes sequentially may run it in three minutes in parallel. They are not cheaper in dollar cost; they are usually more expensive, because the combining step (even when trivial) adds tokens. The weakness of parallel workers is coordination: all the workers have to agree on the shape of the input they were given and the shape of the output they produce, because a combiner that receives four differently-shaped outputs cannot cleanly merge them.

3. Hierarchical supervisor. A supervisor agent is handed a goal; the supervisor decides how to decompose the goal into subtasks, spawns worker agents to execute the subtasks (often in parallel), reads their outputs, and produces the final artifact — sometimes calling another round of workers if the first attempt was insufficient. Hierarchical is the right shape when the decomposition itself is non-trivial and you would rather have an agent do it than write the decomposition by hand. It is also the right shape when you do not know, in advance, how many workers the task will need — a research sweep across an unknown number of sources, say.

Hierarchical orchestration is the most powerful shape and the most dangerous. A supervisor that is allowed to spawn workers that spawn workers is the canonical picture of runaway orchestration — the tree of agents grows unchecked, the token count grows with it, and a simple question quietly costs more than a restaurant dinner. The depth and fan-out caps in the Module 8 rails (depth ≤ 3, fan-out ≤ 4) are aimed squarely at this pattern. A sequential pipeline cannot run away. A parallel-worker pipeline runs away only at its fan-out step. A hierarchical pipeline can run away anywhere, any time, and it usually takes the novice by surprise.

Picking the shape: a two-question test CORE

Before you build, you pick. The two-question test is:

Question 1: Does the work have order dependencies? Is there a clean “first this, then this, then this” story you can tell? If yes, you are looking at a sequential pipeline for the ordered portion. If no — if several pieces of work could start at the same time because none of them needs the others’ output — you are looking at parallel workers.

Question 2: Does the decomposition itself need judgment? If you already know what the subtasks are and how many there will be, you do not need a supervisor to figure it out — parallel or sequential composition is enough. If the number of subtasks is unknown, or the right decomposition depends on what the input actually is, you want a hierarchical supervisor to make that call at runtime.

Apply the test to the canonical examples from earlier modules:

  • Research sweep on a topic, produce a brief, check citations, deliver the final draft. Order dependency is strong (cannot draft before research, cannot check before draft). Decomposition is fixed (three named stages). Sequential pipeline.
  • Given one article, produce a blog post, three social posts, and a newsletter blurb. No order dependency — each output derives from the same source independently. Decomposition is fixed (four named outputs). Parallel workers.
  • Given a topic, do enough research to answer it — however many sources that takes, however many angles it needs — and produce a brief. Order dependency is present (decide decomposition first, then execute). Decomposition is not fixed — the supervisor has to decide how to slice the research at runtime based on what the topic actually requires. Hierarchical supervisor.

Most students’ first real pipeline is sequential. That is the intended starting point. A sequential pipeline is legible; a hierarchical pipeline is not. When in doubt, sequential.

The one-agent-too-many test: when not to orchestrate CORE

The most expensive mistake a Module 8 student can make is orchestrating a job a single agent could have done well. Orchestration has real overhead: more prompts to write, more handoffs to design, more artifacts to read, more cost per run, and — worst of all — more surface area for silent failure. If a single agent can do the job reliably, orchestrating it is not leverage; it is drag.

The one-agent-too-many test is four questions. If the answer to any of them is “no,” do not orchestrate. Tighten the single-agent version first.

1. Does the single-agent version already succeed? Have you actually run this as one prompt, one skill, or one scheduled task, and gotten an acceptable result? If you have not tried the single-agent version, do that first. Half of would-be pipelines collapse into “actually, one agent with a better prompt is fine.”

2. Is the single-agent version failing on separable sub-problems? When it fails, does it fail at the same step every time — “the drafting part is fine but the citation-checking part is always sloppy” — or does it fail diffusely? Orchestration helps when the failure is localized; orchestration does not help when the failure is diffuse (that is a prompt or skill problem, and lives in Modules 3–7).

3. Does each sub-problem get materially better with a dedicated agent? Would a focused draft-checker agent — with its own prompt, its own context, maybe its own model — actually do a cleaner job than “please also check the citations” appended to the drafter? If the failure is localized but the cure is “a sharper prompt in the same agent,” stay with one agent.

4. Is the goal big enough to be worth the orchestration overhead? A pipeline costs at least one handoff contract to design, two invocations’ worth of tokens to run, and a place to live in the capstone register. A goal that a human takes five minutes on and an agent takes thirty seconds on is almost never worth a pipeline. A goal that a human takes an hour on, that happens three times a week, is usually worth one.

Applied to concrete examples:

  • Summarize an article in my own voice. One agent. A better prompt, maybe a skill from Module 7, will finish this cleanly. Do not orchestrate.
  • Write a paper about Napoleon. One agent, mostly. The real work is the student’s thinking; orchestration cannot substitute. Do not orchestrate (for the student’s own writing).
  • Review my week’s calendar, draft three preparation notes, and post them into a “meeting prep” folder. Three separable sub-problems, each materially different (one is calendar-reading, one is drafting, one is filing). Happens weekly. Orchestrate, sequentially.
  • Once a morning: gather the top headlines from three feeds, summarize each, produce a combined briefing. Two separable stages, each with its own failure mode (gathering fails differently from summarizing). Daily. Orchestrate, sequentially.

A student who applies this test honestly ends Module 8 with one well-built pipeline, not three. That is the intended outcome. The capstone requires one frozen pipeline, not a catalog.

The runaway failure mode, in miniature CORE

Lesson 8.5 treats runaway orchestration at full depth. This block is the headline warning, named early so you see it coming through Lessons 8.2, 8.3, and 8.4.

A runaway pipeline is one that consumes materially more tokens, time, or dollars than the student expected when they wrote the blueprint. The three most common shapes are:

  • Depth runaway. A supervisor spawns workers that spawn sub-workers that spawn sub-sub-workers. Each layer looks innocent; the total is not. The depth cap of 3 in Module 8 (workers, one supervisor, optionally one meta-supervisor) exists because students who exceed that depth almost always regret it the first time they read the cost.
  • Fan-out runaway. A supervisor or parallel step spawns twelve workers when four would have done. Each worker’s cost is small; twelve of them is not. The fan-out cap of 4 exists as a default to force the student to notice when they are about to exceed it.
  • Retry runaway. An agent retries on failure, and the retry logic does not have a clear stopping condition. A flaky tool call plus an aggressive retry loop can multiply token usage by five or ten without anyone seeing it happen. The handoff contract (Lesson 8.2) is what makes “failure” something the system can notice clearly; without it, retry loops do not know when to stop.

Your defense against all three is the same three-part rail: a depth limit, a fan-out cap, and a pre-flight cost estimate you run before every invocation. The rail does not need to be complicated; it needs to be present. Write the estimate in the blueprint. Refuse to run the pipeline until the estimate is within the per-pipeline budget. Practice stopping the pipeline mid-run at least once. These rituals are installed in Lesson 8.5, but every lesson between now and then assumes you are operating under them.

Stage 2 of 3

Try & Build

1 recipe + activity

Where the orchestration logic lives, on each path RECIPE

Tools Claude Code CLI (subagent path) · Claude desktop app — Cowork tab (scheduled-task path)
Last verified 2026-04-18
Next review 2026-07-18
OSes (CLI path) macOS, Linux, Windows
OSes (Cowork-tab path) macOS, Windows (the Claude desktop app is required)

This is the only recipe block in Lesson 8.1. The details move as the tools move; the Recipe Book carries the canonical versions.

Claude Code path. The orchestration logic lives in the parent Claude Code session, which uses the Task tool to dispatch work to subagents. A subagent is a folder — typically under ~/.claude/subagents/<subagent-name>/ or <project-root>/.claude/subagents/<subagent-name>/ — with a description (a classifier, same discipline as Module 7), a system prompt, and an explicit tools allowlist. The parent session reads subagent descriptions the same way Claude Code reads skill descriptions, and dispatches to the best match. Inter-agent state passes through files in a shared folder the parent owns. The parent session is, in effect, the supervisor of any hierarchical pipeline; for sequential pipelines, the parent just calls subagents in turn.

Cowork path. The orchestration logic lives in the scheduled-task definitions and the folder convention. There is no single “orchestrator process” — what orchestrates the pipeline is that task A is scheduled to fire at 6:00 a.m. and write its output to /pipeline/<name>/<date>/01-research/, task B is scheduled to fire at 6:10 a.m. and read from that folder, and so on. Each scheduled task can invoke whichever Module 7 plugin skills it needs. Cowork does not (at the time of Module 8 authoring) have a native “run these in parallel as a unit” primitive for scheduled tasks; what you have is independent schedules that read and write shared files. That is enough for every pattern in this module, but the student should know that the coordination is by convention — you are the one who lined up the schedules, and you are the one who has to keep them lined up.

Which path fits which shape. Sequential pipelines work on both paths. Parallel-worker patterns work more cleanly on Claude Code (the parent session can dispatch multiple subagents in parallel) than on Cowork (where parallel means “both scheduled to fire at once against disjoint input folders”). Hierarchical patterns belong on Claude Code, full stop — the supervisor has to be a live agent making decisions, and that is what the parent Claude Code session is. If your goal needs a hierarchical shape, that is a strong signal to pick the Claude Code path when you freeze in Lesson 8.4.

Canonical recipe entries

The Recipe Book entries building-a-subagent-pipeline-in-claude-code.md and chaining-scheduled-tasks-in-cowork-tab.md carry the current file paths, command names, and UI flows. Those entries are dated and versioned. This Core Book lesson stops at the conceptual level.

Try it — Pick your goal, name your shape, set your posture CORE

three deliverables · written to my-first-loop.md and a scratch file

Part 1 — Pattern picker: four scenarios.

For each of the four scenarios below, apply the two-question test from Content Block 3 and name the pattern. Write your answer in a short markdown block in a scratch file (this is working notes, not a capstone artifact).

  • Scenario A. Every weekday at 6 a.m., gather the top five headlines from three news feeds, summarize each in two sentences, and produce a single briefing file you read at breakfast.
  • Scenario B. Given a single source document, produce a blog-post draft, a three-tweet social version, and a newsletter blurb — all from the same source.
  • Scenario C. Given a topic you name, research it until you have enough material to answer it (the “enough” is judgment-dependent on the topic), and produce a two-page brief.
  • Scenario D. Summarize the article I just pasted in, in my voice.

Self-check after you answer: A = sequential (order dependency: gather, then summarize, then combine; decomposition fixed). B = parallel workers (no order dependency among the three outputs; decomposition fixed). C = hierarchical supervisor (decomposition requires runtime judgment about how many angles the topic needs). D = not orchestrated at all (single agent; the one-agent-too-many test fails at question 2 — there are no separable sub-problems).

Part 2 — Add your Orchestration Posture to my-first-loop.md.

Open your running my-first-loop.md capstone file and add a new section titled Orchestration Posture. Write three short paragraphs:

  1. My one-agent-first rule. Name the rule you will apply when tempted to orchestrate: “Before I orchestrate, I will try the single-agent version of this task and confirm it fails on at least one separable sub-problem.” Write it in your own words.
  2. My agent-count ceiling for this module. A soft ceiling for how many agents live in a single pipeline. Pick a number you can defend, and write it down. Most students keep their pipeline small — a handful of agents, not a long list. The ceiling is a commitment to yourself, not a rule the system will enforce.
  3. My depth and fan-out defaults. Restate the module’s defaults (depth ≤ 3, fan-out ≤ 4) and name the condition under which you would exceed them. The right answer for most students is “I will not exceed them in Module 8.”

Part 3 — Pick a goal and sketch a shape.

List three to five goals from your own life that feel like they might need orchestration — things you are doing now as single-agent work (or manually) that are big enough, repeated enough, and have localized failures. Candidates to consider: an extended Module 4 research-and-draft workflow, a Module 5 email triage that also produces reply drafts, a Module 6 scheduled summary that also annotates sources, a homework or project workflow where research → draft → review is a natural shape.

For each candidate, apply the one-agent-too-many test and write down the answer to each of the four questions. Most candidates will fail at least one question; that is a feature of the test, not a bug.

Pick one candidate that passes all four. This is your Module 8 goal. Write it in a new block at the bottom of your Orchestration Posture section titled My Module 8 pipeline. Under it, write:

  • The goal, in one sentence.
  • The pattern you think the goal fits (sequential / parallel / hierarchical), with one sentence of justification.
  • The names of the two or three agents you expect the pipeline to have, each with a one-sentence description of what that agent is responsible for.
  • The input that starts the pipeline and the output it produces.

This is a sketch, not a commitment. You will rewrite it after Lesson 8.2, when the handoff contracts force the agent scopes to be more precise. The point of the sketch here is to have something concrete to write contracts against.

If none of your candidates pass the test

That is a legitimate outcome. Most students’ real lives do not contain three good pipeline candidates. Pick the candidate closest to passing and treat Module 8 as a learning exercise on it, even if you would not choose to build it if this were not a course. The module is not trying to manufacture a pipeline out of thin air; it is trying to give you the skill to recognize one when it shows up. If your goal is on the edge of “worth orchestrating,” the module will still work. You can also revisit your candidate list at the end of Lesson 8.2 — sometimes a vague “not quite a pipeline” goal sharpens once the handoff contracts are written.

Deliverables. Three artifacts: a new Orchestration Posture section in my-first-loop.md with three paragraphs; a My Module 8 pipeline block underneath it with the goal, the pattern, the agent sketch, and the input/output; and the four-scenario pattern-picker sort in a scratch file (working notes, not a capstone artifact).

Done with the hands-on?

When the recipe steps and any activity above are complete, mark this stage to unlock the assessment, reflection, and project checkpoint.

Stage 3 of 3

Check & Reflect

key concepts, quiz, reflection, checkpoint, instructor note

Quick check

Four questions. Tap a question to reveal the answer and the reasoning.

Q1. Which statement best describes the shift Module 8 asks you to make?
  • A You learn to write better prompts for one agent at a time.
  • B You learn to compose several agents into a system, and the quality of the system is set by the interfaces between the agents more than by any single agent.
  • C You learn to run agents on a schedule so they happen without you.
  • D You learn to build plugins that bundle several skills together.
Show explanation

Answer: B. (A) is the shape of Modules 3 and 7. (C) is Module 6. (D) is Module 7. Module 8’s particular move is composition — several agents, with the interfaces between them as the thing you design and own. The second clause of (B) — interfaces matter more than agents — is the headline and reappears in every subsequent lesson.

Q2. You have a task: given a topic, do enough research to answer it (where “enough” depends on the topic), and produce a brief. Which pattern fits?
  • A Sequential pipeline.
  • B Parallel workers.
  • C Hierarchical supervisor.
  • D Do not orchestrate; use a single well-tuned research-sweep skill instead.
Show explanation

Answer: C. The key phrase is “where ‘enough’ depends on the topic” — decomposition requires runtime judgment. A sequential pipeline has a fixed decomposition. Parallel workers have a fixed decomposition. Only a hierarchical supervisor can decide how many workers (or how many rounds of workers) the topic needs. (D) is tempting, and for narrow, well-scoped topics it may still be right — but as stated, the task has runtime-variable decomposition, which is exactly when orchestration earns its keep.

Q3. A student proposes a four-layer hierarchical pipeline: a top supervisor, a middle supervisor that coordinates four workers, each worker calling two helper agents, and each helper calling one verifier. What is the correct Module 8 response?
  • A Approve as designed.
  • B Reject: exceeds the depth cap of 3 and the fan-out cap at multiple levels; redesign to flatten.
  • C Approve if the student agrees to pay for a larger budget.
  • D Approve the shape but reduce each layer’s prompt length.
Show explanation

Answer: B. Depth is four or more layers (top sup → middle sup → workers → helpers → verifiers), exceeding the cap of 3. Fan-out is at the edge at the middle supervisor and again at each worker. Redesigning to flatten — pulling verification up into the worker’s own scope, having the top supervisor dispatch workers directly — is the default move. (C) misses the point of the rail: the cap is not primarily a budget concern; it is a legibility and failure-surface concern. (D) does not address the runaway risk.

Q4. Your single-agent weekly-report prompt fails every week at the same step — it summarizes meetings well but always produces a weak “decisions captured” section. What is the right next move?
  • A Build a three-agent pipeline: a meeting-summarizer, a decisions-extractor, and a combiner.
  • B Try a sharper prompt for the decisions step inside the single agent; orchestrate only if that does not fix it.
  • C Orchestrate immediately — this is what Module 8 is for.
  • D Abandon the report; it is not worth orchestrating.
Show explanation

Answer: B. The one-agent-too-many test requires you to try the single-agent version first. “Failing at the same step every time” is a classic localized failure — a sign orchestration could help, but also a sign a better prompt for that step inside the same agent might fix it at a fraction of the overhead. Module 8 prefers you exhaust the one-agent fix first. If the sharper prompt fails, the failure is now diagnosed more precisely, and the pipeline in (A) is a reasonable next step with tight scope. (C) is the mistake this module is designed to prevent. (D) is an overreaction.

Reflection prompt

From directing to composing: where did you under- and over-estimate the overhead?

Write 4–6 sentences in your journal or my-first-loop.md in response to the following: Before this lesson, if someone had asked you “when would you use more than one AI agent on the same job?”, what would you have said? Compare that to what you now think. Where did you under-estimate the overhead of orchestration? Where did you over-estimate it? Which of your candidate goals felt most naturally like a pipeline to you, and which felt like it was trying too hard to be one?

No length minimum past four sentences; no length maximum. The purpose is to notice the move from “I will use whatever agents are available” to “orchestration is a deliberate design choice with real costs, and the default is not to.”

Project checkpoint

Stand up your Orchestration Posture and pick the Module 8 goal.

By the end of this lesson, your my-first-loop.md capstone file should contain a new section titled Orchestration Posture. Use this template:

Orchestration Posture. My one-agent-first rule: before I orchestrate, I will try the single-agent version of this task and confirm it fails on at least one separable sub-problem.

My agent-count ceiling for this module is [3 / 4]. Over that, I redesign, not reach for a bigger budget.

My depth and fan-out defaults: depth ≤ 3, fan-out ≤ 4. I will not exceed them in Module 8.

My Module 8 pipeline.

Goal: [one sentence].

Pattern: [sequential / parallel / hierarchical] — because [one sentence].

Agents (2–3): [name — one-sentence responsibility, x2–3].

Input that starts the pipeline: [one line]. Output it produces: [one line].

Also keep the four-scenario pattern-picker sort from Part 1 in a scratch file — it is working notes, not a capstone artifact.

Do not proceed to Lesson 8.2 until the posture is written and your Module 8 goal passes the one-agent-too-many test. The shape comes before the contracts.

Next in Module 8

Lesson 8.2 — Handoff contracts: every arrow is a specification.

The headline technical insight of the module. Turn the agent sketch you just wrote into a set of handoff contracts — the four-field specification (shape, what counts as success, what to do on failure, what changes require a re-audit) that says what each agent owes the next. The contracts land in /capstone/pipeline-v1-draft/blueprint.md. The HTML activity Handoff contract drill ships with this lesson.

Continue to Lesson 8.2 →