| name | review-prompt-plan |
|---|---|
| description | Review a prompt plan against its source spec: launches an internal prompt plan review and `/peer-review-prompt-plan` in parallel and returns combined findings. Use when the user asks to "review my prompt plan", "review my prompts", "check my prompt plan", "critique my prompts", or wants prompt plan feedback before implementation. |
Run two AI prompt plan reviews in parallel and return combined findings. The core principle: a prompt plan is a "broken down spec." Following the full prompt chain should implement the entire spec with nothing missing and nothing disconnected.
Determine the prompt plan to review:
- If prompt plan text is in conversation context, use it
- If a file path was provided, read the file
- If neither was provided, check for a prompt plan at
.turbo/prompts.md
Also read the source spec (path listed in the prompt plan's Source: field).
Launch two agents in a single message so they run concurrently (model: "opus", do not set run_in_background):
Spawn a subagent with the full prompt plan text and the source spec. Instruct it to:
- Apply the prompt plan review dimensions below
- Return findings in the output format below
Spawn a subagent to run the /peer-review-prompt-plan skill with the prompt plan text.
Wait for both agents to complete. Aggregate their findings with attribution (reviewer: "internal" or "peer") and return them to the caller.
The caller determines what to do with the findings (evaluate, apply, or present to the user).
For every requirement, feature, acceptance criterion, and constraint in the spec, verify it appears in at least one prompt. Flag any requirement not assigned to a prompt.
Each prompt produces artifacts (files, modules, endpoints, types) that later prompts may depend on. Verify the chain is connected. Check for:
- Dead ends — A prompt creates a module, API, or capability that no subsequent prompt consumes or integrates
- Missing prerequisites — A prompt assumes an artifact exists that no prior prompt creates
- Implicit dependencies — A prompt's
Depends onfield omits a prompt it actually relies on
Walk through the prompts in order and simulate what exists after each one completes. After the final prompt, verify that every spec feature is reachable from the project's entry points and no component is orphaned.
Verify each spec requirement is assigned to exactly one prompt. Duplication causes conflicting implementations. If two prompts intentionally touch the same area (one creates, the other extends), confirm they have clear boundaries.
If prompts reference external resources (other codebases, documentation, APIs, search terms), spot-check that referenced projects or docs actually exist and are relevant.
- P0 — Spec requirement missing from all prompts, or prompt chain produces a broken system
- P1 — Dead end, missing prerequisite, or implicit dependency that will cause implementation problems
- P2 — Moderate issue: duplicated requirement, unclear prompt boundary, or ordering improvement
- P3 — Minor improvement
Return findings as a numbered list. For each finding:
### [P<N>] <title (imperative, ≤80 chars)>
**Prompt(s):** <prompt number(s) affected>
**Reviewer:** <internal | peer>
<one paragraph explaining the problem, what impact it has on implementation, and a suggested fix>
After all findings, add:
## Overall Verdict
**Readiness:** <ready | needs revision>
<1-3 sentence assessment>
If there are no qualifying findings, state that the prompt plan looks ready for implementation and explain briefly.
- If any reviewer is unavailable or returns malformed output, proceed with findings from the remaining reviewer.
- Present findings grouped by priority, then by reviewer.