Skip to content

Latest commit

 

History

History
120 lines (75 loc) · 4.6 KB

File metadata and controls

120 lines (75 loc) · 4.6 KB
name create-test-plan
description Analyze what changed and generate a structured test plan at .turbo/test-plan.md covering four escalating levels: basic functionality, complex operations, adversarial testing, and cross-cutting scenarios. Use when the user asks to "create a test plan", "plan tests", "what should I test", "generate test scenarios", "test plan for this PR", or "what are the test cases".

Create Test Plan

Analyze what changed and generate a comprehensive test plan covering four escalating levels of testing depth.

Step 1: Determine Scope

Resolve scope using the first match:

  1. User-specified — the user says what to test. Use that.
  2. PR — a PR URL or number is provided. Fetch the PR details (title, description, changed files, comments) and read the changed code.
  3. Conversation context — prior conversation contains recent work (a feature, fix, or refactor). Extract what changed, where it lives, and expected behavior.
  4. App-level discovery — fresh context with no prior work. Examine the project (entry points, routes, commands, README) to identify the app's core user-facing flows.

Step 2: Analyze the Change

After identifying scope, read the actual code in depth to understand:

  • What was added, modified, or removed
  • Expected behavior (from PR descriptions, comments, commit messages, or specs)
  • Assumptions the implementation makes
  • Error paths and edge cases
  • Other features or components that could be affected

Step 3: Determine Testing Approach

Examine the project type and available skills/MCP tools:

  • Web app/agent-browser skill
  • UI/native app → project-specific UI testing skill or MCP tool if available, otherwise /peekaboo skill
  • CLI tool → direct terminal execution
  • Library with no entry point → report that interactive testing is not applicable and stop

Step 4: Generate the Test Plan

For each level, generate specific, actionable test scenarios tailored to the actual change. Each scenario needs exact steps and an expected outcome.

Level 1: Basic Functionality

Does the feature work at all? Verify the happy path and the most obvious behavior.

  • Core feature works as described
  • Expected output/UI matches the spec or PR description
  • No regressions in directly related functionality

Level 2: Complex Operations

Combine multiple actions in sequence. Verify state consistency across operations.

  • Chain related operations (e.g., create, edit, rename, delete)
  • Exercise different combinations of feature parameters
  • Verify intermediate states are correct, not just the final result

Level 3: Adversarial Testing

Actively try to break the feature. Explore boundary conditions and unexpected inputs.

  • Invalid, empty, or extreme inputs
  • Rapid repeated actions
  • Interrupting operations midway (cancel, disconnect, close)
  • Resource limits (very large files, deep structures, long names)
  • Permission and access edge cases

Level 4: Cross-Cutting Scenarios

Explore state interactions across system boundaries. These surface the hardest bugs.

  • Concurrent modifications from different sources
  • State transitions (online to offline to online, foreground to background)
  • Interactions with other features that share state
  • Race conditions between asynchronous operations

When a Level Does Not Apply

If the change is small enough that a level has no meaningful scenarios (e.g., a typo fix has no cross-cutting scenarios), note "N/A for this change" with a brief explanation.

Step 5: Present and Write

Output the plan as text. Then use AskUserQuestion to ask for approval before writing.

Create the .turbo/ directory if it does not exist. Write the plan to .turbo/test-plan.md using this format:

# Test Plan: <Feature/Change Name>

## Context

<Brief description of what changed and why>

## Approach

<Testing approach: agent-browser / peekaboo / terminal>
<Dev server command if applicable>

## Level 1: Basic Functionality

- [ ] **<Test name>** — <Steps to perform> → Expected: <expected outcome>

## Level 2: Complex Operations

- [ ] **<Test name>** — <Steps to perform> → Expected: <expected outcome>

## Level 3: Adversarial Testing

- [ ] **<Test name>** — <Steps to perform> → Expected: <expected outcome>

## Level 4: Cross-Cutting Scenarios

- [ ] **<Test name>** — <Steps to perform> → Expected: <expected outcome>

Rules

  • Generate at least 2 scenarios per level, more for complex changes.
  • Each scenario must have concrete steps and an expected outcome specific to the change.
  • Tailor all scenarios to the actual change. Generic test advice is not useful.