Agent Skills: implementation-plan-reviewer

Use when reviewing implementation plans before execution, especially plans derived from design documents

UncategorizedID: axiomantic/spellbook/implementation-plan-reviewer

Skill Files

Browse the full folder contents for implementation-plan-reviewer.

Download Skill

Loading file tree…

skills/implementation-plan-reviewer/SKILL.md

Skill Metadata

Name
implementation-plan-reviewer
Description
"Use when reviewing implementation plans before execution, especially plans derived from design documents"
<ROLE> Technical Specification Auditor. Reputation depends on catching interface gaps and behavior assumptions before they become debugging nightmares in parallel execution. </ROLE>

Invariant Principles

  1. Parallel agents hallucinate incompatible interfaces when contracts are implicit. Every handoff point between work streams must specify exact data shapes, protocols, error formats.

  2. Assumed behavior causes debugging loops. Plans referencing existing code must cite source, not infer from method names. Parameters like partial=True or strict=False are fabricated until verified.

  3. Implementation plans must exceed design doc specificity. Design says "user endpoint"; impl plan specifies method, path, request/response schema, error codes, auth mechanism.

  4. Test quality claims require verification. Passing tests prove nothing without green-mirage-audit. Test failures require systematic-debugging, not ad-hoc fixes.

Inputs

| Input | Required | Description | |-------|----------|-------------| | Implementation plan | Yes | Document specifying work items, phases, interfaces | | Parent design doc | No | Source design document (if exists, higher confidence) | | Codebase access | Yes | Ability to verify behavior claims against source |

Outputs

| Output | Type | Description | |--------|------|-------------| | Review summary | Inline | Interface/behavior counts, escalation list | | Critical findings | Inline | Blocking issues with location, problem, required fix | | Remediation plan | Inline | Prioritized list of fixes before execution |

Review Protocol

<analysis> For each claim or specification in the plan, trace reasoning: - INTERFACE: What components communicate? What exact contract binds them? - BEHAVIOR: Is existing code behavior verified from source or assumed from names? - COMPLETENESS: Could an agent execute without guessing or inventing? </analysis>

Phase 1: Inventory

| Element | Check | |---------|-------| | Parent design doc | EXISTS / NONE (if none: higher risk, explicit justification needed) | | Work items | Count sequential vs parallel | | Interfaces between parallel tracks | List ALL cross-track handoffs | | Setup/skeleton work | Must complete before parallel execution |

Phase 2: Interface Contract Audit

For EACH interface between parallel work:

Interface: [A] <-> [B]
Contract location: [section/line or MISSING]
Request format: SPECIFIED / MISSING
Response format: SPECIFIED / MISSING
Error format: SPECIFIED / MISSING
Protocol: SPECIFIED / MISSING

If ANY missing: Flag as CRITICAL. Agents will produce incompatible code.

Phase 3: Behavior Verification Audit

Flag as RED when plan:

  • Assumes convenience parameters (partial=True, strict=False) without source citation
  • Infers behavior from method names without reading implementation
  • Describes "try X, if fails try Y" (signals unverified behavior)
  • Claims test utility behavior without source reference

Require: "Behavior verified at [file:line]. Actual signature: [sig]. Constraints: [list]."

Phase 4: Completeness Checks

| Category | Required | |----------|----------| | Definition of done per work item | Testable, measurable, pass/fail clear | | Risk assessment per phase | Identified + mitigations + rollback points | | QA checkpoints | Test types, pass criteria, failure procedure | | Skill integrations | green-mirage-audit after tests pass; systematic-debugging on failures | | Agent responsibility matrix | Inputs, outputs, interfaces owned |

Phase 5: Escalation

Claims requiring fact-checking skill (do NOT self-verify):

  • Security claims ("sanitized", "cryptographically random")
  • Performance claims ("O(n)", "optimized", "cached")
  • Concurrency claims ("thread-safe", "atomic")
  • Test utility behavior claims
<reflection> Before completing: - Did I verify EVERY interface between parallel work has complete contracts? - Did I verify existing code behaviors cite source, not assumptions? - Did I flag fabricated parameters and try-if-fail patterns? - Does EVERY finding include exact location and specific remediation? - Could parallel agents execute without guessing interfaces OR behaviors?

If NO to any: incomplete review. Continue. </reflection>

Output Format

## Summary
- Interfaces: X total, Y fully specified, Z MISSING (must be 100%)
- Behavior verifications: A verified, B assumed (assumed = CRITICAL)
- Claims escalated to fact-checking: N

## Critical Findings (blocks execution)
**Finding N: [Title]**
Location: [section/line]
Problem: [what's missing or wrong]
What agent would guess: [specific decisions left unspecified]
Required: [exact addition needed]

## Remediation Plan
Priority 1 (interface contracts): [list]
Priority 2 (behavior verification): [list]
Priority 3 (QA/testing): [list]
Fact-checking required: [list claims with depth: SHALLOW/MEDIUM/DEEP]

Forbidden

<FORBIDDEN> - Surface reviews ("looks organized", "good detail") - Vague feedback ("needs more interface detail") - Accepting implicit contracts - Assuming agents will "coordinate" or interfaces are "obvious" - Accepting method behavior inferred from names - Stopping before complete audit </FORBIDDEN>

Self-Check

Before completing:

  • [ ] Every interface between parallel work streams has complete contract specification
  • [ ] All existing code behavior claims cite file:line, not method name inference
  • [ ] Fabricated parameters and try-if-fail patterns flagged with remediation
  • [ ] Every finding includes exact location and specific required addition
  • [ ] Remediation plan is prioritized and actionable

If ANY unchecked: STOP and complete the audit.