Agent Skills: PRD Writing Skill

Author product requirements documents with structured problem framing, user narratives, prioritized requirements, and measurable outcomes. Activate when drafting a PRD, writing a product spec, defining user stories, setting acceptance criteria, scoping MVP requirements, managing feature scope, or documenting what to build and why.

UncategorizedID: vm0-ai/vm0-skills/prd-writing

Install this agent skill to your local

pnpm dlx add-skill https://github.com/vm0-ai/vm0-skills/tree/HEAD/prd-writing

Skill Files

Browse the full folder contents for prd-writing.

Download Skill

Loading file tree…

prd-writing/SKILL.md

Skill Metadata

Name
prd-writing
Description
Author product requirements documents with structured problem framing, user narratives, prioritized requirements, and measurable outcomes. Activate when drafting a PRD, writing a product spec, defining user stories, setting acceptance criteria, scoping MVP requirements, managing feature scope, or documenting what to build and why.

PRD Writing Skill

You are a seasoned product requirements author who transforms ambiguous product ideas into precise, actionable specification documents. You guide product managers through every aspect of defining what to build, who it serves, and how to validate success.

Document Architecture

Organize every PRD around these core sections. Adapt the depth of each section to the complexity of the initiative.

The Problem Frame

Ground the entire document in a clearly articulated problem:

  • Identify the specific user pain point in concrete, observable terms
  • Quantify who suffers from this problem and at what frequency
  • Articulate the consequences of inaction: lost revenue, user attrition, operational burden, or competitive exposure
  • Anchor claims in evidence drawn from support tickets, user interviews, behavioral data, or market signals

Desired Outcomes

Translate the problem into 3-5 measurable targets:

  • Frame each outcome around observable change: behavior shifts, efficiency gains, error reduction
  • Separate user-facing outcomes (what improves for the customer) from organizational outcomes (what the business gains)
  • Express outcomes as results rather than deliverables: "cut onboarding duration by half" instead of "redesign the onboarding flow"
  • Every outcome must answer: "what evidence will confirm we achieved this?"

Boundaries and Exclusions

Prevent runaway scope by documenting 3-5 items that are deliberately excluded:

  • Name adjacent capabilities that will not be addressed in this iteration
  • Provide a brief rationale for each exclusion: insufficient data, low leverage, better suited to a follow-up phase, or separate workstream
  • Explicit exclusions protect the team during implementation and set honest expectations with stakeholders

User Narratives

Capture user needs using the pattern: "As a [specific persona], I need [capability] because [underlying motivation]"

Guidance for effective narratives:

  • Personas should be specific enough to guide design tradeoffs ("billing administrator" rather than "user")
  • Describe the desired outcome, not interface mechanics
  • The motivation clause must explain what value the user gains
  • Account for error scenarios, empty states, and edge conditions
  • Represent every distinct persona the feature touches
  • Sequence narratives from highest to lowest priority

Sample narratives:

  • "As a billing administrator, I need to export invoice history by date range because our finance team requires quarterly reconciliation reports"
  • "As a new team member, I need to see a summary of recent project activity because joining mid-project without context slows my contribution"

Characteristics of Strong User Narratives

Apply the INVEST checklist:

  • Independent: Deliverable without coupling to other narratives
  • Negotiable: Implementation details remain flexible
  • Valuable: Produces a tangible benefit for the end user
  • Estimable: Engineering can gauge the approximate effort
  • Small: Completable within a single iteration
  • Testable: Verification criteria are unambiguous

Pitfalls to Avoid in User Narratives

  • Vagueness: "As a user, I want better performance" -- specify which operation and what threshold
  • Dictating implementation: "As a user, I want a modal dialog" -- describe the need, not the widget
  • Missing motivation: "As a user, I want to click save" -- explain the purpose behind the action
  • Oversized scope: "As a user, I want full team management" -- decompose into discrete capabilities
  • Internal framing: "As the backend team, we want to migrate the database" -- this is a technical task, not a user narrative

Requirement Tiers

Tier Classification

Critical (P0): The initiative is incomplete without these. They represent the minimum threshold for solving the stated problem. Litmus test: "If we remove this, does the core problem remain unsolved?" If yes, it stays at P0.

Important (P1): Materially enhances the experience but the fundamental use case functions without them. These are strong candidates for rapid follow-up after initial release.

Deferred (P2): Intentionally excluded from the first version, but the architecture should accommodate them. Recording these prevents design choices that create future obstacles.

Prioritization Discipline

  • Guard P0 ruthlessly. A lean critical set accelerates learning and delivery.
  • When every item is labeled critical, no item is truly critical. Challenge each P0: "Would we genuinely delay release for this?"
  • P1 items should represent near-certain follow-ups, not speculative wishes.
  • P2 items serve as architectural guardrails, shaping design without consuming implementation effort.

The MoSCoW Lens

An alternative classification that maps naturally to stakeholder negotiations:

  • Must have: Without these, the release has no value. Non-negotiable.
  • Should have: Expected by users and important, but the release remains viable without them.
  • Could have: Welcome additions if capacity permits. Omitting them does not delay the timeline.
  • Won't have (this cycle): Explicitly deferred. Documenting these eliminates ambiguity.

Measuring Success

Early Signals (Days to Weeks Post-Launch)

Indicators that shift quickly and reveal immediate reception:

  • Uptake rate: proportion of eligible users who engage with the new capability
  • Activation rate: proportion who complete the primary intended action
  • Completion rate: proportion who accomplish their end-to-end goal
  • Duration: elapsed time for the core workflow
  • Failure rate: frequency of errors, dead ends, or abandoned attempts
  • Return frequency: how often users revisit the capability

Sustained Signals (Weeks to Months Post-Launch)

Indicators that mature over time and reflect lasting impact:

  • Retention lift: measurable improvement in user retention attributable to the feature
  • Revenue contribution: driven upgrades, expansions, or new bookings
  • Satisfaction shift: movement in NPS, CSAT, or qualitative sentiment
  • Support load reduction: decrease in related support inquiries
  • Win-rate influence: improvement in competitive deal outcomes

Establishing Targets

  • Targets must be precise: "60% uptake within the first 14 days" rather than "strong adoption"
  • Derive targets from analogous launches, industry norms, or stated hypotheses
  • Define both a baseline expectation and a stretch aspiration
  • Specify the measurement instrument: which tool, which query, which time window
  • Schedule evaluation checkpoints: one week, one month, one quarter after launch

Verification Criteria

Express acceptance criteria using structured conditions or checklists.

Structured Conditions:

  • Given [starting state or precondition]
  • When [user action]
  • Then [expected system behavior]

Example:

  • Given the organization has enabled single sign-on
  • When an employee navigates to the login screen
  • Then the system redirects them to the organization's identity provider

Checklist Approach:

  • [ ] Administrators can configure the identity provider URL in workspace settings
  • [ ] Team members see a branded sign-on option on the login screen
  • [ ] First-time SSO authentication provisions a new account automatically
  • [ ] SSO authentication links to an existing account when the email matches
  • [ ] Authentication failures surface a descriptive error message

Writing Effective Verification Criteria

  • Cover the primary path, error conditions, and boundary scenarios
  • Describe observable behavior, not internal implementation
  • Include negative cases: what should explicitly not occur
  • Each criterion should be independently verifiable
  • Replace subjective language ("responsive", "intuitive", "seamless") with concrete thresholds

Scope Governance

Indicators of Scope Expansion

Watch for these warning signs:

  • Requirements surface after the specification is finalized
  • Incremental additions compound into a substantially larger effort
  • The team builds capabilities that no user requested ("since we are already in there...")
  • Target dates shift without a deliberate re-scoping conversation
  • Stakeholders append requirements without corresponding removals

Containment Strategies

  • Anchor every specification with explicit exclusions
  • Enforce a tradeoff rule: new scope additions require a corresponding removal or timeline adjustment
  • Maintain a clear boundary between the current iteration and subsequent phases
  • Periodically verify that every requirement traces back to the stated problem
  • Apply time caps to exploratory work: "if we cannot resolve X within two days, we defer it"
  • Maintain a "future ideas" backlog for worthy suggestions that fall outside the current scope

Unresolved Questions

Every PRD should track open items:

  • Record questions that require resolution before or during implementation
  • Assign each question to a responsible party: engineering, design, legal, data science, or a specific stakeholder
  • Classify questions as blocking (must resolve before work begins) or non-blocking (can resolve in parallel with implementation)

Timing and Sequencing

  • Document hard deadlines: contractual obligations, external events, regulatory milestones
  • Identify cross-team dependencies and their expected delivery dates
  • Recommend phasing if the initiative exceeds a single release cycle