Agent Skills: Prioritization

Prioritization techniques including MoSCoW, Kano model, weighted scoring, and value-effort matrices. Ranks requirements, features, backlog items, and investment decisions.

UncategorizedID: melodic-software/claude-code-plugins/prioritization

Install this agent skill to your local

pnpm dlx add-skill https://github.com/melodic-software/claude-code-plugins/tree/HEAD/plugins/business-analysis/skills/prioritization

Skill Files

Browse the full folder contents for prioritization.

Download Skill

Loading file tree…

plugins/business-analysis/skills/prioritization/SKILL.md

Skill Metadata

Name
prioritization
Description
Prioritization techniques including MoSCoW, Kano model, weighted scoring, and value-effort matrices. Ranks requirements, features, backlog items, and investment decisions.

Prioritization

Systematically rank and prioritize requirements, features, backlog items, and initiatives using proven prioritization frameworks. Supports MoSCoW, Kano model, weighted scoring, and value-effort analysis.

What is Prioritization?

Prioritization is the process of determining relative importance and ordering of items to focus resources on what matters most. Effective prioritization balances:

  • Value: Benefit to customers or business
  • Effort: Cost, time, and resources required
  • Risk: Uncertainty and potential downsides
  • Dependencies: Constraints and sequencing

Prioritization Techniques

MoSCoW Method

Categorical prioritization for timeboxed delivery:

| Category | Definition | Guidance | |----------|------------|----------| | Must | Non-negotiable, required for success | Without these, delivery is a failure | | Should | Important but not critical | Significant value, workarounds exist | | Could | Desirable if resources permit | Nice to have, enhances experience | | Won't | Explicitly excluded this time | Not now, maybe later |

When to Use: Sprint planning, release scoping, MVP definition, timeboxed projects

Rules:

  • Musts should be ~60% of capacity (leave room for unknowns)
  • Won'ts are explicitly stated (not silently dropped)
  • Categories are relative to the timebox, not absolute

Kano Model

Customer satisfaction-based classification:

| Category | If Present | If Absent | Detection | |----------|------------|-----------|-----------| | Basic (Must-Be) | No increase in satisfaction | Major dissatisfaction | Customers assume these exist | | Performance (Linear) | Proportional satisfaction | Proportional dissatisfaction | Customers explicitly request | | Delighter (Excitement) | High satisfaction | No dissatisfaction | Customers don't expect | | Indifferent | No impact | No impact | No reaction either way | | Reverse | Dissatisfaction | Satisfaction | Segment prefers absence |

When to Use: Product feature prioritization, understanding customer needs, differentiating from competitors

Kano Questionnaire:

  • Functional: "How would you feel if this feature was present?"
  • Dysfunctional: "How would you feel if this feature was absent?"

Responses: Like it, Expect it, Neutral, Can tolerate, Dislike it

Weighted Scoring Matrix

Multi-criteria quantitative comparison:

Step 1: Define Criteria

| Criterion | Weight | Description | |-----------|--------|-------------| | Customer Value | 40% | Impact on customer satisfaction | | Strategic Fit | 25% | Alignment with goals | | Effort | 20% | Development cost (inverse) | | Risk | 15% | Uncertainty/failure potential (inverse) |

Step 2: Score Items

| Item | Customer Value (1-5) | Strategic Fit (1-5) | Effort (1-5) | Risk (1-5) | Weighted Score | |------|---------------------|---------------------|--------------|------------|----------------| | A | 5 | 4 | 3 | 4 | 4.15 | | B | 3 | 5 | 4 | 3 | 3.75 |

Step 3: Calculate Weighted Score

Score = Σ (Weight × Score)
Item A = (0.40×5) + (0.25×4) + (0.20×3) + (0.15×4) = 4.20

When to Use: Complex trade-offs, multiple stakeholders, defensible decisions

Value vs Effort Matrix

2×2 prioritization for quick decisions:

quadrantChart
    title Value vs Effort
    x-axis Low Effort --> High Effort
    y-axis Low Value --> High Value
    quadrant-1 Big Bets (Plan carefully)
    quadrant-2 Quick Wins (Do first)
    quadrant-3 Fill-ins (Do if time permits)
    quadrant-4 Money Pits (Avoid)

| Quadrant | Value | Effort | Action | |----------|-------|--------|--------| | Quick Wins | High | Low | Do first | | Big Bets | High | High | Plan carefully | | Fill-ins | Low | Low | Do if time permits | | Money Pits | Low | High | Avoid or deprioritize |

When to Use: Fast initial triage, backlog grooming, stakeholder alignment

RICE Scoring

Product management prioritization:

| Factor | Definition | Calculation | |--------|------------|-------------| | Reach | Users/customers affected | Number per time period | | Impact | Effect on each user | 0.25 (minimal) to 3 (massive) | | Confidence | Certainty of estimates | 0.5 (low) to 1 (high) | | Effort | Person-months required | Number |

RICE Score = (Reach × Impact × Confidence) / Effort

When to Use: Product roadmap prioritization, feature comparison

WSJF (Weighted Shortest Job First)

SAFe/Lean prioritization for flow:

WSJF = Cost of Delay / Job Duration

Cost of Delay = User/Business Value + Time Criticality + Risk Reduction

| Factor | Score (1-20) | Description | |--------|--------------|-------------| | User/Business Value | 1-20 | Benefit to users or business | | Time Criticality | 1-20 | Urgency, deadlines, decay | | Risk Reduction | 1-20 | Risk/opportunity addressed | | Job Duration | 1-20 | Relative size (inverted) |

When to Use: Continuous flow environments, maximizing value delivery

Workflow

Phase 1: Prepare

Step 1: Gather Items to Prioritize

## Prioritization Session

**Date:** [ISO date]
**Scope:** [What's being prioritized]
**Stakeholders:** [Who's involved]
**Constraint:** [Timebox, budget, capacity]

### Items

| ID | Description | Owner |
|----|-------------|-------|
| 1 | [Item 1] | [Name] |
| 2 | [Item 2] | [Name] |

Step 2: Select Prioritization Technique

| Situation | Recommended Technique | |-----------|----------------------| | Sprint/release planning | MoSCoW | | Product feature decisions | Kano + RICE | | Trade-off decisions | Weighted Scoring | | Quick triage | Value vs Effort | | Continuous flow | WSJF | | Multiple criteria | Weighted Scoring |

Phase 2: Execute

Step 1: Apply Selected Technique

Follow the specific technique workflow (see above).

Step 2: Validate Results

  • Do top priorities align with strategy?
  • Are dependencies respected?
  • Does the team have capacity?
  • Are stakeholders aligned?

Step 3: Document Rationale

## Prioritization Rationale

### Top Priorities

1. **[Item A]** - Score: X
   - Rationale: [Why this is top priority]
   - Dependencies: [What it depends on]

2. **[Item B]** - Score: Y
   - Rationale: [Why this is second]
   - Dependencies: [What it depends on]

### Deferred Items

- **[Item C]** - Reason: [Why deferred]

Phase 3: Communicate

Step 1: Create Prioritized Backlog

## Prioritized Backlog

| Rank | Item | Priority/Score | Owner | Target |
|------|------|----------------|-------|--------|
| 1 | [Item A] | Must / 4.5 | [Name] | Sprint 1 |
| 2 | [Item B] | Must / 4.2 | [Name] | Sprint 1 |
| 3 | [Item C] | Should / 3.8 | [Name] | Sprint 2 |

Step 2: Communicate Decisions

  • Share prioritization results with stakeholders
  • Explain rationale for key decisions
  • Address concerns about deprioritized items
  • Set expectations for what's not included

Output Formats

Narrative Summary

## Prioritization Summary

**Session:** [Scope/context]
**Date:** [ISO date]
**Technique:** [MoSCoW/Kano/Weighted Scoring/etc.]
**Facilitator:** prioritization-analyst

### Results Overview

- **Total Items:** N
- **Top Priority:** [Count]
- **Deferred:** [Count]

### Priority Distribution

| Category | Count | % |
|----------|-------|---|
| Must/Quick Wins | X | Y% |
| Should/Big Bets | X | Y% |
| Could/Fill-ins | X | Y% |
| Won't/Money Pits | X | Y% |

### Key Decisions

1. **[Top Item]**: Prioritized because [reason]
2. **[Deferred Item]**: Deferred because [reason]

### Next Steps

1. Begin work on top priority items
2. Re-prioritize at [next review point]

Structured Data (YAML)

prioritization:
  version: "1.0"
  date: "2025-01-15"
  scope: "Q1 Feature Backlog"
  technique: "weighted_scoring"
  facilitator: "prioritization-analyst"

  criteria:
    - name: "Customer Value"
      weight: 0.40
    - name: "Strategic Fit"
      weight: 0.25
    - name: "Effort"
      weight: 0.20
      inverse: true
    - name: "Risk"
      weight: 0.15
      inverse: true

  items:
    - id: "FEAT-001"
      name: "User Dashboard"
      scores:
        customer_value: 5
        strategic_fit: 4
        effort: 3
        risk: 4
      weighted_score: 4.20
      priority: 1
      rationale: "Highest customer value, manageable effort"

    - id: "FEAT-002"
      name: "API Integration"
      scores:
        customer_value: 3
        strategic_fit: 5
        effort: 4
        risk: 3
      weighted_score: 3.75
      priority: 2
      rationale: "Strong strategic alignment"

  moscow_summary:
    must: ["FEAT-001"]
    should: ["FEAT-002", "FEAT-003"]
    could: ["FEAT-004"]
    wont: ["FEAT-005"]

Mermaid Visualizations

Value-Effort Matrix:

quadrantChart
    title Prioritization Matrix
    x-axis Low Effort --> High Effort
    y-axis Low Value --> High Value
    quadrant-1 Big Bets
    quadrant-2 Quick Wins
    quadrant-3 Fill-ins
    quadrant-4 Money Pits
    "Feature A": [0.2, 0.9]
    "Feature B": [0.3, 0.7]
    "Feature C": [0.7, 0.8]
    "Feature D": [0.8, 0.3]
    "Feature E": [0.2, 0.2]

MoSCoW Distribution:

pie title MoSCoW Distribution
    "Must" : 3
    "Should" : 4
    "Could" : 5
    "Won't" : 2

When to Use Each Technique

| Technique | Best For | Team Size | Time Required | |-----------|----------|-----------|---------------| | MoSCoW | Sprint/release planning | Any | 30-60 min | | Kano | Product features | Product team | 2-4 hours | | Weighted Scoring | Complex trade-offs | Cross-functional | 1-2 hours | | Value vs Effort | Quick triage | Any | 15-30 min | | RICE | Product roadmap | Product team | 1-2 hours | | WSJF | Continuous flow | SAFe teams | 30-60 min |

Common Pitfalls

| Pitfall | Prevention | |---------|------------| | Everything is "Must" | Enforce category limits (60% capacity) | | HiPPO (highest paid person's opinion) | Use objective scoring criteria | | Ignoring effort | Always consider cost/effort dimension | | Static prioritization | Re-prioritize regularly as context changes | | Overcomplicating | Start simple, add complexity only if needed | | Ignoring dependencies | Map dependencies before finalizing order |

Integration

Upstream

  • Requirements - Items to prioritize
  • stakeholder-analysis - Stakeholder input on value
  • swot-pestle-analysis - Strategic context

Downstream

  • Sprint planning - Ordered backlog
  • Roadmaps - Prioritized initiatives
  • decision-analysis - Detailed option evaluation

Related Skills

  • decision-analysis - For complex option evaluation
  • stakeholder-analysis - Stakeholder input on priorities
  • risk-analysis - Risk dimension of prioritization
  • capability-mapping - Capability investment prioritization

User-Facing Interface

When invoked directly by the user, this skill operates as follows.

Arguments

  • <items-or-context>: Items to prioritize (inline list, file reference, or context description)
  • --mode: Prioritization method (default: moscow)
    • moscow: Must/Should/Could/Won't categorization (~4K tokens)
    • kano: Customer satisfaction categorization (~5K tokens)
    • weighted: Multi-criteria weighted scoring (~6K tokens)
    • all: All three methods for comparison (~12K tokens)
  • --output: Output format (default: both)
    • yaml: Structured YAML for downstream processing
    • markdown: Formatted markdown tables
    • both: Both formats
  • --dir: Output directory (default: docs/analysis/)

Execution Workflow

  1. Parse Arguments - Extract items, mode, and output format. If no items provided, ask the user what to prioritize.
  2. Gather Items - Collect from inline list, file reference, or context-based exploration.
  3. Execute Based on Mode:
    • MoSCoW: Categorize into Must/Should/Could/Won't with stakeholder input on business criticality, dependencies, compliance, and user impact.
    • Kano: Classify by satisfaction impact (Basic, Performance, Delighter, Indifferent, Reverse) considering customer expectations and competitive baseline.
    • Weighted: Define criteria with weights, score each item 1-5, calculate weighted scores, and rank.
    • All: Run all three methods, compare for consistency, highlight conflicts, and synthesize final priority.
  4. Generate Output - Produce YAML structure, markdown tables (MoSCoW summary, weighted scoring matrix), Mermaid visualizations (quadrantChart, pie chart), and summary report.
  5. Save Results - Save to docs/analysis/prioritization.yaml and/or docs/analysis/prioritization.md (or custom --dir).
  6. Suggest Follow-Ups - Recommend effort estimation for high-priority items, risk analysis for high-risk items, and capability-mapping for alignment.

Version History

  • v1.0.0 (2025-12-26): Initial release