Agent Skills: Options Comparator

Structured comparison of competing options with weighted scoring matrices, trade-off analysis, decision frameworks, and recommendation templates. Use when evaluating alternatives, making purchase decisions, or comparing strategies.

UncategorizedID: travisjneuman/.claude/options-comparator

Install this agent skill to your local

pnpm dlx add-skill https://github.com/travisjneuman/.claude/tree/HEAD/skills/options-comparator

Skill Files

Browse the full folder contents for options-comparator.

Download Skill

Loading file tree…

skills/options-comparator/SKILL.md

Skill Metadata

Name
options-comparator
Description
Structured comparison of competing options with weighted scoring matrices, trade-off analysis, decision frameworks, and recommendation templates. Use when evaluating alternatives, making purchase decisions, or comparing strategies.

Options Comparator

Structured frameworks for systematically comparing alternatives, scoring options, and producing defensible recommendations.

Weighted Scoring Matrix

Standard Weighted Matrix Template

WEIGHTED SCORING MATRIX:

STEP 1: Define criteria and weights (must sum to 100%)

| Criterion       | Weight | Option A | Option B | Option C |
|-----------------|--------|----------|----------|----------|
| [Criterion 1]   |   25%  |   [1-5]  |   [1-5]  |   [1-5]  |
| [Criterion 2]   |   20%  |   [1-5]  |   [1-5]  |   [1-5]  |
| [Criterion 3]   |   20%  |   [1-5]  |   [1-5]  |   [1-5]  |
| [Criterion 4]   |   15%  |   [1-5]  |   [1-5]  |   [1-5]  |
| [Criterion 5]   |   10%  |   [1-5]  |   [1-5]  |   [1-5]  |
| [Criterion 6]   |   10%  |   [1-5]  |   [1-5]  |   [1-5]  |
|-----------------|--------|----------|----------|----------|
| WEIGHTED TOTAL  |  100%  |  [sum]   |  [sum]   |  [sum]   |

STEP 2: Calculate weighted scores
  Weighted score = Raw score x Weight
  Total = Sum of all weighted scores

STEP 3: Interpret results
  4.5-5.0: Excellent fit
  3.5-4.4: Good fit
  2.5-3.4: Acceptable with trade-offs
  1.5-2.4: Poor fit — significant concerns
  1.0-1.4: Disqualified

SCORING RUBRIC:
  5 = Exceeds requirements / best in class
  4 = Fully meets requirements
  3 = Partially meets requirements
  2 = Significant gaps
  1 = Does not meet requirements / disqualifying

Weight Assignment Methods

| Method | How It Works | Best For | | --- | --- | --- | | Direct assignment | Stakeholders allocate 100 points across criteria | Small groups, quick decisions | | Pairwise comparison | Compare criteria two at a time, derive weights | Rigorous prioritization | | MoSCoW ranking | Must/Should/Could/Won't, then assign within tiers | Requirements-driven decisions | | Swing weighting | Rate criteria by how much best-to-worst matters | Complex multi-attribute decisions | | Stakeholder voting | Each stakeholder distributes 10 votes | Democratic team decisions |

Weight Validation Checklist

BEFORE FINALIZING WEIGHTS:

- [ ] Weights sum to exactly 100%
- [ ] No single criterion exceeds 40% (unless justified)
- [ ] No criterion is below 5% (drop it if irrelevant)
- [ ] Weights reflect stated priorities (not just habit)
- [ ] Stakeholders reviewed and approved weights
- [ ] Weights were set BEFORE scoring options
      (prevents reverse-engineering to a preferred choice)

Pairwise Comparison

Pairwise Comparison Matrix

PAIRWISE COMPARISON TEMPLATE:

Compare criteria A through E. For each pair, indicate
which is more important (mark the winner):

         A      B      C      D      E    WINS  WEIGHT
A  [--]   [ ]    [A]    [ ]    [A]     2     25%
B  [B]    [--]   [B]    [B]    [B]     4     40%
C  [ ]    [ ]    [--]   [ ]    [C]     1     10%
D  [D]    [ ]    [D]    [--]   [D]     3     25%
E  [ ]    [ ]    [ ]    [ ]    [--]    0      0%

Weight = Wins / Total comparisons x 100
Total comparisons = n(n-1)/2 = 5(4)/2 = 10

INSTRUCTIONS:
1. Compare each pair: "Is criterion X more important than Y?"
2. Mark the winner in the matrix
3. Count wins for each criterion
4. Calculate weights from win percentages
5. Adjust if any criterion has 0% but should remain

Forced Ranking

FORCED RANKING METHOD:

List all options and rank from best to worst on each criterion.
No ties allowed (forces differentiation).

| Criterion     | Rank 1 (Best) | Rank 2 | Rank 3 | Rank 4 (Worst) |
|---------------|---------------|--------|--------|-----------------|
| Price         | Option C      | Option A| Option D| Option B       |
| Quality       | Option B      | Option D| Option A| Option C       |
| Speed         | Option A      | Option B| Option C| Option D       |
| Support       | Option D      | Option C| Option B| Option A       |

SCORING:
  Rank 1 = 4 points, Rank 2 = 3, Rank 3 = 2, Rank 4 = 1
  (Or weight the ranking scores by criterion importance)

Pros/Cons with Weights

Structured Pros/Cons Template

WEIGHTED PROS/CONS ANALYSIS:

OPTION: [Name]

PROS:
| # | Advantage                    | Impact | Certainty | Score |
|---|------------------------------|--------|-----------|-------|
| 1 | [Pro description]            | H/M/L  | H/M/L    | [1-9] |
| 2 | [Pro description]            | H/M/L  | H/M/L    | [1-9] |
| 3 | [Pro description]            | H/M/L  | H/M/L    | [1-9] |

CONS:
| # | Disadvantage                 | Impact | Certainty | Score |
|---|------------------------------|--------|-----------|-------|
| 1 | [Con description]            | H/M/L  | H/M/L    | [1-9] |
| 2 | [Con description]            | H/M/L  | H/M/L    | [1-9] |
| 3 | [Con description]            | H/M/L  | H/M/L    | [1-9] |

SCORING GUIDE:
  Impact:    High=3, Medium=2, Low=1
  Certainty: High=3, Medium=2, Low=1
  Score = Impact x Certainty (range: 1-9)

NET SCORE = Sum of Pro scores - Sum of Con scores
  Positive: Pros outweigh cons
  Negative: Cons outweigh pros
  Near zero: Trade-off decision (needs judgment)

Comparative Pros/Cons

| Factor | Option A | Option B | Option C | | --- | --- | --- | --- | | Best for | [ideal use case] | [ideal use case] | [ideal use case] | | Worst for | [poor fit scenario] | [poor fit scenario] | [poor fit scenario] | | Top Pro | [strongest advantage] | [strongest advantage] | [strongest advantage] | | Top Con | [biggest drawback] | [biggest drawback] | [biggest drawback] | | Risk level | Low / Medium / High | Low / Medium / High | Low / Medium / High | | Reversibility | Easy / Hard / Impossible | Easy / Hard / Impossible | Easy / Hard / Impossible |

Decision Matrix Template

Comprehensive Decision Matrix

DECISION MATRIX:

DECISION: [Clear statement of what you are deciding]
DATE:     [Date of analysis]
OWNER:    [Decision maker(s)]
DEADLINE: [When decision must be made]

OPTIONS UNDER CONSIDERATION:
  A. [Option name and brief description]
  B. [Option name and brief description]
  C. [Option name and brief description]
  D. [Status quo / do nothing]

MUST-HAVE CRITERIA (pass/fail — eliminates options):
| Requirement          | Option A | Option B | Option C | Option D |
|----------------------|----------|----------|----------|----------|
| [Hard requirement 1] | Pass/Fail| Pass/Fail| Pass/Fail| Pass/Fail|
| [Hard requirement 2] | Pass/Fail| Pass/Fail| Pass/Fail| Pass/Fail|
| [Hard requirement 3] | Pass/Fail| Pass/Fail| Pass/Fail| Pass/Fail|

NICE-TO-HAVE CRITERIA (scored and weighted):
| Criterion    | Weight | Opt A | Opt B | Opt C | Opt D |
|-------------|--------|-------|-------|-------|-------|
| [Criterion] |   X%   | [1-5] | [1-5] | [1-5] | [1-5] |
| [Criterion] |   X%   | [1-5] | [1-5] | [1-5] | [1-5] |
| [Criterion] |   X%   | [1-5] | [1-5] | [1-5] | [1-5] |
|-------------|--------|-------|-------|-------|-------|
| TOTAL       |  100%  | [sum] | [sum] | [sum] | [sum] |

RECOMMENDATION: [Option letter] because [1-2 sentence rationale]

RISKS OF CHOSEN OPTION:
1. [Risk and mitigation plan]
2. [Risk and mitigation plan]

NEXT STEPS:
1. [Action item, owner, deadline]
2. [Action item, owner, deadline]

Trade-Off Analysis Framework

Trade-Off Mapping

TRADE-OFF ANALYSIS:

STEP 1: Identify the key trade-off dimensions
  Common trade-offs:
  - Cost vs Quality
  - Speed vs Thoroughness
  - Flexibility vs Standardization
  - Control vs Convenience
  - Short-term vs Long-term
  - Risk vs Reward
  - Simplicity vs Capability

STEP 2: Map options on trade-off axes

              HIGH QUALITY
                  |
                  |    Option B
                  |       *
                  |
  LOW COST -------+--------- HIGH COST
                  |
          Option A|
              *   |
                  |    Option C
                  |       *
              LOW QUALITY

STEP 3: Identify the efficient frontier
  Options on the frontier are rationally competitive.
  Options below the frontier are dominated
  (another option is better on all axes).

STEP 4: Choose based on priorities
  "We are optimizing for [dimension] while keeping
  [other dimension] above [minimum threshold]."

Trade-Off Decision Rules

| Rule | When to Use | How It Works | | --- | --- | --- | | Maximize one, threshold others | Clear primary objective | Set minimums for secondary criteria, then pick highest on primary | | Satisfice | Time-pressured, good enough is fine | Pick first option that meets all minimum thresholds | | Lexicographic | Clear priority ordering | Sort by most important criterion first, break ties with second | | Minimax regret | High uncertainty | Choose option that minimizes worst-case disappointment | | Expected value | Quantifiable outcomes and probabilities | Probability x payoff for each scenario, pick highest EV |

Sensitivity Analysis for Decisions

Weight Sensitivity Testing

SENSITIVITY ANALYSIS:

PURPOSE: Test if the recommendation changes when weights shift.

BASELINE WEIGHTS:
  Cost: 30% | Quality: 25% | Speed: 20% | Support: 15% | Risk: 10%
  Winner: Option B (score: 3.85)

SCENARIO 1 — Cost-focused (Cost +15%, others proportionally reduced):
  Cost: 45% | Quality: 20% | Speed: 16% | Support: 12% | Risk: 7%
  Winner: [recalculate]

SCENARIO 2 — Quality-focused (Quality +15%):
  Cost: 24% | Quality: 40% | Speed: 16% | Support: 12% | Risk: 8%
  Winner: [recalculate]

SCENARIO 3 — Risk-averse (Risk +20%):
  Cost: 22% | Quality: 19% | Speed: 15% | Support: 14% | Risk: 30%
  Winner: [recalculate]

INTERPRETATION:
  If the same option wins in all scenarios → ROBUST decision
  If winner changes in 1 scenario → Note the sensitivity
  If winner changes in 2+ scenarios → Decision depends on priorities

Score Sensitivity Testing

BREAKEVEN ANALYSIS:

"How much would Option A's score on [criterion] need
to improve to overtake Option B?"

Current:
  Option A total: 3.45
  Option B total: 3.85
  Gap: 0.40

Criterion X (weight 25%):
  Option A score: 2
  Required score to close gap: 2 + (0.40 / 0.25) = 3.6 → round to 4
  Is this plausible? [Yes/No]
  If yes → decision is sensitive to this criterion
  If no → decision is robust on this dimension

Recommendation Memo Template

Executive Decision Memo

DECISION RECOMMENDATION MEMO

TO:       [Decision maker(s)]
FROM:     [Analyst / Team]
DATE:     [Date]
RE:       Recommendation: [Decision topic]

─────────────────────────────────────────────

EXECUTIVE SUMMARY:
We recommend [Option X] for [one-sentence rationale].
This option scores highest across our evaluation criteria,
particularly in [top 2 criteria]. Estimated [cost/timeline]:
[key number]. Key risk: [top risk and mitigation].

─────────────────────────────────────────────

BACKGROUND:
[2-3 sentences on why this decision is needed now]

OPTIONS EVALUATED:
  A. [Option and one-line description]
  B. [Option and one-line description]
  C. [Option and one-line description]

EVALUATION CRITERIA AND WEIGHTS:
  [Criterion 1] (X%) | [Criterion 2] (X%) | [Criterion 3] (X%)

SCORING SUMMARY:
| Option | Score | Rank | Key Strength         | Key Weakness        |
|--------|-------|------|----------------------|---------------------|
| A      | 3.45  | 2    | [strength]           | [weakness]          |
| B      | 3.85  | 1    | [strength]           | [weakness]          |
| C      | 2.90  | 3    | [strength]           | [weakness]          |

RECOMMENDATION: Option B
Rationale: [3-5 sentences explaining why, addressing trade-offs]

SENSITIVITY: This recommendation holds under all tested scenarios
except [edge case], which would require [condition].

RISKS AND MITIGATIONS:
1. [Risk]: [Mitigation plan]
2. [Risk]: [Mitigation plan]

IMPLEMENTATION PLAN:
1. [Step, owner, date]
2. [Step, owner, date]
3. [Decision review checkpoint, date]

─────────────────────────────────────────────

APPENDIX: Detailed scoring matrix, sensitivity analysis

Vendor Evaluation Scorecard

Vendor Assessment Template

VENDOR EVALUATION SCORECARD:

VENDOR:          [Company name]
EVALUATED BY:    [Names]
DATE:            [Date]
PRODUCT/SERVICE: [What you are evaluating]

CATEGORY 1: PRODUCT FIT (30% weight)
| Criterion                  | Score (1-5) | Notes                |
|----------------------------|-------------|----------------------|
| Feature completeness       |             |                      |
| Integration capability     |             |                      |
| Scalability                |             |                      |
| Customization options      |             |                      |
| User experience / UI       |             |                      |
| Category subtotal          |     /25     |                      |

CATEGORY 2: COMMERCIAL (25% weight)
| Criterion                  | Score (1-5) | Notes                |
|----------------------------|-------------|----------------------|
| Total cost of ownership    |             |                      |
| Pricing transparency       |             |                      |
| Contract flexibility       |             |                      |
| Payment terms              |             |                      |
| ROI timeline               |             |                      |
| Category subtotal          |     /25     |                      |

CATEGORY 3: SUPPORT AND SERVICE (20% weight)
| Criterion                  | Score (1-5) | Notes                |
|----------------------------|-------------|----------------------|
| Implementation support     |             |                      |
| Training resources         |             |                      |
| Ongoing customer support   |             |                      |
| SLA commitments            |             |                      |
| Account management         |             |                      |
| Category subtotal          |     /25     |                      |

CATEGORY 4: COMPANY VIABILITY (15% weight)
| Criterion                  | Score (1-5) | Notes                |
|----------------------------|-------------|----------------------|
| Financial stability        |             |                      |
| Market position            |             |                      |
| Product roadmap            |             |                      |
| Customer references        |             |                      |
| Industry reputation        |             |                      |
| Category subtotal          |     /25     |                      |

CATEGORY 5: RISK (10% weight)
| Criterion                  | Score (1-5) | Notes                |
|----------------------------|-------------|----------------------|
| Data security / compliance |             |                      |
| Vendor lock-in risk        |             |                      |
| Migration complexity       |             |                      |
| Business continuity plan   |             |                      |
| Reference check results    |             |                      |
| Category subtotal          |     /25     |                      |

OVERALL WEIGHTED SCORE: [calculated] / 5.0
RECOMMENDATION: Proceed / Shortlist / Reject

Technology Selection Framework

Technology Evaluation Criteria

TECHNOLOGY SELECTION MATRIX:

FUNCTIONAL FIT:
- [ ] Meets core requirements (pass/fail list)
- [ ] Handles expected scale (users, data, transactions)
- [ ] Integrates with existing stack
- [ ] Supports required platforms/environments

DEVELOPER EXPERIENCE:
- [ ] Documentation quality and completeness
- [ ] Community size and activity (GitHub stars, forums)
- [ ] Learning curve for the team
- [ ] Tooling and IDE support
- [ ] Error messages and debugging experience

OPERATIONAL:
- [ ] Deployment model fits infrastructure
- [ ] Monitoring and observability support
- [ ] Backup and disaster recovery
- [ ] Security track record and patching cadence

STRATEGIC:
- [ ] Aligned with technology direction
- [ ] Vendor/project longevity (not abandonware)
- [ ] Hiring market (can you find people who know it?)
- [ ] Exit strategy (migration path if you switch later)

TOTAL COST:
- [ ] License / subscription fees
- [ ] Infrastructure costs
- [ ] Training and ramp-up time
- [ ] Maintenance and operations
- [ ] Opportunity cost of alternatives

Build vs Buy Decision

| Factor | Build | Buy | Hybrid | | --- | --- | --- | --- | | Core differentiator? | Yes — build it | No — buy it | Customize a platform | | Team has expertise? | Yes | No | Partial | | Time to value | Months | Weeks | Weeks-Months | | Long-term cost | Higher (maintenance) | Predictable (subscription) | Mixed | | Control | Full | Limited | Moderate | | Risk | Technical debt | Vendor dependency | Both | | Best when | Unique requirements, strategic IP | Commodity functionality | 80/20 fit |

Decision Anti-Patterns

COMMON DECISION MISTAKES:

1. ANALYSIS PARALYSIS
   Symptom: Endless evaluation, no decision made
   Fix: Set a decision deadline and "good enough" threshold

2. ANCHORING TO FIRST OPTION
   Symptom: First option evaluated becomes the default
   Fix: Evaluate all options before scoring any

3. CONFIRMATION BIAS
   Symptom: Seeking data that supports preferred option
   Fix: Assign a devil's advocate for each option

4. SUNK COST FALLACY
   Symptom: Sticking with an option because of past investment
   Fix: Evaluate options on future value only

5. RECENCY BIAS
   Symptom: Overweighting the last demo or reference call
   Fix: Standardize evaluation timing and criteria

6. GROUPTHINK
   Symptom: Team converges without genuine debate
   Fix: Independent scoring before group discussion

7. FEATURE COUNTING
   Symptom: Most features = best option (ignoring fit)
   Fix: Weight criteria by importance, not count

8. IGNORING STATUS QUO
   Symptom: Not comparing options against doing nothing
   Fix: Always include "do nothing" as Option D

See Also