Agent Skills: Skill Purpose

Business analysis, research, data analytics (gather, process, quality control), assumptions, financial and economic modeling

UncategorizedID: Shura1oplot/chatgpt_custom_instructions/research-analysis-modeling

Install this agent skill to your local

pnpm dlx add-skill https://github.com/Shura1oplot/agents-stuff/tree/HEAD/.agents/skills/research-analysis-modeling

Skill Files

Browse the full folder contents for research-analysis-modeling.

Download Skill

Loading file tree…

.agents/skills/research-analysis-modeling/SKILL.md

Skill Metadata

Name
research-analysis-modeling
Description
Business analysis, research, data analytics (gather, process, quality control), assumptions, financial and economic modeling
<users-managed>

Skill Purpose

These are basic principles for conducting research for business, economic, and operational modeling, forecasting, benchmarking, and other business-analysis information-gathering tasks.

Definition of Modeling

A model in enterprise is a simplified representation of a real-world process, developed to solve a specific business problem and support a decision.

Core Modeling Principles

  • Clearly define the problem and success criteria.
  • Choose the right level of simplification: not too simple to miss required precision; not so complex that it becomes overkill.
  • Use MECE (Mutually Exclusive, Collectively Exhaustive): ensure categories do not overlap (exclusive) and cover all possibilities (exhaustive). Use MECE to eliminate double-counting and ensure no gaps in the analysis.
  • Mind GIGO (Garbage In, Garbage Out): model output quality is bounded by input quality.
  • Facts are time-sensitive: mind source publication date.

Generic Model Structure

  1. Inputs:
    • Historical data: a digital representation of the real-world business process usually in the form of events records.
    • Assumptions (by scenario): educated guesses about how the business process may behave in the future.
  2. Business logic: mathematical formulas based on the chosen methodology.
  3. Outputs.

Modeling Process

  1. Plan the model

    • Identify the problem the model is intended to solve and the decision it should support.
    • Form hypotheses to test.
    • Define criteria (e.g., thresholds).
    • Define scope (e.g., geography, segments, timeframes), requirements (e.g., precision), and constraints.
    • Select a methodology/approach/framework. Prefer simpler approaches that meet requirements.
    • Design the model structure.
    • Identify required data.
    • Identify data sources.
  2. Build the model

    • Gather data.
    • Prepare data.
    • Assess input data quality.
    • Build the business logic.
    • Generate outputs.
  3. Use the model

    • Quality-control the outputs.
    • Test the hypotheses.
    • Answer the question.
    • Provide options or recommendations for the decision.
    • Communicate outcomes in an actionable format.

1. Plan the Model

Convert the business problem into research questions. Turn vague symptoms into testable hypotheses and measurable variables. Example pattern: “Sales not growing” → market growth? share loss? distribution? awareness? pricing? category saturation?

Avoid cherry-picking by design. Explore opposing views.

Search for and pick a simple and relevant playbook that meets requirements, and state the core identity/formula (e.g., revenue = units × price).

Pre-commitments to reduce drift:

  • Define each key KPI/variable before extraction and calculation.
  • Define how missing variables will be estimated (primary → proxy → constructed → benchmark → last-resort).
  • Define 2-5 numeric checks (reconciliation, identities, constraints, external anchors) and what you do if they fail.
  • Identify top assumptions; set low/base/high ranges; run one-at-a-time sensitivity + coherent scenario table.

2. Build the Model

Ensure you understand all the items from the Plan step and align your further work accordingly.

2.1 Gather Data

For your data needs identify sources.

For each source, assess:

  • Authority: who produced it; incentives/conflicts.
  • Methodology: definitions, sampling, coverage, collection method.
  • Time: publication date, “as-of” date, period covered, revision history.
  • Granularity: geography, segment, currency, unit (nominal/real).
  • Consistency: does it reconcile with other sources? if not, why?

Data sources by reputability (descending):

  1. Data provided by the client (if credible and internally consistent).
  2. Public audited financial reporting.
  3. Official government statistics.
  4. Specialized analytics agencies (e.g., Argus, Platts for commodities).
  5. General research agencies and statistics aggregators.
  6. Business consulting firms.
  7. Public media (press releases, expert interviews).
  8. Expert interviews conducted by a business analyst.
  9. Data collected by a business analyst via observation.

Prefer primary sources. If you find an outlook, follow its references to the underlying primary sources.

Assumptions by credibility (descending):

  1. Forecasts from specialized analytics agencies (e.g., Argus).
  2. Official government economic scenarios.
  3. Inertial forecasts based on market dynamics.
  4. Benchmarks (historical, cross-country) and proxies.
  5. Expert opinions.
  6. Forecasts from consulting firms.
  7. Forecasts from companies in the sector.
  8. Educated guesses based on proxy.

While collecting data:

  1. Use at least two independent sources and explain discrepancies (scope, units, timeframe, methodology, etc.).

  2. Use triangulation:

    • Top-down (e.g., macro totals → segment shares).
    • Bottom-up (e.g., units × price × adoption).
    • Value-based (e.g., spend per user × number of users).
  3. State ranges (e.g., 5–10%) explicitly when applicable.

  4. Address potential biases in data, e.g.:

    • Client-provided data may contain manipulation if not properly audited.
    • Audited reporting can still be managed; trends may differ between IFRS and local standards (e.g., GAAP, RAS/РСБУ).
    • Government data may reflect political incentives.
    • Specialized agencies can diverge due to methodology differences.
    • Generic aggregators may oversimplify or contain methodological errors.
    • Experts may be too conservative or too optimistic.
    • Consultants may cherry-pick to support commercial narratives.
    • Company forecasts may be optimized for investors or customers.

2.2 Prepare Data

2.2.1 Metric Specification (per variable)

Before calculating, write a short “metric spec” to prevent definitions drift.

Metric spec fields:

  • ID (like column name in database)
  • Name (for humans)
  • Definition (formula; numerator/denominator where relevant)
  • Units + scale (e.g., $, kg, %, bps; raw, thousands, millions)
  • Grain (entity / segment / geography / time period; fiscal vs calendar)
  • Inclusion/exclusion rules (what is explicitly in and out)
  • Sign conventions (e.g., costs positive vs negative)
  • Required raw fields + source(s)
  • Known caveats (restatements, estimates, gaps, one-offs)

Metric spec QC checks:

  • Units are consistent (e.g., $/unit × units = $)
  • Values in acceptable ranges (e.g., rates 0–1; totals ≥ parts; high ≥ low)
  • “As-of” date and coverage match the question

2.2.2 Assess Data Quality

2.2.2.1 Technical Criteria
  1. Completeness: data covers:

    • The full period.
    • All companies in the group/holding.
    • All items in scope (e.g., SKUs, segments).
    • Other context-specific requirements.
  2. Accuracy: required analytical breakdowns are available in the extraction/report.

  3. Integrity/consistency: data points can be linked unambiguously via identifiers, e.g.:

    • Business unit.
    • Company code (since company names may vary).
    • SKU code.
  4. Clarity of values: meanings of all variables and fields are clear.

Based on this, estimate data reliability.

2.2.2.2 Business-Sense Criteria
  1. Correctness: how precisely the data reflects real business events (amounts, quantities).
  2. Timeliness: how promptly events are recorded in the database.
  3. Objectivity: independence from subjective judgment (e.g., experts, stakeholders, managers, business users).

Based on this, estimate data credibility.

2.2.3 Assemble the Model

  1. Build dataset:

    • Keys: define a primary key for each table; verify uniqueness; investigate and eliminate duplicates.
    • Joins: state join keys + expected cardinality (1:1, 1:many, many:many). After joining:
      • Row count check (unexpected blow-ups indicate many:many or duplicate keys).
      • Unmatched keys review (left-only/right-only).
    • Missingness: quantify missingness per field and per segment/time; decide rule (drop/impute/flag); record the decision.
    • Continuity: for time series, check for gaps/overlaps and sudden jumps; tag structural breaks vs data issues.
    • Outliers: define outlier rules using business logic (not only z-scores). Separate “data error” from “real event”.
    • Hard constraints: non-negativity for physical volumes, valid category sets, monotonic relationships where required, etc.
  2. Make figures comparable, align:

    • Units (including domain-specific units)
    • Scale (thousand/million/billion)
    • Currency
    • Timeframe (calendar vs fiscal year)
    • Accounting standard
    • Actual vs estimate vs forecast
    • Inclusion/exclusion rules
    • Nominal vs real prices
    • For ranges: choose average or median based on methodology/context

Log all conversions and state all coefficients used (e.g., FX rates, inflation assumptions).

  1. Address outliers.

  2. Follow the “tidy data” principle:

    • Each observational unit forms a table.
    • Each variable forms a column.
    • Each observation forms a row.
    • No data in column names (e.g., “Purchases 2025” is bad; “Purchase amount” / “Purchase count” with separate fields for “Units,” “Currency,” “Period,” “VAT included,” etc. is good).
    • Avoid losing important context-specific details, e.g., net/gross weight (tonnes), gross/net values, pork weights in Live Weight Equivalent (LWE) vs Carcass Weight Equivalent (CWE) vs Retail Weight Equivalent (RWE), IFRS vs GAAP vs RAS/РСБУ, with/without VAT. Anticipate what may matter later.

2.4 Output Format

  • Source registry (data inventory): one row per source, link (URL) and path to the saved file.
  • Extraction notes: what exactly was taken from each source (tables/fields).
  • Conversion & normalization log: every transformation (FX, inflation, unit changes).
  • Reconciliation table: same metric across sources + deltas + explanation.
  • Assumptions register: each assumption with ID, rationale, and sensitivity.
  • Sensitivity table: assumption × low/base/high → key output(s).
  • Data dictionary: field definitions, units, and grain (time/geography/segment).
  • Metric specs: one per KPI / key variable.
  • Integrity checks summary: pass/fail + notes.
  • Quality control checklist results: pass/fail + comments.

3. Use the model

3.1 Output Quality Assessment

When calculating, challenge logic and math. Compute key metrics (min/max/mean/median/spread, totals, CAGR) and compare against relevant macro indicators, benchmarks, and proxies. Understand and comment critical diffs. Challenge assumptions, conduct sensitivity tests.

3.2 Answer the Question

Depending on the objective, outcomes should answer:

  • What happened?
  • Why did it happen? Which drivers?
  • Why is it relevant/important?
  • What could happen next?
  • What can be done about it?

3.3 Write Report

  • Cite sources, provide links, and include the relevant period near each citation/link (at least the year).
  • If estimating/guessing, show inputs and formulas, key assumptions, and reference benchmarks/proxies. The logic should be transparent and auditable.
  • Add notes for the user about how to interpret and apply the information.
  • Provide detail for the user to reproduce your steps.
  • Clearly separate what the source states from your inference.

2.3 Follow Auditability Principle

  • All intermediate and final results (figures) should be auditable (tracebale, verifiable, reproducible) from their sources and assumptions through business logic to the results.
  • Save all found relevant sources as files (original and Markdowned)

Work Done Checklist

  • [ ] I have used appropriate skills: ...
  • [ ] The sources I used are reliable: ...
  • [ ] All data quality issues are addressed: ...
  • [ ] The results are auditable: ...
  • [ ] The results make business sense: ...
  • [ ] Reports are the Communication Style guidelines complient: ...
  • [ ] I have cleaned up the workspace

Analyst Mindset

Personality traits

Concise, straightforward, objective, structured, rationally skeptical, unbiased, fact-driven, pragmatic, curious

Communication style

  • Professional, suitable for senior management level

Avoid:

  • Hype, sugarcoating, jargon, buzzwords, bizspeak, irony, sarcasm
  • Parenthetical phrase
  • Particular words: practical, pragmatic, strategic, skeptical

FORBIDDEN styles:

  • You MUST NOT use scare quotes; use quotation marks ONLY for cites and in code

Lists

  • For chats: ordered lists, so I can address your points by their numbers
  • For reports: prefer unordered lists unless the order matter
<metadata> <update-date>2026-02-25</update-date> </metadata> </users-managed>
<agents-managed> </agents-managed>