Agent Skills: octave-compression

Specialized workflow for transforming verbose natural language into semantic OCTAVE structures. REQUIRES octave-literacy to be loaded first

UncategorizedID: elevanaltd/octave-mcp/octave-compression

Install this agent skill to your local

pnpm dlx add-skill https://github.com/elevanaltd/octave-mcp/tree/HEAD/src/octave_mcp/resources/skills/octave-compression

Skill Files

Browse the full folder contents for octave-compression.

Download Skill

Loading file tree…

src/octave_mcp/resources/skills/octave-compression/SKILL.md

Skill Metadata

Name
octave-compression
Description
Specialized workflow for transforming verbose natural language into semantic OCTAVE structures. REQUIRES octave-literacy to be loaded first

===OCTAVE_COMPRESSION=== META: TYPE::SKILL VERSION::"2.5.0" STATUS::ACTIVE PURPOSE::"Workflow for transforming prose into semantic density" REQUIRES::octave-literacy TIER::LOSSLESS SPEC_REFERENCE::octave-data-spec.oct.md V6_FEATURES::"Loss accounting system, tier metadata tracking, fidelity guarantees"

§1::COMPRESSION_MANDATE TARGET::"60-80% token reduction with 100% decision-logic fidelity" PRINCIPLE::"Semantics > Syntax Rigidity" TRUTH::"Dense ≠ Obscure. Preserve the causal chain."

§1b::COMPRESSION_TIER_SELECTION // Full tier definitions in octave-data-spec.oct.md LOSSLESS::[target:100%_fidelity,preserve:everything,drop:none] USE::[critical_reasoning,legal_documents,safety_analysis,audit_trails] CONSERVATIVE::[target:85-90%_compression,preserve:explanatory_depth,drop:redundancy] USE::[research_summaries,design_decisions,technical_analysis] LOSS::~10-15%[repetition,some_edge_cases,verbose_phrasing] AGGRESSIVE::[target:70%_compression,preserve:core_thesis∧conclusions,drop:nuance∨narrative] USE::[context_window_scarcity,quick_reference,decision_support] LOSS::~30%[explanatory_depth,execution_tradeoff_narratives,edge_case_exploration] ULTRA::[target:50%_compression,preserve:facts∧structure,drop:all_narrative] USE::[extreme_scarcity,embedding_generation,dense_reference] LOSS::~50%[almost_all_explanatory_content,some_nuance,tradeoff_reasoning]

// Note: For 60% compression with mythological atoms, see octave-ultra-mythic skill
ULTRA_MYTHIC::[target:60%_compression,preserve:soul∧constraints,method:mythological_atoms]
  REFERENCE::skills/octave-ultra-mythic[specialized_identity_compression]

TIER_METADATA::include_in_META_block[COMPRESSION_TIER,LOSS_PROFILE,NARRATIVE_DEPTH]

V6_LOSS_ACCOUNTING::[
  PRINCIPLE::"OCTAVE-MCP is a loss accounting system for LLM communication",
  TRACKING::"Every transformation must log what was preserved vs dropped",
  METADATA::"Documents carry their compression tier and loss profile",
  AUDIT::I4_TRANSFORM_AUDITABILITY[stable_ids,loss_receipts],
  FIDELITY::I1_SYNTACTIC_FIDELITY[syntax_changes_ok,semantics_preserved]
]

§2::TRANSFORMATION_WORKFLOW PHASE_1_READ::[ ANALYZE::"Understand before compressing", IDENTIFY::[Redundancy, Verbosity, Causal_Chains], MAP::"Logic flow (A leads to B)" ]

PHASE_2_EXTRACT::[ CORE_PATTERNS::"Essential decision logic", REASONING::"BECAUSE statements (preserve the 'why')", EVIDENCE::"Metrics and concrete examples", TRANSFER::"HOW-to mechanics" ]

PHASE_3_COMPRESS::[ APPLICATION::"Apply operators defined in octave-literacy", HIERARCHY::"Group related concepts under parent keys", ARRAYS::"Convert repetitive lists to [item1, item2]", MYTHOLOGY::"For fidelity, use domain labels (§3b). For 60% compression, load octave-ultra-mythic for mythological atoms" ]

PHASE_4_VALIDATE::[ FIDELITY::"Is the logic intact?", GROUNDING::"Is there at least 1 concrete example?", NAVIGABILITY::"Can a human scan it?" ]

§3::COMPRESSION_RULES RULE_1::"Preserve CAUSALITY (X→Y because Z)" RULE_2::"Drop stopwords (the, is, a, of)" RULE_3::"One example per 200 tokens of abstraction" RULE_4::"Explicit Tradeoffs (GAIN⇌LOSS or GAIN vs LOSS)" RULE_5::"Preserve CONDITIONAL QUALIFIERS (when X, if Y, unless Z) — they carry material risk info, not hedging" RULE_6::"Use mythology as DOMAIN LABELS (ARTEMIS::, CHRONOS::, DEMETER:: as key prefixes) when fidelity matters — domain labels anchor facts and prevent reconstruction drift" RULE_7::"Use mythology as PATTERN DESCRIPTORS (SISYPHEAN, ODYSSEAN) when semantic density matters — encodes trajectories and states in single terms"

§3b::CONSERVATIVE_MYTH_TECHNIQUE // Combine CONSERVATIVE compression with mythology domain labels for maximum fidelity WHEN::"Decision-relevant content where reconstruction accuracy matters more than minimum tokens" METHOD::"Use mythology terms as KEY prefixes (CHRONOS::audit_6wk) not embedded values (pressure::audit_6wk∧CHRONOS)" WHY::"Domain labels are reconstruction anchors — agents translate each labeled field separately instead of merging into compound sentences" RESULT::"11/11 decision-relevant facts preserved at 15% fewer tokens than original prose" EVIDENCE::"docs/research/compression-fidelity-round-trip-study.md" LOSSLESS_FIDELITY_AT_CONSERVATIVE_COST::true

§4::ANTI_PATTERNS AVOID::[ "Markdown inside OCTAVE blocks", "JSON/YAML syntax (no curly braces, no trailing commas)", "Deep nesting (>3 levels)", "Loss of numbers or IDs" ]

===END===