Agent Skills: LLM Tuning Patterns

LLM Tuning Patterns

UncategorizedID: parcadei/continuous-claude-v3/llm-tuning-patterns

Install this agent skill to your local

pnpm dlx add-skill https://github.com/parcadei/Continuous-Claude-v3/tree/HEAD/.claude/skills/llm-tuning-patterns

Skill Files

Browse the full folder contents for llm-tuning-patterns.

Download Skill

Loading file tree…

.claude/skills/llm-tuning-patterns/SKILL.md

Skill Metadata

Name
llm-tuning-patterns
Description
LLM Tuning Patterns

LLM Tuning Patterns

Evidence-based patterns for configuring LLM parameters, based on APOLLO and Godel-Prover research.

Pattern

Different tasks require different LLM configurations. Use these evidence-based settings.

Theorem Proving / Formal Reasoning

Based on APOLLO parity analysis:

| Parameter | Value | Rationale | |-----------|-------|-----------| | max_tokens | 4096 | Proofs need space for chain-of-thought | | temperature | 0.6 | Higher creativity for tactic exploration | | top_p | 0.95 | Allow diverse proof paths |

Proof Plan Prompt

Always request a proof plan before tactics:

Given the theorem to prove:
[theorem statement]

First, write a high-level proof plan explaining your approach.
Then, suggest Lean 4 tactics to implement each step.

The proof plan (chain-of-thought) significantly improves tactic quality.

Parallel Sampling

For hard proofs, use parallel sampling:

  • Generate N=8-32 candidate proof attempts
  • Use best-of-N selection
  • Each sample at temperature 0.6-0.8

Code Generation

| Parameter | Value | Rationale | |-----------|-------|-----------| | max_tokens | 2048 | Sufficient for most functions | | temperature | 0.2-0.4 | Prefer deterministic output |

Creative / Exploration Tasks

| Parameter | Value | Rationale | |-----------|-------|-----------| | max_tokens | 4096 | Space for exploration | | temperature | 0.8-1.0 | Maximum creativity |

Anti-Patterns

  • Too low tokens for proofs: 512 tokens truncates chain-of-thought
  • Too low temperature for proofs: 0.2 misses creative tactic paths
  • No proof plan: Jumping to tactics without planning reduces success rate

Source Sessions

  • This session: APOLLO parity - increased max_tokens 512->4096, temp 0.2->0.6
  • This session: Added proof plan prompt for chain-of-thought before tactics