Eval Creator CI
Install
npx skills add pskoett/pskoett-ai-skills/skills/eval-creator-ci
For interactive sessions, use:
npx skills add pskoett/pskoett-ai-skills/skills/eval-creator
Purpose
Runs the outer loop's regress-test step in CI. Executes all eval cases in .evals/, reports pass/fail results, and optionally blocks merges on regressions. Can also create new eval cases from promotion candidates flagged by learning-aggregator-ci.
The interactive eval-creator skill is designed for in-session use where the user creates evals and runs them with immediate feedback. This CI variant runs on schedule or per-PR and posts results as check annotations.
Context Limitation (Important)
CI agents do not have implementation context. They execute mechanical verification methods (grep checks, command checks, file checks, rule checks) defined in eval case files. They do not interpret results beyond pass/fail — nuanced judgment is left to human review of the posted report.
Prerequisites
- GitHub Actions enabled on the repository
ghCLI authenticated with repo accessgh-awextension installed (gh extension install github/gh-aw, v0.40.1+).evals/directory with eval cases (created byeval-creatororeval-creator-ci).evals/EVAL_INDEX.mdwith eval case index
CI Contract
Hard rules for headless execution:
- Eval execution is read-only for code — eval cases read files and run check commands but do not modify source code
- Eval case creation writes to
.evals/only — when creating new evals from promotion candidates - Headless — no interactive prompts, no approval gates
- Structured output — emit results as YAML under
eval_creator_cikey - Gate policy — can fail the check run on eval regressions (configurable)
- Single comment — post one consolidated results comment per run
Authoring Workflow (gh-aw)
- Copy
references/workflow-example.mdinto.github/workflows/eval-creator-ci.md - Customize trigger and gate policy
- Validate:
gh aw compile(add--actionlint --zizmorfor security scan) - Push to enable
Persistence and Chaining
cache-memory:stores eval run history (last-run dates, result trends) across runs. Avoids re-running evals that haven't changed.workflow_call:in create mode, triggered bylearning-aggregator-civiacall-workflow. Receives promotion candidates as workflow inputs.upload-artifact:persists eval results YAML for downstream consumption.
Workflow Rules
The CI agent follows these rules in order:
Mode: Run Evals (default)
- Read
.evals/EVAL_INDEX.mdto get the list of all eval cases - For each eval case file in
.evals/cases/: a. Read the eval case metadata and verification method b. Check preconditions — if not met, mark asskipc. Execute the verification method:grep-check: Search target files for pattern, compare to expected (found/not_found)command-check: Run the command, check exit code and/or outputfile-check: Verify file or section existsrule-check: Read target file, search for expected content d. Compare result to expected outcome e. Record pass/fail/skip
- Update
.evals/EVAL_INDEX.mdwithlast-rundate andlast-resultfor each case - Emit structured YAML under key
eval_creator_ci - Post results summary as a PR comment or check annotation
- If gate policy is enabled and any eval fails: fail the check run
Mode: Create Evals (from promotion candidates)
- Read the
learning_aggregator_ciartifact or gap report from the most recent learning-aggregator-ci run - For each promotion-ready pattern with
eval_candidate: true: a. Determine the appropriate verification method based on the pattern type b. Create the eval case file in.evals/cases/with proper frontmatter c. Add the entry to.evals/EVAL_INDEX.md - Commit the new eval cases (if running with write permissions)
- Report created evals in the output
Output Schema
eval_creator_ci:
version: "0.1.0"
source:
run_id: "<workflow run ID>"
trigger: "pull_request | schedule | workflow_dispatch"
run_date: "YYYY-MM-DD"
mode: "run | create | both"
run_results:
total: 12
passed: 10
failed: 1
skipped: 1
failures:
- id: "eval-20260301-001"
pattern_key: "harden.input_validation"
rule_summary: "Always validate external inputs"
expected: "not_found"
actual: "found"
target: "src/api/handler.ts"
recovery_action: "Add input validation to new handler endpoint"
skips:
- id: "eval-20260315-003"
reason: "Precondition not met: project does not use TypeScript"
create_results:
created: 2
cases:
- id: "eval-20260411-001"
pattern_key: "simplify.dead_code"
verification_method: "grep-check"
source_learning: "LRN-20260301-001"
- id: "eval-20260411-002"
pattern_key: "harden.authorization"
verification_method: "rule-check"
source_learning: "LRN-20260315-003"
summary:
regressions: 1
new_evals_created: 2
gate_result: "fail"
followup_required: true
Recommended Outputs
| Output | Destination | Content |
|--------|------------|---------|
| Eval results | PR comment or check annotation | Pass/fail summary with failure details |
| YAML artifact | Workflow artifact | Machine-readable eval_creator_ci payload |
| Check status | Check run | Pass or fail based on gate policy |
| New eval files | .evals/cases/ (if create mode) | Eval case markdown files |
Gate Policy
Configure blocking behavior:
| Policy | Behavior |
|--------|----------|
| strict | Any eval failure blocks the check run |
| advisory | Failures are reported but do not block |
| critical-only | Only evals from critical or high severity patterns block |
Default: advisory (report but don't block). Teams should escalate to strict once eval coverage stabilizes.
Trigger Configuration
Recommended: per-PR + weekly schedule
on:
pull_request:
types: [opened, synchronize, reopened, ready_for_review]
schedule:
- cron: '0 10 * * 1' # Monday 10am UTC (after learning-aggregator-ci)
workflow_dispatch:
Per-PR runs catch regressions before merge. Weekly runs catch drift in the harness itself. Schedule after learning-aggregator-ci so new evals from promotions are available.
Integration with Other Skills
Upstream (feeds from)
eval-creator(interactive) — creates eval cases manuallylearning-aggregator-ci— produces promotion candidates witheval_candidate: trueharness-updater(interactive) — flags eval candidates after promoting patterns
Downstream (feeds into)
- self-improvement / self-improvement-ci — regression failures become new error entries in
.learnings/ - Human review — failure report posted for team triage
- PR merge gate — can block merge on regressions (configurable)
Data Flow
learning-aggregator-ci → promotion candidates (eval_candidate: true)
↓
eval-creator-ci (create mode)
↓
.evals/cases/
↓
eval-creator-ci (run mode, per-PR)
↓
pass/fail report → PR comment + check annotation
↓
regression failures → self-improvement-ci → .learnings/
Differences from Interactive Version
| Aspect | Interactive (eval-creator) | CI (eval-creator-ci) |
|--------|------|------|
| Trigger | Manual invocation | PR events, cron schedule, workflow_dispatch |
| Eval creation | User-driven with immediate feedback | Automated from learning-aggregator-ci candidates |
| Eval execution | In-session with inline results | Headless with PR comment output |
| Human interaction | User reviews results inline | Async review via GitHub |
| Gate behavior | No blocking — informational | Configurable: advisory, critical-only, strict |
| File modification | Updates eval case metadata | Updates eval index + creates new cases (in create mode) |