Agent Skills: llm-evaluation

Implement comprehensive evaluation strategies for LLM applications using automated metrics, human feedback, and benchmarking. Use when testing LLM performance, measuring AI application quality, or establishing evaluation frameworks.

UncategorizedID: Microck/ordinary-claude-skills/llm-evaluation

Install this agent skill to your local

pnpm dlx add-skill https://github.com/Microck/ordinary-claude-skills/llm-evaluation

Skill Files

Browse the full folder contents for llm-evaluation.

Download Skill

Loading file tree…

Select a file to preview its contents.