Agent Skills: llm-evaluation
Implement comprehensive evaluation strategies for LLM applications using automated metrics, human feedback, and benchmarking. Use when testing LLM performance, measuring AI application quality, or establishing evaluation frameworks.
UncategorizedID: Microck/ordinary-claude-skills/llm-evaluation
12615
Install this agent skill to your local
Skill Files
Browse the full folder contents for llm-evaluation.
Loading file tree…
Select a file to preview its contents.