advanced-evaluation
This skill should be used when the user asks to "implement LLM-as-judge", "compare model outputs", "create evaluation rubrics", "mitigate evaluation bias", or mentions direct scoring, pairwise comparison, position bias, evaluation pipelines, or automated quality assessment.
Files5 files
SKILL.mdreadonly
Select a file to preview
Install
Copy into Claude Code to install this skill.
/plugin marketplace add muratcankoylan/Agent-Skills-for-Context-Engineering