advanced-evaluation

Category: Data & AI | Uploader: guanyangguanyang | Downloads: 0 | Version: v1.0(Latest)

This skill should be used when the user asks to "implement LLM-as-judge", "compare model outputs", "create evaluation rubrics", "mitigate evaluation bias", or mentions direct scoring, pairwise comparison, position bias, evaluation pipelines, or automated quality assessment.

Changelog: Source: GitHub https://github.com/guanyang/antigravity-skills

Directory Structure

Current level: tree/main/skills/advanced-evaluation/

  • 📁 references/
    • 📄 bias-mitigation.md 8.9 KB
    • 📄 evaluation-pipeline.md 2.8 KB
    • 📄 implementation-patterns.md 8.8 KB
    • 📄 metrics-guide.md 9.1 KB
  • 📁 scripts/
    • 📄 evaluation_example.py 13.4 KB
  • 📄 SKILL.md 15.9 KB

SKILL.md

Login to download/like/favorite ❤ 573 | ★ 0
Comments 0

Please login before commenting.

Loading comments...