Back to Skills

anthropic-evaluations

verified

This skill should be used when the user asks to "create evals", "evaluate an agent", "build evaluation suite", or mentions agent testing, graders, or benchmarks. Also suggest when building coding agents, conversational agents, or research agents that need quality assurance.

View on GitHub

Marketplace

claude-toolkit

dwmkerr/claude-toolkit

Plugin

toolkit

Repository

dwmkerr/claude-toolkit
4stars

plugins/toolkit/skills/anthropic-evaluations/SKILL.md

Last Verified

January 20, 2026

Install Skill

Select agents to install to:

Scope:
npx add-skill https://github.com/dwmkerr/claude-toolkit/blob/main/plugins/toolkit/skills/anthropic-evaluations/SKILL.md -a claude-code --skill anthropic-evaluations

Installation paths:

Claude
.claude/skills/anthropic-evaluations/
Powered by add-skill CLI

Instructions

# Anthropic Evaluations

Build rigorous evaluations for AI agents using Anthropic's proven patterns.

## Quick Reference

You MUST read the reference files for detailed guidance:

- [Grader Types](./references/grader-types.md) - Code-based, model-based, human graders
- [Agent Type Patterns](./references/agent-type-patterns.md) - Coding, conversational, research, computer use
- [Roadmap](./references/roadmap.md) - Steps 0-8 for building evals from scratch
- [Frameworks](./references/frameworks.md) - Harbor, Promptfoo, Braintrust, etc.

**YAML Templates:**
- [coding-agent-eval.yaml](./references/coding-agent-eval.yaml) - Coding agent template
- [conversational-agent-eval.yaml](./references/conversational-agent-eval.yaml) - Support agent template

**Annotated Examples:**
- [Example: Coding Agent](./references/example-coding-agent.md) - Auth bypass fix walkthrough
- [Example: Conversational](./references/example-conversational.md) - Refund handling walkthrough

## Core Definitions

| Term | Definition |
|------|------------|
| **Task** | Single test with defined inputs and success criteria |
| **Trial** | One attempt at a task (run multiple for consistency) |
| **Grader** | Logic that scores agent performance; tasks can have multiple |
| **Transcript** | Complete record of a trial (outputs, tool calls, reasoning) |
| **Outcome** | Final state in environment (not just what agent said) |
| **Evaluation harness** | Infrastructure that runs evals end-to-end |
| **Agent harness** | System enabling model to act as agent (scaffold) |
| **Evaluation suite** | Collection of tasks measuring specific capabilities |

## Grader Types (Quick Reference)

| Type | Methods | Best For |
|------|---------|----------|
| **Code-based** | String match, unit tests, static analysis, state checks | Fast, cheap, objective verification |
| **Model-based** | Rubric scoring, assertions, pairwise comparison | Nuanced, open-ended tasks |
| **Human** | SME review, A/B testing, spot-check sampling | Go

Validation Details

Front Matter
Required Fields
Valid Name Format
Valid Description
Has Sections
Allowed Tools
Instruction Length:
3157 chars