Eval Relevance
Eval Relevance assesses how well a given query matches a document’s content, boosting search accuracy and information retrieval effectiveness.
Install on your platform
We auto-selected Claude Code based on this skill’s supported platforms.
Run in terminal (recommended)
claude mcp add eval-relevance npx -- -y @trustedskills/eval-relevance
Or manually add to ~/.claude/settings.json
{
"mcpServers": {
"eval-relevance": {
"command": "npx",
"args": [
"-y",
"@trustedskills/eval-relevance"
]
}
}
}Requires Claude Code (claude CLI). Run claude --version to verify your install.
About This Skill
What it does
The eval-relevance skill assesses the relevance of AI agent responses to given prompts. It provides a numerical score representing how well the response addresses the prompt's intent and requirements, allowing for quantitative evaluation of agent performance. This skill is designed to be integrated into automated testing and feedback loops for AI agents.
When to use it
- Automated Testing: Evaluate an AI agent’s responses across a suite of test prompts to identify areas needing improvement.
- A/B Testing: Compare the relevance scores of different AI agent versions or configurations.
- Prompt Engineering Evaluation: Determine how changes to prompt wording impact response quality and relevance.
- Feedback Loop Optimization: Use relevance scores to automatically adjust training data or fine-tuning parameters for an AI agent.
Key capabilities
- Relevance Scoring: Assigns a numerical score (likely on a scale) indicating the degree of relevance.
- Automated Evaluation: Designed for integration into automated workflows and testing pipelines.
- Quantitative Assessment: Provides objective, measurable data about response quality.
Example prompts
- "Evaluate the following agent response to this prompt: [prompt text] Response: [agent response]"
- "Score the relevance of this AI assistant's reply: Prompt: [prompt text], Reply: [agent response]"
- "Assess the relevance of the following conversation turn: User: [user message], Agent: [agent response]"
Tips & gotchas
The accuracy of the eval-relevance skill depends on the clarity and specificity of the prompts provided to both the agent being evaluated and to the evaluation skill itself. Ensure prompts are well-defined to obtain meaningful relevance scores.
Tags
TrustedSkills Verification
Unlike other registries that point to live repositories, TrustedSkills pins every skill to a verified commit hash. This protects you from malicious updates — what you install today is exactly what was reviewed and verified.
Security Audits
| Gen Agent Trust Hub | Pass |
| Socket | Pass |
| Snyk | Pass |
🌐 Community
Passed automated security scans.