Llm Evaluation
Provides LLMs guidance and assistance for building AI and machine learning applications.
Install on your platform
We auto-selected Claude Code based on this skill’s supported platforms.
Run in terminal (recommended)
claude mcp add yonatangross-llm-evaluation npx -- -y @trustedskills/yonatangross-llm-evaluation
Or manually add to ~/.claude/settings.json
{
"mcpServers": {
"yonatangross-llm-evaluation": {
"command": "npx",
"args": [
"-y",
"@trustedskills/yonatangross-llm-evaluation"
]
}
}
}Requires Claude Code (claude CLI). Run claude --version to verify your install.
About This Skill
What it does
This skill, yonatangross-llm-evaluation, provides the capability to evaluate Large Language Models (LLMs) based on provided criteria. It can assess LLM outputs for qualities like helpfulness, accuracy, and relevance to a given prompt. The evaluation process allows for nuanced feedback beyond simple pass/fail assessments, offering insights into model performance.
When to use it
- Model Selection: Compare the output of different LLMs on identical prompts to determine which best suits your needs.
- Prompt Engineering Refinement: Evaluate how changes to a prompt impact an LLM's response quality and adjust accordingly.
- Content Quality Assurance: Assess generated content for factual accuracy, clarity, and overall suitability before publishing or using it in production.
- Bias Detection: Identify potential biases present in LLM outputs by evaluating responses across diverse prompts and scenarios.
Key capabilities
- LLM output evaluation
- Criteria-based assessment
- Nuanced feedback beyond pass/fail
- Performance insights
Example prompts
- "Evaluate the following response: [LLM Response] based on helpfulness, accuracy, and clarity."
- "Assess this LLM output for bias given the prompt: [Prompt], Response: [LLM Response]"
- "Compare these two responses to the same prompt and rate them on relevance: [Response 1], [Response 2]."
Tips & gotchas
The quality of the evaluation depends heavily on the clarity and specificity of the criteria provided. Vague or poorly defined criteria will lead to less useful results.
Tags
TrustedSkills Verification
Unlike other registries that point to live repositories, TrustedSkills pins every skill to a verified commit hash. This protects you from malicious updates — what you install today is exactly what was reviewed and verified.
Security Audits
| Gen Agent Trust Hub | Pass |
| Socket | Pass |
| Snyk | Pass |
🌐 Community
Passed automated security scans.