Microsoft AI Evaluations
Overview
Section titled “Overview”The Microsoft.Extensions.AI.Evaluation library is an open-source .NET framework that automates quality checks for AI-generated responses. Deploy evaluators as CI/CD gates to catch quality regressions before production.
Evaluators Reference
Section titled “Evaluators Reference”| Evaluator | Metric | What It Checks |
|---|---|---|
| RelevanceEvaluator | Relevance | Response addresses user query/intent |
| TruthEvaluator | Truth | Factual correctness against ground truth |
| CompletenessEvaluator | Completeness | All required aspects covered |
| FluencyEvaluator | Fluency | Grammar, vocabulary, readability |
| CoherenceEvaluator | Coherence | Logical flow and organization |
| RetrievalEvaluator | Retrieval | Context retrieval effectiveness |
| EquivalenceEvaluator | Equivalence | Similarity to reference output |
| GroundednessEvaluator | Groundedness | Response grounded in provided context |
Why Use This
Section titled “Why Use This”- Automated quality gates — Catch regressions before production; integrate into CI/CD workflows
- CI/CD-ready — Available as NuGet package:
Microsoft.Extensions.AI.Evaluation.Quality - Customizable — Extend interfaces to build domain-specific evaluators
- LLM-powered consistency — Scalable assessments without manual review bottlenecks