Skip to content

Microsoft AI Evaluations

The Microsoft.Extensions.AI.Evaluation library is an open-source .NET framework that automates quality checks for AI-generated responses. Deploy evaluators as CI/CD gates to catch quality regressions before production.

EvaluatorMetricWhat It Checks
RelevanceEvaluatorRelevanceResponse addresses user query/intent
TruthEvaluatorTruthFactual correctness against ground truth
CompletenessEvaluatorCompletenessAll required aspects covered
FluencyEvaluatorFluencyGrammar, vocabulary, readability
CoherenceEvaluatorCoherenceLogical flow and organization
RetrievalEvaluatorRetrievalContext retrieval effectiveness
EquivalenceEvaluatorEquivalenceSimilarity to reference output
GroundednessEvaluatorGroundednessResponse grounded in provided context
  • Automated quality gates — Catch regressions before production; integrate into CI/CD workflows
  • CI/CD-ready — Available as NuGet package: Microsoft.Extensions.AI.Evaluation.Quality
  • Customizable — Extend interfaces to build domain-specific evaluators
  • LLM-powered consistency — Scalable assessments without manual review bottlenecks

📚 Microsoft Evaluations Documentation