GEval evaluation metrics support

validmind-library
26.01.02
enhancement
Published

January 7, 2026

ValidMind now supports GEval LLM evaluation metrics from DeepEval, allowing you to define custom criteria for evaluations that are logged directly in your documentation via ValidMind tests.

GEval enables “LLM-as-judge” metrics, where a language model scores outputs according to explicit, user-defined standards, eliminating dependence on string matching or reference responses.