GEval evaluation metrics support
validmind-library
26.01.02
enhancement
ValidMind now supports GEval LLM evaluation metrics from DeepEval, allowing you to define custom criteria for evaluations that are logged directly in your documentation via ValidMind tests.
GEval enables “LLM-as-judge” metrics, where a language model scores outputs according to explicit, user-defined standards, eliminating dependence on string matching or reference responses.