Add new deepeval
tests in library
validmind-library
2.10.0
enhancement
This update enhances the integration between ValidMind and DeepEval by introducing primary scorers within a dedicated deepeval
namespace. You can now evaluate large language model (LLM) outputs using standardized metrics. Primary scorers are essential tools for evaluating LLM outputs. We have added several DeepEval-based LLM scorers, including:
- Hallucination
- Contextual precision/recall
- Summarization
- Task completion
Additionally, a demo notebook is available to guide you through end-to-end usage. Improvements have also been made to plots, such as boxplots, and examples.