Project Icon

deepeval

Comprehensive Framework for Large-Language Model Evaluation

Product DescriptionDeepEval is an open-source tool that evaluates large-language models (LLMs) using metrics such as G-Eval and answer relevancy. It operates locally, supports CI/CD workflows, and offers integration with platforms like Hugging Face. DeepEval helps determine hyperparameters for optimal LLM performance and facilitates transitions between systems such as OpenAI and self-hosted Llama2.
Project Details