Docs
Confident AI
Guides
Tutorials
Github
Blog

$ the open-source LLM evaluation framework

Get StartedTry the DeepEval Platform
Delivered by
Confident AI
Unit-Testing for LLMs

Native integration with Pytest, that fits right in your workflow.

LLM-as-a-Judge Metrics

40+ research-backed metrics, including custom G-Eval and deterministic metrics.

Single and Multi-Turn Evals

Covering any use cases, any system architecture, including multi-modality.

Documentation
  • Introduction
  • Confident AI
  • Tutorials
  • Guides
Articles You Must Read
  • LLM evaluation metrics
  • LLM-as-a-judge
  • LLM testing
  • LLM chatbot evaluation
Evaluation Community
  • GitHub
  • Discord
  • Newsletter
Copyright © 2025 Confident AI Inc. Built with ❤️ and confidence.