Skip to main content
ToolAI Tools & APIsv1.3

DeepEval

by Confident AI · open-source · Last verified 2026-03-17

Open-source evaluation framework for testing LLM applications like unit tests. Provides 14+ metrics including hallucination, answer relevancy, and bias detection with Pytest-like syntax and CI/CD integration.

https://docs.confident-ai.com
C+
C+Average
Adoption: C+Quality: AFreshness: A+Citations: C+Engagement: F

Specifications

License
Apache-2.0
Pricing
open-source
Capabilities
llm-unit-testing, hallucination-detection, bias-detection, ci-cd-integration, benchmark-evaluation
Integrations
langchain, llamaindex, openai, anthropic
Use Cases
llm-testing, regression-testing, ci-cd-evaluation, quality-assurance
API Available
Yes
SDK Languages
python
Deployment
self-hosted, confident-ai-cloud
Rate Limits
N/A (open-source)
Data Privacy
Self-hosted by default; Confident AI Cloud for collaboration
Tags
evaluation, llm-testing, unit-testing, ci-cd
Added
2026-03-17
Completeness
100%

Index Score

50.9
Adoption
55
Quality
82
Freshness
90
Citations
50
Engagement
0

Explore the full AI ecosystem on Agents as a Service