DeepEval
by Confident AI · open-source · Last verified 2026-03-17
Open-source evaluation framework for testing LLM applications like unit tests. Provides 14+ metrics including hallucination, answer relevancy, and bias detection with Pytest-like syntax and CI/CD integration.
https://docs.confident-ai.com ↗C+
C+—Average
Adoption: C+Quality: AFreshness: A+Citations: C+Engagement: F
Specifications
- License
- Apache-2.0
- Pricing
- open-source
- Capabilities
- llm-unit-testing, hallucination-detection, bias-detection, ci-cd-integration, benchmark-evaluation
- Integrations
- langchain, llamaindex, openai, anthropic
- Use Cases
- llm-testing, regression-testing, ci-cd-evaluation, quality-assurance
- API Available
- Yes
- SDK Languages
- python
- Deployment
- self-hosted, confident-ai-cloud
- Rate Limits
- N/A (open-source)
- Data Privacy
- Self-hosted by default; Confident AI Cloud for collaboration
- Tags
- evaluation, llm-testing, unit-testing, ci-cd
- Added
- 2026-03-17
- Completeness
- 100%
Index Score
50.9Adoption
55
Quality
82
Freshness
90
Citations
50
Engagement
0