These tools integrates with

DeepEvalvsLangfuse

LLM evaluation framework — 14+ metrics versus OSS LLM engineering platform

Compare interactively in Explore →

Choose DeepEval when…

  • You want a pytest-style framework for LLM testing
  • Unit-test-like evals for LLM outputs fit your workflow
  • You need RAG-specific metrics like faithfulness and relevancy

Choose Langfuse when…

  • You want open-source LLM observability
  • Self-hosting your tracing stack is important
  • You need cost tracking across models and users

Side-by-side comparison

Field
DeepEval
Langfuse
Category
Prompt & Eval
LLM Infrastructure
Type
Open Source
Open Source
Free Tier
✓ Yes
✓ Yes
Pricing Plans
Cloud: $59/mo
GitHub Stars
5,500
7,000
Health
80 Active
80 Active

DeepEval

Open-source evaluation framework with 14+ metrics including faithfulness, relevancy, and hallucination detection. Integrates with CI/CD.

Langfuse

Open-source platform for tracing, evaluations, and prompt management. Self-hostable alternative to LangSmith with clean UX.

Shared Connections3 tools both integrate with

Only DeepEval (4)

LangfuseTruLensInspectGalileo

Only Langfuse (27)

Claude CodeCrewAILangGraphLangChainCursorLlamaIndexOpenHandsDifyMastraAutoGen

Explore the full AI landscape

See how DeepEval and Langfuse fit into the bigger picture — 207 tools, 452 relationships, all mapped.

Open in Explore →