

Langfuse is an MIT-licensed LLM engineering platform providing distributed tracing, prompt management, evaluation, and dataset tooling for teams building LLM-powered applications. The platform captures LLM usage via SDK instrumentation or gateway integration rather than proxying traffic itself.
Langfuse records each LLM call as a trace span with token counts, latency, cost attribution, and arbitrary metadata, assembling them into nested trace trees that mirror application call stacks. Prompt management supports versioning, labels, and remote prompt loading so prompt edits do not require code deploys. The evaluation framework runs offline batch evaluations and online LLM-as-judge scoring against specified datasets.
Native SDKs exist for Python, TypeScript, and Java. OpenTelemetry integration feeds traces into existing observability stacks. Direct integrations with LiteLLM, Portkey, LangChain, LlamaIndex, and other orchestration frameworks capture traces without manual instrumentation.
Teams pairing Langfuse with an LLM gateway for rich trace analysis and prompt management, or running Langfuse directly against application code that uses provider SDKs without a gateway.
Core tracing, prompt management, evaluation, dataset tooling, and OIDC SSO are MIT in the self-hostable repo. Langfuse Cloud is a hosted option. Enterprise-tier features cover advanced RBAC, audit log retention policies, SCIM provisioning, and enterprise-grade support.
Docker Compose reference stack with Postgres and ClickHouse for trace storage. Helm chart available for Kubernetes. Operational footprint is comparable to other ClickHouse-backed observability tools.