

LangDB is an AI gateway from the LangDB team that routes LLM traffic across providers with structured observability, semantic caching, and virtual-key management. The gateway ships as a Rust binary and Docker image and has been rebranded several times (LangDB, then vLLora) as the project evolved.
LangDB exposes OpenAI-compatible endpoints for OpenAI, Anthropic, Google, Azure, AWS Bedrock, and other OpenAI-compatible providers, with routing, retries, and fallbacks. A virtual-key system scopes usage per user or team with cost tracking. Semantic caching uses embeddings to deduplicate similar requests and return cached responses.
Structured telemetry captures every inference with token counts, latency, cost, and metadata, and the gateway includes MCP client support for calling MCP tools inside completions.
Deployments that want the gateway and observability layers combined in one tool, with semantic caching as a first-class feature. The Rust runtime keeps the operational footprint small.
The gateway is published under the Elastic License v2 (ELv2), a source-available license that permits internal use and modification but restricts offering the software as a managed service or reselling it. ELv2 is not an OSI-approved open-source license.
Docker image on Docker Hub. Newer vLLora-branded builds publish compose references as those stabilize.