

Envoy AI Gateway is an Apache 2.0 project from the Envoy Proxy community that extends Envoy with LLM-aware traffic management. It builds on Envoy Gateway and the Kubernetes Gateway API to add AI routing, token-based metrics, JWT authentication, and MCP OAuth 2.1 support on top of Envoy's filter chain.
The gateway routes LLM traffic across OpenAI, Anthropic, AWS Bedrock, Azure, Vertex, and OpenAI-compatible upstreams, using CEL expressions for dynamic routing decisions based on request content, headers, or upstream health. Token-based usage metrics are emitted to Prometheus and OpenTelemetry with per-backend cost attribution.
JWT authentication runs through Envoy's built-in JWT filter with automatic JWKS refresh against any OIDC provider's discovery endpoint. MCP support covers the OAuth 2.1 authorization server flow and CEL-driven policy decisions. Request transformation uses Envoy's filter chain — request and response payloads can be rewritten, injected, or blocked inline.
Apache 2.0 under the Envoy Proxy community governance, CNCF-adjacent.
Kubernetes-native. Installed via Helm or raw manifests; the Envoy Gateway controller reconciles Gateway API resources into Envoy configuration. A standalone non-Kubernetes mode is available but earlier in stabilization than the Kubernetes path.