Skip to main content
Open Source AI Gateway

The AI Gateway Your Infrastructure Deserves

Route, secure, and optimize all your LLM traffic through a single, unified gateway. 50+ providers, enterprise-grade guardrails, and full cost control — powered by Otoroshi.

Otoroshi LLM Extension
50+LLM Providers
20+Guardrail Rules
17+Workflow Functions
16+Embedding Providers

Everything You Need to Manage LLMs in Production

From routing to guardrails, from caching to cost control — a complete toolkit for production AI infrastructure.

🔗
Unified API

One OpenAI-compatible API to rule them all. Connect any LLM provider without changing a single line of client code.

⚖️
Load Balancing & Failover

Distribute workloads across providers and automatically switch during outages. Zero downtime, always.

🛡️
20+ Guardrails

Block prompt injections, PII leakage, toxic content, and more. Validate every request and response with built-in rules.

💰
Cost Management

Real-time cost tracking, budget limits per provider, token quotas per consumer. Never get an unexpected bill again.

🧠
Semantic Caching

Embedding-based similarity matching reduces costs and latency. Exact-match and semantic caches working together.

🔐
Enterprise Security

Granular API keys, role-based access control, secret vault integration. LLM access as secure as your APIs.

🤖
AI Agents & MCP

Build agentic workflows with tool calling, agent handoffs, persistent memory, and Model Context Protocol support.

📊
Full Observability

Audit every request, track environmental impact, export metrics to your favorite dashboards and SIEM tools.

🎨
Multi-Modal

Text, images, audio, video — generate and process any modality through the same gateway with dedicated APIs.

Why Otoroshi LLM Extension?

Not just another proxy. A full-featured AI gateway built for teams that take production seriously.

🏗️
Built on Otoroshi

Leverage a battle-tested, cloud-native API gateway. Get mTLS, service mesh, plugins, and admin UI out of the box. Your LLM gateway inherits enterprise-grade infrastructure.

🔄
Hot-Reload Everything

Change providers, update guardrails, adjust budgets — all without restart or downtime. Configuration changes apply instantly across your entire LLM fleet.

🧩
Extensible by Design

Custom guardrails via WASM, webhooks, or LLM-based validation. Function calling through QuickJS, HTTP, or Otoroshi workflows. Extend everything without forking.

🌍
Sovereign & Open Source

Run on your infrastructure, keep your data where it belongs. Support for EU/French providers like OVH, Scaleway, Cloud Temple. Fully open source under Apache 2.0.

50+ LLM Providers, One API

Connect to any major LLM provider through a single, consistent OpenAI-compatible interface. Switch providers in seconds.

OpenAIAnthropicGoogle GeminiMistralAzure OpenAIOllamaGroqCohereDeepseekX.aiCloudflareOVH AIScalewayHugging FaceMeta LlamaOpenRouterPerplexityTogether AIFireworks AIDeepInfraSambaNovaLambda Labs+30 more

Built for Real-World Scenarios

From startups to enterprises, deploy AI with confidence.

🏢
Enterprise AI Gateway

Centralize all LLM access for your organization with security, quotas, and compliance built in.

🔀
Multi-Provider Strategy

Avoid vendor lock-in by routing requests to the best provider for each use case with automatic failover.

🛡️
Compliance & Governance

Enforce content policies, audit every interaction, track costs, and meet regulatory requirements.

Performance at Scale

Cache responses, balance load, rate-limit consumers, and keep latency low across millions of requests.

Ready to Take Control of Your AI Infrastructure?

Get started in minutes. Open source, free forever.