OxiGateOxiGate

Zero-overhead LLM Gateway.
FinOps Infrastructure for AI.

Engineered in Rust. Hard and soft budget caps. Per-identity, team, and tag spend tracking. Fallbacks and load balancing across 100+ models. Extensible via Rust and Python plugins. All at sub-100µs overhead — because governance shouldn't slow you down.

View on GitHub

Enterprise-Grade FinOps Infrastructure

Everything you need to manage LLM costs and reliability at scale.

Built in Rust

Sub-100µs overhead, memory-safe, no GC pauses. Performance that never sleeps.

Financial Correctness

Integer nano-USD accounting, zero drift over billions of requests. Every cent tracked.

Budget & Rate Limits

Per-user, per-team, per-tag enforcement with soft warnings and hard caps.

OpenAI Compatible

Drop-in replacement supporting OpenAI, Anthropic, Gemini, AWS Bedrock.

Smart Fallbacks

Automatic retry, weighted load balancing, provider failover. Zero downtime.

Deep Observability

Structured JSON logs, Prometheus metrics, OpenTelemetry export. See everything.

Powerful Plugin System

Extend with Rust or Python plugins. Add custom providers, middleware, or FinOps logic without forking.

Built in Rust for speed. Built for control.

How It Works

Simple integration, powerful results. One gateway to rule all your LLM providers.

Your App
OxiGate
Rust
OpenAI
Anthropic
Gemini
Open AI Compatible

Simple, Transparent Pricing

Start free with our Free Tier gateway. Scale with support when you need it.

Free Tier

$0

Free to use

  • Full gateway functionality
  • Unlimited requests
  • All provider integrations
  • Community support

Pro

Get in touch

For growing teams

  • Everything in Free Tier
  • Priority support
  • Advanced analytics
  • Custom integrations

Enterprise

Get in touch

For large organizations

  • Everything in Pro
  • Dedicated support
  • SLA guarantees
  • On-premise deployment