One API key, all models
Unified billing, unified observability. Switch models with a single line change — no new SDKs, no new credentials.
Built-in failovers
Automatic fallback during provider outages. Define failover chains per model to keep your app running through any disruption.
No markup, list price
Zero platform fee on model usage. You pay the exact same price as going direct to each provider — we make money on the platform, not the gateway.
Drop-in compatible
OpenAI-compatible API. Works with any SDK that supports custom base URLs.
Every modality, one endpoint
Text, code, images, embeddings — all routed through the same gateway.
Chat, completion, reasoning. 20+ text models from all major providers.
Code generation, completion, and refactoring with specialized models.
Generate and edit images from text prompts with state-of-the-art models.
Semantic search, clustering, and classification with multimodal embeddings.
Built for production
Everything you need to run AI in production, without the complexity.
Usage analytics
Track token usage, costs, latency, and error rates per model, per team, per project.
Rate limiting
Set per-team or per-project rate limits and spending caps. No surprise bills.
Automatic retries
Retry failed requests with exponential backoff. Configure fallback models for resilience.
Zero data retention
We never store prompts or completions. Requests are proxied and forgotten.
Streaming support
Full SSE streaming for all text and code models. Same stream format across all providers.
Semantic caching
Cache identical or semantically similar requests to reduce costs and latency.
No markup, no platform fee
The AI Gateway is free to use. You pay the upstream provider's list price — nothing more. We make money from the Proxifai platform, not the gateway.
Frequently asked questions
How is the gateway free?
The gateway itself has no platform fee. You pay the exact list price from each provider. We monetize through the Proxifai platform (agent compute, collaboration, etc.), and offer the gateway as infrastructure for developers using any part of our ecosystem.
Do you impose rate limits?
We don't impose our own rate limits on top of provider limits. You get the full rate limits of your upstream provider tier. You can optionally configure spending caps and per-team limits for cost control.
Do you store my prompts or completions?
No. Zero data retention by default. Requests are proxied to the provider and immediately discarded. We only store aggregate metadata (token counts, latency, error codes) for analytics — never prompt content.
Can I bring my own API keys?
Yes. You can use our pooled access (pay-as-you-go via Proxifai billing) or bring your own provider API keys. BYOK mode routes through the gateway for unified analytics and failovers while billing directly to your provider account.
How many models are supported?
We currently support 20+ models from 10 providers including Anthropic, OpenAI, Google, Meta, Mistral, DeepSeek, xAI, and Cohere. New models are added within days of launch.
Start routing requests
Get an API key and start using any model in under a minute.