Capability

Run every AI request through one secure, observable control plane

Outgate AI Gateway sits between your applications and the model providers you use.

Give your team one place to connect providers, route traffic, manage access, enforce limits, apply guardrails, control tools, and monitor usage across production AI workloads.

API Keys / IdentityWeb AppFrontendBackendAPI ServerAgentsCLI / WorkersAI GatewayGuardrailsRoutingLimitsLogsRoute + protect every request automaticallyOpenAIAnthropicSelf-hostedOllama, vLLM, etc.

Connect OpenAI, Anthropic, Ollama-compatible services, or your own custom upstream providers. Instead of wiring every application directly to every model API, your teams send traffic through a consistent gateway layer.

Each provider can use stored credentials or forward the caller's own authentication upstream. Endpoints, logging, model discovery, rate limits, token limits, and guardrail policies are all configured from the Console.

What you get

01

One gateway for every provider

Connect OpenAI, Anthropic, Ollama-compatible services, or custom upstream providers behind a consistent gateway layer.

02

Smarter routing for production AI

Failover routing for reliability, weighted routing for distribution, and Smart Router for quality, speed, and cost-aware selection.

03

Scoped endpoints with shares

Create named sub-endpoints under a provider with their own URL, key, request limits, and token budgets for teams, apps, or customers.

04

Access control built in

Gateway API keys are managed through reusable access policies. Restrict each key to all endpoints in a region or to a specific list of providers and shares.

05

Rate limits and token budgets

Hourly and daily request limits or token quotas by hour, day, or month. Prevent runaway usage, separate budgets, stay within plan limits.

06

Guardrails before the model

Detect and anonymize personal info and credentials before upstream model calls, then restore placeholders on the response path.

07

MCP tool governance

Monitor MCP tool usage, allow or deny tools individually or by server, and reduce the tool list sent to the model with Smart Tool Selection.

08

Full observability

Per-request logs with metadata, body inspection, latency, token usage, costs. Metrics on volume, errors, percentiles, cache hits, and consumption heatmaps.

Request / response transformation

Let every client speak to any LLM API

Outgate translates between client and provider API flavors in both directions. Claude Code can send Anthropic-shaped messages, Codex can use OpenAI Responses, apps can call Chat Completions, and the gateway still routes the request to the upstream you choose.

Detect Anthropic, OpenAI Chat Completions, OpenAI Responses, and Ollama-style requests automatically.
Normalize requests through one internal message shape before rewriting the provider-bound body, URL, and headers.
Map streaming and non-streaming responses back to the exact format the client expects.
Clients
Claude Code
Anthropic Messages
Outgate Gateway
Format transform layer
Upstreams
Anthropic
/v1/messages

Common use cases

Standardize AI access

Give every application one consistent way to call AI providers, instead of managing provider-specific configuration across many codebases.

Improve reliability

Route around provider failures with fallback chains and multiple upstream options.

Control spend

Apply token and request budgets per provider, share, app, team, or customer.

Protect sensitive data

Detect and anonymize PII and credentials before they reach an upstream model.

Govern agent tools

See which MCP tools agents are using and control which tools each provider is allowed to call.

Debug production issues

Trace requests, inspect responses, compare latency, monitor errors, and attribute token usage or cost to the right source.

Operate AI traffic with confidence

Connect a provider, attach a policy, and run every AI request through Outgate Gateway.

Frequently asked questions

OpenAI, Anthropic, Ollama-compatible services, and any custom upstream provider you configure. Each provider can use stored credentials or forward the caller's own auth upstream.
Named sub-endpoints under a provider, each with its own gateway URL, API key, request limits, and token budgets. Use them to give teams, apps, or customers isolated access without duplicating provider integrations.
Gateway API keys are bound to access policies. A key can call every endpoint in a region or only a specific list of providers and shares. Keys carry a gw_ prefix, are shown once, and tracked over time.
Request metadata, response status, latency, provider, model, token usage, request and response sizes, API key prefix, and correlation ID. Bodies can be expanded and inspected; sensitive headers and anonymized content remain redacted.
Yes. Providers, shares, keys, access policies, guardrails, logs, metrics, and usage views are managed within the active region, giving clearer operational boundaries when running across multiple regions.