1 min read


AI Gateway is now generally available, providing a single unified API to access hundreds of AI models with transparent pricing and built-in observability.
With sub-20ms latency routing across multiple inference providers, AI Gateway delivers:
Transparent pricing with no markup on tokens (including Bring Your Own Keys)
Automatic failover for higher availability
High rate limits
Detailed cost and usage analytics
You can use AI Gateway with the AI SDK or through the OpenAI-compatible endpoint. With the AI SDK, it’s just a simple model string switch.
Get started with a single API call:
import { streamText } from 'ai'
const result = streamText({ model: 'openai/gpt-5', prompt: 'How can AI Gateway not have a markup on tokens?'})
Read more about the announcement, learn more about AI Gateway, or get started now.