The AI Toolkit for TypeScript
One endpoint, all your models
An agent that knows your stack
Helping teams ship 6× faster
Fast, scalable, and reliable
Servers, in serverless form
Trace every step
Scalable bot protection
Invisible CAPTCHA
DDoS Protection, Firewall
Granular, custom protection
Trusted by the best teams
The latest posts and changes
See what shipped
Read the latest news
Join us at an event
The native Next.js platform
The progressive web framework
The web's efficient UI framework
Speed with Enterprise scale
Learn the ins and outs of Vercel
Extend and automate workflows
Jumpstart app development
Find help quickly
Get help from solution partners
Deploy at the speed of AI
Power storefronts that convert
Launch campaigns fast
Scale apps with one codebase
Ship features, not infrastructure
Automate away repetition
Deploy for every idea
AI Gateway
English-focused text embedding model optimized for code and English language tasks.
1import { embed } from 'ai';2 3const result = await embed({4 model: 'google/text-embedding-005',5 value: 'Sunny day at the beach',6})
Vertex AI
Input Tokens $0.03/M
Gemini 2.0 Flash
Gemini 2.0 Flash delivers next-gen features and improved capabilities, including superior speed, built-in tool use, multimodal generation, and a 1M token context window.
Gemini 2.0 Flash Lite
Gemini 2.5 Flash
Gemini 2.5 Flash is a thinking model that offers great, well-rounded capabilities. It is designed to offer a balance between price and performance with multimodal support and a 1M token context window.
Gemini 2.5 Flash Image (Code name: Nano Banana)
Gemini 2.5 Flash Image is our first fully hybrid reasoning model, letting developers turn thinking on or off and set thinking budgets to balance quality, cost, and latency. Upgraded for rapid creative workflows, it can generate interleaved text and images and supports conversational, multi‑turn image editing in natural language. It’s also locale‑aware, enabling culturally and linguistically appropriate image generation for audiences worldwide.
Gemini 2.5 Flash Lite
Gemini 2.5 Flash-Lite is a balanced, low-latency model with configurable thinking budgets and tool connectivity (e.g., Google Search grounding and code execution). It supports multimodal input and offers a 1M-token context window.
Gemini 2.5 Flash Lite Preview 09-2025
Gemini 2.5 Flash Preview 09-2025
Gemini 2.5 Pro
Gemini 2.5 Pro is our most advanced reasoning Gemini model, capable of solving complex problems. It features a 2M token context window and supports multimodal inputs including text, images, audio, video, and PDF documents.
Gemini Embedding 001
State-of-the-art embedding model with excellent performance across English, multilingual and code tasks.
Gemma 2 9B IT
9 billion parameter open source model by Google fine-tuned for chat purposes. Served by Groq with their custom Language Processing Units (LPUs) hardware to provide fast and efficient inference.
Text Multilingual Embedding 002
Multilingual text embedding model optimized for cross-lingual tasks across many languages.