AI Gateway
AI Gateway is the Databricks solution for governing and monitoring LLM endpoints, coding agents, and model serving endpoints. Use AI Gateway to analyze usage, configure permissions, and manage capacity across providers.
AI Gateway (Beta)
The new AI Gateway experience is available in Beta. AI Gateway (Beta) is the enterprise control plane for governing LLM endpoints and coding agents with enhanced features, including a rich UI, improved observability, and expanded API coverage.
Topic | Description |
|---|---|
Learn about AI Gateway (Beta), its supported features, and how to get started. | |
Create and configure AI Gateway endpoints for your LLMs and coding agents. | |
Query AI Gateway endpoints using the OpenAI client and other supported APIs. | |
Monitor usage and costs for AI Gateway endpoints using system tables. | |
Monitor and audit requests and responses in Unity Catalog Delta tables. | |
Integrate coding agents like Cursor, Gemini CLI, Codex CLI, and Claude Code with AI Gateway. |
AI Gateway features don't incur charges during Beta.
Serving endpoints
AI Gateway features are also available for model serving endpoints, including external model endpoints, Foundation Model API endpoints, and custom model endpoints.
Topic | Description |
|---|---|
Learn about AI Gateway features for serving endpoints, including supported features and limitations. | |
Configure AI Gateway features such as usage tracking, payload logging, rate limits, and guardrails on a model serving endpoint. | |
Monitor served models using AI Gateway-enabled inference tables | Monitor served models using AI Gateway-enabled inference tables. |