Skip to main content

AI Gateway

AI Gateway is the Databricks solution for governing and monitoring LLM endpoints, coding agents, and model serving endpoints. Use AI Gateway to analyze usage, configure permissions, and manage capacity across providers.

AI Gateway (Beta)

The new AI Gateway experience is available in Beta. AI Gateway (Beta) is the enterprise control plane for governing LLM endpoints and coding agents with enhanced features, including a rich UI, improved observability, and expanded API coverage.

Topic

Description

AI Gateway (Beta)

Learn about AI Gateway (Beta), its supported features, and how to get started.

Configure AI Gateway endpoints

Create and configure AI Gateway endpoints for your LLMs and coding agents.

Query AI Gateway endpoints

Query AI Gateway endpoints using the OpenAI client and other supported APIs.

Monitor usage for AI Gateway endpoints

Monitor usage and costs for AI Gateway endpoints using system tables.

Monitor models using inference tables

Monitor and audit requests and responses in Unity Catalog Delta tables.

Integrate with coding agents

Integrate coding agents like Cursor, Gemini CLI, Codex CLI, and Claude Code with AI Gateway.

note

AI Gateway features don't incur charges during Beta.

Serving endpoints

AI Gateway features are also available for model serving endpoints, including external model endpoints, Foundation Model API endpoints, and custom model endpoints.

Topic

Description

AI Gateway for serving endpoints

Learn about AI Gateway features for serving endpoints, including supported features and limitations.

Configure AI Gateway on model serving endpoints

Configure AI Gateway features such as usage tracking, payload logging, rate limits, and guardrails on a model serving endpoint.

Monitor served models using AI Gateway-enabled inference tables

Monitor served models using AI Gateway-enabled inference tables.