Mosaic AI capabilities for GenAI
Mosaic AI provides a platform for building, evaluating, deploying, and monitoring generative AI applications (GenAI apps). It brings together a suite of tools that tackle the challenges of developing enterprise-grade GenAI apps. Mosaic AI integrates with popular open source frameworks, adding enterprise-grade governance, observability, and operational tooling, collectively known as LLMOps.
This page lists major features for GenAI, organized by GenAI workflow stages.
Query GenAI
Databricks makes state-of-the-art GenAI models from top model providers readily available through Databricks-hosted Foundation Models. You can also query models from external providers and your custom models. All of these models can be queried through UI, API/SDK, and SQL interfaces. This optionality lets you use GenAI for all use cases---general chat, complex agents, automated data pipelines, interactive data analytics, and more.
Method | Features |
|---|---|
UI |
|
API and SDK |
|
SQL |
|
Build GenAI
Databricks provides a flexible set of tools for building GenAI apps, agents, tools, and models. These include UI and code-based frameworks, all of which can optimize GenAI systems based on your data. You can leverage any open-source GenAI framework and can integrate custom tools and MCP servers.
Category | Features |
|---|---|
Apps |
|
Agents |
|
Tools |
|
Models and prompts |
|
Prepare and serve data
Databricks simplifies data for AI by unifying governance of traditional data and AI workloads. With all data managed under Unity Catalog with fine-grained access controls, it is easy to adjust data engineering and AI boundaries to fit your organization. Data can be prepared for GenAI using any data engineering tools such as Lakeflow Spark Declarative Pipelines. A table in Unity Catalog can be served for GenAI, using a vector index for unstructured data or a feature table for structured data.
Type of data | Features |
|---|---|
Unstructured (text, images, etc.) |
|
Structured (tables) |
|
Deploy and serve GenAI
Databricks provides production-ready serving systems for GenAI apps, agents, and models backed by Databricks Apps and Model Serving. These scalable deployments can be used for both real-time serving and batch inference. All deployments integrate with observability and evaluation and monitoring tooling.
The framework you use to develop your GenAI system will guide the deployment path:
Development method | Deployment and serving |
|---|---|
Apps and agents can be deployed using Databricks Apps, which provides UI- and API-based deployment. | |
Custom agent code and open-source frameworks | If you develop an agent using custom code or frameworks, you can use Agent Framework to simplify deployment to Model Serving. |
Fine-tuned models | Deploy fine-tuned models to Foundation Model APIs. |
Trace, evaluate, and monitor GenAI
Databricks provides managed MLflow for GenAI observability, evaluation, and monitoring. Open-source APIs make integration and portability simple, while the managed service provides production-ready endpoints. Databricks-managed MLflow can be used for GenAI apps and agents hosted on Databricks and hosted elsewhere.
Category | Features |
|---|---|
Tracing and observability |
|
Evaluation |
|
Monitoring |
|
Human feedback |
|
LLMOps
Databricks provides a full suite of tools for GenAI operations, or "LLMOps." Unified governance of data and AI assets under Unity Catalog simplifies and secures deployment of AI across an organization. AI Gateway simplifies managing models from many GenAI model providers. MLflow and Databricks Asset Bundles provide versioning and infrastructure-as-code for implementing robust LLMOps processes.
Category | Features |
|---|---|
Data and AI asset governance | Unity Catalog provides unified governance for data and AI assets. Data assets include files, tables, vector indexes, and feature stores. AI assets include models, tools, and connections for MCP servers and other APIs. |
Model endpoint governance | AI Gateway provides central governance and monitoring for GenAI model endpoints. |
App, agent, and prompt versioning | MLflow supports app and agent versioning and provides a Prompt Registry, as well as experiment tracking. |
Infrastructure as code | MLOps Stacks, which is built on top of Databricks Assets Bundles, provides code-based management and deployment of infrastructure and workflows. |
Open source support in Mosaic AI
Mosaic AI provides full support for the rapidly growing open-source ecosystem for GenAI.
For development, you can use any open-source framework and deploy it with Agent Framework to Model Serving. Databricks services can be used by third-party GenAI tools and apps by using MCP servers or the REST API or SDKs.
For observability, evaluation, and monitoring, MLflow Tracing provides native autologging for 20+ open-source GenAI frameworks, and you can add custom tracing to any other frameworks or code. The traces can then be used with MLflow Evaluation and production monitoring. The traces follow OpenTelemetry trace specs and can be exported to third-party tools.