Skip to main content

SAP Databricks user guide

The articles in this user guide explain how to get started with Unity Catalog, Databricks SQL, notebooks, and machine learning in SAP Databricks.

Data governance with Unity Catalog

Unity Catalog is a fine-grained governance solution for data and AI on the Databricks platform. It helps simplify security and governance of your data and AI assets. You can browse your data assets through the Catalog Explorer, receive Delta Shares, and access third party data through external locations.

To get started, see Data governance with Unity Catalog.

Business intelligence with Databricks SQL

In SAP Databricks, users can use the SQL editor to connect to serverless SQL warehouses set up by administrators to efficiently run SQL queries against data in your lakehouse. The SAP Databricks UI includes a SQL editor that you can use to author queries, browse available data, and create visualizations.

To get started, see Business intelligence with Databricks SQL.

Data analysis with notebooks

Notebooks are a common tool in data science and machine learning for developing code and presenting results. In SAP Databricks, notebooks are the primary tool for writing code, creating data science workflows, and collaborating with colleagues. Databricks notebooks provide real-time coauthoring in multiple languages, automatic versioning, and built-in data visualizations.

To get started, see Data analysis with notebooks.

Machine learning

SAP Databricks provides various products to help you implement your end-to-end workflows to build and deploy AI and ML systems. Mosaic AI unifies the AI lifecycle from data collection and preparation, to model development and LLMOps, to serving and monitoring. With Mosaic AI, a single platform serves every step of ML development and deployment, from raw data to inference tables that save every request and response for a served model. Data scientists, data engineers, ML engineers and DevOps can do their jobs using the same set of tools and a single source of truth for the data.

To get started, see Machine learning with Mosaic AI.

Serverless compute

You can quickly connect to serverless compute to run your queries and code. With serverless compute, SAP Databricks automatically allocates and manages the necessary compute resources. This enables you to focus on writing code and analyzing data, without worrying about compute management or resource utilization.

Serverless compute offers the following benefits:

  • Cloud resources are managed by Databricks, reducing management overhead and providing instant compute to enhance user productivity.
  • Rapid start-up and scaling times for serverless compute resources minimize idle time and ensure you only pay for the compute you use.
  • Because capacity handling, security, patching, and upgrades are managed automatically, you can worry less about reliability, security policies, and capacity shortages.

SAP Databricks currently offers the following types of serverless compute: