May 2024

These features and Databricks platform improvements were released in May 2024.


Releases are staged. Your Databricks account might not be updated until a week or more after the initial release date.

New compliance and security settings APIs (Public Preview)

May 23, 2024

Databricks has introduced new APIs for the compliance security profile, enhanced security monitoring, and automatic cluster update settings in workspaces. See:

Databricks Runtime 15.2 is GA

May 22, 2024

Databricks Runtime 15.2 and Databricks Runtime 15.2 ML are now generally available.

See Databricks Runtime 15.2 and Databricks Runtime 15.2 for Machine Learning.

New Tableau connector for Delta Sharing

May 22, 2024

The new Tableau Delta Sharing Connector simplifies Tableau Desktop access to data that was shared with you using the Delta Sharing open sharing protocol. See Tableau: Read shared data.

New deep learning recommendation model examples

May 22, 2024

Databricks has published two new examples illustrating modern deep learning recommendation models, including the two-tower model and Meta’s DLRM. To learn more about deep learning recommendation models, see Train recommender models.

Bind storage credentials and external locations to specific workspaces (Public Preview)

May 22, 2024

You can now bind storage credentials and external locations to specific workspaces, preventing access to those objects from other workspaces. This feature is especially helpful if you use workspaces to isolate user data access, for example, if you have separate production and development workspaces or a dedicated workspace for handling sensitive data.

For more information, see (Optional) Assign an external location to specific workspaces and (Optional) Assign a storage credential to specific workspaces.

Git folders are GA

May 22, 2024

Git folders are now generally available. See Git integration with Databricks Git folders. If you are a user of the former “Repos” feature, see What happened to Databricks Repos?.

Pre-trained models in Unity Catalog (Public Preview)

May 21, 2024

Databricks now includes a selection of high-quality, pre-trained GenAI models in Unity Catalog. These pre-trained models allow you to access state-of-the-art AI capabilities for your inference workflows, saving you the time and expense of building your own custom models. See Pre-trained models in Unity Catalog and Marketplace.

Databricks Vector Search is GA

May 21, 2024

Databricks Vector Search is now generally available. See Databricks Vector Search.

The compliance security profile now supports AWS Graviton instance types

May 21, 2024

You can now use AWS Graviton instance types in workspaces that enable the compliance security profile or enhanced security monitoring. See Compliance security profile and Enhanced security monitoring.

Databricks Assistant autocomplete (Public Preview)

May 20, 2024

Databricks Assistant autocomplete provides AI-powered suggestions in real-time as you type in notebooks, queries, and files. To enable it, go to Settings > Developer > Experimental Features and toggle Databricks Assistant autocomplete. For details, see AI-based autocomplete.

Meta Llama 3 support in Foundation Model Training

May 20, 2024

Foundation Model Training now supports Meta Llama 3. See Foundation Model Training.

New changes to Git folder UI

May 17, 2024

You may notice some changes to the user interface for Git folder interactions. We’ve added the following:

  • When you share a Git folder, you will see a new alert in a banner that prompts you to Copy link to create Git folder. When you click the button, a URL is copied to your local clipboard, which you can send to another user. When that recipient user loads that URL in a browser, the user is taken to the workspace where they can create their own Git folder cloned from the same remote Git repository. When the recipient accesses the URL, they will see a Create Git folder dialog in the UI that is pre-populated with the values taken from your Git folder.

    Click the Copy link to Git folder button the banner to share the Git repo configuration for the folder with another user in your Databricks organization
  • Similarly, a new button, Create Git folder, appears on a new alert banner when you view a Git folder created by another user. Click this button to create your own Git folder for the same Git repository, based on the pre-populated values in the Create Git folder dialog.

    When viewing another user's Git folder, click the **Create Git folder** button in the banner to make a copy of that folder in your own workspace

Compute now uses EBS GP3 volumes for autoscaling local storage

May 15, 2024

Starting May 15, 2024, Databricks began updating new and existing compute to use Amazon EBS GP3 volumes for autoscaling local storage. This change may lead to compute launch failures for customers that have not increased GP3 EBS volume quota to match their quotas for older-generation ST1 EBS volumes.

To avoid compute launch failures, ensure that your AWS accounts have sufficient quota to use GP3 volumes. You can request quota increases from AWS if necessary.

Foundation Model Training (Public Preview)

May 13, 2024

Databricks now supports Foundation Model Training. With Foundation Model Training, you use your own data to customize a foundation model to optimize its performance for your specific application. By fine-tuning or continuing training of a foundation model, you can train your own model using significantly less data, time, and compute resources than training a model from scratch. The training data, checkpoints, and fine-tuned model all reside on the Databricks platform and are integrated with its governance and productivity tools.

For details, see Foundation Model Training.

Attribute tag values for Unity Catalog objects can now be 1000 characters long (Public Preview)

May 8, 2024

Attribute tag values in Unity Catalog can now be up to 1000 characters long. The character limit for tag keys remains 255. See Apply tags to Unity Catalog securable objects.

New Previews page

May 8, 2024

Enable and manage access to Databricks previews on the new Previews page. See Manage Databricks Previews.

Credential passthrough and Hive metastore table access controls are deprecated

May 7, 2024

Credential passthrough and Hive metastore table access controls are deprecated on Databricks Runtime 15.0 and support will be removed in an upcoming DBR version.

Upgrade to Unity Catalog to simplify the security and governance of your data by providing a central place to administer and audit data access across multiple workspaces in your account. See What is Unity Catalog?.

Databricks JDBC driver 2.6.38

May 6, 2024

We have released version 2.6.38 of the Databricks JDBC driver (download). This release adds the following new features and enhancements:

  • Native Parameterized Query support if the server uses SPARK_CLI_SERVICE_PROTOCOL_V8. The limit of the number of parameters in a query is 256 in the native query mode.

  • Data ingestion using a Unity Catalog volume support. See more about Unity Catalog volumes in Connect to cloud object storage using Unity Catalog. To use this, set UseNativeQuery to 1.

  • QueryProfile interface added to IHadoopStatement allows applications to retrieve a query’s query id. The query id can be be used to fetch the query’s metadata using Databricks REST APIs.

  • Async operations for metadata Thrift calls if the server uses SPARK_CLI_SERVICE_PROTOCOL_V9. To use this feature, set EnableAsyncModeForMetadataOperation property to 1.

  • JWT assertion support. The connector now supports JWT assertion OAuth using client credentials. To do this, set the UseJWTAssertion property to 1.

This release also resolves the following issues:

  • Jackson libraries updates. The connector now uses the following libraries for the Jackson JSON parser: jackson-annotations 2.16.0 (previously 2.15.2), jackson-core 2.16.0 (previously 2.15.2), jackson-databind-2.16.0 (previously 2.15.2)

  • The connector contains unshaded class files in META-INF directory.

Databricks Runtime 15.2 (Beta)

May 2, 2024

Databricks Runtime 15.2 and Databricks Runtime 15.2 ML are now available as Beta releases.

See Databricks Runtime 15.2 and Databricks Runtime 15.2 for Machine Learning.

Notebooks now detect and auto-complete column names for Spark Connect DataFrames

May 1, 2024

Databricks notebooks now automatically detect and display the column names in Spark Connect DataFrames and allow you to use auto-complete to select columns.