Use the UCX utilities to upgrade your workspace to Unity Catalog
This article introduces UCX, a Databricks Labs project that provides tools to help you upgrade your non-Unity-Catalog workspace to Unity Catalog.
Note
UCX, like all projects in the databrickslabs GitHub account, is provided for your exploration only, and is not formally supported by Databricks with service-level agreements (SLAs). It is provided as-is. We make no guarantees of any kind. Do not submit a Databricks support ticket relating to issues that arise from the use of this project. Instead, file a GitHub issue. Issues will be reviewed as time permits, but there are no formal SLAs for support.
The UCX project provides the following migration tools and workflows:
Assessment workflow to help you plan your migration.
Group migration workflow to help you upgrade group membership from your workspace to your Databricks account and migrate permissions to the new account-level groups.
Table migration workflow to help you upgrade tables that are registered in your workspace’s Hive metastore to the Unity Catalog metastore. This workflow also helps you migrate storage locations and the credentials required to access them.
This diagram shows the overall migration flow, identifying migration workflows and utilities by name:
Note
The code migration workflow that is depicted in the diagram remains under development and is not yet available.
Before you begin
Before you can install UCX and run the UCX workflows, your environment must meet the following requirements.
Packages installed on the computer where you run UCX:
Databricks CLI v0.213 or above. See Install or update the Databricks CLI.
You must have a Databricks configuration file with configuration profiles for both the workspace and the Databricks account.
Python 3.10 or above.
If you want to run the UCX workflow that identifies storage locations used by Hive tables in your workspace (recommended, but not required), you must have the CLI for your cloud storage provider (Azure CLI or AWS CLI) installed on the computer where you run the UCX workflows.
Network access:
Network access from the computer that runs the UCX installation to the Databricks workspace that you are migrating.
Network access to the internet from the computer that runs the UCX installation. This is required for access to pypi.org and github.com.
Network access from your Databricks workspace to pypi.org to download the
databricks-sdk
andpyyaml
packages.
Databricks roles and permissions:
Databricks account admin and workspace admin roles for the user who runs the UCX installation. You cannot run the installation as a service principal.
Other Databricks prerequisites:
A Unity Catalog metastore created for every region that hosts a workspace that you want to upgrade, with each of those Databricks workspaces attached to a Unity Catalog metastore.
To learn how to determine whether you already have a Unity Catalog metastore in the relevant workspace regions, how to create a metastore if you don’t, and how to attach a Unity Catalog metastore to a workspace, see Step 1: Confirm that your workspace is enabled for Unity Catalog in the Unity Catalog setup article. As an alternative, UCX provides a utility for assigning Unity Catalog metastores to workspaces that you can use after UCX is installed.
Attaching a Unity Catalog metastore to a workspace also enables identity federation, in which you centralize user management at the Databricks account level, which is also a prerequisite for using UCX. See Enable identity federation.
If your workspace uses an external Hive metastore (such as AWS Glue) instead of the default workspace-local Hive metastore, you must perform some prerequisite setup. See External Hive Metastore Integration in the databrickslabs/ucx repo.
A Pro or Serverless SQL warehouse running on the workspace where you run UCX workflows, required to render the report generated by the assessment workflow.
Install UCX
To install UCX, use the Databricks CLI:
databricks labs install ucx
You are prompted to select the following:
The Databricks configuration profile for the workspace that you want to upgrade. The configuration file must also include a configuration profile for the workspace’s parent Databricks account.
A name for the inventory database that will be used to store the output of the migration workflows. Typically it’s fine to select the default, which is
ucx
.A SQL warehouse to run the installation process on.
A list of workspace-local groups you want to migrate to account-level groups. If you leave this as the default (
<ALL>
), any existing account-level group whose name matches a workspace-local group will be treated as the replacement for that workspace-local group and will inherit all of its workspace permissions when you run the group migration workflow after installation.You do have the opportunity to modify the workspace-group-to-account-group mapping after you run the installer and before you run group migration. See Group Name Conflict Resolution in the UCX repo.
If you have an external Hive metastore, such as AWS Glue, you have the option to connect to it or not. See External Hive Metastore Integration in the databrickslabs/ucx repo.
Whether to open the generated README notebook.
When the installation is done, it deploys a README notebook, dashboards, databases, libraries, jobs, and other assets in your workspace.
For more information, see the installation instructions in the project readme. You can also install UCX on all of the workspaces in your Databricks account.
Open the README notebook
Every installation creates a README notebook that provides a detailed description of all workflows and tasks, with quick links to the workflows and dashboards. See Readme notebook.
Step 1. Run the assessment workflow
The assessment workflow assesses the Unity Catalog compatibility of group identities, storage locations, storage credentials, access controls, and tables in the current workspace and provides the information necessary for planning the migration to Unity Catalog. The tasks in the assessment workflow can be executed in parallel or sequentially, depending on specified dependencies. After the assessment workflow finishes, an assessment dashboard is populated with findings and common recommendations.
The output of each workflow task is stored in Delta tables in the $inventory_database
schema that you specify during installation. You can use these tables to perform further analysis and decision-making using an assessment report. You can run the assessment workflow multiple times to ensure that all incompatible entities are identified and accounted for before you start the migration process.
You can trigger the assessment workflow from the UCX-generated README notebook and the Databricks UI (Workflows > Jobs > [UCX] Assessment), or run the following Databricks CLI command:
databricks labs ucx ensure-assessment-run
For detailed instructions, see Assessment workflow.
Step 2. Run the group migration workflow
The group migration workflow upgrades workspace-local groups to account-level groups to support Unity Catalog. It ensures that the appropriate account-level groups are available in the workspace and replicates all permissions. It also removes any unnecessary groups and permissions from the workspace. The tasks in the group migration workflow depend on the output of the assessment workflow.
The output of each workflow task is stored in Delta tables in the $inventory_database
schema that you specify during installation. You can use these tables to perform further analysis and decision-making. You can run the group migration workflow multiple times to ensure that all groups are upgraded successfully and that all necessary permissions are assigned.
For information about running the group migration workflow, see your UCX-generated README notebook and Group migration workflow in the UCX readme.
Step 3. Run the table migration workflow
The table migration workflow upgrades tables from the Hive metastore to the Unity Catalog metastore. External tables in the Hive metastore are upgraded as external tables in Unity Catalog, using SYNC. Managed tables in the Hive metastore that are stored in workspace storage (also known as DBFS root) are upgraded as managed tables in Unity Catalog, using DEEP CLONE.
Hive managed tables must be in Delta or Parquet format to be upgraded. External Hive tables must be in one of the data formats listed in Work with external tables.
Run the preparatory commands
Table migration includes a number of preparatory tasks that you run before you run the table migration workflow. You perform these tasks using the following Databricks CLI commands:
The
create-table-mapping
command, which creates a CSV file that maps a target Unity Catalog catalog, schema, and table to each Hive table that will be upgraded. You should review and update the mapping file before proceeding with the migration workflow.The
create-uber-principal
command, which creates a service principal with read-only access to all storage used by the tables in this workspace. The workflow job compute resource uses this principal to upgrade the tables in the workspace. Deprovision this service principal when you are done with your upgrade.(Optional) The
principal-prefix-access
command, which identifies the storage accounts and storage access credentials used by the Hive tables in the workspace.(Optional) The
migrate-credentials
command, which creates Unity Catalog storage credentials from the storage access credentials identified byprincipal-prefix-access
.(Optional) The
migration locations
command, which creates Unity Catalog external locations from the storage locations identified by the assessment workflow, using the storage credentials created bymigrate-credentials
.(Optional) The
create-catalogs-schemas
command, which creates Unity Catalog catalogs and schemas that will hold the upgraded tables.
For details, including additional table migration workflow commands and options, see Table migration commands in the UCX readme.
Run the table migration
Once you’ve run the preparatory tasks, you can run the table migration workflow from the UCX-generated README notebook or from Workflows > Jobs in the workspace UI.
The output of each workflow task is stored in Delta tables in the $inventory_database
schema that you specify during installation. You can use these tables to perform further analysis and decision-making. You might need to run the table migration workflow multiple times to ensure that all tables are upgraded successfully.
For complete table migration instructions, see your UCX-generated README notebook and the Table Migration Workflow in the UCX readme.
Additional tools
UCX also includes:
Utilities for enabling Hive metastore federation, the Databricks integration tool that enables Unity Catalog to govern tables registered in a Hive metastore:
enable-hms-federation
create-federated-catalog
Hive metastore federation aids in migration by enabling you to run workloads on both your legacy Hive metastore and its mirror in Unity Catalog, easing the transition to Unity Catalog. For more information about using Hive metastore federation in a migration scenario, see How do you use Hive metastore federation during migration to Unity Catalog?.
Debugging tools and other utilities to help you succeed with your migration.
For more information, see your UCX-generated README notebook and the UCX project readme.
Upgrade your UCX installation
The UCX project is updated regularly. To upgrade your UCX installation to the latest version:
Verify that UCX is installed.
databricks labs installed Name Description Version ucx Unity Catalog Migration Toolkit (UCX) 0.20.0
Run the upgrade:
databricks labs upgrade ucx
Get help
For help with the UCX CLI, run:
databricks labs ucx --help
For help with a specific UCX command, run:
databricks labs ucx <command> --help
To troubleshoot issues:
Run
--debug
with any command to enable debug logs.Use the debug notebook that is generated automatically by UCX.
To file an issue or feature request, file a GitHub issue.
UCX release notes
See the changelog in the UCX GitHub repo.