Your best companion for upgrading to Unity Catalog. It helps you to upgrade all Databricks workspace assets: Legacy Table ACLs, Entitlements, AWS instance profiles, Clusters, Cluster policies, Instance Pools, Databricks SQL warehouses, Delta Live Tables, Jobs, MLflow experiments, MLflow registry, SQL Dashboards & Queries, SQL Alerts, Token and Password usage permissions that are set on the workspace level, Secret scopes, Notebooks, Directories, Repos, Files.
See contributing instructions to help improve this project.
UCX will guide you, the Databricks customer, through the process of upgrading your account, groups, workspaces, jobs etc. to Unity Catalog.
- The upgrade process will first install code, libraries, and workflows into your workspace.
- After installation, you will run a series of workflows and examine the output.
UCX leverages Databricks Lakehouse platform to upgrade itself. The upgrade process includes creating jobs, notebooks, and deploying code and configuration files. The install.sh
guides you through the installation.
By running the installation you install the assessment job and several upgrade jobs. The assessment and upgrade jobs are outlined in the custom-generated README.py that is created by the installer and displayed to you by the install.sh
. See interactive installation tutorial here.
The custom-generated README.py
, config.yaml
, and other assets are placed into your Databricks workspace home folder, into a subfolder named .ucx
. See interactive tutorial.
Once the custom Databricks jobs are installed, begin by triggering the assessment job. The assessment job can be found under your workflows or via the active link in the README.py. Once the assessment job is complete, you can review the results in the custom-generated Databricks dashboard (linked to by the custom README.py found in the workspace folder created for you).
You will need an account, unity catalog, and workspace administrative authority to complete the upgrade process. To run the installer, you will need to setup databricks-cli
and a credential, following these instructions. Additionally, the interim metadata and config data being processed by UCX will be stored into a Hive Metastore database schema generated at install time.
For questions, troubleshooting or bug fixes, please see your Databricks account team or submit an issue to the Databricks UCX github repo
- Get trained on UC [free instructor-led training 2x week] [full training schedule]
- [AWS] [Azure] [GCP] Account level Identity Setup
- [AWS] [Azure] [GCP] Unity Catalog Metastore Created (per region)
As a customer, download the latest release from github onto your laptop/desktop machine. Unzip or untar the release.
The ./install.sh
script will guide you through installation process.
Make sure you have Python 3.10 (or greater) installed on your workstation, and you've configured authentication for
the Databricks Workspace.
The easiest way to install and authenticate is through a Databricks configuration profile:
export DATABRICKS_CONFIG_PROFILE=ABC
./install.sh
You can also specify environment variables in a more direct way, like in this example for installing on an Azure Databricks Workspace using the Azure CLI authentication:
az login
export DATABRICKS_HOST=https://adb-123....azuredatabricks.net/
./install.sh
Please follow the instructions in ./install.sh
, which will deploy UCX to your workspace and open a notebook with the description of all jobs to trigger. The journey starts with assessment.
If your Databricks Workspace relies on an external Hive Metastore (such as glue), make sure to read the External HMS Document.
Please note that all projects in the /databrickslabs github account are provided for your exploration only, and are not formally supported by Databricks with Service Level Agreements (SLAs). They are provided AS-IS and we do not make any guarantees of any kind. Please do not submit a support ticket relating to any issues arising from the use of these projects.
Any issues discovered through the use of this project should be filed as GitHub Issues on the Repo. They will be reviewed as time permits, but there are no formal SLAs for support.