Skip to main content

Databricks Unity Catalog

Connection Type

ODBC (64-bit)

Driver Configuration Requirements

The host must be a Databricks Unity Catalog cluster JDBC/ODBC Server hostname.

Supported for both AWS and Azure.

Type of Support

Read & Write, In-Database

Validated On

Databricks Cluster and SQL Warehouse, Simba Apache Spark Driver 2.6.23

Driver Details

In-Database processing requires 64-bit database drivers.

Alteryx Tools Used to Connect

Standard Workflow Processing

Input Data ToolInput Data Tool

Input Data Tool Icon

Link 

In-Database Workflow Processing

Connect In-DB Tool

Blue icon with database being plugged in.

Link 

Data Stream In Tool

Blue icon with a stream-like object flowing into a database.

Link 

Notice

Databricks Unity Catalog support is only supported using DCM.

Databricks Unity Catalog is only supported using DSN-less connections.

Writing to Databricks Unity Catalog is only supported using the In-DB tools.

We support MergeInDB for Databricks Unity Catalog, go to Write Data In-DB Tool.

Configure Input tool

The tool uses the Apache Spark ODBC DSN-less with Simba Databricks Unity Catalog connection Technology in DCM.

  1. Make sure DCM is enabled.

  2. In the Input tool, select Set up a Connection.

  3. Select the Data Sources tab.

  4. Select the Quick Connect option under Databricks Unity Catalog.

  5. The DCM Connection Manager is pre-filtered to show only Apache Spark ODBC DSN-less with Simba Databricks Unity Catalog connections.

  6. Choose an existing DCM connection or select +New to create a new connection. See below for configuring a new connection using DCM.

  7. The Choose Table or Specify Query window loads and allow the selection of tables.

Configure In-DB Connection

  1. Open the In-DB Connections Manager.

  2. Select Databricks Unity Catalog in the Data Source dropdown.

  3. Select New to create a new connection.

  4. Enter a Connection Name.

  5. On the Read tab, select Setup Connection to open the DCM connection manager for Databricks Unity Catalog. The DCM Connection Manager is pre-filtered to show only Apache Spark ODBC DSN-less with Simba Databricks Unity Catalog connections.

  6. Select an existing connection or click +New to create a new connection. See below for configuring a new connection using DCM.

  7. On the Write tab, select Setup Connection to open the DCM connection manager for the Databricks Connection. The DCM Connection Manager is pre-filtered to show only Apache Spark ODBC Bulk DSN-less with Simba Databricks Unity Catalog connections.

  8. Select an existing connection or select +New to create a new connection. See below for configuring a new connection using DCM.

  9. On the Write tab, select Setup Connection to open the DCM connection manager for the Delta Lake Connection. The DCM Connection Manager is pre-filtered to show only Delta Lake connections.

  10. Select an existing connection or select +New to create a new connection. See below for configuring a new connection using DCM.

  11. Select Apply and OK to save the connection and close the window.

  12. If the In-DB Connections Manager was accessed through the Connect In-DB tool, the Choose Table or Specify Query window loads and allow the selection of tables.

Configure Apache Spark ODBC DSN-less with Simba Databricks Unity Catalog in DCM

This connection is used for reading data from Databricks Unity Catalog.

  1. Open Data Connection Manager and navigate to Apache Spark ODBC DSN-less with Simba Databricks Unity Catalog

    • From an Input tool or the In-DB Connection Manager, DCM is pre-filtered

    • From the File Menu, go to File > Manage Connections > +New > Apache Spark > Apache Spark ODBC DSN-less with Simba Databricks Unity Catalog

  2. Enter a Data Source Name.

  3. Enter the Databricks Unity Catalog Host name.

  4. The Port is set to 443 by default. Change as needed.

  5. Enter the http path. The http path is the Databricks compute resources URL.

  6. Select Save to save the Data Source.

  7. Select +Connect Credential.

    • Select an Authentication Method.

    • To use a Personal Access Token, select Username and password as the authentication method and make the username “token”.

    • To use Azure AD, see Databricks Azure OAuth Authentication.

    • Select an Existing Credential or select Create New Credential to create a new credential and enter the Personal Access Token or the information for Azure AD.

  8. Select Link to link the credential to the Data Source and select Connect.

Configure Apache Spark ODBC Bulk DSN-less with Simba Databricks Unity Catalog in DCM

This connection is used for writing data to Databricks Unity Catalog.

  1. Open Data Connection Manager and navigate to Apache Spark ODBC Bulk DSN-less with Simba Databricks Unity Catalog.

    • From an Input tool or the In-DB Connection Manager, DCM is pre-filtered.

    • From the File Menu, go to File > Manage Connections > +New > Apache Spark > Apache Spark ODBC DSN-less with Simba Databricks Unity Catalog.

  2. Enter a Data Source Name.

  3. Enter the Databricks Unity Catalog Host name.

  4. The Port is set to 443 by default. Change as needed.

  5. Enter the http path. The http path is the Databricks compute resources URL.

  6. Enter the Catalog. This sets the catalog that is used for writing data and creating tables.

  7. Enter the Schema. This sets the schema that is used for writing data and creating tables.

  8. Select Save to save the Data Source.

  9. Select +Connect Credential to add a Credential.

    • Select an Authentication Method.

    • To use a Personal Access Token, select Username and password as the authentication method and make the username “token”.

    • To use Azure AD, see Databricks Azure OAuth Authentication.

    • Select an Existing Credential or select Create New Credential to create a new credential and enter the Personal Access Token or the information for Azure AD.

  10. Select Link to link the credential to the Data Source.

  11. Select Connect to connect.

Configure Delta Lake Connection in DCM

This connection is used for staging data in AWS S3 or ADLS.

  1. Open Data Connection Manager and navigate to Delta Lake on AWS or Delta Lake on Azure

    1. From an Input tool or the In-DB Connection Manager, DCM is pre-filtered.

    2. From the File Menu, go to File > Manage Connections > +New > Delta Lake > Delta Lake on AWS/Delta Lake on Azure.

  2. For Delta Lake on AWS

    1. Enter a Data Source Name.

    2. Enter an Endpoint or leave it as Default. When Default is used, Amazon determines the endpoint based on the selected bucket.

    3. Make sure Use Signature V4 for Authentication is selected unless specifically instructed otherwise. If unchecked Signature V2 is used. Regions created after January 30, 2014 support only Signature Version 4. These regions require Signature Version 4 authentication:

      • US East (Ohio) Region,

      • Canada (Central) Region,

      • Asia Pacific (Mumbai) Region,

      • Asia Pacific (Seoul) Region,

      • EU (Frankfurt) Region,

      • EU (London) Region,

      • China (Beijing) Region.

    4. Select the level of Server-Side Encryption needed, None is the default.

      • None (Default): An encryption method is not used.

      • SSE-KMS: Use server-side encryption with AWS KMS-managed keys. You can also provide a KMS Key ID. When you select this method, Use Signature V4 for Authentication must be selected.

    5. Enter the Bucket name to be used for staging. The user needs to have read, write, and delete permissions for the bucket.

  3. For Delta Lake on Azure

    1. Enter Data Source Name, ADLS Container, and Storage Account.

    2. Storage Temporary Directory is optional. When entering the Temp Directory, don’t repeat the Container name. If the directory entered here does not already exist, Alteryx creates one. Alteryx creates one sub-folder with the table name for each table that is staged.

  4. Select Save to save the Data Source.

  5. Select +Connect Credential.

    1. Select an Authentication Method.

      • For Delta Lake on AWS, the only Authentication Method is AWS IAM access keys.

      • For Delta Lake on Azure, you may select between Shared Access Key or Azure AD authentication.

    2. Select an Existing Credential or Create New Credential to create a new credential.

      • Enter a Credential Name and the AWS IAM access keys or the Azure Shared Key.

  6. Select Link to link the credential to the Data Source.

  7. Select Connect.