Skip to main content

Databricks

databricks.png

Databricks

Nexla's bi-directional connectors can both send data to and receive data from any data system. This means that once a user has created or gained access to a credential for any data system, building any data flow to ingest data from or send data to a location within that data system requires only a few simple steps.

Authentication

This section provides step-by-step instructions for creating a new Databricks credential that will allow Nexla to authenticate to your Databricks account.

  • To create a new Databricks credential, after selecting the data source/destination type, click the Add Credential tile to open the Add New Credential overlay.

New Credential Overlay – Databricks

DatabricksCred.png

Credential Name & Description

  1. Enter a name for the credential in the Credential Name field and a short, meaningful description in the Credential Description field.

    Resource Descriptions

    Resource descriptions should provide information about the resource purpose, data freshness, etc. that can help the owner and other users efficiently understand and utilize the resource.

Authentication Details

  1. Select the authentication method that will be used to connect to the Databricks account from the Authentication Type pulldown menu, and follow the corresponding instructions in the box below to configure associated credential settings.

Configure Authentication Settings

Authenticate with Databricks using the OAuth 2.0 protocol – recommended for use with production environments

  1. Databricks authentication information can be entered as a JDBC URL or as HTTP path parts that Nexla will automatically combine to create the connection string. Select the preferred option for this credential from the URL Format pulldown menu, and follow the instructions below to enter the authentication information.

  2. Enter the JDBC URL or HTTP path parts required for authentication:

    • Enter the JDBC URL of the Databricks location in the JDBC URL field. The URL should be in the form of jdbc:databricks:/... or jdbc:spark:/....

      Ensure that the URL does not include the AuthMech and Auth_Flow parameters. Connection strings copied from the Databricks console often include these parameters, and they should be removed before saving the credential.

  3. Enter the service principal UUID or application ID of the Databricks warehouse this credential will be used to access in the Service Principal UUID/Application ID field.

  4. Enter the service principal secret of the Databricks warehouse this credential will be used to access in the Service Principal Secret field.

Database Information

  1. Enter the name of the Databricks database that will be accessed with this credential in the Database Name field.

  2. If this credential will be used to access data that is also contained in a schema (typically used to organize individual use cases, projects, or team sandboxes), enter the name of the schema in the Schema Name field.

Save the Credential

  1. Once all of the relevant steps in the above sections have been completed, click the Save button at the bottom of the overlay to save the configured credential.

  2. The newly added credential will now appear in a tile on the Authenticate screen during data source/destination creation and can be selected for use with a new data source or destination.

Data Source

To ingest data from a Databricks location, follow the instructions in Section 2 of Common Setup for Databases & Data Warehouses.

Data Destination

To send data to a Databricks location, follow the instructions in Section 3 of Common Setup for Databases & Data Warehouses.