Skip to end of banner
Go to start of banner

Configuring Databricks Accounts

Skip to end of metadata
Go to start of metadata

You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 14 Next »

In this article


Articles in this section

Overview

You must create Databricks accounts to connect to Databricks Snaps in your Pipelines with the source or target CDWs (databases)This account enables you to write-and-transform data in the target databases hosted in the cloud locations listed in the following table. The JDBC URL you configure for your target database indicates the respective cloud location where the database is hosted. You can configure your Databricks accounts in SnapLogic using either the Designer or the Manager.

Target Database

Supported Cloud Location

Cloud Location in JDBC URL

Databricks Lakehouse Platform (DLP)

AWS

jdbc:spark://<your_instance_code>.cloud.databricks.com or jdbc:databricks://<your_instance_code>.cloud.databricks.com

Microsoft Azure

jdbc:spark://<your_instance_code>.azuredatabricks.net or jdbc:databricks://<your_instance_code>.azuredatabricks.net

Supported JDBC JAR Version

You can configure your Databricks Account to automatically use the recommended JDBC JAR file - databricks-jdbc-2.6.25-1.jar for connecting to your target DLP instance and performing the load and transform operations. 

Using Alternate JDBC JAR File Versions

We recommend that you let the Snaps use the listed JAR file versions. However, you may choose a different JAR file version.

Snap-Account Compatibility

Snaps in the Databricks Snap Pack work with the different accounts and protocols in the following table:

Snap

Databricks Account (Source-wise)#

ADLS Gen2

ADLS Blob Storage

AWS S3

GCS

JDBC (Any database)

Databricks File System (DBFS)

Databricks - Select

Databricks - Insert

Databricks - Delete

Databricks - Bulk Load

Databricks - Merge Into

Databricks - Multi Execute

Databricks - Unload

# Source type is required in

Configuring Databricks Accounts Using SnapLogic Designer

Open the SnapLogic Designer and drag a Databricks Snap to the Canvas and click the Snap to open its settings. Click the Account tab. You can now either use an existing account or create a new one.

Selecting an existing account

SnapLogic organizes and displays all accounts to which you have access, sorting them by account type and location. To select an existing account:

  1. In the Account tab, click the List (blue star)  icon to view the accounts to which you have access, and select the account that you want to use. 

  2. Click the Save (blue star) icon.

Creating an account

  1. In the Account tab, click Add Account below the Account Reference field.

  2. Select the Location in which you want to create the account, select the Account Type, and click ContinueThe Add Account dialog window associated with the account type appears, as shown:

  3. Enter the required account details. Learn more about how to provide the information required for each account type in the following articles:

  4. Click Validate to verify the account, if the account type supports validation.

  5. Click Apply to complete configuring the Databricks account.

You can choose the Source/Target Location as None to perform operations within the Databricks tables.

Configuring Databricks Accounts Using SnapLogic Manager

You can use Manager to create accounts without associating them immediately with Pipelines.

Accounts in SnapLogic are associated with projects. You can use accounts created in other projects only if you have at least Read access to them.

  1. In the left pane, browse to the project in which you want to create the account and click  Create > Account Databricks, followed by the appropriate account type. The Create Account dialog associated with the selected account type appears, as shown:

  2. Repeat steps 3 through 5 in the Creating an account section.

Avoid updating account credentials when Pipelines using that account are executing. Doing so may lead to unexpected results, including your account getting locked.

Snap Pack History

 Click here to expand...

Release

Snap Pack Version

Date

Type

Updates

November 2024

main29029

Stable

Updated and certified against the current SnapLogic Platform release.

August 2024

main27765

Stable

Upgraded the org.json.json library from v20090211 to v20240303, which is fully backward compatible.

May 2024

437patches27246

Latest

Added Databricks - Run Job. This Snap executes a job, checks its status in Databricks, and, based on the job's status, completes or fails the pipeline.

May 2024

437patches26400

Latest

Fixed an invalid session handle issue with the Databricks Snap Pack that intermittently triggered an error message when the Snaps failed to connect with Databricks to execute the SQL statement.

May 2024

main26341

Stable

Updated the Delete Condition (Truncates a Table if empty) field in the Databricks - Delete Snap to Delete condition (deletes all records from a table if left blank) to indicate that all entries will be deleted from the table when this field is blank, but no truncate operation is performed.

February 2024

main25112

Stable

Updated and certified against the current SnapLogic Platform release.

November 2023

main23721

Stable

Updated and certified against the current SnapLogic Platform release.

August 2023

main22460

Stable

Updated and certified against the current SnapLogic Platform release.

May 2023

433patches21630

Latest

Enhanced the performance of the Databricks - Insert Snap to improve the amount of time it takes for validation.

May 2023

main21015

Stable

Upgraded with the latest SnapLogic Platform release.

February 2023

main19844

Stable

Upgraded with the latest SnapLogic Platform release.

November 2022

main18944

Stable

The Databricks - Insert Snap now creates the target table only from the table metadata of the second input view when the following conditions are met:

  • The Create table if not present checkbox is selected.

  • The target table does not exist.

  • The table metadata is provided in the second input view.

September 2022

430patches18305

Latest

The following fields are added to each Databricks Snap as part of this enhancement:

  • Number of Retries: The number of attempts the Snap should make to perform the selected operation when the Snap account connection fails or times out.

  • Retry Interval (seconds): The time interval in seconds between two consecutive retry attempts.

September 2022

430patches17796

Latest

The Manage Queued Queries property in the Databricks Snap Pack enables you to decide whether a given Snap should continue or cancel executing the queued Databricks SQL queries.

August 2022

main17386

Stable

Upgraded with the latest SnapLogic Platform release.

4.29.2.0

42920rc17045

Latest

A new Snap Pack for Databricks Lakehouse Platform (Databricks or DLP) introduces the following Snaps:


  • No labels