Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

Info
  • Asterisk ( * ): Indicates a mandatory field.

  • Suggestion icon ( (blue star) ): Indicates a list that is dynamically populated based on the configuration.

  • Expression icon ( (blue star) ): Indicates the value is an expression (if enabled) or a static value (if disabled). Learn more about Using Expressions in SnapLogic.

  • Add icon ( (blue star) ): Indicates that you can add fields in the fieldsetfield set.

  • Remove icon ( (blue star) ): Indicates that you can remove fields from the fieldsetfield set.

Field Name

Field Type

Field Dependency

Description

Label*

 

Default Value: N/A
Example: STD DB Acc DeltaLake AWS S3

String

None.

Specify a unique label for the account.

 

Download JDBC Driver Automatically

 

 

 

 

 

 

 

 

 

 

 

Default Value: Not Selected

Example: Selected

Checkbox

None

Select this checkbox to allow the Snap account to download the certified JDBC Driver for DLP. The following fields are disabled when this checkbox is selected.

  • JDBC JAR(s) and/or ZIP(s) : JDBC Driver

  • JDBC driver class

To use a JDBC Driver of your choice, clear this checkbox, upload (to SLDB), and choose the required JAR files in the JDBC JAR(s) and/or ZIP(s): JDBC Driver field. 

Use of Custom JDBC JAR version

You can use a different JAR file version outside of the recommended listed JAR file versions.

Spark JDBC and Databricks JDBC

If you do not select this checkbox and use an older JDBC JAR file (older than version 2.6.25), ensure that you use: 

  • The old format JDBC URL ( jdbc:spark:// ) instead of the new one ( jdbc:databricks:// )

    • For JDBC driver prior to version 2.6.25, the JDBC URL starts with jdbc:spark://

    • For JDBC driver version 2.6.25 or later, the JDBC URL starts with jdbc:databricks://

  • The older JDBC Driver Class com.simba.spark.jdbc.Driver instead of the new com.databricks.client.jdbc.Driver.

 

JDBC URL*

 

 

Default Value: N/A
Example: jdbc:spark://adb-2409532680880038.18.azuredatabricks.net:443/default;transportMode=http;ssl=1;httpPath=sql/protocolv1/o/2409532680880038/0326-212833-drier754;AuthMech=3;

String

None

Enter the JDBC driver connection string that you want to use in the syntax provided below, for connecting to your DLP instance. Learn more about Microsoft's JDBC and ODBC drivers and configuration parameters.

jdbc:spark://dbc-ede87531-a2ce.cloud.databricks.com:443/default;transportMode=http;ssl=1;httpPath=
sql/protocolv1/o/6968995337014351/0521-394181-guess934;AuthMech=3;UID=token;PWD=<personal-access-token> 

Avoid passing Password inside the JDBC URL

If you specify the password inside the JDBC URL, it is saved as it is and is not encrypted. We recommend passing your password using the Password field provided, instead, to ensure that your password is encrypted.

 

Use Token Based Authentication

Default value: Selected
Example: Not selected

Checkbox

None

Select this checkbox to use token-based authentication for connecting to the target database (DLP) instance. Activates the Token field.

 

Token*

Default value: N/A
Example: <Encrypted>

String

When Use Token Based Authentication checkbox is selected.

Enter the token value for accessing the target database/folder path.

 

Database name*

Default value: N/A
Example: Default

String

None

Enter the name of the database to use by default. This database is used if you do not specify one in the Databricks Select or Databricks Insert Snaps.

 

Source/Target Location*

Dropdown

None

Select the source or target data warehouse into which the queries must be loaded, that is AWS S3. This activates the following fields:

  • S3 Bucket

  • S3 Folder

  • AWS Authorization type

  • S3 Access Key ID

  • S3 Secret Key

S3 Bucket*

String

None

Specify the name of the S3 bucket that you want to use for staging data to Databricks. 

Default Value: N/A

Examplesl-bucket-ca

S3 Folder*

String

None

Specify the relative path to a folder in the S3 bucket listed in the S3 Bucket field. This is used as a root folder for staging data to Databricks.

Default Value: N/A

Example:  https://sl-bucket-ca.s3.<ca>.amazonaws/<sf>

Aws Authorization type

Default value: Source Location Credentials for S3 and Azure, Storage Integration for Google Cloud Storage.

Example: Storage Integration

Dropdown

None

Select the authentication method to use for accessing the source data.

Available options are:

  • Source/Target Location Credentials. Select this option when you do not have a storage integration setup in your S3. Activates the Access Key and Secret Key fields for S3.

  • Source/Target Location Session Credentials. Select this option if you have session credentials to access the source location in S3. Activates the Session Access KeySession Secret Key, and Session Token fields.

  • Storage Integration. Select this option when you want to use the storage integration to access the selected source location. Activates the Storage Integration Name field.

S3 Access-key ID*

Default Value: N/A

ExampleNAVRGGRV7EDCFVLKJH

String

None

Specify the S3 access key ID that you want to use for AWS authentication.

S3 Secret key*

Default Value: N/A

Example2RGiLmL/6bCujkKLaRuUJHY9uSDEjNYr+ozHRtg

String

None

Specify the S3 secret key associated with the S3 Access-ID key listed in the S3 Access-key ID field.

S3 AWS Token*

Default Value: None
ExampleAQoDYXdzEJr

String

Appears when Source/Target Location Session Credentials is selected in Aws Authorization type

Specify the S3 AWS Token to connect to private and protected Amazon S3 buckets.

Info

The temporary AWS Token is used when:

  • Data is staged in S3 location.

  • Data is coming from the input view and the files are staged in an external staging location

.

Troubleshooting

Error

Reason

Resolution

Account validation failed.

The Pipeline ended before the batch could complete execution due to a connection error
  • .

Verify that the Refresh token field is configured to handle the inputs properly. If you are not sure when the input data is available, configure this field as zero to keep the connection always open.

Snap Pack History

Expand

Insert excerpt
Databricks Snap Pack
Databricks Snap Pack
nameDatabricks Snap Pack History
nopaneltrue

...