Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

Field Name

Field Type

Description

Label*

Default ValueDatabricks - Execute
ExampleDb_MultiQuery

String

The name for the Snap. You can modify this to be more specific, especially if you have more than one of the same Snap in your Pipeline.

SQL Statements*

Use this fieldset to define your SQL statements, one in each row. You can add as many SQL statements as you need.

SQL statement*

Default Value: None.
Examplecreate table employee (name String, age BigInt)

String/Expression

Specify the Databricks SQL statement you want the Snap to execute. The SQL statement must follow the SQL syntax as stipulated in DLP. We recommend you to add a single query in the SQL Statement field. The SQL statement must follow the SQL syntax as stipulated in DLP.

Number of Retries

Default Value: 0
Example: 3

Minimum value: 0

 

Integer

Specify the maximum number of retry attempts when the Snap fails to write.

Retry Interval (Seconds)

Default value: 1
Example: 3

Minimum value: 1

 

Integer

Specify the minimum number of seconds the Snap must wait before each retry attempt.

Use Result Query

Checkbox

Select this checkbox to write the SQL statement execution result to the Snap's output view for each successful execution. The output of the Snap is enclosed in the key ResultQuery, and the value will be the actual output produced by the SQL statement. The example Pipeline below depicts the difference between the output previews when this checkbox is selected and when it is not.

This option allows you to effectively track the SQL statement's execution by clearly indicating the successful execution and the number of records affected, if any, after the execution.

For DDL statements

Because the Databricks JDBC driver does not return a result set for Data Definition Language (DDL) statements such as DROP, CREATE, and ALTER, the Snap displays a standard message: SQL statement executed successfully but did not produce any result set.

Manage Queued Queries

Default ValueContinue to execute queued queries when pipeline is stopped or if it fails.
Example: Cancel queued queries when pipeline is stopped or if it fails

Dropdown list

Select either of the following options from the dropdown list to handle queued SQL queries:

  • Continue to execute queued queries when pipeline is stopped or if it fails: Continues the execution of the queued Databricks SQL queries when you stop the Pipeline.

  • Cancel queued queries when pipeline is stopped or if it fails: Cancels the execution of the queued Databricks SQL queries when you stop the Pipeline

If you select Cancel queued queries when pipeline is stopped or if it fails, then the read queries under execution are cancelled, whereas the write type of queries under execution are not cancelled. Databricks internally determines which queries are safe to be cancelled and cancels those queries.

Note

Due to an issue with DLP, aborting an ELT Pipeline validation (with preview data enabled) causes only those SQL statements that retrieve data using bind parameters to get aborted while all other static statements (that use values instead of bind parameters) persist.

  • For example, select * from a_table where id = 10 will not be aborted while select * from test where id = ? gets aborted.

To avoid this issue, ensure that you always configure your Snap settings to use bind parameters inside its SQL queries.

Snap Execution

Default ValueExecute only
Example: Validate & Execute

Dropdown list

Select one of the three modes in which the Snap executes. Available options are:

  • Validate & Execute: Performs limited execution of the Snap, and generates a data preview during Pipeline validation. Subsequently, performs full execution of the Snap (unlimited records) during Pipeline runtime.

  • Execute only: Performs full execution of the Snap during Pipeline execution without generating preview data.

  • Disabled: Disables the Snap and all Snaps that are downstream from it.

...