Appsplus panel macro confluence macro | ||
---|---|---|
| ||
**Delete Before Publishing**
|
In this article
Table of Contents | ||||||
---|---|---|---|---|---|---|
|
Overview
You can use this Snap to execute a Databricks SQL DELETE statement based on specific conditions. ..Ensure adequate discretion while using this Snap as it can truncate the table if run without specifying a WHERE condition for the DELETE statement.
...
Snap Type
<Snap name> Databricks - Delete Snap is a READ/WRITE/PARSE/TRANSFORM/FLOW/FORMATwrite-type Snap that reads/fetches/writes/parses/executes/transforms/calls/creates… deletes rows from a target DLP table.
Prerequisites
Valid client ID.
A valid account with the required permissions.
None.
Support for Ultra Pipelines
Works in Ultra Pipelines.
Works in Ultra Pipelines if....access credentials to a DLP instance with adequate access permissions to perform the action in context.
Valid access to the external source data in one of the following: Azure Blob Storage, ADLS Gen2, DBFS, GCP, AWS S3, or another database (JDBC-compatible).
Support for Ultra Pipelines
Does not support Ultra Pipelines.
Limitations
NoneSnaps in the Databricks Snap Pack do not support array, map, and struct data types in their input and output documents.
Known Issues
NoneWhen you add an input view to this Snap, ensure that you configure the Batch size as 1 in the Snap’s account configuration. For any other batch size, the Snap fails with the exception: Multi-batch parameter values are not supported for this query type
.
Snap Views
Type | Format | Number of Views | Examples of Upstream and Downstream Snaps | Description |
---|---|---|---|---|
Input | Document |
Binary
Binary or Document
|
|
..
Requires the EDI data and internal ID as a JSON document.
Output
Document
Binary
Binary or
| A JSON document containing the reference to the table and rows to be deleted. | |
Output | Document |
|
..
..
| A JSON document containing the result of the delete operation on the target table. | |||
Error | Error handling is a generic way to handle errors without losing data or failing the Snap execution. You can handle the errors that the Snap might encounter while running the |
pipeline by choosing one of the following options from the When errors occur list under the Views tab. The available options are:
Learn more about Error handling in Pipelines. |
Snap Settings
Appsplus panel macro confluence macro | ||
---|---|---|
| ||
**Delete Before Publishing** Choose from the following sentences to document specific field types. Drop-down lists/Option Buttons (radio buttons): You must list the LoV and describe them if their meaning isn't apparent. In this case, format the LoV in italics, regular font for the LoV's description. In either case, list the LoVs as a bullet list.
Check boxes:
Text Fields
Numeric Text Fields
Notes in field descriptions
|
Info |
---|
|
Field Name | Field Type |
---|
Field Dependency
Description |
---|
Label* Default Value: |
Databricks - Delete |
Db_Del_Duplicates | String |
None.
The name for the Snap. You can modify this to be more specific, especially if you have more than one of the same Snap in your Pipeline. |
Database name
Default Value: |
None. | String/Expression |
Sampling Type is Number of records.
Enter the number of records to output.
Field set Name
Specify advanced parameters that you want to include in the request.
Field 1*
Default Value: <value> or None.
Example: <value>
String
Debug mode check box is not selected.
Field 2
Default Value: <value> or None.
Example: <value>
String
/Suggestion | Enter your corresponding DLP database name for the DELETE statement to delete existing rows from the table. | |
Table name*
Default Value: None. | String/Expression/Suggestion | Enter your table name for the DELETE statement to delete existing rows from. |
Delete condition (deletes all records from table if left blank) Default Value: N/A | String/Expression/Suggestion | Specify the condition for the DELETE statement to filter the rows to delete from the target table. Specify a valid WHERE clause for the delete condition. If you leave this field blank, the Snap deletes all the records from the table. |
Number of Retries Minimum value: 0 Default value: 0 | Integer | Specifies the maximum number of retry attempts when the Snap fails to write. |
Retry Interval (seconds) Minimum value: 1 Default value: 1 | Integer | Specifies the minimum number of seconds the Snap must wait before each retry attempt. |
Manage Queued Queries Default value: Continue to execute queued queries when pipeline is stopped or if it fails. Example: Cancel queued queries when pipeline is stopped or if it fails | Dropdown list | Select this property to determine whether the Snap should continue or cancel the execution of the queued Databricks SQL queries when you stop the Pipeline. If you select Cancel queued queries when pipeline is stopped or if it fails, then the read queries under execution are cancelled, whereas the write type of queries under execution are not cancelled. Databricks internally determines which queries are safe to be cancelled and cancels those queries. Due to an issue with DLP, aborting an ELT Pipeline validation (with preview data enabled) causes only those SQL statements that retrieve data using bind parameters to get aborted while all other static statements (that use values instead of bind parameters) persist.
To avoid this issue, ensure that you always configure your Snap settings to use bind parameters inside its SQL queries. |
Snap Execution Default Value: Execute only | Dropdown list |
N/A
Select one of the three modes in which the Snap executes. Available options are:
|
Troubleshooting
Error | Reason | Resolution |
---|
Account validation failed.
The Pipeline ended before the batch could complete execution due to a connection error.
Verify that the Refresh token field is configured to handle the inputs properly. If you are not sure when the input data is available, configure this field as zero to keep the connection always open.
Examples
Excluding Fields from the Input Data Stream
We can exclude the unrequired fields from the input data stream by omitting them in the Input schema field set. This example demonstrates how we can use the <Snap Name> to achieve this result:
...
Missing property value | You have not specified a value for the required field where this message appears. | Ensure that you specify valid values for all required fields. |
Examples
Delete employee information from DLP table
Consider the scenario where we want to delete information of certain employees from an intermediate data location that runs on DLP. We can achieve this through a Pipeline containing the Databricks - Delete Snap.
...
We configure this Snap (Pipeline) to delete the employee rows from the company_employees table in our DLP instance if their joining date is before Jan 01, 2010. We also configure an appropriate account for the Snap to connect to the target DLP instance.
...
Upon validation, the Pipeline deletes the rows satisfying the condition specified and returns the status of the operation in the Snap’s output.
...
Downloads
Info |
---|
|
Attachments | ||||||
---|---|---|---|---|---|---|
|
Snap Pack History
Expand | ||
---|---|---|
| ||
|
...
|
...
|
...
Related Links
Endpoint Doc Link 1
Endpoint Doc Link 2
Endpoint Doc Link 3