Skip to end of banner
Go to start of banner

WIP: Databricks - Run Job

Skip to end of metadata
Go to start of metadata

You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 2 Next »

the

In this article

Overview

You can use this Snap to execute a job, check its status in Databricks, and, based on the job's status, complete or fail the pipeline. The Snap triggers the task to execute and then periodically checks its status. The Snap stops after the job is executed. However, if the pipeline is canceled before the task is finished, the Snap sends a request to stop the task.

Snap Type

The Databricks - Run Job Snap is a Write-type Snap.

Prerequisites

  • Valid client ID.

  • A valid account with the required permissions.

Support for Ultra Pipelines

Limitations and Known Issues

None.

Snap Views

Type

Format

Number of Views

Examples of Upstream and Downstream Snaps

Description

Input 

Document

Binary

Binary or Document

  • Min: 0

  • Max: 1

  • Mapper

  • Copy

Requires a valid task name, notebook path, and cluster-info.

Output

Document

Binary

Binary or Document

  • Min: 1

  • Max: 1

  • Mapper

  • Filter

Executes the selected notebook.

Error

Error handling is a generic way to handle errors without losing data or failing the Snap execution. You can handle the errors that the Snap might encounter when running the pipeline by choosing one of the following options from the When errors occur list under the Views tab:

  • Stop Pipeline Execution: Stops the current pipeline execution if the Snap encounters an error.

  • Discard Error Data and Continue: Ignores the error, discards that record, and continues with the remaining records.

  • Route Error Data to Error View: Routes the error data to an error view without stopping the Snap execution.

Learn more about Error handling in Pipelines.

Snap Settings

  • Asterisk ( * ): Indicates a mandatory field.

  • Suggestion icon ((blue star)): Indicates a list that is dynamically populated based on the configuration.

  • Expression icon ((blue star) ): Indicates the value is an expression (if enabled) or a static value (if disabled). Learn more about Using Expressions in SnapLogic.

  • Add icon ( (blue star) ): Indicates that you can add fields in the field set.

  • Remove icon ( (blue star)): Indicates that you can remove fields from the field set.

  • Upload icon ((blue star) ): Indicates that you can upload files.

Field Name

Field Type

Description

Label*

Default ValueDatabricks - Run Job
ExampleRun Job

String

The name for the Snap. You can modify this to be more specific, especially if you have more than one of the same Snap in your pipeline.

Task name*

Default Value: N/A
Example: Test username and password

String/Expression

Specify the name of the task to perform the job.

Notebook path*

Default Value: N/A
Example: /Users/johndoe@snaplogic.com/notebook

String/Expression

Specify the path of the saved notebook that will run in this job.

Cluster*

Default Value: N/A
Example: Code Ammonite - Shared Compute Cluster - V2

Specify the cluster to run the job.

Parameter(s)

Use this field set to specify the parameters to run the job.

Key*

Default Value: N/A
Example

String/Expression

Specify the parameter key.

Value*

Default Value: N/A
Example

String/Expression

Specify the parameter value.

Interval check (seconds)*

Default Value10
Example15

Integer/Expression

Specify the number of seconds to wait before checking the status of the task.

Snap Execution

Default ValueExecute only
Example: Validate & Execute

Dropdown list

Select one of the following three modes in which the Snap executes:

  • Validate & Execute: Performs limited execution of the Snap and generates a data preview during pipeline validation. Subsequently, performs full execution of the Snap (unlimited records) during pipeline runtime.

  • Execute only: Performs full execution of the Snap during Pipeline execution without generating preview data.

  • Disabled: Disables the Snap and all Snaps that are downstream from it.

Troubleshooting

Error

Reason

Resolution

Account validation failed.

The Pipeline ended before the batch could complete execution due to a connection error.

Verify that the Refresh token field is configured to handle the inputs properly. If you are not sure when the input data is available, configure this field as zero to keep the connection always open.

Examples

Excluding Fields from the Input Data Stream

We can exclude the unrequired fields from the input data stream by omitting them in the Input schema fieldset. This example demonstrates how we can use the <Snap Name> to achieve this result:

<screenshot of Pipeline/Snap and description>

Download this Pipeline. 

Downloads

  1. Download and import the Pipeline into SnapLogic.

  2. Configure Snap accounts, as applicable.

  3. Provide Pipeline parameters, as applicable.

  File Modified
No files shared here yet.

Snap Pack History

 Click here to expand...
Error rendering macro 'excerpt-include' : No link could be created for 'XYZ Snap Pack'.

Related Content

  • No labels