Teradata Execute

Teradata Execute

 

In this article

Overview

This Snap allows you to execute a Teradata statement. It works only with single queries and supports simple DML (SELECT, INSERT, UPDATE, DELETE) type statements. If a select query is executed, the query's results are merged into the incoming document, and any existing keys will have their values overwritten. On the other hand, the original document is written if there are no results from the query.

You can drop your database with it, so be cautious.

teradata-execute-overview.png

 

Snap Type

The Teradata Execute Snap is a Write-type Snap.

Prerequisites

  • A valid account with the required permissions.

Support for Ultra Pipelines  

Known Issues

Creating stored procedures is not supported with Execute Snap. However, you can only invoke the existing stored procedures.

Behavior Change

  • In 4.26, when the stored procedures were called using the Database Execute Snaps, the queries were treated as write queries instead of read queries. So the output displayed message and status keys after executing the stored procedure.
    In 4.27, all the Database Execute Snaps run stored procedures correctly, that is, the queries are treated as read queries. The output now displays message key, and OUT params of the procedure (if any). The status key is not displayed.

  • If the stored procedure has no OUT parameters then only the message key is displayed with value success. If you have any existing Pipelines that are mapped with status key or previous description then those Pipelines will fail. So, you might need to revisit your Pipeline design.

Snap Views

Type

Format

Number of Views

Examples of Upstream and Downstream Snaps

Description

Input 

Document

 

  • Min: 0

  • Max: 1

  • Mapper

  • Copy

This Snap has at most one document input view. If the input view is defined, then the where clause can substitute incoming values for a given expression.

Output

Document

 

  • Min: 0

  • Max: 1

  • Mapper

  • Router

This Snap has at most one document output view. If an output view is available and an update/insert/merge/delete statement was executed, then the original document that was used to create the statement will be output with the status of the statement executed.

Valid JSON paths that are defined in the where clause for queries/statements will be substituted with values from an incoming document.

Error

Error handling is a generic way to handle errors without losing data or failing the Snap execution. You can handle the errors that the Snap might encounter when running the pipeline by choosing one of the following options from the When errors occur list under the Views tab:

  • Stop Pipeline Execution: Stops the current pipeline execution if the Snap encounters an error.

  • Discard Error Data and Continue: Ignores the error, discards that record, and continues with the remaining records.

  • Route Error Data to Error View: Routes the error data to an error view without stopping the Snap execution.

Learn more about Error handling in Pipelines.

Snap Settings

Label*

 

Specify the name for the Snap. You can modify this to be more specific, especially if you have more than one of the same Snap in your Pipeline.

Default ValueTeradata Execute
ExampleTeradata Execute

SQL statement*

 

Specify the SQL statement to execute on the server. 

There are two possible scenarios that you encounter when working with SQL statements in SnapLogic. You must understand the following scenarios to successfully execute your SQL statements: 

Scenarios to successfully execute your SQL statements

Scenario 1: Executing SQL statements without expressions

If the expression toggle of the SQL statement field is disabled:

  • The SQL statement must not be within quotes. 

  • The $<variable_name> parts of the SQL statement are expressions. In the below example, $id and $book.

For example:

Additionally, the JSON path is allowed only in the WHERE clause. If the SQL statement starts with SELECT (case-insensitive), the Snap regards it as a select-type query and executes once per input document. If not, it regards it as write-type query and executes in batch mode.

Scenario 2: Executing SQL queries with expressions 

If the expression toggle of the SQL statement field is selected: 

  • The SQL statement must be within quotes. 

  • The + $<variable_name> + parts of the SQL statement are expressions, and must not be within quotes. In the below example, $tablename.

  • The $<variable_name> parts of the SQL statement are bind parameter, and must be within quotes. In the below example, $id and $book.

Table name and column names must not be provided as bind parameters. Only values can be provided as bind parameters.

For example:

  • We recommend that you add a single query in the SQL Statement field.

  • When the SQL statement property is an expression, the Pipeline parameters are shown in the suggest, but not the input schema.

  • The non-expression form uses bind parameters, making it significantly faster than executing N arbitrary SQL expressions.
    With the SQL statement property set as an expression, the Snap is vulnerable to SQL injection. Please use this feature with caution.

  • The '$' sign and identifier characters, such as double quotes (“), single quotes ('), or back quotes (`), are reserved characters and should not be used in comments or for purposes other than their originally intended purpose.

Single quotes in values must be escaped

Any relational database (RDBMS) treats single quotes (') as special symbols. So, single quotes in the data or values passed through a DML query may cause the Snap to fail when the query is executed. Ensure that you pass two consecutive single quotes in place of one within these values to escape the single quote through these queries.

For example:

If String 

To pass this value

Use

Has no single quotes

Schaum Series

'Schaum Series'

Contains single quotes

O'Reilly's Publication

'O''Reilly''s Publication'

Default Value: [None]

Query type

Select the type of query for your SQL statement (Read or Write).

When Auto is selected, the Snap tries to determine the query type automatically.
If the execution result of the query is not as expected, you can change the query type to Read or Write.

Default Value: Auto
ExampleRead

Query band

Specify the name-value pairs to use in the session's generated query band statement. The query band is passed to the Teradata database as a list of name-value pairs separated by semi-colons.

Default Value: N/A
Example: ProjectName=dstage1;user=test;

Pass through

 

Select this checkbox to pass the input document to the output view under the key 'original'. This property applies only to the Execute Snaps with SELECT statement.

Default Value: Selected 

Ignore empty result

 

Select this checkbox to ignore empty result; no document will be written to the output view when a SELECT operation does not produce any result.
If you deselect this checkbox and select the Pass through checkbox, the input document will be passed to the output view.

Default Value: Not selected 

Number of retries

Specify the maximum number of retry attempts the Snap must make in case of network failure. 

When you set the Number of retries to more than 0, the Snap generates duplicate records when the connection is not established. To prevent duplicate records, we recommend that you follow one of the following:

  • Set the Number of retries to 0 (default value) to prevent duplicate records from being passed downstream while executing a pipeline.

  • Use a Primary key to prevent records from being inserted into the database.

  • Use an Upsert instead of an Insert statement.

Default Value: 0

Example: 3

Retry interval

Specify the time interval between two successive retry requests. A retry happens only when the previous attempt resulted in an exception. 

Auto commit

Select one of the options for this property to override the state of the Auto commit property on the account. The Auto commit at the Snap-level has three values: TrueFalse, and Use account setting. The expected functionality for these modes are:

  •  True - The Snap will execute with auto-commit enabled regardless of the value set for Auto commit in the Account used by the Snap.

  •  False - The Snap will execute with auto-commit disabled regardless of the value set for Auto commit in the Account used by the Snap.

  • Use account setting - The Snap will execute with Auto commit property value inherited by the Account used by the Snap.

Default Value: Use account setting
ExampleFalse

'Auto commit' may be enabled for certain use cases if PostgreSQL jdbc driver is used in either Redshift, PostgreSQL or generic JDBC Snap. But the JDBC driver may cause out of memory issues when Select statements are executed. In those cases, “Auto commit" in Snap property should be set to ‘False’ and the Fetch size in the “Account setting" can be increased for optimal performance.

Behavior of DML Queries in Database Execute Snap when auto-commit is false

DDL queries used in the Database Execute Snap will be committed by the Database itself, regardless of the Auto-commit setting.
When Auto commit is set to false for the DML queries, the commit is called at the end of the Snap's execution.
The Auto commit needs to be true in a scenario where the downstream Snap does depend on the data processed on an Upstream Database Execute Snap containing a DML query.
When the Auto commit is set to the Use account setting on the Snap, the account level commit needs to be enabled.

Advanced options

Select the option ‘Timestamp with microsecond precision’ to support the microsecond precision for TIMESTAMP data type. The SELECT query for TIMESTAMP columns produces string values with microsecond precision in the output documents.

Default Value: None

Troubleshooting

Error

Reason 

Resolution

Failure: Failed to execute query

error code: 5992

A syntax error was found in the QUERY_BAND.

Verify that the query band is in the format specified in the Query band field above.

For example, if priority=1;workload=high in the Query band, add a semi-colon and update the value to priority=1;workload=high;.

Example

This example demonstrates how to execute various commands using the Teradata Execute Snap. This pipeline creates a table, inserts the data into the table, and reads the data using the respective execute commands (CREATE, INSERT, and SELECT).

1. Create a table, Snaplogic.character_datatype  under the schema name Snaplogic. The table has 4 columns: INTEGER, CHAR, VARCHAR and CLOB .

2. Insert the required data into the table Snaplogic.character_datatype using the Teradata Execute Snap.

3. We use the Teradata Execute Snap again and select the data from the table Snaplogic.character_datatype1 using the SELECT command.

Successful execution of the Pipeline displays the following preview: