TPT FastExport

On this Page

Snap type:

Read


Description:

The TPT (Teradata Parallel Transport) FastExport Snap allows you to export data from a Teradata database by executing the script generated using the fields provided by the user in the Snap fields. The Snap writes the exported data to the specified file in the local file system and provides the console output and status code on the output view for any downstream Snap checking for successful execution. 

Queries produced by the Snap have the format:

SELECT TOP [limit] * FROM [table] WHERE [where clause] ORDER BY [ordering] .. and uses QUALIFY() for LIMIT
  • Expected downstream Snaps: A Snap monitoring the exit code to check result status.
  • Expected input: None or documents with values to evaluate expression properties in the Snap. Each input document triggers one execution of the script.
  • Expected output: A single document containing the console output and return status from the external TPT application for each input document


Prerequisites:

TPT Snaps (TPT Insert, TPT Update, TPT Delete, TPT  Upsert and TPT Load) uses the 'tbuild' utility for the respective operations.

In order to use these Snaps in the pipelines, the respective TPT utilities should be installed/available on the Snapplexes. 

All the required TPT utilities must be installed on the node where JCC is running or the Snap simply executes but may not perform the operation. Any utility/library file missing on the node may cause the Snap to perform a failed operation.

Basic steps for installing the TPT utilities:

Procedure to install the Teradata tools and utilities:

  • Download the Teradata Tools and Utilities base such as TTU 16.20.25.00 Linux - Base for Linux environment.
  • If the OS is 64 bit,  install 32 bit support packages on the environment and also any further dependencies
  • Install the Teradata Tools and Utilities (that includes tbuild, fastload, mload etc)
  • The default installation also configures the PATH in the OS

Flow:  If the tbuild or fastload is already available on the OS path then the corresponding TPT Snap can invoke the required utility, however, if these binaries are not available on the OS path, then we need to provide the absolute path of the respective binary in the Snap.

Support and limitations:Works in Ultra Task Pipelines.
Account: 

This Snap uses account references created on the Accounts page of  SnapLogic Manager to handle access to this endpoint. See Configuring Teradata Database Accounts for information on setting up this type of account.


Views:
InputThis Snap has at most one document input view. 
Output

This Snap has one or two output views. The first output view produces one Document containing the console output and return status from the external Teradata application. The second output view provides the table and column metadata.

ErrorThis Snap has at most one document error view and produces zero or more documents in the view.

Settings

Label


Required. The name for the Snap. You can modify this to be more specific, especially if you have more than one of the same Snap in your pipeline.


Schema name



The database schema name. In case it is not defined, the suggestion for the table name will retrieve all the table names of all the schemas. The property is suggestible and will retrieve available database schemas during suggest values.

Example: SYS

Default value: [None] 

Table name

Required. Name of the table to execute a select query on.

Example: people

Default value: [None] 

Where clause
 

Where clause of the select statement. This supports document value substitution (such as $person.firstname will be substituted with the value found in the incoming document at the path). However, you may not use a value substitution after "IS" or "is" word. 

Examples

Without using expressions

Using expressions

  • "EMPNO=$EMPNO and ENAME=$EMPNAME"

  • email = $email 

  • emp=$emp

  • "emp='" + $emp + "'"

  • "EMPNO=" + $EMPNO + " and ENAME='" + $EMPNAME+ "'"

Caution

Using expressions that join strings together to create SQL queries or conditions has a potential SQL injection risk and is hence unsafe. Ensure that you understand all implications and risks involved before using concatenation of strings with '=' Expression enabled. 


Default value: [None] 

Order by 

Enter in the columns in the order in which you want to order by. The default database sort order will be used.

Example
     name
     email

Default value: [None] 

Limit offset


The starting row for the query.

Default value: [None] 

Limit rows


The maximum number of rows the query should return.

Default value: [None] 

Output fields



Enter or select output field names for SQL SELECT statement. To select all fields, leave it at default.

Example: email, address, first, last, etc.

Default value: [None] 

Number of retries

Specify the maximum number of reconnection attempts that the Snap must perform, in case of connection failure or timeout.

Default Value: 0

Retry interval (seconds)


Enter in seconds the duration for which the Snap must wait between two reconnection attempts, until the number of retries is reached.

Default Value: 1

File Action


Required. Select an action when the specified file already exists in the directory. The options available include: OVERWRITE, IGNORE and ERROR.

Default value: ERROR  

TBUILD location


Required. Location of the Teradata application.

The TBUILD location will be available on the Snaplex.

Default value: /usr/bin/tbuild 

Output directory



Required. Attribute specifies the full path to the output directory. Specify a local directory path. An absolute path should be entered.
Use the suggest bubble to get the list of subdirectories in the current directory. If this property is blank, it will list all subdirectories in the top directory of the local file system.  

Example:  /home/Snap/tpt/

         file:///home/Snap/tpt/  (local file system path where the file should be stored)

Default value: ERROR 

Output File


Required. Location of exported data file.

If there is a white space, then the path should be in quotes as per the Teradata documentation.

Default value: [None] 

File Format


Required. Attribute that specifies the logical record format of the exported data file. The  options available include: Binary, Delimited, Formatted, Unformatted and Text. Click TPT Reference, to know more about the details of each format. 

Text format requires all column data types to be CHAR or DATE in the selected Output fields. If DATE data type is included, the Date format property should be set to Text.

Default value: Delimited 

Text Delimiter Type


Required. Attribute that specifies Type of Text Delimiter that is to be used. The options available include: TextDelimiter and TextDelimiterHEX.

Default value: TextDelimiter  

Text Delimiter  


Attribute that specifies the bytes that separate fields in delimited records. Any number of characters can be defined via the attribute assignment. The default delimiter character is the pipe character ( | ). To embed a pipe delimiter character in your data, precede the pipe character with a backslash ( \ ). Use Hex code for Hex Type and Text for the other.

 To use the tab character as the delimiter character, specify TextDelimiter = 'TAB'. Use uppercase “TAB” not lowercase “tab”. The backslash is required if you want to embed a tab character in your data.

Default value: ( | ) 

Escape Text Delimiter Type


Required. Attribute that specifies Type of Escape Text Delimiter that is to be used. The options available include: EscapeDelimiterText and EscapeDelimiterHEX.

Default value: EscapeTextDelimiter 

Escape Text Delimiter


Attribute that allows you to define the delimiter escape character within delimited data. There is no default data. When processing data in delimited format, if the backslash precedes the delimiter, that instance of the delimiter is included in the data rather than marking the end of the column. Use Hex code for Hex Type and Text for the other. 

Default value:  ( \ ) 

Date format


Required. Date format used in the output file in text mode. The options available include: Integer and Text. 

If this property is set to 'Integer' and the File Format property is set to Binary/Formatted /Unformatted, the DATE type values are written to the output file as an integer. 
If this property is 'Integer', and the File Format property is 'Text' and a DATE data type output field is included, the Snap will fail with an error "... TPT19108 Data Format 'TEXT' requires all 'CHAR/ANSIDATE' schema...."  

Default value: Integer 

Quoted Data


Required. Determines if data is expected to be enclosed within quotation marks.

Default value: Not Selected 

Snap execution

Page lookup error: page "Anaplan Read" not found.

If you're experiencing issues please see our Troubleshooting Guide.

For the Suggest in the Order by columns and the Output fields properties, the value of the Table name property should be an actual table name instead of an expression. If it is an expression, it will display an error message "Could not evaluate accessor:  ..." when the Suggest button is pressed. This is because, at the time the Suggest button is pressed, the input document is not available for the Snap to evaluate the expression in the Table name property. The input document is available to the Snap only during the preview or execution time.

Temporary Files

During execution, data processing on Snaplex nodes occurs principally in-memory as streaming and is unencrypted. When larger datasets are processed that exceeds the available compute memory, the Snap writes Pipeline data to local storage as unencrypted to optimize the performance. These temporary files are deleted when the Snap/Pipeline execution completes. You can configure the temporary data's location in the Global properties table of the Snaplex's node properties, which can also help avoid Pipeline errors due to the unavailability of space. For more information, see Temporary Folder in Configuration Options

Example


In this example, the pipeline exports the data from a table to a flat file using the TPT FastExport Snap.The data thus exported is read and parsed to an out put view using the File Reader and the CSV Parser Snaps respectively.

TPT FastExport exports the data from the table DATETIMETABLE under the schema, Snaplogic to a flat file, Teradata_Date.txt under the Output directory path, file:///opt/Snaplogic.

 


The preview of the data read by the File Reader Snap is:

The data is parsed using the CSV Parser Snap. Successful execution of the pipeline displays the below output preview:  

Snap Pack History

 Click to view/expand
Release Snap Pack VersionDateType  Updates

February 2024

main25112

 

Stable

Updated and certified against the current SnapLogic Platform release.

November 2023main23721 StableUpdated and certified against the current SnapLogic Platform release.

August 2023

main22460

 


Stable

The Teradata Execute Snap now includes a new Query type field. When Auto is selected, the Snap tries to determine the query type automatically.

May 2023

main21015 

Stable

Upgraded with the latest SnapLogic Platform release.

February 2023main19844 StableUpgraded with the latest SnapLogic Platform release.
November 2022main18944 Stable

The TPT Insert Snap now creates the target table only from the table metadata of the second input view when the following conditions are met:

  • The Create table if not present checkbox is selected.

  • The target table does not exist.

  • The table metadata is provided in the second input view.

August 2022main17386 StableUpgraded with the latest SnapLogic Platform release.
4.29 Patch429patches16235 Latest

Fixed an issue with Teradata Execute Snap where the Snap did not display valid error message when the delete condition is invalid.

4.29main15993 Stable

Enhanced the Teradata FastExport Snap with Character Set dropdown list to support encoding of data when you export data from the Teradata Database.

4.28main14627 StableUpgraded with the latest SnapLogic Platform release.
4.27main12833 Stable

Enhanced the Teradata Execute Snap to invoke stored procedures.

4.26main11181 StableUpgraded with the latest SnapLogic Platform release.
4.25 Patch425patches11008 Latest

Improved the error messages for all the Snaps in the Teradata Snap Pack where the Snaps fail with a null pointer exception error when the account reference provided is invalid. 

4.25main9554
 
StableUpgraded with the latest SnapLogic Platform release.

4.24 Patch

424patches8799 Latest
  • Enhanced the Teradata Execute Snap by adding a new field, Advanced options, which extends support with microsecond precision for TIMESTAMP data type.
  • Fixed an issue in the TPT Load Snap where the Snap failed to load data into the table while creating an output.
  • Fixed an issue with the TPT Load Snap wherein now a null value is loaded as null and empty string is loaded as empty string.

    • This fix may cause existing pipelines to fail if empty string values are expected to be loaded as null.

    • Following are the new configurations:

      • VARCHAR QuotedData= Optional

      • VARCHAR OpenQuoteMark= \”

      • VARCHAR NullColumns= Yes

4.24main8556
Stable

Enhanced the Teradata Select Snap to return only the selected output fields or columns in the output schema (second output view) using the Fetch Output Fields In Schema check box. If the Output Fields field is empty all the columns are visible.

4.23main7430
 
Stable

Fixes the multi-line value issue and issue in the TPT Load Snap where the Snap writes null for both an empty string and null values in the input data. The fix for this issue was to add the following three lines in the script and wrap all the values in double quotes unless they are null while writing the input data into a temp CSV file. 
VARCHAR QuotedData = 'Optional',
VARCHAR OpenQuoteMark = '\"'
VARCHAR NullColumns = 'Yes'

This fix may cause existing pipelines to fail if empty string values are expected to be loaded as null.

4.22main6403
 
StableUpgraded with the latest SnapLogic Platform release.

4.21 Patch

421patches6272 Latest

Fixed the issue where Snowflake SCD2 Snap generates two output documents despite no changes to Cause-historization fields with DATE, TIME and TIMESTAMP Snowflake data types, and with Ignore unchanged rows field selected.

4.21 Patch

421patches6144 Latest

Fixed the following issues with DB Snaps:

  • The connection thread waits indefinitely causing the subsequent connection requests to become unresponsive.
  • Connection leaks occur during Pipeline execution.

4.21 Patch

MULTIPLE8841 Latest

Fixed the connection issue in Database Snaps by detecting and closing open connections after the Snap execution ends. 

4.21snapsmrc542

 

StableUpgraded with the latest SnapLogic Platform release.
4.20snapsmrc535
 
StableUpgraded with the latest SnapLogic Platform release.

4.19 Patch 

db/teradata8414 LatestFixed an issue with the TPT Update Snap wherein the Snap is unable to perform operations when:
  • An expression is used in the Update condition property.
  • Input data contain the character '?'.
4.19snaprsmrc528
 
StableUpgraded with the latest SnapLogic Platform release.
4.18snapsmrc523
 
StableUpgraded with the latest SnapLogic Platform release.
4.17ALL7402
 
Latest

Pushed automatic rebuild of the latest version of each Snap Pack to SnapLogic UAT and Elastic servers.

4.17snapsmrc515
 
Latest
  • Fixed an issue with the Teradata Execute Snap wherein the Snap would send the input document to the output view even if the Pass through field is not selected in the Snap configuration. With this fix, the Snap sends the input document to the output view, under the key original, only if you select the Pass through field.
  • Added the Snap Execution field to all Standard-mode Snaps. In some Snaps, this field replaces the existing Execute during preview check box.
4.16snapsmrc508
 
StableUpgraded with the latest SnapLogic Platform release.

4.15 Patch 

db/teradata6338 Latest

Replaced Max idle time and Idle connection test period properties with Max life time and Idle Timeout properties respectively, in the Account configuration. The new properties fix the connection release issues that were occurring due to default/restricted DB Account settings.

4.15snapsmrc500
 
StableUpgraded with the latest SnapLogic Platform release.
4.14snapsmrc490
 
StableUpgraded with the latest SnapLogic Platform release.
4.13

snapsmrc486

 
StableUpgraded with the latest SnapLogic Platform release.
4.12

snapsmrc480

 
StableUpgraded with the latest SnapLogic Platform release.
4.11snapsmrc465
 
StableUpgraded with the latest SnapLogic Platform release.
4.10

snapsmrc414

 
Stable

Added Auto commit property to the Select and Execute Snaps at the Snap level to support overriding of the Auto commit property at the Account level.

4.9.0 Patch

teradata3077 Latest

Fixed an issue regarding connection not closed after login failure; Expose autocommit for "Select into" statement in PostgreSQL Execute Snap and Redshift Execute Snap

4.9snapsmrc405
 
Stable
  • Enhanced the Output view of the Snap with Order Summary field that displays the output values (added to the out, err, status, classpath and env fields of the single output view that display the subprocess only).
  • Teradata Export to HDFS Snap supported with Dynamic account.
4.8

snapsmrc398

 
Stable
  • Introduced the TPT Delete, Insert, Load, Upsert and Update Snaps in this release.
  • Introduced Teradata Export to HDFS Snap in this release.
  • Info tab added to accounts.
  • Database accounts now invalidate connection pools if account properties are modified and login attempts fail.
4.7

snapsmrc382

 
Stable
  • Introduced the Teradata FastLoad and Execute Snaps in this release.
  • Migration impact: In Teradata FastExport, the values of the Data Format field have been standardized to be in all caps. Existing pipelines that use the values of Binary, Text, or Unformatted will fail unless the new value of BINARY, TEXT, or UNFORMAT are used.
4.6snapsmrc362
 
Stable

Snap Pack introduced in 4.6.0. This includes only Teradata extract functionality to move data out of Teradata database using the FastExport Utility. It does not include Snaps for load, select, insert, delete, execute or others at this time. It also does not utilize the Teradata Parallel Transporter to extract data.