Skip to end of banner
Go to start of banner

Snowflake - Bulk Upsert

Skip to end of metadata
Go to start of metadata

You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 94 Current »

On this Page

Overview

You can use the Snowflake - Bulk Upsert Snap to bulk update existing records or insert records to a target Snowflake table.

Depending on the account that you configure with the Snap, you can use the Snowflake Bulk Upsert Snap to upsert data into AWS S3 buckets or Microsoft Azure Storage Blobs or Google Cloud Storage.

The Snap creates temporary files in the JCC when the staging location is internal and the data source is Input view. These temporary files are removed automatically after the pipeline completes execution.

Snap Type

The Snowflake - Bulk Upsert Snap is a Write-type Snap that performs a bulk upsert operation.

PrerequisitesYou must have minimum permissions on the database to execute Snowflake Snaps. To understand if you already have them, you must retrieve the current set of permissions. The following commands enable you to retrieve those permissions:
SHOW GRANTS ON DATABASE <database_name>
SHOW GRANTS ON SCHEMA <schema_name>
SHOW GRANTS TO USER <user_name>

Security Prerequisites: 

You should have the following permission in your Snowflake account to execute this Snap:

  • Usage (DB and Schema): Privilege to use the database, role, and schema.

The following commands enable minimum privileges in the Snowflake Console:

grant usage on database <database_name> to role <role_name>;
grant usage on schema <database_name>.<schema_name>;

For more information on Snowflake privileges, refer to Access Control Privileges.

The below are mandatory when using an external staging location:

When using an Amazon S3 bucket for storage:

  • The Snowflake account should contain S3 Access-key ID, S3 Secret key, S3 Bucket and S3 Folder.

  • The Amazon S3 bucket where the Snowflake will write the output files must reside in the same region as your cluster.

When using a Microsoft Azure storage blob:

  • A working Snowflake Azure database account.

Internal SQL Commands

This Snap uses the following commands internally:

  • SELECT FROM TABLE: SELECT as a statement enables you to query the database and retrieve a set of rows. SELECT as a clause enables you to define the set of columns returned by a query.

  • COPY INTO <table>: Enables loading data from staged files to an existing table. 

  • CREATE TEMPORARY TABLE: Enables creating a temporary table in the database.

  • MERGE: Enables inserting, updating, and deleting values in a table based on values in a second table or a subquery.

Requirements for External Storage Location

The following are mandatory when using an external staging location:

When using an Amazon S3 bucket for storage:

  • The Snowflake account should contain S3 Access-key ID, S3 Secret key, S3 Bucket and S3 Folder.

  • The Amazon S3 bucket where the Snowflake will write the output files must reside in the same region as your cluster.

When using a Microsoft Azure storage blob:

  • A working Snowflake Azure database account.

When using a Google Cloud Storage:

  • Provide permissions such as Public access and Access control to the Google Cloud Storage bucket on the Google Cloud Platform.

Support for Ultra Pipelines

Works in Ultra Pipelines. However, we recommend that you not use this Snap in an Ultra Pipeline.

Limitations

The special character tilde (~) is not supported if it exists in the temp directory name for Windows. It is reserved for the user's home directory.

Known Issues

Because of performance issues, all Snowflake Snaps now ignore the Cancel queued queries when pipeline is stopped or if it fails option for Manage Queued Queries, even when selected. Snaps behave as though the default Continue to execute queued queries when the Pipeline is stopped or if it fails option were selected.

Behavior Change

In the 4.31 main18944 release and later, the Snowflake - Bulk Upsert Snap no longer supports values from an upstream input document in the Key columns field when the expression button is enabled.

To avoid breaking your existing Pipelines that use a value from the input schema in the Key columns field, update the Snap settings to use an input value from Pipeline parameters instead.

Snap Views

Type

Format

Number of Views

Examples of Upstream and Downstream Snaps

Description

Input

Document

  • Min: 1

  • Max: 1

  • JSON Generator

  • Binary to Document

Incoming documents are first written to a staging file on Snowflake's internal staging area. A temporary table is created on Snowflake with the contents of the staging file. An update operation is then run to update existing records in the target table and/or an insert operation is run to insert new records into the target table.

Output

Document

  • Min: 0

  • Max: 1

  • Mapper

  • Snowflake Execute

If an output view is available, then the output document displays the number of input records and the status of the bulk upload.

Error

Error handling is a generic way to handle errors without losing data or failing the Snap execution. You can handle the errors that the Snap might encounter when running the Pipeline by choosing one of the following options from the When errors occur list under the Views tab:

  • Stop Pipeline Execution: Stops the current pipeline execution if the Snap encounters an error.

  • Discard Error Data and Continue: Ignores the error, discards that record, and continues with the remaining records.

  • Route Error Data to Error View: Routes the error data to an error view without stopping the Snap execution.

Learn more about Error handling in Pipelines.

Snap Settings

  • Asterisk (*): Indicates a mandatory field.

  • Suggestion icon ((blue star)): Indicates a list that is dynamically populated based on the configuration.

  • Expression icon ((blue star)): Indicates whether the value is an expression (if enabled) or a static value (if disabled). Learn more about Using Expressions in SnapLogic.

  • Add icon ((blue star)): Indicates that you can add fields in the field set.

  • Remove icon ((blue star)): Indicates that you can remove fields from the field set.

Field Name

Field Type

Field Dependency

Description

Label*

Default ValueSnowflake - Bulk Upsert
Example: Snowflake - Bulk Upsert

String

N/A

Specify the name for the instance. You can modify this to be more specific, especially if you have more than one of the same Snap in your Pipeline.

Schema Name

Default Value: N/A
Example: schema_demo

String/Expression/Suggestion

N/A

Specify the database schema name. In case it is not defined, then the suggestion for the Table Name retrieves all tables names of all schemas. The property is suggestible and will retrieve available database schemas during suggest values.

The values can be passed using the Pipeline parameters but not the upstream parameter.

Table Name*

Default Value: N/A
Example: employees_table

String/Expression/Suggestion

N/A

Specify the name of the table to execute bulk load operation on.

The values can be passed using the Pipeline parameters but not the upstream parameter.

Staging location

Default Value: Internal
Example: External

Dropdown list/Expression

N/A

Select the type of staging location that is to be used for data loading:

  • External: Location that is not managed by Snowflake. The location should be an AWS S3 Bucket or Microsoft Azure Storage Blob or Google Cloud Storage. These credentials are mandatory while validating the Account.

  • Internal: Location that is managed by Snowflake.

Vector support:

  • This Snap supports loading data in vector data type only when the Staging location is Internal.

  • The input vector data must be an array.

  • If you use a SELECT query, you must parse the vector.

Flush chunk size (in bytes)

String/Expression

Appears when you select Input view for Data source and Internal for Staging location.

When using internal staging, data from the input view is written to a temporary chunk file on the local disk. When the size of a chunk file exceeds the specified value, the current chunk file is copied to the Snowflake stage and then deleted. A new chunk file simultaneously starts to store the subsequent chunk of input data. The default size is 100,000,000 bytes (100 MB), which is used if this value is left blank.

File format object

Default ValueNone
Example: jsonPath()

String/Expression

N/A

Specify an existing file format object to use for loading data into the table. The specified file format object determines the format type such as CSV, JSON, XML, AVRO, or other format options for data files.

File format type

Default Value: None
Example: CSV

String/Expression/Suggestion

N/A

Specify a predefined file format object to use for loading data into the table. The available file formats include CSV, JSON, XML, and AVRO.

This Snap supports only CSV and NONE file format types.

File format option

Default value:  N/A
Example: BINARY_FORMAT=UTF8

String/Expression

N/A

Specify the file format option. Separate multiple options by using blank spaces and commas.

You can use various file format options including a binary format which passes through in the same way as other file formats. Learn more: File Format Type Options.

Before loading binary data into Snowflake, you must specify the binary encoding format, so that Snap can decode the string type to binary types before loading it into Snowflake. This can be done by selecting the following binary file format:

BINARY_FORMAT=xxx (Where XXX = HEX|BASE64|UTF8)

However, the file you upload and download must be in similar formats. For instance, if you load a file in HEX binary format, you should specify the HEX format for download as well.

Encryption type

Default Value: None
Example: Server-Side Encryption

Dropdown list

N/A

Specify the type of encryption to be used on the data. The available encryption options are:

  • None: Files do not get encrypted.

  • Server Side Encryption: The output files on Amazon S3 are encrypted with server-side encryption.

  • Server-Side KMS Encryption: The output files on Amazon S3 are encrypted with an Amazon S3-generated KMS key. 

The KMS Encryption option is available only for S3 Accounts (not for Azure Accounts) with Snowflake.

If Staging Location is set to Internal, and when Data source is Input view, the Server Side Encryption and Server-Side KMS Encryption options are not supported for Snowflake snaps:

This happens because Snowflake encrypts loading data in its internal staging area and does not allow the user to specify the type of encryption in the PUT API. Learn more: Snowflake PUT Command Documentation.

KMS key

Default Value: N/A
Example: <Encrypted>

String/Expression

N/A

Specify the KMS key that you want to use for S3 encryption. Learn more about the KMS key: AWS KMS Overview and Using Server Side Encryption.

This property applies only when you select Server-Side KMS Encryption in the Encryption Type field above.

Buffer size (MB)

Default Value: 10MB
Example: 20MB

String/Expression

N/A

Specify the data in MB to be loaded into the S3 bucket at a time. This property is required when bulk loading to Snowflake using AWS S3 as the external staging area.

Minimum value: 6 MB

Maximum value: 5000 MB

S3 allows a maximum of 10000 parts to be uploaded so this property must be configured accordingly to optimize the bulk load. Refer to Upload Part for more information on uploading to S3. 

Key columns*

Default Value: None 
Example: EmpID

Specify the column to use for existing entries in the target table.

The Key columns field does not support values from the upstream input document when the expression button is enabled. The value must be defined for all inputs and remain constant throughout the Snap execution. Therefore, the Snap supports expressions to include only constant pipeline parameters. If you are using a function to get or transform the value from a pipeline parameter, that value must evaluate to a value that is not null and is not empty at the time of pipeline execution.

Delete Upsert Condition

Default: N/A

String

N/A

Delete Upsert Condition when true, causes the case to be executed.

Preserve case sensitivity

Default Value: Deselected

Checkbox

N/A

Select this check box to preserve the case sensitivity of the column names.

  • If you do not select Preserve case sensitivity, the input documents are loaded to the target table if the key names in the input documents match the target table column names ignoring the case.

  • If you include a second input view, selecting Preserve case sensitivity has no effect on the column names of the target table, because Snap uses the metadata from the second input view.

Manage Queued Queries

Default Value: Continue to execute queued queries when the Pipeline is stopped or if it fails
Example: Cancel queued queries when the Pipeline is stopped or if it fails

Dropdown list

N/A

Choose an option to determine whether the Snap should continue or cancel the execution of the queued queries when the pipeline stops or fails.

If you select Cancel queued queries when the Pipeline is stopped or if it fails, then the read queries under execution are canceled, whereas the write type of queries under execution are not canceled. Snowflake internally determines which queries are safe to be canceled and cancels those queries.

Load empty strings

Default value:  Selected

Checkbox

N/A

If selected, empty string values in the input documents are loaded as empty strings to the string-type fields. Otherwise, empty string values in the input documents are loaded as null. Null values are loaded as null regardless.

Additional Options

On Error

Default Value: ABORT_STATEMENT
Example: CONTINUE

Dropdown list

N/A

Select an action to perform when errors are encountered in a file. The available actions are:

  • ABORT_STATEMENT: Aborts the COPY statement if any error is encountered. The error will be thrown from the Snap or routed to the error view.

  • CONTINUE: Continues loading the file. The error will be shown as a part of the output document.

  • SKIP_FILE: Skips file if any errors encountered in the file.

  • SKIP_FILE_*error_limit*: Skips file when the number of errors in the file exceeds the number specified in Error Limit.

  • SKIP_FILE_*error_percent_limit*%: Skips file when the percentage of errors in the file exceeds the percentage specified in Error percentage limit.

Error Limit

Default Value: 0
Example: 3

Integer

Appears when you select SKIP_FILE_*error_limit* for On Error.

Specify the error limit to skip file. When the number of errors in the file exceeds the specified error limit or when SKIP_FILE_number is selected for On Error.

Error Percentage Limit

Default Value: 0
Example: 1

Integer

Appears when you select SKIP_FILE_*error_percent_limit*% 
for On Error.

Specify the percentage of errors  to skip file. If the file exceeds the specified percentage when SKIP_FILE_number% is selected for On Error

Snap Execution

Default Value: Execute only
Example: Validate & Execute

Dropdown list

N/A

Select one of the three modes in which the Snap executes. Available options are:

  • Validate & Execute: Performs limited execution of the Snap, and generates a data preview during Pipeline validation. Subsequently, performs full execution of the Snap (unlimited records) during Pipeline runtime.

  • Execute only: Performs full execution of the Snap during Pipeline execution without generating preview data.

  • Disabled: Disables the Snap and all Snaps that are downstream from it.

Troubleshooting

Error

Reason

Resolution

Cannot lookup a property on a null value.

The value referenced in the Key Column field is null.

This Snap does not support values from an upstream input document in the Key columns field when the expression button is enabled.

Update the Snap settings to use an input value from pipeline parameters and run the pipeline again.

Data can only be read from Google Cloud Storage (GCS) with the supplied account credentials (not written to it).

Snowflake Google Storage Database accounts do not support external staging when the Data source is the Input view.

Data can only be read from GCS with the supplied account credentials (not written to it).

Use internal staging if the data source is the input view or change the data source to staged files for Google Storage external staging.

Example

Upsert records

This example Pipeline demonstrates how you can efficiently update and delete data (rows) using the Key Column ID field and Upsert Delete condition. We use Snowflake - Bulk Upsert Snap to accomplish this task.

First, we configure the Mapper Snap with the required details to pass them as inputs to the downstream Snap.

After validation, the Mapper Snap prepares the output as shown below to pass to the Snowflake Bulk - Upsert Snap.

Next, we configure the Snowflake - Bulk Upsert Snap to:

  • Upsert the existing row for P_ID column, (so, we provide P_ID in the Key column field).

  • Delete the rows where the FIRSTNAME is snaplogic in the target table, (so, we specify FIRSTNAME = 'snaplogic' in the Delete Upsert Condition field).

After execution, this Snap inserts a new record into the existing row for the P_ID key column in Snowflake.

Inserted Records Output in JSON

Inserted Records in Snowflake

Upon execution, if the Delete Upsert condition is true, the Snap deletes the records in the target table as shown below.

Output in JSON

Deleted Record in Snowflake

Download this Pipeline.

Bulk load records

In the following example, we update a record using the Snowflake Bulk Upsert Snap. The invalid records which cannot be inserted will be routed to an error view. 

In the pipeline execution:

  1. The Mapper (Data) Snap maps the record details to the input fields of Snowflake Upsert Snap:

  2. Snowflake Upsert Snap updates the record using the "PRASANNA"."Persons" table object:

  3. Successful execution of the pipeline displays the following data preview (Note that one input has been made and two records with the name P_ID have been updated within the Persons table.):

Downloads 

Important steps to successfully reuse Pipelines

  1. Download and import the pipeline into the SnapLogic application.
  2. Configure Snap accounts as applicable.
  3. Provide pipeline parameters as applicable.

  File Modified
No files shared here yet.

Snap Pack History

 Click here to expand...

Release 

Snap Pack Version

Date

Type

  Updates

November 2024

main29029

Stable

Updated and certified against the current SnapLogic Platform release.

August 2024

438patches28811

Latest

Fixed an issue with the Snowflake Snaps where the Snaps intermittently displayed an error requesting to rollback/commit.

August 2024

438patches28667

Latest

Added the Snowflake - Snowpipe Streaming Snap that enables data insertion into Snowflake using the Snowpipe Streaming API. It allows for continuous data ingestion into Snowflake tables as soon as data becomes available.

August 2024

main27765

Stable

Upgraded the org.json.json library from v20090211 to v20240303, which is fully backward compatible.

May 2024

437patches27531

Latest

Enhanced the Snowflake - Select Snap with Time Travel query support that enables you to access historical data at any point within a defined period. This includes restoring data-related objects, duplicating and backing up data, and analyzing data usage.

May 2024

437patches27156

Latest

May 2024

437patches26821

Latest

  • Added Snowflake - Vector Search Snap to enable advanced vector-based queries as part of the Select statement.

  • Enhanced the Snowflake Snap Pack with full vector support, including schema generation and output compatibility with vector data types.

May 2024

437patches26508

Latest

May 2024

main26341

Stable

  • Enhanced the Snowflake Bulk Load and Snowflake Insert Snaps to support iceberg tables. Iceberg tables partition and organize data across multiple nodes to distribute the workload and speed up data processing.

  • Updated the Delete Condition (Truncates a Table if empty) field in the Snowflake - Delete Snap to Delete condition (deletes all records from a table if left blank) to indicate that all entries will be deleted from the table when this field is blank, but no truncate operation is performed.

February 2024

436patches25630

Latest

  • Enhanced the Snowflake Bulk Load Snap to fetch the file format types based on the selected Datasource. The Snap now supports only CSV and NONE file format types (as a Suggestions list) when the Datasource is Input view.

  • Enhanced the Snowflake Bulk Upsert Snap to support only CSV and NONE file format types (as a Suggestions list).

February 2024

main25112

Stable

Updated and certified against the current SnapLogic Platform release.

November 2023

435patches24865

Latest

Fixed an issue across the Snowflake Snaps that populated all suggestions for the Schema and Table Names existing in the configured Snowflake Account. Now, the Snaps only populate suggestions related to the database configured in the Account.

November 2023

435patches24110

Latest

Added a lint warning to the Snowflake-Bulk Load Snap that recommends users to select the Purge checkbox when the Data source is input view and the Staging location is External.

November 2023

main23721

Stable

The Snowflake Snap Pack is now bundled with the default Snowflake JDBC driver v3.14.

August 2023

434patches23541

Latest

Fixed an issue with the Snowflake-Bulk Load Snap where the Snap wrote irrelevant errors to the error view when both of the following conditions occurred:

  • The inserted data was not in the same order as the table schema.

  •  The Staging location was Internal.

Now, the Snap  writes the correct errors to the error view.

August 2023

main22460

Stable

The Snowflake - Execute Snap now includes a new Query type field. When Auto is selected, the Snap tries to determine the query type automatically.

May 2023

433patches21890

Latest

  • Enhanced the Snowflake - Bulk Load and Snowflake - Bulk Upsert Snaps with a new Flush chunk size (in bytes) field. When using internal staging, data is written to a temporary chunk file. When the file exceeds the specified size, it is copied to the Snowflake stage and deleted.

  • Added validation of external storage credentials for Snowflake S3 Accounts.

May 2023

433patches21370

Latest

  • Fixed an issue where the output document was missing information about the error records count and the reason for the error.

  • Fixed an issue that caused stored procedure executions to fail in the Snowflake - Multi Execute Snap.

May 2023

main21015

Stable

  • Snaplogic is specified as a partner tag in all requests going to Snowflake, making it easier for Snowflake to identify the requests coming from Snaplogic.

  • The default JDBC driver JAR for the Snowflake Snap Pack is upgraded to version 3.13.28 to support the GEOMETRY data type.

February 2023

432patches20906

Latest

  • The default JDBC driver JAR for the Snowflake Snap Pack is upgraded to version 3.13.28 to support the GEOMETRY data type.

  • Fixed an issue with the Snowflake Multi Execute Snap to enable processing of large amounts of data (more than 2 billion records).

  • Fixed a column mismatch issue that was caused by an assumed null AWS token that could potentially be an empty string.

February 2023

432patches20266

Latest

Fixed an issue with the Snowflake - Bulk Load Snap that resulted in lowercase (or mixed case) column names when creating a new table under specific conditions. The new Create table with uppercase column names checkbox addresses this issue.

February 2023

432patches20120

 

Latest

The Snowflake Bulk Load, Bulk Upsert, and Unload Snaps now support expressions for the Staging location field.

February 2023

main19844

 

Stable

  • The JDBC Driver (JAR file) setting in all the Snowflake Accounts is optional. If you do not provide a JDBC driver, the account uses the default Snowflake JDBC driver JAR V3.13.25 that is bundled with the Snap Pack.

November 2022

431patches19581

 

Latest

  • The Snowflake SCD2 Snap works as expected and does not display the Cannot find input data related to the output record error when you use the TIMESTAMPNTZ data type columns as Natural key fields.

  • Fixed the memory leakage issue ( because of Key/pair authentication that occupied some memory)  in the Snowflake Snaps. The performance of the Snowflake Snaps is optimized—there is less memory consumption, and exact memory dumps before and after the Pipeline execution.

November 2022

431patches19454

 

Latest

The Snowflake Snap Pack supports geospatial data types. As the Snowflake Snap Pack requires using our custom Snowflake JDBC driver for full support of all data types, contact support@snaplogic.com for details.

November 2022

431patches19220

 

Latest

The Snowflake S3 OAuth2 Account now support expressions for external staging fields.

November 2022

431patches19220

 

Latest

  • The Snowflake - Bulk Upsert Snap no longer supports values from an upstream input document in the Key columns field when the expression button is enabled. The value must remain constant throughout the Snap’s execution and now limits support for expressions to include only Pipeline parameters.

    To avoid breaking your existing Pipelines that use a value from the input schema in the Key columns field, update the Snap settings to use an input value from Pipeline parameters instead.

  • The Snowflake Bulk Load Snap no longer displays the validation warning in the following cases:

    • When you configure the Validation Mode as NONE and Validation Errors Type as Do not show errors in the Snap, or

    • When Validation Mode is RETURN_n_ROWS, RETURN_ERRORS, or RETURN_ALL_ERRORS.

November 2022

main18944

 

Stable

  • The Snowflake - Bulk Load Snap now routes the 403-Access Denied error to the error view (when the error view is enabled) when you do not have DeleteObject permission to delete the temporary files in the S3 bucket.

  • Fixed the following issues in the Snowflake - Bulk Load Snap:

    • The Snap failed when the Preserve case sensitivity field was not selected and the column names in the input document and the target table were the same but the cases were different. Now, the Snap does not fail and loads the data correctly to the target table.

    • The Snap created a table when it did not exist with multiple columns in the target table when there were two input views, but the column name cases were different in both views. Now, the Snap creates the table from the metadata provided by the second input view and loads the data into the target table from the first input view.

  • The Snowflake Insert Snap now creates the target table only from the table metadata of the second input view when the following conditions are met:

    • The Create table if not present checkbox is selected.

    • The target table does not exist.

    • The table metadata is provided in the second input view.

November 2022

430patches18911

 

Latest

Because of performance issues, all Snowflake Snaps now ignore the Cancel queued queries when pipeline is stopped or if it fails option for Manage Queued Queries, even when selected. Snaps behave as though the default Continue to execute queued queries when the Pipeline is stopped or if it fails option were selected.

October 2022

430patches18781

 

Latest

The Snowflake Insert and Snowflake Bulk Upsert Snaps now do not fail with the SQL Operation failed error, when you insert the uppercase key name into a mixed-case column if the Preserve case sensitivity checkbox is deselected.

The Snowflake Bulk Load Snap now works as expected when you configure On Error with SKIP_FILE_*error_percent_limit*% and set the Error Percent Limit to more than the percentage of rows with invalid data in the CSV file.

October 2022

430patches18432

 

Latest

The Snowflake Bulk Load Snap now has a Validation Errors Type dynamic field, which provides options for displaying validation errors. You can now choose Aggregate errors per row to display a summary view of errors.

September 2022

430patches17962

 

Latest

The Snowflake Bulk Load Snap now triggers the metadata query only once even for invalid input, thereby improving the performance of Snap.

September 2022

430patches17894

 

Latest

The Snowflake Select Snap now works as expected when the table name is dependent on an upstream input.

August 2022

430patches17748

 

Latest

Fixes in the Snowflake Bulk Load Snap:

  • The Snap now routes the 403-Access Denied error to the error view (when the error view is enabled) when you do not have DeleteObject permission to delete the temporary files in the S3 bucket.

  • The Snap now checks for the expiration time whenever it receives a new access token for refreshing the token in Snowflake OAuth2 account.

August 2022

430patches17377

 

Latest

  • The following issues are fixed in the Snowflake - Bulk Load Snap:

    • The Snap successfully loads the data even when you deselect the Preserve case- sensitivity checkbox and the source and target column names are similar, but with different casing.

    • When the Snap has two input views, it creates the target table from the metadata provided by the second input view and loads the data into the target table from the first input view.

  • The Snowflake Select Snap now validates successfully even when you set the Table Name field with an expression.

August 2022

main17386

 

Stable

The following Snowflake Accounts support Key Pair Authentication.

4.29 Patch

429patches16478

 

Latest

Fixed an issue with Snowflake - Bulk Load Snap where Snap failed with 301 status code - SnapDataException.

4.29 Patch

429patches16458

 

Latest

  • Fixed the following issues with the Snaps that use Snowflake Accounts Key Pair Authentication.

    • The Snaps were not able to read the private keys provided as Pipeline parameters, because the key was malformed by replacing the line breaks with whitespace characters. Now, the private keys are properly formed and read when provided as Pipeline parameters.

    • The Snap displayed a Null Point Exception error when you pass an invalid key into the Private Key field like test. Now, the Snap displays a valid error message.

  • Fixed an issue with the Snaps due to an upgrade made to HikariCP that caused a conditional failure and prevented the downstream Snaps to supply the parameters to the stored procedure.  

  • Fixed an issue with Snowflake Execute and Snowflake Select Snaps where the Snap displayed an exception error when the FLOAT or DOUBLE datatype column has the value as NaN (Not a Number). Now the output value is written as-is (NaN) for FLOAT or DOUBLE datatype column and does not display an exception.

  • Enhanced the following Snowflake Accounts with support for Key Pair Authentication.

  • Removed support for unencrypted Private Keys to enforce security in Snowflake accounts.

4.29

main15993

 

Stable

Upgraded with the latest SnapLogic Platform release.

4.28 Patch

428patches15236

 

Latest

  • Updated the Snowflake - Bulk Upsert Snap for the following:

    • The Snap displayed a incorrect resolution when the length of the value exceeded the value defined in a column.

    • The Snap failed with a NullPointer Exception when no value was provided for Error Limit field and Error percentage limit fields.

    • The Snap displayed an incorrect error message when S3 details were not provided.

    • Added the On Error dropdown list, where you can select an action to perform when the Snap encounters errors in a file.

  • Removed the Username field in the following accounts to allow reusing of accounts among different users:

  • The following enhancements were done for Snowflake - Bulk Load Snap:

    • The Snap skips AWS account validation if S3 Storage Integration property is provided.

    • The Snap skips the validate command if S3 Storage Integration property is not provided

    • The Snap overrides the storage integration specified in the account settings with the storage integration specified in the Snap settings.

  • Enhanced the Snowflake - Unload Snap to skip AWS account validation if S3 Storage Integration property is provided.

4.28

main14627

 

Stable

4.27

427patches12999

 

Latest

Enhanced the Snowflake SCD2 Snap to support Pipeline parameters for Natural key and Cause-historization fields.

4.27

main12833

 

Stable

  • Enhanced the Snowflake S3 Database Account to skip batch execution, when the Batch size value is one. When the Batch size is greater than one, the batch is executed.

  • Added the following truncate options to Snowflake - Bulk Load Snap to truncate existing data before performing data load to more efficiently transfer data where possible.

    • Truncate data: Truncates existing data before performing data load. 

    • Truncate columns: Truncates column values that are larger than the maximum column length in the table.

  • Enhanced the Snowflake - Execute Snap to invoke stored procedures.

4.26 Patch

426patches11469

 

Latest

Fixed an issue with Snowflake Insert and Snowflake Bulk Load Snaps where the schema names or database names containing underscore (_) caused the time out of Pipelines.

4.26

main11181

 

Stable

  • Enhanced Snowflake - Lookup and Snowflake SCD2 Snaps with the Input Date Format field to select from the following two options:

    • Continue to execute the snap with the given input Date format

    • Auto Convert the format to Snowflake default format

  • Added a new account type Snowflake Google Storage Database to connect to Google Cloud Storage to load data.

  • Added support for all existing Snowflake Snap accounts to connect to a Snowflake instance hosted on the Google Cloud Platform

4.25

425patches10190

Latest

Enhanced the Snowflake S3 Database and Snowflake S3 Dynamic accounts with a new field S3 AWS Token that allows you to connect to private and protected Amazon S3 buckets.

4.25

main9554

 

Stable

Upgraded with the latest SnapLogic Platform release.

4.24 Patch

424patches8905

 

Latest

Enhanced the Snowflake - Bulk Load Snap to allow transforming data using a new field Select Query before loading data into the Snowflake database. This option enables you to query the staged data files by either reordering the columns or loading a subset of table data from a staged file. This Snap supports CSV and JSON file formats for this data transformation.

4.24

main8556

Stable

Enhanced the Snowflake - Select Snap to return only the selected output fields or columns in the output schema (second output view) using the Fetch Output Fields In Schema check box. If the Output Fields field is empty all the columns are visible.

4.23 Patch

423patches7905

 

Latest

Fixed the performance issue in the Snowflake - Bulk Load Snap while using External Staging on Amazon S3.

4.23

main7430

 

Stable

4.22 Patch 

422patches7246

 

Latest

Fixed an issue with the Snowflake Snaps that failed while displaying similar error in the Snowflake URL connection: message, javax.management.MalformedObjectNameException: Invalid character '=' in value part of property, repeatedly when there is "=" or ":" .

4.22 Patch

422patches6849

 

Latest

4.22

main6403

 

Stable

Updated with the latest SnapLogic Platform release.

4.21 Patch

421patches6272

 

Latest

Fixes the issue where Snowflake SCD2 Snap generates two output documents despite no changes to Cause-historization fields with DATE, TIME and TIMESTAMP Snowflake data types, and with Ignore unchanged rows field selected.

4.21 Patch

421patches6144

 

Latest

  • Fixed the following issues with DB Snaps:

    • The connection thread waits indefinitely causing the subsequent connection requests to become unresponsive.

    • Connection leaks occur during Pipeline execution.

  • Fixed the exception RefCnt has gone negative across the Snaps in the Snowflake Snap Pack.

4.21 Patch 

db/snowflake8860

 

Latest

Added a new field, Handle Timestamp and Date Time Data, to Snowflake Lookup Snap. This field enables you to decide whether the Snap should translate UTC time to your local time and the format of the Date Time data.

4.21 Patch

MULTIPLE8841

 

Latest

Fixed the connection issue in Database Snaps by detecting and closing open connections after the Snap execution ends. 

4.21

snapsmrc542

 

Stable

Upgraded with the latest SnapLogic Platform release.

4.20 Patch 

db/snowflake8800

 

Latest

  • Certifies the Snowflake Snap Pack against JDBC Driver version 3.12.3. Snowflake Execute and Multi-Execute Snaps may break existing Pipelines if the JDBC Driver is updated to a newer version.

  • With the updated JDBC driver (version 3.12.3), the Snowflake Execute and Multi-Execute Snaps' output displays a Status of "-1" instead of "0" without the Message field upon successfully executing DDL statements. If your Pipelines use these Snaps and downstream Snaps use the Status field's value from these, you must modify the downstream Snaps to proceed on a status value of -1 instead of 0.

  • This change in the Snap behavior follows from the change introduced in the Snowflake JDBC driver in version 3.8.1:
    "Statement.getUpdateCount() and PreparedStatement.getUpdateCount() return the number of rows updated by DML statements. For all other types of statements, including queries, they return -1."

4.20 Patch 

db/snowflake8758

 

Latest

Re-release of fixes from db/snowflake8687 for 4.20: Fixes the Snowflake Bulk Load snap where the Snap fails to load documents containing single quotes when the Load empty strings checkbox is not selected.

4.20

snapsmrc535

 

Stable

Upgraded with the latest SnapLogic Platform release.

4.19 Patch 

db/snowflake8687

 

Latest

Fixed the Snowflake Bulk Load snap where the Snap fails to load documents containing single quotes when the Load empty strings checkbox is not selected.

4.19 Patch 

db/snowflake8499

 

Latest

Added the property Handle Timestamp and Date Time Data to Snowflake - Execute and Snowflake - Select Snaps. This property enables you to decide whether the Snap should translate UTC time to your local time.

4.19 Patch 

db/snowflake8412

 

Latest

Fixed an issue with the Snowflake - Update Snap wherein the Snap is unable to perform operations when:

  • An expression is used in the Update condition property.

  • Input data contain the character '?'.

4.19

snaprsmrc528

 

Stable

  • Added a new field-set, Auto Historization Query, in the Snowflake SCD2 Snap to support auto-historization of column data. With this enhancement you can detect whether the incoming record is a historical event or a current event.  

  • Raised the minimum buffer size in the Snowflake - Bulk Upsert and Snowflake - Bulk Load Snaps to 6 MB.

4.18 Patch 

db/snowflake8044

 

Latest

Fixed an issue with the Snowflake - Select Snap wherein the Snap converts the Snowflake-provided timestamp value to the local timezone of the account.

4.18 Patch 

db/snowflake8044

 

Latest

Enhanced the Snap Pack to support AWS SDK 1.11.634 to fix the NullPointerException issue in the AWS SDK. This issue occurred in AWS-related Snaps that had HTTP or HTTPS proxy configured without a username and/or password. 

4.18 Patch

MULTIPLE7884

 

Latest

Fixed an issue with the PostgreSQL grammar to better handle the single quote characters.

4.18 Patch 

db/snowflake7821

 

Latest

Fixed an issue with the Snowflake - Execute Snap wherein the Snap is unable to support the '$' character in query syntax.

4.18 Patch

MULTIPLE7778

 

Latest

Updated the AWS SDK library version to default to Signature Version 4 Signing process for API requests across all regions.

4.18 Patch 

db/snowflake7739

 

Latest

  • Fixed an issue with the Snowflake - Bulk Upsert Snap wherein the Snap fails when using a pipeline parameter in Key columns.

  • Fixed an issue with the Snowflake - Unload Snap wherein the Snap does not abort the query when you stop the Pipeline execution.

4.18

snapsmrc523

 

Stable

Added the Use Result Query property to the Multi Execute Snap, which enables you to write results to an output view.

4.17

ALL7402

 

Latest

Pushed automatic rebuild of the latest version of each Snap Pack to SnapLogic UAT and Elastic servers.

4.17 Patch 

db/snowflake7396

 

Latest

Fixed an issue wherein bit data types in the Snowflake - Select table convert to true or false instead of 0 or 1.

4.17 Patch 

db/snowflake7334

 

Latest

Added AWS Server-Side Encryption support for AWS S3 and AWS KMS (Key Management Service) for Snowflake Bulk Load, Snowflake Bulk Upsert, and Snowflake Unload Snaps.

4.17

snapsmrc515

 

Latest

  • Fixed an issue with the Snowflake Execute Snap wherein the Snap would send the input document to the output view even if the Pass through field is not selected in the Snap configuration. With this fix, the Snap sends the input document to the output view, under the key original, only if you select the Pass through field.

  • Added the Snap Execution field to all Standard-mode Snaps. In some Snaps, this field replaces the existing Execute during preview check box.

4.16 Patch 

db/snowflake6945

 

Latest

Fixed an issue with the Snowflake Lookup Snap failing when Date datatype is used in JavaScript functions.

4.16 Patch 

db/snowflake6928

 

Latest

Added support for file format options for input data from upstream Snaps, to the Snowflake Bulk Load Snap.

4.16 Patch 

db/snowflake6819

 

Latest

  • Snowflake Bulk Load: Added new property, Buffer size (MB). Configure this to specify the size limit of each buffer when writing to external staging systems such as S3. 

  • Fixed an issue with the Lookup Snap passing data simultaneously to output and error views when some values contained spaces at the end.

4.16

snapsmrc508

 

Stable

  • Snowflake Account: Added the ability to use SnapLogic to securely connect to and query a Snowflake instance using Azure Blob as its storage layer. 

  • Snowflake Account: Added support for Snowflake JDBC driver JAR version 3.6.17. 

  • Snowflake Unload, Bulk Load, and Bulk Upsert: Updated the Snaps to enable SnapLogic users to successfully connect to a Snowflake instance to query, bulk load, and unload data from Azure Blob storage.

4.15

snapsmrc500

 

Stable

  • Added two new Snaps, Snowflake - Multi Execute, and Snowflake SCD2. Snowflake - Multi Execute is used for executing multiple DDL and DML queries on the Snowflake DB. Snowflake SCD2 is used for Type 2 field historization. 

  • Enhanced the Snowflake Bulk Upsert Snaps to improve the Snaps performance.

  • Enhanced the Snowflake Snap Pack to reflect Azure certification.

4.14

snapsmrc490

 

Stable

Upgraded with the latest SnapLogic Platform release.

4.13

snapsmrc486

 

Stable

Upgraded with the latest SnapLogic Platform release.

4.12

snapsmrc480

 

Stable

Upgraded with the latest SnapLogic Platform release.

4.11 Patch

MULTIPLE4377

 

Latest

Fixed a document call issue that was slowing down the Snowflake Bulk Load Snap.

4.11 Patch 

db/snowflake4283

 

Latest

Snowflake Bulk Load - Fixed an issue by adding PUT command to the list of DDL command list for Snowflake.

4.11 Patch 

db/snowflake4273

 

Latest

Snowflake Bulk Load - Resolved an issue with Snowflake Bulk Load Delimiter Consistency (comma and newline).

4.11

snapsmrc465

 

Stable

Upgraded with the latest SnapLogic Platform release.

4.10 Patch

snowflake4133

Latest

Updated the Snowflake Bulk Load Snap with Preserve case sensitivity property to preserve the case sensitivity of column names.

4.10

snapsmrc414

 

Stable

  • Updated the Snowflake Bulk Load Snap with Load empty strings property for the empty string values in the input documents to be loaded as empty strings to the string-type fields.

  • Updated the Snowflake Bulk Load Snap with Table Columns to support the order of the entries on the staged files that contain a subset of the columns in the Snowflake table.

  • Added the property Use Result Query to view the output preview field with a result statement.

  • Tested for the JDBC driver JAR version 3.1.1 on the Database and the Dynamic accounts.

4.9.0 Patch

snowflake3234

 

Latest

Enhanced Snowflake - Execute Snap results to include additional details

4.9.0 Patch

snowflake3125

 

Latest

Addressed an issue in Snowflake Bulk Load where the comma character in a value is not escaped.

4.9

snapsmrc405

 

Stable

JDBC Driver Class property added to enable the user to custom configure the JDBC driver in the Database and the Dynamic accounts.

4.8.0 Patch

snowflake2760

 

Latest

Potential fix for JDBC deadlock issue.

4.8.0 Patch

snowflake2739

 

Latest

Addressed an issue with the Snowflake schema not correctly represented in the Mapper Snap.

4.8

snapsmrc398

 

Stable

  • Info tab added to accounts.

  • Database accounts now invalidate connection pools if account properties are modified and login attempts fail.

  • Enhanced the default count of input and output view (UI) behavior of the Snaps for better user experience. Note that the Snowflake Select, Table List and Lookup Snap views remain unchanged.


  • No labels