Skip to end of banner
Go to start of banner

Snowflake - Bulk Load

Skip to end of metadata
Go to start of metadata

You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 76 Next »

On this Page

Snap type:

Write


Description:

This Snap executes a Snowflake bulk load, writing data into an Amazon S3 bucket or a Microsoft Azure Storage Blob.

The Snap creates temporary files in JCC when the staging location is internal and data source is Input view. These temporary files are removed automatically once the Pipeline completes execution.

  • Expected input: Documents containing the data to be uploaded to the target location.
  • Expected output: A document containing the original document provided as input and the status of the bulk upload 
  • Expected upstream Snaps: Any document that offers documents. For example, JSON Generator or Binary to Document Snap.
  • Expected downstream Snaps: Any Snap that accepts document. For example, Mapper or Snowflake Execute Snap.

Prerequisites:


You should have minimum permissions on the database to execute Snowflake Snaps. To understand if you already have them, you must retrieve the current set of permissions. The following commands enable you to retrieve those permissions.

SHOW GRANTS ON DATABASE <database_name>
 
SHOW GRANTS ON SCHEMA <schema_name>
 
SHOW GRANTS TO USER <user_name>
Security Prerequisites: You should have the following permissions in your Snowflake account to execute this Snap: 

  • Usage (DB and Schema): Privilege to use the database, role, and schema.
  • Create table: Privilege to create a temporary table within this schema.

The following commands enable minimum privileges in the Snowflake Console:

grant usage on database <database_name> to role <role_name>;
grant usage on schema <database_name>.<schema_name>;
 
grant "CREATE TABLE" on database <database_name> to role <role_name>;
grant "CREATE TABLE" on schema <database_name>.<schema_name>;

For more information on Snowflake privileges, refer to Access Control Privileges.

The below are mandatory when using an external staging location:

When using an Amazon S3 bucket for storage:

  • The Snowflake account should contain S3 Access-key ID, S3 Secret key, S3 Bucket and S3 Folder.
  • The Amazon S3 bucket where the Snowflake will write the output files must reside in the same region as your cluster.

When using a Microsoft Azure storage blob:

  • A working Snowflake Azure database account.
Internal SQL Commands

This Snap uses the following Snowflake commands internally:

  • COPY INTO - Enables loading data from staged files to an existing table.
  • PUT - Enables staging the files internally in a table or user stage.
Support and limitations:
  • Ultra Pipelines: Works in Ultra Pipelines if batching is disabled.
  • Special character '~': Not supported if it is there in temp directory name for Windows. It is reserved for user's home directory.
  • Snowflake provides the option to use the Cross Account IAM into the external staging. You can adopt the cross account access through option Storage Integration. With this setup, you don’t need to pass any credentials around, and access to the storage only using the the named stage or integration object. For more details: Configuring Cross Account IAM Role Support for Snowflake Snaps

  • Snowflake Bulk Load expects column order should be as like table from upstream snaps otherwise it will result in failure of data validation.
Account: 

This Snap uses account references created on the Accounts page of SnapLogic Manager to handle access to this endpoint. See Snowflake Account for information on setting up this type of account.


Views:
Input

This Snap has one document input view by default. 

A second view can be added for metadata for the table as a document so that the target absent table can be created in the database with a similar schema as the source table. This schema is usually from the second output of a database Select Snap. If the schema is from a different database, there is no guarantee that all the data types would be properly handled.

Output

This Snap has at most one output view. If an output view is available, then the original document that was used to create the statement will be output with the status of the insert executed.

Error

This Snap has at most one error view and produces zero or more documents in the view.


Settings

Label


Required. The name provided for the instance. You can modify this to be more specific, especially if you have more than one of the same Snap in your pipeline.

Schema name



Required. The database schema name. In case it is not defined, then the suggestion for the Table Name will retrieve all tables names of all schemas. The property is suggestible and will retrieve available database schemas during suggest values.

The values can be passed using the pipeline parameters but not the upstream parameter.

Default value:  [None]


Table name

Required. The name of the table to execute bulk load operation on.

The values can be passed using the pipeline parameters but not the upstream parameter.

Default value:  [None]


Create table if not present


Whether the table should be automatically created if not already present.

Default value:  Not selected

  • Using this option creates a table with all columns of type STRING.
  • Due to implementation details a newly created table is not visible to subsequent database Snaps during runtime validation. If you wish to immediately use the newly updated data you must use a child pipeline invoked via a Pipe Exec Snap.
  • Create table if not present works only when the set datasource = inputview in the Snap. When we enable Create table if not present and the datasource is set to a stage file, then the Snowflake Bulk Load Snap will throw a configuration exception saying "Failure: Invalid snap configuration, Reason: Table creation not supported for uploading existing files, Resolution: Please create the table if the table does not exist and retry".

This should not be used in production since there are no indexes or integrity constraints on any column and the default varchar() column is over 30k bytes.


Data source


The source from where the data is loaded from. The options available are Input view and Staged files.

When the option 'Input View' is selected, leave the Table Coumns field empty, and if the 'Staged files' option is selected, provide the column names for the Table Coumns to which the records are to be added.

Default value:  Input view

Preserve case sensitivity

If selected, the case sensitivity of the column names will be preserved.

Default value:  Not selected

Load empty strings

If selected, empty string values in the input documents are loaded as empty strings to the string-type fields. Otherwise, empty string values in the input documents are loaded as null. Null values are loaded as null regardless.

Default value:  Not selected

Staging Location


The type of staging location that is used for data loading.

The options available include:

  • External: Location that is not managed by Snowflake. The location should be an AWS S3 Bucket or Microsoft Azure Storage Blob. These credentials are mandatory while validating the Account.
  • Internal: Location that is managed by Snowflake.

Default value:  Internal


Target


An internal or external location where data can be loaded. If you selected External in the Staging Location field above, a staging area is created in S3 or Azure as required; else a staging area is created in Snowflake's internal location.

The staging field accepts the following input:

  • Named Stage: The name for user-defined named stage. This should be used only when a Staging location is set as Internal. 

Format: @<Schema>.<StageName>[/path]

  • Internal Path: The staging represent with path.

Format: @~/[path]

  • S3 Url: The external S3 URL that specifies an S3 storage.

Format: s3://[path]

  • Microsoft Azure Storage Blob URL: The external URL required to connect to the Microsoft Azure Storage.

  • Folder Name: Anything else including none input. This is regarded as a Folder name under the Internal Home Path(@~) if using internal staging or under the S3 bucket and folder specified in the Snowflake account.

The value for the expression has to be provided as a pipeline parameter and cannot be provided from the Upstream Snap for performance reasons when the property is used as an expression parameter.

Default value:  [None]


Storage Integration

The pre-defined storage integration which is used to authenticate the external stages.

The value for the expression has to be provided as a pipeline parameter and cannot be provided from the Upstream Snap for performance reasons when the property is used as an expression parameter.

Default value: None

Staged file list

List of the staged file(s) loaded to the target file. 


File name pattern


A regular expression pattern string, enclosed in single quotes, specifying the file names and /or path to match.

Default value:  [None]


File format object


Specify an existing file format object to use for loading data into the table. The specified file format object determines the format type such as CSV, JSON,XMLand AVRO, or other format options for data files.

Default value:  [None]


File Format type


Specify a predefined file format object to use for loading data into the table. The available file formats include CSV, JSON, XML and AVRO.
Default value: NONE


File Format option


Specify the file format option. Separate multiple options by using blank spaces and commas.

You can use various file format options including binary format which passes through in the same way as other file formats. See File Format Type Options for additional information.

Before loading binary data into Snowflake, you must specify the binary encoding format, so that the Snap can decode the string type to binary types before loading into Snowflake. This can be done by specifying the following binary file format:

BINARY_FORMAT=xxx (Where XXX = HEX|BASE64|UTF-8)

However, the file you upload and download must be in similar formats. For instance, if you load a file in HEX binary format, you should specify the HEX format for download as well.

Example: BINARY_FORMAT=UTF-8

Default value:  [None]

When using external staging locations

  • When loading numeric data from staged files, you must provide file format options depending upon the user's data.
  • Do not use the following file format options if you selected Input view in the Data source field.

    • FIELD_DELIMITER

    • RECORD_DELIMITER

    • FIELD_OPTIONALLY_ENCLOSED_BY

Table Columns

Conditional. This specifies the table columns to use in the Snowflake COPY INTO query. This only applies when the Data source is 'Staged files'. This configuration is useful when the staged files contain a subset of the columns in the Snowflake table. For example, if the Snowflake table contains columns A, B, C and D and the staged files contain columns A and D then the Table Columns field would be have two entries with values A and D. The order of the entries should match the order of the data in the staged files.

Default value:  [None]

If the Data source is selected with 'Input view', the snap would throw an error as displayed:

Select Query

Activates when you select Staged files in the Data source field.

Specify the SELECT query to transform data before loading into Snowflake database. 

The SELECT statement transform option enables querying the staged data files by either reordering the columns or loading a subset of table data from a staged file. For example, select $1:location, $1:dimensions.sq_ft, $1:sale_date, $1:price from @mystage/sales.json.gz t
This query loads the file sales.json from the internal stage mystage, (which stores the data files internally); wherein locationdimensions.sq_ft, and sale_date are the objects.

(OR)

select substr(t.$2,4), t.$1, t.$5, t.$4 from @mystage t
This query reorders the column data from the internal stage mystage before loading it into a table. The (SUBSTR), SUBSTRING function removes the first few characters of a string before inserting it.

We recommend you not to use temporary stage while loading your data.

Default value: N/A

Encryption type

Specifies the type of encryption to be used on the data. The available encryption options are:

  • None: Files do not get encrypted.
  • Server Side Encryption: The output files on Amazon S3 are encrypted with the server-side encryption.
  • Server-Side KMS Encryption: The output files on Amazon S3 are encrypted with Amazon S3-generated KMS key. 

Default value: No default value.

The KMS Encryption option is available only for S3 Accounts (not for Azure Accounts) with Snowflake.

If Staging Location is set to Internal, and when Data source is Input view, the Server Side Encryption and Server-Side KMS Encryption options are not supported for Snowflake snaps:

This happens because Snowflake encrypts loading data in its internal staging area and does not allow the user to specify the type of encryption in the PUT API (see Snowflake PUT Command Documentation.)

KMS key

The KMS key that you want to use for S3 encryption. For more information about the KMS key, see AWS KMS Overview and Using Server Side Encryption.

Default value: No default value.

This property applies only when you select Server-Side KMS Encryption in the Encryption Type field above.

Additional Options



Buffer size (MB)

This property is required when bulk loading to Snowflake using AWS S3 as the external staging area. It specifies the data in MB to be loaded into the S3 bucket, at a time. 

Minimum value: 5 MB

Maximum value: 5000 MB

Default value: 10 MB


S3 allows a maximum of 10000 parts to be uploaded so this property must be configured accordingly to optimize the bulk load. Refer to Upload Part for more information on uploading to S3. 

Manage Queued Queries

Select this property to decide whether the Snap should continue or cancel the execution of the queued Snowflake Execute SQL queries when you stop the pipeline.

If you select Cancel queued queries when pipeline is stopped or if it fails, then the read queries under execution are cancelled, whereas the write type of queries under execution are not cancelled. Snowflake internally determines which queries are safe to be cancelled and cancels those queries.

Default value: Continue to execute queued queries when pipeline is stopped or if it fails

On Error



Specifies an action to be performed when errors are encountered in a file. The available actions include:

  • ABORT_STATEMENT: Abort the COPY statement if any error is encountered. The error will be thrown from the snap or routed to the error view.
  • CONTINUE: Continue loading the file. The error will be shown as a part of the output document.
  • SKIP_FILE: Skip file if any errors encountered in the file.
  • SKIP_FILE_*error_limit*: Skip file when the number of errors in the file exceeds the number specified in Error Limit.
  • SKIP_FILE_*error_percent_limit*%: Skip file when the percentage of errors in the file exceeds the percentage specified in Error percentage limit.

Default value: ABORT_STATEMENT

Error Limit


Specifies to skip file when the number of errors in the file exceeds the specified error limit on when SKIP_FILE_number is selected for On Error. 

Default value: 0

Error percentage limit

 Specifies to skip file when the percentage of errors in the file exceeds the specified percentage when SKIP_FILE_number% is selected for On Error

Default value: 0

Size limit


Specifies the maximum size (in bytes) of data to be loaded.

 At least one file is loaded regardless of the value specified for SIZE_LIMIT unless there is no file to be loaded. A null value indicates no size limit.

Default value: 0

Purge


Specifies whether to purge the data files from the location automatically after the data is successfully loaded. 

Default value: Not Selected

Return Failed Only

Specifies whether to return only files that have failed to load while loading.

Default value: Not Selected 

Force


Specifies to load all files, regardless of whether they have been loaded previously and have not changed since they were loaded.

Default value:  Not Selected

Validation mode


This mode is useful for visually verifying the data before unloading it. The available options are:

NONE

RETURN_n_ROWS

RETURN_ERRORS

RETURN_ALL_ERRORS

Default value: NONE

Rows to return


Specifies the number of rows not loaded into the corresponding table. Instead, the data is validated to be loaded and returns results based on the validation option specified. It can be one of the following values: RETURN_n_ROWS | RETURN_ERRORS | RETURN_ALL_ERRORS

Default value: 0

Snap execution

Select one of the three modes in which the Snap executes. Available options are:

  • Validate & Execute: Performs limited execution of the Snap, and generates a data preview during Pipeline validation. Subsequently, performs full execution of the Snap (unlimited records) during Pipeline runtime.
  • Execute only: Performs full execution of the Snap during Pipeline execution without generating preview data.
  • Disabled: Disables the Snap and all Snaps that are downstream from it.

Instead of building multiple Snaps with inter dependent DML queries, it is recommended to use the Stored Procedure or the Multi Execute Snap.
In a scenario where the downstream Snap does depends on the data processed on an Upstream Database Bulk Load Snap, use the Script Snap to add delay for the data to be available.

For example, when performing a create, insert and a delete function sequentially on a pipeline, using a Script Snap helps in creating a delay between the insert and delete function or otherwise it may turn out that the delete function is triggered even before inserting the records on the table.

Examples

Loading Binary Data Into Snowflake

The following example Pipeline demonstrates how you can convert the staged data into binary data using the binary file format before loading into Snowflake database.

To begin with, we configure the Snowflake Execute Snap with this query: select * from "PUBLIC"."EMP2" limit 25 – this query reads 25 records from the Emp2 table.

Then, we configure the Mapper Snap with the output from the upstream Snap by mapping the employee details to the target table. Note that the Bio column is binary data type and the Text column is varbinary data type. Upon validation, the Mapper Snap passes the output with the given mappings (employee details) in the table.

Next, we configure the Snowflake - Bulk Load Snap to load the records into Snowflake. We set the File format option as BINARY_FORMAT=UTF-8 to enable the Snap to encode the binary data before loading.

Upon validation, the Snap loads the database with 25 employee records.

Finally, we connect the JSON Formatter Snap to the Snowflake - Bulk Load Snap to transform the binary data to JSON format, and finally write this output in S3 using the File Writer Snap.

Download this Pipeline.

Transforming Data Using Select Query Before Loading Into Snowflake

The following example Pipeline demonstrates how you can reorder the columns using the SELECT statement transform option before loading data into Snowflake database. We use the Snowflake - Bulk Load Snap to accomplish this task.

Prerequisite: You must create an internal or external stage in Snowflake before you transform your data. This stage is used for loading data from source files into the tables of Snowflake database.

To begin with, we create a stage using a query in the following format. Snowflake supports both internal (Snowflake) and external (Microsoft Azure and AWS S3) stages for this transformation. 

"CREATE STAGE IF NOT EXISTS "+_stageName+" url='"+_s3Location+"' CREDENTIALS = (AWS_KEY_ID='string' AWS_SECRET_KEY='string') "

This query creates an external stage in Snowflake pointing to S3 location with AWS credentials (Key ID and Secrete Key).

We recommend you not to use a temporary stage to prevent issues while loading and transforming your data.

Now, we add the Snowflake - Bulk Load Snap to the canvas and configure it to transform the data in the staged file SNAP7517_EXT_CSV.csv by providing the following query in the Select Query field:
"select t.$1,t.$4,t.$3,t.$4,t.$5,t.$6,t.$7 from @"+_stageName+" t"

You must provide the stage name along with schema name in the Select Query, else the Snap displays an error. For instance,
SELECT t.$1,t.$4,t.$3,t.$4,t.$5,t.$6,t.$7 from @mys3stage t", displays an error.
SELECT t.$1,t.$4,t.$3,t.$4,t.$5,t.$6,t.$7 from @<Schema Name>.<stagename> t", executes correctly.

Next, we connect a Snowflake Select Snap with the Snowflake - Bulk Load Snap to select the data from the Snowflake database. Upon validation we can view the transformed data in the output view.

Download this Pipeline.

 Snowflake Bulk Load Snap removes spaces from input documents

You can remove empty spaces from input documents. When you select Input view as Data Source, enter TRIM_SPACE=TRUE in the File Format Option field to remove empty spaces, if any.

Download this pipeline.

In this example, the Pipeline uses the following Snaps:

  • Data (JSON Generator): Generates a JSON document for the Snowflake - Bulk Load Snap in the Pipeline. In this example, the JSON document contains an empty space.
  • BL: Executes a Snowflake Bulk Load, writing data into an Amazon S3 bucket or a Microsoft Azure Storage Blob. Also, it enables you to remove all spaces from the input document.
  • Schema (JSON Generator): Provides the schema to interpret the document passed to the Snowflake - Bulk Load Snap.
  • Snowflake - Execute: Reads the newly uploaded document and enables you to check whether the spaces were removed as expected.

Data (JSON Generator)

Data streams from your database source, and you do not necessarily need a Snap to provide input documents. In this example, however, we use the JSON Generator Snap to provide the input document.

Input:

Output:

As you can see, the value listed against the key ACCOUNT_NAME has empty spaces in it.

Snowflake - Bulk Load

Input:

Notice that the File Format Option is TRIM_SPACE=TRUE.

Output:

Schema (JSON Generator)

Table schema is taken from your database source, and you do not necessarily need a Schema Snap to provide the table schema. In this example, however, we use the JSON Generator Snap to provide the table schema.

Input:


Output:

Snowflake - Execute

Input:

Output:

As you can see, the data no longer contains any spaces.


 Snowflake Bulk Load with Table Columns on Staged files

In this pipeline, the Snowflake Bulk Load Snap loads the records from a staged file 'employee.csv content' on to a table on Snowflake.

The staged file 'employee.csv content' is passed via the Upstream Snap:

The Snowflake Bulk Load Snap is configured with Data source as Staged files and the Table Columns added as ID, FIRSTNAME, CITY, to be loaded into a table "PRASANNA"."EMPLOYEE" on Snowflake.


The successful execution of the pipeline displays the below output preview:

If the 'employee.csv content' (Staged file) has the below details:

1,PRASANNA,Hyderabad
2,Aparna,hyderabad

Table Columns added are:
ID, FIRSTNAME, CITY

then the table, "PRASANNA"."EMPLOYEE" on Snowflake is as below:

The Snowflake table "PRASANNA"."EMPLOYEE"
Create table "PRASANNA"."EMPLOYEE" (ID int, FIRSTNAME varchar(30), LASTNAME varchar(30), CITY varchar(30), ADDRESS varchar(30), JOIN_DATE date)

Note the columns ID, FIRSTNAME and CITY are populated as provided and the LASTNAME, ADDRESS and JOIN DATE are null.

 Snowflake Bulk Load with Load empty strings property selected

In this pipeline, the Snowflake Bulk Load Snap loads the input document with empty string values as empty strings to the string-type fields. The values to be updated are passed via the upstream Mapper Snap and the Snowflake Execute Snap displays the output view of the table records.

The Mapper Snap passes the values to be updated to the table on Snowflake:

The Snowflake Bulk Load Snap with Load empty strings property selected:

The Snowflake Execute Snap runs the query to select the table 'TEST BULKLOAD' for the 'ID=1002':

The successful execution of the pipeline displays the below output preview:

The Name field is left empty. It means the empty strings are loaded as empty strings only.

The below screenshot displays the output preview (on a table with an ID=1003), wherein the Load empty strings property is not selected, and the the Name field has a value 'Null' in it:



 Snowflake Bulk Load Snap loads data from an Oracle Database

In this example, the data is loaded from the table "TECTONIC"."ADOBEPRASANNA_NEW1124"  using Oracle Select Snap to the target table "PRASANNA"."ADOBEDATA123"  using Snowflake Bulk Load Snap.

The Pipeline:

The Oracle Select Snap gets records in the table "TECTONIC"."ADOBEPRASANNA_NEW1124" and passes them to the Snowflake Bulk Load Snap: 

The Snowflake Bulk Load Snap loads records to table "PRASANNA"."ADOBEDATA123":

 

The output of the Snowflake Bulk Load Snap after executing the pipeline:  

 Loading Data from S3

This example demonstrates how you can use the Snowflake Bulk Snap to load files from an external staging location such as S3. It further shows the configuration required when loading numeric data.

Download this Pipeline. 

The Snowflake Bulk Load Snap has a minimum of 1 Input view. This is useful when the data source is Input view. Even though, within the scope of this Pipeline, the Snap does not require any input from an upstream Snap, the view cannot be disabled. 

Data is to be loaded from the staged file threecolumns.csv present in a S3 folder into the table "EXTR_PERF_01_SC"."THREECOLUMNS".

Below is a screenshot of the data in threecolumns.csv:

The Snowflake Bulk Load Snap is configured accordingly as follows:

Furthermore, since this data has numeric values, the Snowflake Bulk Load Snap is configured with the following file format options to handle any string/NULL values that may be present in the dataset: 

  • SKIP_HEADER = 1: Specifies that the first row is a header row so that all rows after the first row are loaded.
  • NULL_IF = "": Convert empty spaces to SQL NULL.
  • FIELD_OPTIONALLY_ENCLOSED_BY ="": Specifies the character used to enclose strings.

See Format Type Options for a detailed explanation of the above file format options.

Upon execution, the Snowflake Bulk Load Snap loads three rows:

To confirm that three rows were loaded, we use a Snowflake Execute Snap configured to count the number of rows in the target table:

Below is a preview of the output from the Snowflake Execute Snap. 

We can see that the count is 3, thereby confirming a successful bulk load.

You can also modify the SELECT query in the Snowflake Execute Snap to read the data in the table and thus verify the data loaded into the target table. 

 

Downloads

  File Modified
No files shared here yet.

Release 

Snap Pack Version

Date

Type

  Updates

November 2024

main29029

Stable

Updated and certified against the current SnapLogic Platform release.

August 2024

438patches28811

Latest

Fixed an issue with the Snowflake Snaps where the Snaps intermittently displayed an error requesting to rollback/commit.

August 2024

438patches28667

Latest

Added the Snowflake - Snowpipe Streaming Snap that enables data insertion into Snowflake using the Snowpipe Streaming API. It allows for continuous data ingestion into Snowflake tables as soon as data becomes available.

August 2024

main27765

Stable

Upgraded the org.json.json library from v20090211 to v20240303, which is fully backward compatible.

May 2024

437patches27531

Latest

Enhanced the Snowflake - Select Snap with Time Travel query support that enables you to access historical data at any point within a defined period. This includes restoring data-related objects, duplicating and backing up data, and analyzing data usage.

May 2024

437patches27156

Latest

May 2024

437patches26821

Latest

  • Added Snowflake - Vector Search Snap to enable advanced vector-based queries as part of the Select statement.

  • Enhanced the Snowflake Snap Pack with full vector support, including schema generation and output compatibility with vector data types.

May 2024

437patches26508

Latest

May 2024

main26341

Stable

  • Enhanced the Snowflake Bulk Load and Snowflake Insert Snaps to support iceberg tables. Iceberg tables partition and organize data across multiple nodes to distribute the workload and speed up data processing.

  • Updated the Delete Condition (Truncates a Table if empty) field in the Snowflake - Delete Snap to Delete condition (deletes all records from a table if left blank) to indicate that all entries will be deleted from the table when this field is blank, but no truncate operation is performed.

February 2024

436patches25630

Latest

  • Enhanced the Snowflake Bulk Load Snap to fetch the file format types based on the selected Datasource. The Snap now supports only CSV and NONE file format types (as a Suggestions list) when the Datasource is Input view.

  • Enhanced the Snowflake Bulk Upsert Snap to support only CSV and NONE file format types (as a Suggestions list).

February 2024

main25112

Stable

Updated and certified against the current SnapLogic Platform release.

November 2023

435patches24865

Latest

Fixed an issue across the Snowflake Snaps that populated all suggestions for the Schema and Table Names existing in the configured Snowflake Account. Now, the Snaps only populate suggestions related to the database configured in the Account.

November 2023

435patches24110

Latest

Added a lint warning to the Snowflake-Bulk Load Snap that recommends users to select the Purge checkbox when the Data source is input view and the Staging location is External.

November 2023

main23721

Stable

The Snowflake Snap Pack is now bundled with the default Snowflake JDBC driver v3.14.

August 2023

434patches23541

Latest

Fixed an issue with the Snowflake-Bulk Load Snap where the Snap wrote irrelevant errors to the error view when both of the following conditions occurred:

  • The inserted data was not in the same order as the table schema.

  •  The Staging location was Internal.

Now, the Snap  writes the correct errors to the error view.

August 2023

main22460

Stable

The Snowflake - Execute Snap now includes a new Query type field. When Auto is selected, the Snap tries to determine the query type automatically.

May 2023

433patches21890

Latest

  • Enhanced the Snowflake - Bulk Load and Snowflake - Bulk Upsert Snaps with a new Flush chunk size (in bytes) field. When using internal staging, data is written to a temporary chunk file. When the file exceeds the specified size, it is copied to the Snowflake stage and deleted.

  • Added validation of external storage credentials for Snowflake S3 Accounts.

May 2023

433patches21370

Latest

  • Fixed an issue where the output document was missing information about the error records count and the reason for the error.

  • Fixed an issue that caused stored procedure executions to fail in the Snowflake - Multi Execute Snap.

May 2023

main21015

Stable

  • Snaplogic is specified as a partner tag in all requests going to Snowflake, making it easier for Snowflake to identify the requests coming from Snaplogic.

  • The default JDBC driver JAR for the Snowflake Snap Pack is upgraded to version 3.13.28 to support the GEOMETRY data type.

February 2023

432patches20906

Latest

  • The default JDBC driver JAR for the Snowflake Snap Pack is upgraded to version 3.13.28 to support the GEOMETRY data type.

  • Fixed an issue with the Snowflake Multi Execute Snap to enable processing of large amounts of data (more than 2 billion records).

  • Fixed a column mismatch issue that was caused by an assumed null AWS token that could potentially be an empty string.

February 2023

432patches20266

Latest

Fixed an issue with the Snowflake - Bulk Load Snap that resulted in lowercase (or mixed case) column names when creating a new table under specific conditions. The new Create table with uppercase column names checkbox addresses this issue.

February 2023

432patches20120

 

Latest

The Snowflake Bulk Load, Bulk Upsert, and Unload Snaps now support expressions for the Staging location field.

February 2023

main19844

 

Stable

  • The JDBC Driver (JAR file) setting in all the Snowflake Accounts is optional. If you do not provide a JDBC driver, the account uses the default Snowflake JDBC driver JAR V3.13.25 that is bundled with the Snap Pack.

November 2022

431patches19581

 

Latest

  • The Snowflake SCD2 Snap works as expected and does not display the Cannot find input data related to the output record error when you use the TIMESTAMPNTZ data type columns as Natural key fields.

  • Fixed the memory leakage issue ( because of Key/pair authentication that occupied some memory)  in the Snowflake Snaps. The performance of the Snowflake Snaps is optimized—there is less memory consumption, and exact memory dumps before and after the Pipeline execution.

November 2022

431patches19454

 

Latest

The Snowflake Snap Pack supports geospatial data types. As the Snowflake Snap Pack requires using our custom Snowflake JDBC driver for full support of all data types, contact support@snaplogic.com for details.

November 2022

431patches19220

 

Latest

The Snowflake S3 OAuth2 Account now support expressions for external staging fields.

November 2022

431patches19220

 

Latest

  • The Snowflake - Bulk Upsert Snap no longer supports values from an upstream input document in the Key columns field when the expression button is enabled. The value must remain constant throughout the Snap’s execution and now limits support for expressions to include only Pipeline parameters.

    To avoid breaking your existing Pipelines that use a value from the input schema in the Key columns field, update the Snap settings to use an input value from Pipeline parameters instead.

  • The Snowflake Bulk Load Snap no longer displays the validation warning in the following cases:

    • When you configure the Validation Mode as NONE and Validation Errors Type as Do not show errors in the Snap, or

    • When Validation Mode is RETURN_n_ROWS, RETURN_ERRORS, or RETURN_ALL_ERRORS.

November 2022

main18944

 

Stable

  • The Snowflake - Bulk Load Snap now routes the 403-Access Denied error to the error view (when the error view is enabled) when you do not have DeleteObject permission to delete the temporary files in the S3 bucket.

  • Fixed the following issues in the Snowflake - Bulk Load Snap:

    • The Snap failed when the Preserve case sensitivity field was not selected and the column names in the input document and the target table were the same but the cases were different. Now, the Snap does not fail and loads the data correctly to the target table.

    • The Snap created a table when it did not exist with multiple columns in the target table when there were two input views, but the column name cases were different in both views. Now, the Snap creates the table from the metadata provided by the second input view and loads the data into the target table from the first input view.

  • The Snowflake Insert Snap now creates the target table only from the table metadata of the second input view when the following conditions are met:

    • The Create table if not present checkbox is selected.

    • The target table does not exist.

    • The table metadata is provided in the second input view.

November 2022

430patches18911

 

Latest

Because of performance issues, all Snowflake Snaps now ignore the Cancel queued queries when pipeline is stopped or if it fails option for Manage Queued Queries, even when selected. Snaps behave as though the default Continue to execute queued queries when the Pipeline is stopped or if it fails option were selected.

October 2022

430patches18781

 

Latest

The Snowflake Insert and Snowflake Bulk Upsert Snaps now do not fail with the SQL Operation failed error, when you insert the uppercase key name into a mixed-case column if the Preserve case sensitivity checkbox is deselected.

The Snowflake Bulk Load Snap now works as expected when you configure On Error with SKIP_FILE_*error_percent_limit*% and set the Error Percent Limit to more than the percentage of rows with invalid data in the CSV file.

October 2022

430patches18432

 

Latest

The Snowflake Bulk Load Snap now has a Validation Errors Type dynamic field, which provides options for displaying validation errors. You can now choose Aggregate errors per row to display a summary view of errors.

September 2022

430patches17962

 

Latest

The Snowflake Bulk Load Snap now triggers the metadata query only once even for invalid input, thereby improving the performance of Snap.

September 2022

430patches17894

 

Latest

The Snowflake Select Snap now works as expected when the table name is dependent on an upstream input.

August 2022

430patches17748

 

Latest

Fixes in the Snowflake Bulk Load Snap:

  • The Snap now routes the 403-Access Denied error to the error view (when the error view is enabled) when you do not have DeleteObject permission to delete the temporary files in the S3 bucket.

  • The Snap now checks for the expiration time whenever it receives a new access token for refreshing the token in Snowflake OAuth2 account.

August 2022

430patches17377

 

Latest

  • The following issues are fixed in the Snowflake - Bulk Load Snap:

    • The Snap successfully loads the data even when you deselect the Preserve case- sensitivity checkbox and the source and target column names are similar, but with different casing.

    • When the Snap has two input views, it creates the target table from the metadata provided by the second input view and loads the data into the target table from the first input view.

  • The Snowflake Select Snap now validates successfully even when you set the Table Name field with an expression.

August 2022

main17386

 

Stable

The following Snowflake Accounts support Key Pair Authentication.

4.29 Patch

429patches16478

 

Latest

Fixed an issue with Snowflake - Bulk Load Snap where Snap failed with 301 status code - SnapDataException.

4.29 Patch

429patches16458

 

Latest

  • Fixed the following issues with the Snaps that use Snowflake Accounts Key Pair Authentication.

    • The Snaps were not able to read the private keys provided as Pipeline parameters, because the key was malformed by replacing the line breaks with whitespace characters. Now, the private keys are properly formed and read when provided as Pipeline parameters.

    • The Snap displayed a Null Point Exception error when you pass an invalid key into the Private Key field like test. Now, the Snap displays a valid error message.

  • Fixed an issue with the Snaps due to an upgrade made to HikariCP that caused a conditional failure and prevented the downstream Snaps to supply the parameters to the stored procedure.  

  • Fixed an issue with Snowflake Execute and Snowflake Select Snaps where the Snap displayed an exception error when the FLOAT or DOUBLE datatype column has the value as NaN (Not a Number). Now the output value is written as-is (NaN) for FLOAT or DOUBLE datatype column and does not display an exception.

  • Enhanced the following Snowflake Accounts with support for Key Pair Authentication.

  • Removed support for unencrypted Private Keys to enforce security in Snowflake accounts.

4.29

main15993

 

Stable

Upgraded with the latest SnapLogic Platform release.

4.28 Patch

428patches15236

 

Latest

  • Updated the Snowflake - Bulk Upsert Snap for the following:

    • The Snap displayed a incorrect resolution when the length of the value exceeded the value defined in a column.

    • The Snap failed with a NullPointer Exception when no value was provided for Error Limit field and Error percentage limit fields.

    • The Snap displayed an incorrect error message when S3 details were not provided.

    • Added the On Error dropdown list, where you can select an action to perform when the Snap encounters errors in a file.

  • Removed the Username field in the following accounts to allow reusing of accounts among different users:

  • The following enhancements were done for Snowflake - Bulk Load Snap:

    • The Snap skips AWS account validation if S3 Storage Integration property is provided.

    • The Snap skips the validate command if S3 Storage Integration property is not provided

    • The Snap overrides the storage integration specified in the account settings with the storage integration specified in the Snap settings.

  • Enhanced the Snowflake - Unload Snap to skip AWS account validation if S3 Storage Integration property is provided.

4.28

main14627

 

Stable

4.27

427patches12999

 

Latest

Enhanced the Snowflake SCD2 Snap to support Pipeline parameters for Natural key and Cause-historization fields.

4.27

main12833

 

Stable

  • Enhanced the Snowflake S3 Database Account to skip batch execution, when the Batch size value is one. When the Batch size is greater than one, the batch is executed.

  • Added the following truncate options to Snowflake - Bulk Load Snap to truncate existing data before performing data load to more efficiently transfer data where possible.

    • Truncate data: Truncates existing data before performing data load. 

    • Truncate columns: Truncates column values that are larger than the maximum column length in the table.

  • Enhanced the Snowflake - Execute Snap to invoke stored procedures.

4.26 Patch

426patches11469

 

Latest

Fixed an issue with Snowflake Insert and Snowflake Bulk Load Snaps where the schema names or database names containing underscore (_) caused the time out of Pipelines.

4.26

main11181

 

Stable

  • Enhanced Snowflake - Lookup and Snowflake SCD2 Snaps with the Input Date Format field to select from the following two options:

    • Continue to execute the snap with the given input Date format

    • Auto Convert the format to Snowflake default format

  • Added a new account type Snowflake Google Storage Database to connect to Google Cloud Storage to load data.

  • Added support for all existing Snowflake Snap accounts to connect to a Snowflake instance hosted on the Google Cloud Platform

4.25

425patches10190

Latest

Enhanced the Snowflake S3 Database and Snowflake S3 Dynamic accounts with a new field S3 AWS Token that allows you to connect to private and protected Amazon S3 buckets.

4.25

main9554

 

Stable

Upgraded with the latest SnapLogic Platform release.

4.24 Patch

424patches8905

 

Latest

Enhanced the Snowflake - Bulk Load Snap to allow transforming data using a new field Select Query before loading data into the Snowflake database. This option enables you to query the staged data files by either reordering the columns or loading a subset of table data from a staged file. This Snap supports CSV and JSON file formats for this data transformation.

4.24

main8556

Stable

Enhanced the Snowflake - Select Snap to return only the selected output fields or columns in the output schema (second output view) using the Fetch Output Fields In Schema check box. If the Output Fields field is empty all the columns are visible.

4.23 Patch

423patches7905

 

Latest

Fixed the performance issue in the Snowflake - Bulk Load Snap while using External Staging on Amazon S3.

4.23

main7430

 

Stable

4.22 Patch 

422patches7246

 

Latest

Fixed an issue with the Snowflake Snaps that failed while displaying similar error in the Snowflake URL connection: message, javax.management.MalformedObjectNameException: Invalid character '=' in value part of property, repeatedly when there is "=" or ":" .

4.22 Patch

422patches6849

 

Latest

4.22

main6403

 

Stable

Updated with the latest SnapLogic Platform release.

4.21 Patch

421patches6272

 

Latest

Fixes the issue where Snowflake SCD2 Snap generates two output documents despite no changes to Cause-historization fields with DATE, TIME and TIMESTAMP Snowflake data types, and with Ignore unchanged rows field selected.

4.21 Patch

421patches6144

 

Latest

  • Fixed the following issues with DB Snaps:

    • The connection thread waits indefinitely causing the subsequent connection requests to become unresponsive.

    • Connection leaks occur during Pipeline execution.

  • Fixed the exception RefCnt has gone negative across the Snaps in the Snowflake Snap Pack.

4.21 Patch 

db/snowflake8860

 

Latest

Added a new field, Handle Timestamp and Date Time Data, to Snowflake Lookup Snap. This field enables you to decide whether the Snap should translate UTC time to your local time and the format of the Date Time data.

4.21 Patch

MULTIPLE8841

 

Latest

Fixed the connection issue in Database Snaps by detecting and closing open connections after the Snap execution ends. 

4.21

snapsmrc542

 

Stable

Upgraded with the latest SnapLogic Platform release.

4.20 Patch 

db/snowflake8800

 

Latest

  • Certifies the Snowflake Snap Pack against JDBC Driver version 3.12.3. Snowflake Execute and Multi-Execute Snaps may break existing Pipelines if the JDBC Driver is updated to a newer version.

  • With the updated JDBC driver (version 3.12.3), the Snowflake Execute and Multi-Execute Snaps' output displays a Status of "-1" instead of "0" without the Message field upon successfully executing DDL statements. If your Pipelines use these Snaps and downstream Snaps use the Status field's value from these, you must modify the downstream Snaps to proceed on a status value of -1 instead of 0.

  • This change in the Snap behavior follows from the change introduced in the Snowflake JDBC driver in version 3.8.1:
    "Statement.getUpdateCount() and PreparedStatement.getUpdateCount() return the number of rows updated by DML statements. For all other types of statements, including queries, they return -1."

4.20 Patch 

db/snowflake8758

 

Latest

Re-release of fixes from db/snowflake8687 for 4.20: Fixes the Snowflake Bulk Load snap where the Snap fails to load documents containing single quotes when the Load empty strings checkbox is not selected.

4.20

snapsmrc535

 

Stable

Upgraded with the latest SnapLogic Platform release.

4.19 Patch 

db/snowflake8687

 

Latest

Fixed the Snowflake Bulk Load snap where the Snap fails to load documents containing single quotes when the Load empty strings checkbox is not selected.

4.19 Patch 

db/snowflake8499

 

Latest

Added the property Handle Timestamp and Date Time Data to Snowflake - Execute and Snowflake - Select Snaps. This property enables you to decide whether the Snap should translate UTC time to your local time.

4.19 Patch 

db/snowflake8412

 

Latest

Fixed an issue with the Snowflake - Update Snap wherein the Snap is unable to perform operations when:

  • An expression is used in the Update condition property.

  • Input data contain the character '?'.

4.19

snaprsmrc528

 

Stable

  • Added a new field-set, Auto Historization Query, in the Snowflake SCD2 Snap to support auto-historization of column data. With this enhancement you can detect whether the incoming record is a historical event or a current event.  

  • Raised the minimum buffer size in the Snowflake - Bulk Upsert and Snowflake - Bulk Load Snaps to 6 MB.

4.18 Patch 

db/snowflake8044

 

Latest

Fixed an issue with the Snowflake - Select Snap wherein the Snap converts the Snowflake-provided timestamp value to the local timezone of the account.

4.18 Patch 

db/snowflake8044

 

Latest

Enhanced the Snap Pack to support AWS SDK 1.11.634 to fix the NullPointerException issue in the AWS SDK. This issue occurred in AWS-related Snaps that had HTTP or HTTPS proxy configured without a username and/or password. 

4.18 Patch

MULTIPLE7884

 

Latest

Fixed an issue with the PostgreSQL grammar to better handle the single quote characters.

4.18 Patch 

db/snowflake7821

 

Latest

Fixed an issue with the Snowflake - Execute Snap wherein the Snap is unable to support the '$' character in query syntax.

4.18 Patch

MULTIPLE7778

 

Latest

Updated the AWS SDK library version to default to Signature Version 4 Signing process for API requests across all regions.

4.18 Patch 

db/snowflake7739

 

Latest

  • Fixed an issue with the Snowflake - Bulk Upsert Snap wherein the Snap fails when using a pipeline parameter in Key columns.

  • Fixed an issue with the Snowflake - Unload Snap wherein the Snap does not abort the query when you stop the Pipeline execution.

4.18

snapsmrc523

 

Stable

Added the Use Result Query property to the Multi Execute Snap, which enables you to write results to an output view.

4.17

ALL7402

 

Latest

Pushed automatic rebuild of the latest version of each Snap Pack to SnapLogic UAT and Elastic servers.

4.17 Patch 

db/snowflake7396

 

Latest

Fixed an issue wherein bit data types in the Snowflake - Select table convert to true or false instead of 0 or 1.

4.17 Patch 

db/snowflake7334

 

Latest

Added AWS Server-Side Encryption support for AWS S3 and AWS KMS (Key Management Service) for Snowflake Bulk Load, Snowflake Bulk Upsert, and Snowflake Unload Snaps.

4.17

snapsmrc515

 

Latest

  • Fixed an issue with the Snowflake Execute Snap wherein the Snap would send the input document to the output view even if the Pass through field is not selected in the Snap configuration. With this fix, the Snap sends the input document to the output view, under the key original, only if you select the Pass through field.

  • Added the Snap Execution field to all Standard-mode Snaps. In some Snaps, this field replaces the existing Execute during preview check box.

4.16 Patch 

db/snowflake6945

 

Latest

Fixed an issue with the Snowflake Lookup Snap failing when Date datatype is used in JavaScript functions.

4.16 Patch 

db/snowflake6928

 

Latest

Added support for file format options for input data from upstream Snaps, to the Snowflake Bulk Load Snap.

4.16 Patch 

db/snowflake6819

 

Latest

  • Snowflake Bulk Load: Added new property, Buffer size (MB). Configure this to specify the size limit of each buffer when writing to external staging systems such as S3. 

  • Fixed an issue with the Lookup Snap passing data simultaneously to output and error views when some values contained spaces at the end.

4.16

snapsmrc508

 

Stable

  • Snowflake Account: Added the ability to use SnapLogic to securely connect to and query a Snowflake instance using Azure Blob as its storage layer. 

  • Snowflake Account: Added support for Snowflake JDBC driver JAR version 3.6.17. 

  • Snowflake Unload, Bulk Load, and Bulk Upsert: Updated the Snaps to enable SnapLogic users to successfully connect to a Snowflake instance to query, bulk load, and unload data from Azure Blob storage.

4.15

snapsmrc500

 

Stable

  • Added two new Snaps, Snowflake - Multi Execute, and Snowflake SCD2. Snowflake - Multi Execute is used for executing multiple DDL and DML queries on the Snowflake DB. Snowflake SCD2 is used for Type 2 field historization. 

  • Enhanced the Snowflake Bulk Upsert Snaps to improve the Snaps performance.

  • Enhanced the Snowflake Snap Pack to reflect Azure certification.

4.14

snapsmrc490

 

Stable

Upgraded with the latest SnapLogic Platform release.

4.13

snapsmrc486

 

Stable

Upgraded with the latest SnapLogic Platform release.

4.12

snapsmrc480

 

Stable

Upgraded with the latest SnapLogic Platform release.

4.11 Patch

MULTIPLE4377

 

Latest

Fixed a document call issue that was slowing down the Snowflake Bulk Load Snap.

4.11 Patch 

db/snowflake4283

 

Latest

Snowflake Bulk Load - Fixed an issue by adding PUT command to the list of DDL command list for Snowflake.

4.11 Patch 

db/snowflake4273

 

Latest

Snowflake Bulk Load - Resolved an issue with Snowflake Bulk Load Delimiter Consistency (comma and newline).

4.11

snapsmrc465

 

Stable

Upgraded with the latest SnapLogic Platform release.

4.10 Patch

snowflake4133

Latest

Updated the Snowflake Bulk Load Snap with Preserve case sensitivity property to preserve the case sensitivity of column names.

4.10

snapsmrc414

 

Stable

  • Updated the Snowflake Bulk Load Snap with Load empty strings property for the empty string values in the input documents to be loaded as empty strings to the string-type fields.

  • Updated the Snowflake Bulk Load Snap with Table Columns to support the order of the entries on the staged files that contain a subset of the columns in the Snowflake table.

  • Added the property Use Result Query to view the output preview field with a result statement.

  • Tested for the JDBC driver JAR version 3.1.1 on the Database and the Dynamic accounts.

4.9.0 Patch

snowflake3234

 

Latest

Enhanced Snowflake - Execute Snap results to include additional details

4.9.0 Patch

snowflake3125

 

Latest

Addressed an issue in Snowflake Bulk Load where the comma character in a value is not escaped.

4.9

snapsmrc405

 

Stable

JDBC Driver Class property added to enable the user to custom configure the JDBC driver in the Database and the Dynamic accounts.

4.8.0 Patch

snowflake2760

 

Latest

Potential fix for JDBC deadlock issue.

4.8.0 Patch

snowflake2739

 

Latest

Addressed an issue with the Snowflake schema not correctly represented in the Mapper Snap.

4.8

snapsmrc398

 

Stable

  • Info tab added to accounts.

  • Database accounts now invalidate connection pools if account properties are modified and login attempts fail.

  • Enhanced the default count of input and output view (UI) behavior of the Snaps for better user experience. Note that the Snowflake Select, Table List and Lookup Snap views remain unchanged.

  • No labels