Skip to end of banner
Go to start of banner

Snowflake SCD2

Skip to end of metadata
Go to start of metadata

You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 28 Next »

On this Page

Overview

This Snap provides the functionality of SCD (Slowly Changing Dimension) Type 2 on a target Snowflake table. The Snap executes one SQL lookup request per set of input documents to avoid making a request for every input record. Its output is typically a stream of documents for the Snowflake - Bulk Upsert Snap, which updates or inserts rows into the target table. Therefore, this Snap must be connected to the Snowflake - Bulk Upsert Snap to accomplish the complete SCD2 functionality.

Input and Output

Expected input: Each document in the input view should contain a data map of key-value entries. The input data must contain data in the Natural Key (primary key) and Cause-historization fields. 

Expected output: Each document in the output view contains a data map of key-value entries for all fields of a row in the target Snowflake table.

Expected upstream Snaps: Any Snap, such as a Mapper or JSON Parser Snap, whose output contains a map of key-value entries.

Expected downstream SnapsSnowflake Bulk Upsert snap must be used as downstream snap since the Snowflake SCD2 snap only generates set of rows to be inserted or updated and it doesn't do any write operation on the table.

Prerequisites

  • Read and write access to the Snowflake instance.
  • The target table should have the following three columns for field historization to work:
    • Column to demarcate whether a row is a current row or not. For example, "CURRENT_ROW". For the current row, the value would be true or 1. For the historical row, the value would be false or 0.
    • Column to denote the starting date of the current row. For example, "START_DATE".
    • Column to denote when the row was historized. For example, "END_DATE". For the active row, it is null. For a historical row, it has the value that indicates it was effective till that date.
  • You must have minimum permissions on the database to execute Snowflake Snaps. To understand if you already have them, you must retrieve the current set of permissions. The following commands enable you to retrieve those permissions:
    SHOW GRANTS ON DATABASE <database_name>
    SHOW GRANTS ON SCHEMA <schema_name>
    SHOW GRANTS TO USER <user_name>
  • Security Prerequisites: You should have the following permissions in your Snowflake account to execute this Snap: 

    • Usage (DB and Schema): Privilege to use database, role and schema.
    • Create table: Privilege to create a table on the database. role and schema.

    For more information on Snowflake privileges, refer to Access Control Privileges.

Internal SQL Commands

This Snap uses the SELECT command internally. It enables querying the database to retrieve a set of rows.

Configuring Accounts

This Snap uses account references created on the Accounts page of SnapLogic Manager to handle access to this endpoint. See Snowflake Account for information on setting up this type of account.

Configuring Views

Input

This Snap has exactly one document input view.
OutputThis Snap has exactly one document output view.
ErrorThis Snap has at most one document error view.

Troubleshooting

None.

Limitations and Known Issues

None.

Modes

Snap Settings


LabelRequired. The name for the Snap. Modify this to be more specific, especially if there are more than one of the same Snap in the pipeline.
Schema name

The name of the schema containing the target table. Providing the schema name along with the table name in the Table name field is sufficient. The suggestible field that lists all available schema in the configured account. 

Default value: None 

Example: "TestSchema"

Table name

Required. The name of the target table. Syntax is "<schema_name>"."<table_name>".  This is a suggestible field that lists all available tables if the schema is provided in the Schema name field. Alternatively, if the Schema name field is blank, it lists all tables within the account if the schema is not provided. 

Default value: None 

Example: "TestSchema"."TestTable"


The target table should have the following three columns for field historization to work:

  • Column to demarcate whether a row is a current row or not. For example, "CURRENT_ROW". For the current row, the value would be true or 1. For the historical row, the value would be false or 0.
  • Column to denote the starting date of the current row. For example, "START_DATE".
  • Column to denote when the row was historized. For example, "END_DATE". For the active row, it's null. For a historical row, it has the value that indicates it was effective till that date.

Use the ALTER table command to add these columns to your target table if they are not present. 

Natural key

Names of fields that identify a unique row in the target table. The identity key cannot be used as the Natural key, since a current row and its historical rows cannot have the same natural key value

Default value: None

Example id (Each record has to have a unique value)

Cause-historization fields 

Names of fields where any change in value causes the historization of an existing row and the insertion of a new current row.

Default value: None

Example gold bullion rate

SCD fields

Required. The historical and updated information for the Cause-historization field. Click + to add SCD fields. By default, there are four rows in this field-set:

  • Current row
  • Historical row
  • Start date of the current row
  • End date of historical row.
Meaning

Specifies the table columns that are to be updated for implementing the SCD2 type transformation.

Default value

  • Current row
  • Historical row
  • Start date of current row
  • End date of historical row
Field

The fields in the table will contain the historical information. 

Default value: None

Below are the values that must be configured for each row:

  • Current row: The name of the column in the target table that holds the flag for the historized field. For example, "CURRENT_ROW".
  • Historical row: The name of the column in the target table that holds the flag for the historized field. It has to be the same as the value configured for the Current row field. For example, "CURRENT_ROW".
  • Start date of current row: The name of the column in the target table for denoting the start date for the current row. For example, "START_DATE".
  • End date of historical row: The name of the column in the target table for denoting the end date for the historical row. For example, "END_DATE".


By default, the start and end date for both Current row and Historical row are null. After the Snap is executed, the start date for the updated row data automatically becomes the end date for the earlier version of the data (Historical row).

Value

The value to be assigned to the current or historical row. For date-related rows, the default is Date.now(). 

Default value

  • Current row and Historical row: None.
  • Start date of current row, and End date of historical row: Date.now().

The Value field should be configured as follows:

  • Current row: 1
  • Historical row: 0
Ignore unchanged rows

Specifies whether the Snap must ignore writing unchanged rows from the source table to the target table. If you enable this option, the Snap generates a corresponding document in the target only if the Cause-historization column in the source row is changed. Else, the Snap does not generate any corresponding document in the target. 

Default value: Not selected

Number of retries

The number of times that the Snap must try to write the fields in case of an error during processing. An error is displayed if the maximum number of tries has been reached.

Default value: 0

Retry interval (seconds)

The time interval, in seconds, between subsequent retry attempts. 

Default value: 1

Auto Historization QueryThis field-set is used to specify the fields that are to be used to historize table data. Historization is in the sort order specified. Care must be taken that the field is sortable. You can also add multiple fields here; historizaton occurs when even of the fields is changed. 
Field

The name of the field. This is a suggestible field and suggests all the fields in the target table.

Example: Invoice_Number

Default value: N/A

If this field has null values in the incoming records, then the value in the Snowflake table is treated as the current value and the incoming record is historized. 

Sort Order
The order in which the selected field is to be historized. Available options are:
  • Ascending Order: The higher value is classified as a current event. For example date of transaction, age, height, etc.
  • Descending Order: The lower value is classified as the current event. For example, rank.

Default value: Ascending Order

Manage Queued Queries

Select this property to decide whether the Snap should continue or cancel the execution of the queued Snowflake Execute SQL queries when you stop the pipeline.

If you select Cancel queued queries when pipeline is stopped or if it fails, then the read queries under execution are canceled, whereas the write queries under execution are not canceled. Snowflake internally determines which queries are safe to be canceled and cancels those queries.

Default value: Continue to execute queued queries when pipeline is stopped or if it fails

Snap Execution

Select one of the three modes in which the Snap executes. Available options are:
  • Validate & Execute: Performs limited execution of the Snap, and generates a data preview during Pipeline validation. Subsequently, performs full execution of the Snap (unlimited records) during Pipeline runtime.
  • Execute only: Performs full execution of the Snap during Pipeline execution without generating preview data.
  • Disabled: Disables the Snap and all Snaps that are downstream from it.

Examples

Historizing Incoming Records

This example demonstrates how you can use the Snowflake SCD2 Snap to auto-historize records. In this example, since the existing record in the Snowflake table is the latest, the incoming records are historized. 

This Pipeline performs the following operations:


The Target Table

Before we start, let us look at the target table and understand some of its columns that are necessary for the Pipeline:

We focus on the highlighted columns above to demonstrate auto-historization and describe their function in the table.

  • POINT ID is the natural key in this table.
  • Changes in SCHEDULEDVOLUME for a natural key are historized.
  • HISTORYSTARTDATE and HISTORYENDDATE are used to identify the current record. 
  • FLAG denotes if a record is a current record with the value TRUE.
  • STARTDATE and ENDDATE are automatically calculated by the Snap. They represent the date range during which a record was the current record. The ENDDATE is blank for the current record.

Input Data: Reading, Parsing and Mapping

This Pipeline is configured to send records into the target table. The File Reader Snap is configured to read a CSV file that contains the records. The downstream CSV Parser Snap parses the CSV file read by the File Reader Snap. Below is a preview of this file:

Based on the values of HistoryStDate and HistoryEndDate, it is clear that the existing record in the target table is the latest (or current) record. 

Since the output from the CSV Parser Snap is a string, it has to be parsed into the appropriate data type. Parsing and data mapping is done using the Mapper Snap, as shown below:

This mapped data is then sent to the Snowflake SCD2 Snap.

Data Processing

The Snowflake SCD2 Snap performs SCD2 operations on the target table. We configure it as shown below:

Let us take a look at the highlighted Snap fields and how they affect the Snap functionality in this example:

  • Natural key: The Snap looks for records with matching POINT ID values in the incoming documents to group the records.
  • Cause-historization fields: For each unique POINT ID, changes in SCHEDULEDVOLUME initiate historization. If a change has not occurred, the incoming records are historized..
  • SCD fields:
    • The state of the current or historical record is marked in the FLAG field, T for current record and F for the historical record.
    • The columns STARTDATE and ENDDATE in the target table are maintained to denote the start and end dates of the current state of the table's data. The ENDDATE is always blank for a current record.
  • Auto Historization Query: The Snap sorts the values in the HISTORYSTARTDATE and HISTORYENDDATE columns for the same POINT ID in the Snowflake table and the incoming documents in ascending order. The record with the highest value in those fields is considered the current record. 

All incoming records pertaining to a POINT ID are historized. The value F is assigned under the FLAG column to these fields and the corresponding STARTDATE and ENDDATE are evaluated by the expression Date.now().

This can be seen in the SCD2 Snap's output preview:

The Snap identifies the current and historical records and this data is now ready to be updated and inserted into the target table.

Upsert Data into the Target Table

We use the Snowflake Bulk Upsert Snap to update the target table with this historized data. We configure the Snowflake Bulk Upsert Snap as shown below:

Download this Pipeline and sample data. This is a compressed file, unzip it to extract its contents before importing them in SnapLogic.

Downloads

  File Modified
You are not logged in. Any changes you make will be marked as anonymous. You may want to Log In if you already have an account.
No files shared here yet.
  • Drag and drop to upload or browse for files
  • Release 

    Snap Pack Version

    Date

    Type

      Updates

    February 2024

    436patches25630

    Latest

    • Enhanced the Snowflake Bulk Load Snap to fetch the file format types based on the selected Datasource. The Snap now supports only CSV and NONE file format types (as a Suggestions list) when the Datasource is Input view.

    • Enhanced the Snowflake Bulk Upsert Snap to support only CSV and NONE file format types (as a Suggestions list).

    February 2024

    main25112

    Stable

    Updated and certified against the current SnapLogic Platform release.

    November 2023

    435patches24865

    Latest

    Fixed an issue across the Snowflake Snaps that populated all suggestions for the Schema and Table Names existing in the configured Snowflake Account. Now, the Snaps only populate suggestions related to the database configured in the Account.

    November 2023

    435patches24110

    Latest

    Added a lint warning to the Snowflake-Bulk Load Snap that recommends users to select the Purge checkbox when the Data source is input view and the Staging location is External.

    November 2023

    main23721

    Stable

    The Snowflake Snap Pack is now bundled with the default Snowflake JDBC driver v3.14.

    August 2023

    434patches23541

    Latest

    Fixed an issue with the Snowflake-Bulk Load Snap where the Snap wrote irrelevant errors to the error view when both of the following conditions occurred:

    • The inserted data was not in the same order as the table schema.

    •  The Staging location was Internal.

    Now, the Snap  writes the correct errors to the error view.

    August 2023

    main22460

    Stable

    The Snowflake - Execute Snap now includes a new Query type field. When Auto is selected, the Snap tries to determine the query type automatically.

    May 2023

    433patches21890

    Latest

    • Enhanced the Snowflake - Bulk Load and Snowflake - Bulk Upsert Snaps with a new Flush chunk size (in bytes) field. When using internal staging, data is written to a temporary chunk file. When the file exceeds the specified size, it is copied to the Snowflake stage and deleted.

    • Added validation of external storage credentials for Snowflake S3 Accounts.

    May 2023

    433patches21370

    Latest

    • Fixed an issue where the output document was missing information about the error records count and the reason for the error.

    • Fixed an issue that caused stored procedure executions to fail in the Snowflake - Multi Execute Snap.

    May 2023

    main21015

    Stable

    • Snaplogic is specified as a partner tag in all requests going to Snowflake, making it easier for Snowflake to identify the requests coming from Snaplogic.

    • The default JDBC driver JAR for the Snowflake Snap Pack is upgraded to version 3.13.28 to support the GEOMETRY data type.

    February 2023

    432patches20906

    Latest

    • The default JDBC driver JAR for the Snowflake Snap Pack is upgraded to version 3.13.28 to support the GEOMETRY data type.

    • Fixed an issue with the Snowflake Multi Execute Snap to enable processing of large amounts of data (more than 2 billion records).

    • Fixed a column mismatch issue that was caused by an assumed null AWS token that could potentially be an empty string.

    February 2023

    432patches20266

    Latest

    Fixed an issue with the Snowflake - Bulk Load Snap that resulted in lowercase (or mixed case) column names when creating a new table under specific conditions. The new Create table with uppercase column names checkbox addresses this issue.

    February 2023

    432patches20120

     

    Latest

    The Snowflake Bulk Load, Bulk Upsert, and Unload Snaps now support expressions for the Staging location field.

    February 2023

    main19844

     

    Stable

    • The JDBC Driver (JAR file) setting in all the Snowflake Accounts is optional. If you do not provide a JDBC driver, the account uses the default Snowflake JDBC driver JAR V3.13.25 that is bundled with the Snap Pack.

    November 2022

    431patches19581

     

    Latest

    • The Snowflake SCD2 Snap works as expected and does not display the Cannot find input data related to the output record error when you use the TIMESTAMPNTZ data type columns as Natural key fields.

    • Fixed the memory leakage issue ( because of Key/pair authentication that occupied some memory)  in the Snowflake Snaps. The performance of the Snowflake Snaps is optimized—there is less memory consumption, and exact memory dumps before and after the Pipeline execution.

    November 2022

    431patches19454

     

    Latest

    The Snowflake Snap Pack supports geospatial data types. As the Snowflake Snap Pack requires using our custom Snowflake JDBC driver for full support of all data types, contact support@snaplogic.com for details.

    November 2022

    431patches19220

     

    Latest

    The Snowflake S3 OAuth2 Account now support expressions for external staging fields.

    November 2022

    431patches19220

     

    Latest

    • The Snowflake - Bulk Upsert Snap no longer supports values from an upstream input document in the Key columns field when the expression button is enabled. The value must remain constant throughout the Snap’s execution and now limits support for expressions to include only Pipeline parameters.

      To avoid breaking your existing Pipelines that use a value from the input schema in the Key columns field, update the Snap settings to use an input value from Pipeline parameters instead.

    • The Snowflake Bulk Load Snap no longer displays the validation warning in the following cases:

      • When you configure the Validation Mode as NONE and Validation Errors Type as Do not show errors in the Snap, or

      • When Validation Mode is RETURN_n_ROWS, RETURN_ERRORS, or RETURN_ALL_ERRORS.

    November 2022

    main18944

     

    Stable

    • The Snowflake - Bulk Load Snap now routes the 403-Access Denied error to the error view (when the error view is enabled) when you do not have DeleteObject permission to delete the temporary files in the S3 bucket.

    • Fixed the following issues in the Snowflake - Bulk Load Snap:

      • The Snap failed when the Preserve case sensitivity field was not selected and the column names in the input document and the target table were the same but the cases were different. Now, the Snap does not fail and loads the data correctly to the target table.

      • The Snap created a table when it did not exist with multiple columns in the target table when there were two input views, but the column name cases were different in both views. Now, the Snap creates the table from the metadata provided by the second input view and loads the data into the target table from the first input view.

    • The Snowflake Insert Snap now creates the target table only from the table metadata of the second input view when the following conditions are met:

      • The Create table if not present checkbox is selected.

      • The target table does not exist.

      • The table metadata is provided in the second input view.

    November 2022

    430patches18911

     

    Latest

    Because of performance issues, all Snowflake Snaps now ignore the Cancel queued queries when pipeline is stopped or if it fails option for Manage Queued Queries, even when selected. Snaps behave as though the default Continue to execute queued queries when the Pipeline is stopped or if it fails option were selected.

    October 2022

    430patches18781

     

    Latest

    The Snowflake Insert and Snowflake Bulk Upsert Snaps now do not fail with the SQL Operation failed error, when you insert the uppercase key name into a mixed-case column if the Preserve case sensitivity checkbox is deselected.

    The Snowflake Bulk Load Snap now works as expected when you configure On Error with SKIP_FILE_*error_percent_limit*% and set the Error Percent Limit to more than the percentage of rows with invalid data in the CSV file.

    October 2022

    430patches18432

     

    Latest

    The Snowflake Bulk Load Snap now has a Validation Errors Type dynamic field, which provides options for displaying validation errors. You can now choose Aggregate errors per row to display a summary view of errors.

    September 2022

    430patches17962

     

    Latest

    The Snowflake Bulk Load Snap now triggers the metadata query only once even for invalid input, thereby improving the performance of Snap.

    September 2022

    430patches17894

     

    Latest

    The Snowflake Select Snap now works as expected when the table name is dependent on an upstream input.

    August 2022

    430patches17748

     

    Latest

    Fixes in the Snowflake Bulk Load Snap:

    • The Snap now routes the 403-Access Denied error to the error view (when the error view is enabled) when you do not have DeleteObject permission to delete the temporary files in the S3 bucket.

    • The Snap now checks for the expiration time whenever it receives a new access token for refreshing the token in Snowflake OAuth2 account.

    August 2022

    430patches17377

     

    Latest

    • The following issues are fixed in the Snowflake - Bulk Load Snap:

      • The Snap successfully loads the data even when you deselect the Preserve case- sensitivity checkbox and the source and target column names are similar, but with different casing.

      • When the Snap has two input views, it creates the target table from the metadata provided by the second input view and loads the data into the target table from the first input view.

    • The Snowflake Select Snap now validates successfully even when you set the Table Name field with an expression.

    August 2022

    main17386

     

    Stable

    The following Snowflake Accounts support Key Pair Authentication.

    4.29 Patch

    429patches16478

     

    Latest

    Fixed an issue with Snowflake - Bulk Load Snap where Snap failed with 301 status code - SnapDataException.

    4.29 Patch

    429patches16458

     

    Latest

    • Fixed the following issues with the Snaps that use Snowflake Accounts Key Pair Authentication.

      • The Snaps were not able to read the private keys provided as Pipeline parameters, because the key was malformed by replacing the line breaks with whitespace characters. Now, the private keys are properly formed and read when provided as Pipeline parameters.

      • The Snap displayed a Null Point Exception error when you pass an invalid key into the Private Key field like test. Now, the Snap displays a valid error message.

    • Fixed an issue with the Snaps due to an upgrade made to HikariCP that caused a conditional failure and prevented the downstream Snaps to supply the parameters to the stored procedure.  

    • Fixed an issue with Snowflake Execute and Snowflake Select Snaps where the Snap displayed an exception error when the FLOAT or DOUBLE datatype column has the value as NaN (Not a Number). Now the output value is written as-is (NaN) for FLOAT or DOUBLE datatype column and does not display an exception.

    • Enhanced the following Snowflake Accounts with support for Key Pair Authentication.

    • Removed support for unencrypted Private Keys to enforce security in Snowflake accounts.

    4.29

    main15993

     

    Stable

    Upgraded with the latest SnapLogic Platform release.

    4.28 Patch

    428patches15236

     

    Latest

    • Updated the Snowflake - Bulk Upsert Snap for the following:

      • The Snap displayed a incorrect resolution when the length of the value exceeded the value defined in a column.

      • The Snap failed with a NullPointer Exception when no value was provided for Error Limit field and Error percentage limit fields.

      • The Snap displayed an incorrect error message when S3 details were not provided.

      • Added the On Error dropdown list, where you can select an action to perform when the Snap encounters errors in a file.

    • Removed the Username field in the following accounts to allow reusing of accounts among different users:

    • The following enhancements were done for Snowflake - Bulk Load Snap:

      • The Snap skips AWS account validation if S3 Storage Integration property is provided.

      • The Snap skips the validate command if S3 Storage Integration property is not provided

      • The Snap overrides the storage integration specified in the account settings with the storage integration specified in the Snap settings.

    • Enhanced the Snowflake - Unload Snap to skip AWS account validation if S3 Storage Integration property is provided.

    4.28

    main14627

     

    Stable

    4.27

    427patches12999

     

    Latest

    Enhanced the Snowflake SCD2 Snap to support Pipeline parameters for Natural key and Cause-historization fields.

    4.27

    main12833

     

    Stable

    • Enhanced the Snowflake S3 Database Account to skip batch execution, when the Batch size value is one. When the Batch size is greater than one, the batch is executed.

    • Added the following truncate options to Snowflake - Bulk Load Snap to truncate existing data before performing data load to more efficiently transfer data where possible.

      • Truncate data: Truncates existing data before performing data load. 

      • Truncate columns: Truncates column values that are larger than the maximum column length in the table.

    • Enhanced the Snowflake - Execute Snap to invoke stored procedures.

    4.26 Patch

    426patches11469

     

    Latest

    Fixed an issue with Snowflake Insert and Snowflake Bulk Load Snaps where the schema names or database names containing underscore (_) caused the time out of Pipelines.

    4.26

    main11181

     

    Stable

    • Enhanced Snowflake - Lookup and Snowflake SCD2 Snaps with the Input Date Format field to select from the following two options:

      • Continue to execute the snap with the given input Date format

      • Auto Convert the format to Snowflake default format

    • Added a new account type Snowflake Google Storage Database to connect to Google Cloud Storage to load data.

    • Added support for all existing Snowflake Snap accounts to connect to a Snowflake instance hosted on the Google Cloud Platform

    4.25

    425patches10190

    Latest

    Enhanced the Snowflake S3 Database and Snowflake S3 Dynamic accounts with a new field S3 AWS Token that allows you to connect to private and protected Amazon S3 buckets.

    4.25

    main9554

     

    Stable

    Upgraded with the latest SnapLogic Platform release.

    4.24 Patch

    424patches8905

     

    Latest

    Enhanced the Snowflake - Bulk Load Snap to allow transforming data using a new field Select Query before loading data into the Snowflake database. This option enables you to query the staged data files by either reordering the columns or loading a subset of table data from a staged file. This Snap supports CSV and JSON file formats for this data transformation.

    4.24

    main8556

    Stable

    Enhanced the Snowflake - Select Snap to return only the selected output fields or columns in the output schema (second output view) using the Fetch Output Fields In Schema check box. If the Output Fields field is empty all the columns are visible.

    4.23 Patch

    423patches7905

     

    Latest

    Fixed the performance issue in the Snowflake - Bulk Load Snap while using External Staging on Amazon S3.

    4.23

    main7430

     

    Stable

    4.22 Patch 

    422patches7246

     

    Latest

    Fixed an issue with the Snowflake Snaps that failed while displaying similar error in the Snowflake URL connection: message, javax.management.MalformedObjectNameException: Invalid character '=' in value part of property, repeatedly when there is "=" or ":" .

    4.22 Patch

    422patches6849

     

    Latest

    4.22

    main6403

     

    Stable

    Updated with the latest SnapLogic Platform release.

    4.21 Patch

    421patches6272

     

    Latest

    Fixes the issue where Snowflake SCD2 Snap generates two output documents despite no changes to Cause-historization fields with DATE, TIME and TIMESTAMP Snowflake data types, and with Ignore unchanged rows field selected.

    4.21 Patch

    421patches6144

     

    Latest

    • Fixed the following issues with DB Snaps:

      • The connection thread waits indefinitely causing the subsequent connection requests to become unresponsive.

      • Connection leaks occur during Pipeline execution.

    • Fixed the exception RefCnt has gone negative across the Snaps in the Snowflake Snap Pack.

    4.21 Patch 

    db/snowflake8860

     

    Latest

    Added a new field, Handle Timestamp and Date Time Data, to Snowflake Lookup Snap. This field enables you to decide whether the Snap should translate UTC time to your local time and the format of the Date Time data.

    4.21 Patch

    MULTIPLE8841

     

    Latest

    Fixed the connection issue in Database Snaps by detecting and closing open connections after the Snap execution ends. 

    4.21

    snapsmrc542

     

    Stable

    Upgraded with the latest SnapLogic Platform release.

    4.20 Patch 

    db/snowflake8800

     

    Latest

    • Certifies the Snowflake Snap Pack against JDBC Driver version 3.12.3. Snowflake Execute and Multi-Execute Snaps may break existing Pipelines if the JDBC Driver is updated to a newer version.

    • With the updated JDBC driver (version 3.12.3), the Snowflake Execute and Multi-Execute Snaps' output displays a Status of "-1" instead of "0" without the Message field upon successfully executing DDL statements. If your Pipelines use these Snaps and downstream Snaps use the Status field's value from these, you must modify the downstream Snaps to proceed on a status value of -1 instead of 0.

    • This change in the Snap behavior follows from the change introduced in the Snowflake JDBC driver in version 3.8.1:
      "Statement.getUpdateCount() and PreparedStatement.getUpdateCount() return the number of rows updated by DML statements. For all other types of statements, including queries, they return -1."

    4.20 Patch 

    db/snowflake8758

     

    Latest

    Re-release of fixes from db/snowflake8687 for 4.20: Fixes the Snowflake Bulk Load snap where the Snap fails to load documents containing single quotes when the Load empty strings checkbox is not selected.

    4.20

    snapsmrc535

     

    Stable

    Upgraded with the latest SnapLogic Platform release.

    4.19 Patch 

    db/snowflake8687

     

    Latest

    Fixed the Snowflake Bulk Load snap where the Snap fails to load documents containing single quotes when the Load empty strings checkbox is not selected.

    4.19 Patch 

    db/snowflake8499

     

    Latest

    Added the property Handle Timestamp and Date Time Data to Snowflake - Execute and Snowflake - Select Snaps. This property enables you to decide whether the Snap should translate UTC time to your local time.

    4.19 Patch 

    db/snowflake8412

     

    Latest

    Fixed an issue with the Snowflake - Update Snap wherein the Snap is unable to perform operations when:

    • An expression is used in the Update condition property.

    • Input data contain the character '?'.

    4.19

    snaprsmrc528

     

    Stable

    • Added a new field-set, Auto Historization Query, in the Snowflake SCD2 Snap to support auto-historization of column data. With this enhancement you can detect whether the incoming record is a historical event or a current event.  

    • Raised the minimum buffer size in the Snowflake - Bulk Upsert and Snowflake - Bulk Load Snaps to 6 MB.

    4.18 Patch 

    db/snowflake8044

     

    Latest

    Fixed an issue with the Snowflake - Select Snap wherein the Snap converts the Snowflake-provided timestamp value to the local timezone of the account.

    4.18 Patch 

    db/snowflake8044

     

    Latest

    Enhanced the Snap Pack to support AWS SDK 1.11.634 to fix the NullPointerException issue in the AWS SDK. This issue occurred in AWS-related Snaps that had HTTP or HTTPS proxy configured without a username and/or password. 

    4.18 Patch

    MULTIPLE7884

     

    Latest

    Fixed an issue with the PostgreSQL grammar to better handle the single quote characters.

    4.18 Patch 

    db/snowflake7821

     

    Latest

    Fixed an issue with the Snowflake - Execute Snap wherein the Snap is unable to support the '$' character in query syntax.

    4.18 Patch

    MULTIPLE7778

     

    Latest

    Updated the AWS SDK library version to default to Signature Version 4 Signing process for API requests across all regions.

    4.18 Patch 

    db/snowflake7739

     

    Latest

    • Fixed an issue with the Snowflake - Bulk Upsert Snap wherein the Snap fails when using a pipeline parameter in Key columns.

    • Fixed an issue with the Snowflake - Unload Snap wherein the Snap does not abort the query when you stop the Pipeline execution.

    4.18

    snapsmrc523

     

    Stable

    Added the Use Result Query property to the Multi Execute Snap, which enables you to write results to an output view.

    4.17

    ALL7402

     

    Latest

    Pushed automatic rebuild of the latest version of each Snap Pack to SnapLogic UAT and Elastic servers.

    4.17 Patch 

    db/snowflake7396

     

    Latest

    Fixed an issue wherein bit data types in the Snowflake - Select table convert to true or false instead of 0 or 1.

    4.17 Patch 

    db/snowflake7334

     

    Latest

    Added AWS Server-Side Encryption support for AWS S3 and AWS KMS (Key Management Service) for Snowflake Bulk Load, Snowflake Bulk Upsert, and Snowflake Unload Snaps.

    4.17

    snapsmrc515

     

    Latest

    • Fixed an issue with the Snowflake Execute Snap wherein the Snap would send the input document to the output view even if the Pass through field is not selected in the Snap configuration. With this fix, the Snap sends the input document to the output view, under the key original, only if you select the Pass through field.

    • Added the Snap Execution field to all Standard-mode Snaps. In some Snaps, this field replaces the existing Execute during preview check box.

    4.16 Patch 

    db/snowflake6945

     

    Latest

    Fixed an issue with the Snowflake Lookup Snap failing when Date datatype is used in JavaScript functions.

    4.16 Patch 

    db/snowflake6928

     

    Latest

    Added support for file format options for input data from upstream Snaps, to the Snowflake Bulk Load Snap.

    4.16 Patch 

    db/snowflake6819

     

    Latest

    • Snowflake Bulk Load: Added new property, Buffer size (MB). Configure this to specify the size limit of each buffer when writing to external staging systems such as S3. 

    • Fixed an issue with the Lookup Snap passing data simultaneously to output and error views when some values contained spaces at the end.

    4.16

    snapsmrc508

     

    Stable

    • Snowflake Account: Added the ability to use SnapLogic to securely connect to and query a Snowflake instance using Azure Blob as its storage layer. 

    • Snowflake Account: Added support for Snowflake JDBC driver JAR version 3.6.17. 

    • Snowflake Unload, Bulk Load, and Bulk Upsert: Updated the Snaps to enable SnapLogic users to successfully connect to a Snowflake instance to query, bulk load, and unload data from Azure Blob storage.

    4.15

    snapsmrc500

     

    Stable

    • Added two new Snaps, Snowflake - Multi Execute, and Snowflake SCD2. Snowflake - Multi Execute is used for executing multiple DDL and DML queries on the Snowflake DB. Snowflake SCD2 is used for Type 2 field historization. 

    • Enhanced the Snowflake Bulk Upsert Snaps to improve the Snaps performance.

    • Enhanced the Snowflake Snap Pack to reflect Azure certification.

    4.14

    snapsmrc490

     

    Stable

    Upgraded with the latest SnapLogic Platform release.

    4.13

    snapsmrc486

     

    Stable

    Upgraded with the latest SnapLogic Platform release.

    4.12

    snapsmrc480

     

    Stable

    Upgraded with the latest SnapLogic Platform release.

    4.11 Patch

    MULTIPLE4377

     

    Latest

    Fixed a document call issue that was slowing down the Snowflake Bulk Load Snap.

    4.11 Patch 

    db/snowflake4283

     

    Latest

    Snowflake Bulk Load - Fixed an issue by adding PUT command to the list of DDL command list for Snowflake.

    4.11 Patch 

    db/snowflake4273

     

    Latest

    Snowflake Bulk Load - Resolved an issue with Snowflake Bulk Load Delimiter Consistency (comma and newline).

    4.11

    snapsmrc465

     

    Stable

    Upgraded with the latest SnapLogic Platform release.

    4.10 Patch

    snowflake4133

    Latest

    Updated the Snowflake Bulk Load Snap with Preserve case sensitivity property to preserve the case sensitivity of column names.

    4.10

    snapsmrc414

     

    Stable

    • Updated the Snowflake Bulk Load Snap with Load empty strings property for the empty string values in the input documents to be loaded as empty strings to the string-type fields.

    • Updated the Snowflake Bulk Load Snap with Table Columns to support the order of the entries on the staged files that contain a subset of the columns in the Snowflake table.

    • Added the property Use Result Query to view the output preview field with a result statement.

    • Tested for the JDBC driver JAR version 3.1.1 on the Database and the Dynamic accounts.

    4.9.0 Patch

    snowflake3234

     

    Latest

    Enhanced Snowflake - Execute Snap results to include additional details

    4.9.0 Patch

    snowflake3125

     

    Latest

    Addressed an issue in Snowflake Bulk Load where the comma character in a value is not escaped.

    4.9

    snapsmrc405

     

    Stable

    JDBC Driver Class property added to enable the user to custom configure the JDBC driver in the Database and the Dynamic accounts.

    4.8.0 Patch

    snowflake2760

     

    Latest

    Potential fix for JDBC deadlock issue.

    4.8.0 Patch

    snowflake2739

     

    Latest

    Addressed an issue with the Snowflake schema not correctly represented in the Mapper Snap.

    4.8

    snapsmrc398

     

    Stable

    • Info tab added to accounts.

    • Database accounts now invalidate connection pools if account properties are modified and login attempts fail.

    • Enhanced the default count of input and output view (UI) behavior of the Snaps for better user experience. Note that the Snowflake Select, Table List and Lookup Snap views remain unchanged.

    • No labels