Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

Snap Pack

Date of Update

Snap Pack Version

Updates

Databricks

430patches17796Learn

more about the Databricks patch updates hereThe September 2022 patch release covers multiple enhancements across different Snaps in the Databricks Snap Pack. The complete list is available in the Snaps section of this document.

ELT

430patches18196Learn

more about the ELT patch updates hereThe September 2022 patch release covers multiple enhancements across different Snaps in the ELT Snap Pack. The complete list is available in the Data Automation section of the Release Notes.

Exchange Online

430patches18099

  • The User.Read.All permission is now not mandatory when configuring an application in the Azure portal for authorizing the Exchange Online OAuth2 User Account to read and send messages and create events. Even if you do not provide this permission, the account does not encounter Access Denied (403) error. The User.Read permission fetches only the current user details.

  • The Send Email Snap supports sending the HTML table in the email body using the following fields:

    • HTML Table option under Email type

    • Table Data Path

    • Batch Size

  • Enhancements in the Exchange Online Read Email Snap:

    • The Snap now downloads the inline attachments correctly when you select the Download attachments checkbox.

    • The Snap now shows the email folder suggestions and includes the subfolder names by prefixing the parent folder name with the delimiter ('/'). For example, custom/subfolder.

MongoDB

430patches18223

The MongoDB Update Snap in a low-latency feed Ultra Pipeline now correctly acknowledges the requests.

Transform

430patches18119

The Transcoder Snap used in a low-latency feed Ultra Pipeline now correctly acknowledges the requests.

Flow

430patches18070

The Pipeline Execute Snap with binary output that is used in a low-latency feed Ultra Pipeline now works as expected. The requests are now correctly acknowledged.

Binary

430patches17933

  • The File Delete Snap now correctly passes the input document to the error view.

  • The AWS S3 and S3 Dynamic accounts now support the maximum session duration of an IAM role defined in AWS.

Salesforce

430patches18036

The Salesforce Read Snap now correctly parses the 2-byte UTF-8 characters in Windows OS in PK chunking mode.

Transform

430patches17802

The Avro Parser Snap now displays the decimal number correctly in the output view if the column’s logical type is defined as a decimal.

Active Directory

430patches17872

The Azure Directory Search Snap does not fetch duplicate records when the Group result checkbox is deselected.

Snowflake

430patches17962

The Snowflake Bulk Load Snap now triggers the metadata query only once, even for invalid input, thereby improving the performance of the Snap.

JDBC

430patches17894

The Generic JDBC Snaps connecting to the DB2 database now take less time to execute, thereby improving the performance.

REST

430patches17851

The REST Post Snap now works without displaying any errors when the Show all headers checkbox is selected and the Content-type is text/xml or application/xml.

MySQL

Oracle

Snowflake

PostgreSQL

430patches17894

The following Snaps now work as expected when the table name is dependent on an upstream input:

Google AlloyDB

430patches17841

Introduced the Google AlloyDB Snap Pack— a fully managed PostgresSQL-compatible database service that you can use for all your database workloads. This Snap Pack offers the following Snaps:

Transform

430patches17737

AutoPrep enables you to handle empty or null values.

...

The following fields are added to each Databricks Snap as part of this enhancement:
Number of Retries: The number of attempts the Snap should make to perform the selected operation when the Snap account connection fails or times out.
Retry Interval (seconds): The time interval in seconds between two consecutive retry attempts.

Known Issues

Databricks Snap Pack: Due to an issue with DLP, aborting a Databricks Pipeline validation (with preview data enabled) causes only those SQL statements that retrieve data using bind parameters to get aborted while all other static statements (that use values instead of bind parameters) persist. For example, select * from a_table where id = 10 will not be aborted while select * from test where id = ? gets aborted.
To avoid this issue, ensure that you always configure your Snap settings to use bind parameters inside its SQL queries.

...

  • Due to an issue with the Redshift COPY INTO SQL for PARQUET input files, the ELT Load Snap fails to perform ALTER TABLE actions before loading data from a PARQUET source file to the target table in Redshift. This issue does not occur when the source file is in CSV or another format.

  • Due to an issue with BigQuery table schema management (the time travel feature), an ALTER TABLE action (Add or Update column) that you attempt after deleting a column (DROP action) in your BigQuery target table causes the table to break and the Snap to fail.

    • As a workaround, you can consider either avoiding ALTER TABLE actions on your BigQuery instance using the Snap or creating (CREATE) a temporary copy of your table and deleting (DROP) it after you use it.

  • Due to an issue with Databricks Runtime Version 11 and above, the ELT Aggregate Snap fails to calculate the value for the linear regression aggregate function REGR_R2 for the target DLP instance and returns a cast exception. As a workaround, you can revert your Databricks Runtime Version to 10.5 or below.

  • ELT Pipelines built using an ELT SCD2 or an ELT Load Snap fail to perform the DROP AND CREATE TABLE and ALTER TABLE operations on Delta tables when working with a Databricks SQL persona on the AWS Cloud with the error Operation not allowed: ALTER TABLE RENAME TO is not allowed for managed Delta tables on S3. However, the same action runs successfully with the Data Science and Engineering persona on the AWS Cloud.

    • This is due to the limitation with the Databricks SQL Admin Console that does not allow adding the configuration parameter spark.databricks.delta.alterTable.rename.enabledOnAWS true to its SQL Warehouse Settings.

  • Due to an issue with the Redshift COPY INTO SQL for PARQUET input files, the ELT Load Snap fails to perform ALTER TABLE actions before loading data from a PARQUET source file to the target table in Redshift. This issue does not occur when the source file is in CSV or another format.

  • The ELT Load Snap cannot automatically infer schema from GZIP-compressed source files located in an ADLS Gen2 storage location for loading the data into Snowflake target tables.

  • Because BigQuery does not support Automatic Schema Inferencing (ASI) from source files in Amazon S3, the ELT Load Snap fails to produce the required output when you choose Drop and Create table as the Load Action and define the Target to Source Column Map in the Snap’s settings. However, the same configuration works fine for Pipeline validation purposes.

  • Snowflake does not support COPY statements that result in data transformation in any of the validation modes it allows. Therefore, the ELT Load Snap fails during Pipeline validation when the number of columns you define in the Target to Source Column Map field set is not the same as the number of columns in the target table. But the same Pipeline runs successfully because this limitation does not apply to Pipeline execution.

    • As a workaround for this limitation, ensure that you specify error_on_column_count_mismatch=false in the Format Option field under the File Format Option List field set.

  • While you specify an SQL statement in the SQL Statement Editor of an ELT Snap as an expression, the dynamic validation for the expression displays inline errors when there is more than one incoming document and without the '__sql__' key to the current Snap, when you select Get Preview Data checkbox in the previous Snap, and when Preview Document Count in your user settings is set to a value more than 1.

    • To prevent this error and similar ones, do not select the Get Preview Data checkbox in the previous Snap, set the Preview Document Count in your user settings to 1, or append a condition where 1 = 0 to the SQL statement with the Get Preview Data checkbox selected.

  • Due to a limitation in Snowflake in inferring the source files (and thereby the source schema) based on file name patterns in a Storage Integration-based session, the ELT Load Snap fails to load data into your Snowflake target instance based on any File Name Pattern that you specify. As a workaround, we recommend that you use the File List field set to define the source data file locations for the Snap to load the data.

  • If you are using Azure Active Directory password--based authentication on the older Snap Pack version, you could encounter an error while using ELT Pipelines in Azure Synapse. This is due to absence of the following two JAR files in the ELT Snap Pack. Add these two JAR files to the list that you want to use to resolve the issue:

    • MSAL4J (Microsoft Authentication Library for Java) during compile time.

    • Nimbusds oauth for runtime.

    However, this error would not occur when you use the latest Snap Pack version (version number will included once released) 430patches18196.

IIP UI Redesigns

The UI for SnapLogic IIP is redesigned for the September release.

...