...
Multiexcerpt include macro | ||||
---|---|---|---|---|
|
Known Issue
- When your Databricks Lakehouse Platform instance uses Databricks Runtime Version 8.4 or lower, ELT operations involving large amounts of data might fail due to the smaller memory capacity of 536870912 bytes (512MB) allocated by default. This issue does not occur if you are using Databricks Runtime Version 9.0.
- ELT Pipelines created prior to 4.24 GA release using one or more of the ELT Insert Select, ELT Merge Into, ELT Load, and ELT Execute Snaps may fail to show expected preview data due to a common change made across the Snap Pack for the current release (4.26 GA). In such a scenario, replace the Snap in your Pipeline with the same Snap from the Asset Palette and configure the Snap's Settings again.
- In case you are writing into a Snowflake target table, this Snap attempts to create the target table even when it exists in the database.
- Suggestions displayed for the Schema Name field in this Snap are from all databases that the Snap account user can access, instead of the specific database selected in the Snap's account or Settings.
...
Info | ||
---|---|---|
| ||
You can use the SQL Expressions and Functions supported for ELT to define your Snap or Account settings with the Expression symbol = enabled, where available. This list is common to all target CDWs supported. You can also use other expressions/functions that your target CDW supports. |
...
Parameter Name | Data Type | Description | Default Value | Example | ||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
Label | String |
| ELT Insert-Select | Insert Employee Records | ||||||||
Get preview data | Check box |
| Not selected | Selected | ||||||||
Database Name | String | Required. Enter the name of the database in which the target table is located. Leave it blank to use the database name specified in the account settings. If your target database is Databricks Lakehouse Platform (DLP), you can, alternatively, mention the file format type for your table path in this field. For example, DELTA, CSV, JSON, ORC, AVRO. See Table Path Management for DLP section below to understand the Snap's behavior towards table paths. | N/A | TESTDB | ||||||||
Schema Name (Not applicable to Databricks Lakehouse Platform) | String | Required. Enter the name of the database schema. In case it is not defined, then the suggestion for the schema name retrieves all schema names in the specified database when you click .
| N/A | "TEST_DATA" | ||||||||
Target Table Name | String | Required. The name of the table into which you want to insert the data. If your target database is Databricks Lakehouse Platform (DLP), you can, alternatively, mention the target table path in this field. Enclose the DBFS table path between two
| N/A | "TEST_DATA"."DIRECT" EMPLOYEE_DATA EMPLOYEE_123_DATA REVENUE"-"OUTLET "net_revenue" | ||||||||
Target Table Hash Distribution Column (Azure Synapse Only) | String/Expression | Specify the Hash distribution column name for the target table (in Azure Synapse), if the Snap creates a target table during the execution of the Snap. If the target table is created outside the Snap, you need not specify the target table column name.
| N/A | var table | ||||||||
Insert Expression | This field set enables you to specify the values for a subset of the columns in the target table. The remaining columns are assigned null values automatically. You must specify each column in a separate row. Click to add rows. This field set consists of the following fields:
| |||||||||||
Insert Column | String | Enter the name of the column in the target table to assign values. | N/A | ORD_AMOUNT | ||||||||
Insert Value | String | Enter the value to assign in the specified column. Repeat the column name if you want to use the values in the source table. You can also use expressions to transform the values. | N/A | ORD_AMOUNT ORD_AMOUNT+20 | ||||||||
Overwrite | Check box | Select to overwrite the data in the target table. If not selected, the incoming data is appended. | Not selected | Selected |
...