In this article
Table of Contents | ||||
---|---|---|---|---|
|
Overview
Use this account to connect ELT Snaps with data sources that use ELT database accounts. This account enables you to write transformed data to target databases hosted in the following cloud locations. The JDBC URL you define for your target database indicates the respective cloud location where the database is hosted.
...
Databricks Lakehouse Platform
...
Note |
---|
The ELT Snap Pack does not support mixed accounts from different types of databases in the same Pipeline. For example, a Pipeline in which some Snaps are connecting to the Snowflake database cannot have other Snaps connecting to the Redshift database. |
Prerequisites
A valid Snowflake, Redshift, Azure Synapse, or Databricks Lakehouse Platform database account.
Limitations
- Connects only to the Snowflake, Redshift, Azure Synapse, or Databricks Lakehouse Platform.
- In case of Azure Synapse database, the ELT database account supports the following:
- Only provisioned clusters of Azure Synapse, but not in Serverless mode.
- External locations:
- Blob Storage: https://.blob.core.windows.net//
- ADLS Gen2: https://.dfs.core.windows.net//
Known Issue
- When ELT and Spark SQL 2.x Snap account credentials—such as user names, passwords, client secrets, auth codes and tokens, secret keys, and keystores—are auto-filled using the Google Chrome browser, the accounts and hence the Pipelines fail. This is because the browser overwrites the field values with its own encrypted values that the SnapLogic Platform cannot read. SnapLogic recommends that you do not auto-save your Snap account credentials in the Chrome browser.
- Ensure that you delete any credentials that the browser has already saved for elastic.snaplogic.com, and then perform ONE of the following actions:
- Option 1: Click that appears in the address bar after you submit your login credentials at elastic.snaplogic.com, and then click Never.
- Option 2: Disable the Offer to save Passwords option at chrome://settings/passwords while working with your SnapLogic Pipelines. If you disable this option, your Chrome browser will not remember your passwords on any other website.
Account Settings
Panel | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|
| ||||||||||
Click the = (Expression) button in the Account's configuration, if available, to define the corresponding field value using expression language and Pipeline parameters. Fields marked with an asterisk ( * ) in the table below are mandatory. |
...
Description
...
Required. Unique user-provided label for the account.
Default Value: N/A
Example: ELT Account
...
Use this field set to configure the information required to establish a JDBC connection with the account.
This field set consists of the following fields:
- Database Type
- Download JDBC Driver Automatically
- JDBC JARs
- JDBC Driver Class
- JDBC URL
- Username
- Password
- Warehouse
- Default Database Name
...
Required. Select the target data warehouse into which the queries must be loaded.
Available options are:
- Snowflake. Activates the Warehouse and Source Location fields.
- Redshift. Activates the Source Location field.
- Azure Synapse. Activates the following fields:
- External Location
- External Storage Endpoint
- Storage Account
- Azure Container
- Azure Folder
- Azure Auth Type
- Azure Identity
- Azure Secret
- Databricks Lakehouse Platform. See Spark settings required for supporting ELT on DLP. Activates the following fields:
- Use Token Based Authentication
- Username
- Password
- Token
- Default Database Name
- DBFS Folder path (source for loading Databricks table)
- BigQuery. Activates the following fields:
- Default Database Name
- BigQuery Authentication Types
Default Value: N/A
Example: Snowflake
...
Select this checkbox to allow the Snap account to download the certified JDBC Driver based on the Database Type you have selected. The following fields are disabled when this checkbox is selected.
- JDBC JAR(s) and/or ZIP(s) : JDBC Driver
- JDBC driver class
To use a JDBC Driver of your choice, clear this checkbox, upload (to SLDB), and choose the required JAR files in the JDBC JAR(s) and/or ZIP(s): JDBC Driver field.
Use of Custom JDBC JAR version
You can use different JAR file versions than those listed here. We recommend that you use the listed JAR file versions. See the latest Release Notes for more details.
Default Value: Not Selected
Example: Selected
...
Upload the JDBC driver and other JAR files that you want to use into SLDB. Click to add a new row. Add each JDBC JAR file in a separate row. See Downloading the JDBC Driver for more information about JDBC drivers and downloading the appropriate driver for your account.
Default Value: N/A
Example: snowflake-jdbc-3.9.1.jar
...
Specify the driver class to use for your application.
We recommend that you use the following classes to suit your database type as other classes and methods may change due to future enhancements:
- Snowflake: net.snowflake.client.jdbc.SnowflakeDriver
- Redshift: com.amazon.redshift.jdbc42.Driver
- Azure Synapse: com.microsoft.sqlserver.jdbc.SQLServerDriver
- Databricks Lakehouse Platform: com.simba.spark.jdbc.Driver
Default Value: N/A
Example: net.snowflake.client.jdbc.SnowflakeDriver
...
Required. Enter the JDBC driver connection string that you want to use, based on the Database you are connecting to. Use the appropriate syntax provided below.
...
Alternatively, for Redshift, Azure Synapse, and Databricks Lakehouse Platform, you can make use of the Username, Password, and Database Name fields, along with the Advanced Properties > URL Properties field set to provide the parameters required for building your JDBC URL. See Passing your JDBC URL for more information.
Note | ||
---|---|---|
| ||
If you specify the password inside the JDBC URL, it is saved as it is and is not encrypted. We recommend passing your password using the Password field provided, instead, to ensure that your password is encrypted. |
Default Value: N/A
Example: jdbc:snowflake://xy12345.us-east-2.aws.snowflakecomputing.com, jdbc:snowflake://xy12345.east-us-2.azure.aws.snowflakecomputing.com
...
Use Token Based Authentication
...
Select this checkbox to use token-based authentication for connecting to the target database (DLP) instance. Activates the Token field.
Default value: Selected
Example: Not selected
...
Enter the username provided for your database account.
...
In this article
Table of Contents | ||||
---|---|---|---|---|
|
Overview
Use this account to connect ELT Snaps with data sources that use ELT database accounts. This account enables you to write transformed data to target databases hosted in the following cloud locations. The JDBC URL you define for your target database indicates the respective cloud location where the database is hosted.
Target Database | Supported Cloud Location | Cloud Location in JDBC URL |
---|---|---|
Snowflake | AWS | jdbc:snowflake://<account_name>.snowflakecomputing.com |
Microsoft Azure | jdbc:snowflake://<account_name>.<region>.azure.snowflakecomputing.com | |
Google Cloud Platform (GCP) | jdbc:snowflake://<account_name>.<region>.gcp.snowflakecomputing.com | |
Redshift | AWS | jdbc:redshift://<redshift-cluster-name>.<region>.redshift.amazonaws.com |
Azure Synapse | Microsoft Azure | jdbc:sqlserver://<yourserver>.database.windows.net |
Databricks Lakehouse Platform | Microsoft Azure | jdbc:spark://<your_instance_code>.cloud.databricks.com |
Google BigQuery | Google Cloud Platform (GCP) | jdbc:bigquery://<host_URL>/bigquery |
Note |
---|
The ELT Snap Pack does not support mixed accounts from different types of databases in the same Pipeline. For example, a Pipeline in which some Snaps are connecting to the Snowflake database cannot have other Snaps connecting to the Redshift database. |
Prerequisites
A valid Snowflake, Redshift, Azure Synapse, or Databricks Lakehouse Platform database account.
Limitations
- Connects only to the Snowflake, Redshift, Azure Synapse, or Databricks Lakehouse Platform.
- In case of Azure Synapse database, the ELT database account supports the following:
- Only provisioned clusters of Azure Synapse, but not in Serverless mode.
- External locations:
- Blob Storage: https://.blob.core.windows.net//
- ADLS Gen2: https://.dfs.core.windows.net//
Known Issue
- When ELT and Spark SQL 2.x Snap account credentials—such as user names, passwords, client secrets, auth codes and tokens, secret keys, and keystores—are auto-filled using the Google Chrome browser, the accounts and hence the Pipelines fail. This is because the browser overwrites the field values with its own encrypted values that the SnapLogic Platform cannot read. SnapLogic recommends that you do not auto-save your Snap account credentials in the Chrome browser.
- Ensure that you delete any credentials that the browser has already saved for elastic.snaplogic.com, and then perform ONE of the following actions:
- Option 1: Click that appears in the address bar after you submit your login credentials at elastic.snaplogic.com, and then click Never.
- Option 2: Disable the Offer to save Passwords option at chrome://settings/passwords while working with your SnapLogic Pipelines. If you disable this option, your Chrome browser will not remember your passwords on any other website.
Account Settings
Panel | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|
| ||||||||||
Click the = (Expression) button in the Account's configuration, if available, to define the corresponding field value using expression language and Pipeline parameters. Fields marked with an asterisk ( * ) in the table below are mandatory. |
Parameter | Field Dependency | Description | ||||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Label* | None. | Required. Unique user-provided label for the account. Default Value: N/A Example: ELT Account | ||||||||||||||||||||||
Account Properties* | Use this field set to configure the information required to establish a JDBC connection with the account. This field set consists of the following fields:
| |||||||||||||||||||||||
Database Type* | None. | Required. Select the target data warehouse into which the queries must be loaded. Available options are:
Default Value: N/A Example: Snowflake | ||||||||||||||||||||||
Download JDBC Driver Automatically | Check box | Select this checkbox to allow the Snap account to download the certified JDBC Driver based on the Database Type you have selected. The following fields are disabled when this checkbox is selected.
To use a JDBC Driver of your choice, clear this checkbox, upload (to SLDB), and choose the required JAR files in the JDBC JAR(s) and/or ZIP(s): JDBC Driver field. Use of Custom JDBC JAR version You can use different JAR file versions than those listed here. We recommend that you use the listed JAR file versions. See the latest Release Notes for more details. Default Value: Not Selected Example: Selected | ||||||||||||||||||||||
JDBC JAR(s) and/or ZIP(s): JDBC Driver | Required when the Download JDBC Driver Automatically check box is not selected. | Upload the JDBC driver and other JAR files that you want to use into SLDB. Click to add a new row. Add each JDBC JAR file in a separate row. See Downloading the JDBC Driver for more information about JDBC drivers and downloading the appropriate driver for your account. Default Value: N/A Example: snowflake-jdbc-3.9.1.jar | ||||||||||||||||||||||
JDBC driver class* | Required when the Download JDBC Driver Automatically check box is not selected. | Specify the driver class to use for your application. We recommend that you use the following classes to suit your database type as other classes and methods may change due to future enhancements:
Default Value: N/A Example: net.snowflake.client.jdbc.SnowflakeDriver | ||||||||||||||||||||||
JDBC URL* | None. | Required. Enter the JDBC driver connection string that you want to use, based on the Database you are connecting to. Use the appropriate syntax provided below.
Alternatively, for Redshift, Azure Synapse, and Databricks Lakehouse Platform, you can make use of the Username, Password, and Database Name fields, along with the Advanced Properties > URL Properties field set to provide the parameters required for building your JDBC URL. See Passing your JDBC URL for more information.
Default Value: N/A Example: jdbc:snowflake://xy12345.us-east-2.aws.snowflakecomputing.com, jdbc:snowflake://xy12345.east-us-2.azure.aws.snowflakecomputing.com | ||||||||||||||||||||||
Use Token Based Authentication | Database Type is Databricks Lakehouse Platform | Select this checkbox to use token-based authentication for connecting to the target database (DLP) instance. Activates the Token field. Default value: Selected Example: Not selected | ||||||||||||||||||||||
Username* | When Use Token Based Authentication check box is not selected. | Enter the username provided for your database account.
Default value: N/A Example: Bigdatasnaplogic | ||||||||||||||||||||||
Password* | When Use Token Based Authentication check box is not selected. | Enter the password for the username specified above.
Default value: N/A Example: <Encrypted> | ||||||||||||||||||||||
Token* | When Use Token Based Authentication check box is selected. | Enter the token value for accessing the target database/folder path. Default value: N/A Example: <Encrypted> | ||||||||||||||||||||||
Warehouse* | Database Type is Snowflake | Enter the name of the Snowflake warehouse that you want to use. Default value: N/A Example: 12345abcdefghikj1km2345no | ||||||||||||||||||||||
Default Database Name* | String | Enter the name of the database to use by default. This database is used if you do not specify one in the ELT Select, ELT Insert-Select, or ELT Merge Into Snaps. Default value: N/A Example: EMPLOYEEDB | ||||||||||||||||||||||
DBFS Folder path (source for loading Databricks table) | Database Type is Databricks Lakehouse Platform | Required for ELT Load Snap. Specify the fully qualified path to a target folder in your DBFS instance. It should begin with / which denotes the DBFS Root folder.
Default value: N/A Example: /my_DBFS/mounted0408_folder | ||||||||||||||||||||||
BigQuery Authentication Types* | Database Type is BigQuery | Select the authentication type that you use to connect to your Google BigQuery instance. It can be one of the following:
Default value: Service Account Example: OAuth 2.0 Access Token Account | ||||||||||||||||||||||
Service Account Email* | Database Type is BigQuery, BigQuery Authentication Types is Service Account | Specify the Google email account that is associated with your Google Cloud project. You would need the associated service account key file to authenticate this account. See Creating a Service Account and Authenticating with a service account key file for more information. Default value: N/A Example: my_service_account@domain.com | ||||||||||||||||||||||
Service Account Key File Path* | Database Type is BigQuery, BigQuery Authentication Types is Service Account | Specify the SLDB location (file path) of the service account key file associated with this service account. Default value: N/A Example: | Bigdatasnaplogic||||||||||||||||||||||
Password* | When Use Token Based Authentication check box is not selected. | Enter the password for the username specified above.
Default value: N/A Example: <Encrypted> | ||||||||||||||||||||||
Token* | When Use Token Based Authentication check box is selected. | Enter the token value for accessing the target database/folder path /my_SA/key_file/ | ||||||||||||||||||||||
OAuth 2.0 Access Token | Database Type is BigQuery, BigQuery Authentication Types is OAuth 2.0 Access Token Account | Enter the OAuth 2.0 Access Token pertaining to your BigQuery instance. See Obtaining Access Tokens for the steps to generate and use this access token.
Default value: N/A Example: | Warehouse*ya29.Ci9deZOky9V36Tz497HY1chAA2sA8J_wM8e5FnY9rJg551153GQWGbleO-y9apjLFg | |||||||||||||||||||||
OAuth2 Refresh Token | Database Type | is SnowflakeEnter the name of the Snowflake warehouse that you want to use. Default value: N/A Example: 12345abcdefghikj1km2345no | Default Database Name* | String | Enter the name of the database to use by default. This database is used if you do not specify one in the ELT Select, ELT Insert-Select, or ELT Merge Into Snaps.is BigQuery, BigQuery Authentication Types is OAuth 2.0 Refresh Token Account | Provide the OAuth2 Refresh Token string that the Snap can use to automatically generate a valid access token—that you needed to run SQL queries in BigQuery—whenever an access token expires. This method of authentication helps you to avoid constantly updating your Snap accounts in BigQuery ELT Snaps.
Default value: N/A | DBFS Folder path (source for loading Databricks table) | Database Type is Databricks Lakehouse Platform | Required for ELT Load Snap. Specify the fully qualified path to a target folder in your DBFS instance. It should begin with / which denotes the DBFS Root folder. Note | mnt_point_to_s3 using DBFS mount command in Databricks Python Notebook.Example: | EMPLOYEEDBAA2sA8J_wM8e5FnY9rJg551153GQWGbleO-y9apjLFgya29.Ci9deZOky9V36Tz497HY1ch | |||||||||||||
OAuth2 Client ID* | Database Type is BigQuery, BigQuery Authentication Types is OAuth 2.0 Refresh Token Account | Enter the OAuth2 client ID associated with your BigQuery application. The Snap uses this value along with Secret Key and OAuth2 Refresh Token to generate the required Access Token. Default value: N/A Example: 1055912624410-6a29rmvnogteqt99781q0gm4vp2k7v9a.apps.googleusercontent.com | ||||||||||||||||||||||
OAuth2 Client Secret* | Database Type is BigQuery, BigQuery Authentication Types is OAuth 2.0 Refresh Token Account | Default value: N/A Example: /my_DBFS/mounted0408_folderbOqUqFQdusfInjfDlh2FDtbD | ||||||||||||||||||||||
Source Location* | Database Type is Snowflake, Redshift or BigQuery | Select the source data warehouse from which to load data into the target database. You must configure this field if you want to use the ELT Load Snap. Available options are:
Default value: None Example: S3, Azure, Google Cloud Storage | ||||||||||||||||||||||
Authentication | Source Location is S3, Azure, or Google Cloud Storage | Select the authentication method to use for the loaded data. Available options are:
Default value: Source Location Credentials Example: Storage Integration | ||||||||||||||||||||||
Storage Integration | Source Location is S3, Azure, or Google Cloud Storage and Authentication is Storage Integration. | Enter the name of the storage integration in your Snowflake database. Navigate to one of the following links for more information on setting up your Storage Integration in Snowflake.
Default value: N/A Example: my_s3_integration, my_azure_integration, my_sf_gcs_integration | ||||||||||||||||||||||
IAM Role | Database Type is Redshift, Source Location is S3, and Authentication is IAM Role. | Enter the IAM role associated with the target Redshift cluster. See Create an IAM Role for details on creating and associating IAM Role with your Redshift cluster. Default value: N/A Example: myiamrole | ||||||||||||||||||||||
Access Key | Source Location is S3 or Amazon Redshift, and Authentication is Source Location Credentials. | Enter the access key ID associated with your AWS S3 account. Default value: N/A Example: ABCDEFGHIJKL1MNOPQRS | ||||||||||||||||||||||
Secret Key | Source Location is S3S3 or Amazon Redshift, and Authentication is Source Location Credentials. | Enter the client secret key associated with your AWS S3 account. Default value: N/A Example: aBcdeFGhiJKLM/N1OPQRS/tUvWxyZABCDEFGHIJKLMN | ||||||||||||||||||||||
SAS Token | Source Location is Azure, and Authentication is Source Location Credentials. | Enter the SAS token part of the SAS URI associated with your Azure storage account. See Getting Started with SAS for details. Default value: N/A Example: ?sv=2020-08-05&st=2020-08-29T22%3A18%3A26Z&se=2020-08-30T02%3A23%3A26Z&sr=b&sp=rw&sip=198.1.2.60-198.1.2.70&spr=https&sig=A%1DEFGH1Ijk2Lm3noI3OlWTjEg2tYkboXr1P9ZUXDtkk%3D | ||||||||||||||||||||||
Session Access Key | Source Location is S3, Database Type is Snowflake, Redshift or Databricks Lakehouse Platform, and Authentication is Source Location Session Credentials. | Enter the access key ID associated with your AWS S3 account. Default value: N/A Example: ABCDEFGHIJKL1MNOPQRS | ||||||||||||||||||||||
Session Secret Key | Source Location is S3, Database Type is Snowflake, Redshift or Databricks Lakehouse Platform, and Authentication is Source Location Session Credentials. | Enter the client secret key associated with your AWS S3 account. Default value: N/A Example: aBcdeFGhiJKLM/N1OPQRS/tUvWxyZABCDEFGHIJKLMN | ||||||||||||||||||||||
Session Token | Source Location is S3, Database Type is Snowflake, Redshift or Databricks Lakehouse Platform, and Authentication is Source Location Session Credentials. | Enter the session token value when the authorization to access the S3 folder is based on Security Token Service (STS) based temporary credentials. Default value: N/A Example: A%1DEFGH1Ijk2Lm3noI3OlWTjEg2tYkboXr1P9ZUXDtkk%3D | ||||||||||||||||||||||
Storage Account | Source Location is S3. | Enter the name of your Azure storage account.
Default value: N/A Example: employeedata | ||||||||||||||||||||||
S3 Bucket | Source Location is S3 or Amazon Redshift. | Enter the name of the bucket from which to load the data. Default value: N/A Example: employeedata | ||||||||||||||||||||||
S3 Bucket Region | Database Type is Redshift and Source Location is S3. | Select the region of your Redshift cluster if it is different from the region of the S3 bucket. Available options are:
Default value: Default Example: us-east-1, us-east-2 | ||||||||||||||||||||||
S3 Folder | Source Location is S3 or Amazon Redshift. | Enter the name of the folder in the S3 bucket specified above where the source files are located. Default value: N/A Example: hrdepartment | ||||||||||||||||||||||
External Location (For Azure Synapse only) | Database Type is Azure Synapse. | Select the source data location from which to load data into the target database. You must configure this field if you want to use the ELT Load Snap. Available options are:
Default value: Blob Storage Example: Azure Data Lake Gen2 | ||||||||||||||||||||||
External Storage Endpoint (For Azure Synapse only) | Database Type is Azure Synapse. | Enter the end point path for the selected external location. You must configure this field if you want to use the ELT Load Snap. Default value: N/A Example: .blob.core.windows.net (for Blob Storage), .dfs.core.windows.net (for ADLS Gen2) | ||||||||||||||||||||||
Storage Account (For Azure Synapse only) | Database Type is Azure Synapse. | Enter your Account name to access the selected Azure external storage location. You must configure this field if you want to use the ELT Load Snap. Default value: N/A Example: adlsgen2v02 | ||||||||||||||||||||||
Azure Container | Source Location is Azure. | Enter the name of the container in the Azure storage account.
Default value: N/A Example: hrdepartment | ||||||||||||||||||||||
Azure Folder | Source Location is Azure. | Enter the relative path for the folder within the container specified above. Leave this field blank to use all the folders in the specified container.
Default value: N/A Example: salaries | ||||||||||||||||||||||
Google Cloud Storage (GCS) Bucket | Source Location is Google Cloud Storage and Authentication is Storage Integration. | Enter the name of the GCS bucket from which to load the data to your Snowflake database. Default value: N/A Example: elt_gcs_bucket_1 | ||||||||||||||||||||||
Google Cloud Storage (GCS) Folder | Source Location is Google Cloud Storage and Authentication is Storage Integration. | Enter the name of the folder in the GCS bucket where the source files are located. Default value: N/A Example: elt_gcs_bucket_1_CSV_Files | ||||||||||||||||||||||
Azure Auth Type (For Azure Synapse only) | Database Type is Azure Synapse. | Select the authentication type to use for accessing the selected Azure external storage location. Then, provide the values for Azure Identity, Azure Secret, Storage Key fields for the selected authentication type. See COPY INTO (Transact-SQL) in Microsoft Docs for more information on these authentication types. Available options are:
Default value: Storage Account Key Example: AAD User | Azure Identity (For Azure Synapse only) | Azure Auth Type is AAD User | .Enter the Azure Identity (the constant value) corresponding to the selected authentication type. See COPY INTO (Transact-SQL) in Microsoft Docs for more information on the suggested values for this field. Default value: N/A Example: Storage Account Key | |||||||||||||||||||
Azure Secret (For Azure Synapse only) | Azure Auth Type is Storage Account Key, Shared Access Signature, or Service Principals. | Enter the Azure Secret corresponding to the selected authentication type. See COPY INTO (Transact-SQL) in Microsoft Docs for more information on the expected values for this field. Default value: N/A Example: UvhVF65Bukk9hVF65BuUvhVF | ||||||||||||||||||||||
Storage Key (For Azure Synapse only) | Azure Auth Type is Shared Access Signature, Managed Identity, or Service Principals. | Enter the storage access key value corresponding to the Azure account mentioned in Storage Account field. See COPY INTO (Transact-SQL) in Microsoft Docs for more information on the expected values for this field.
Default value: N/A Example: ufv!befDIuf#fnb$KH&_hweuf | ||||||||||||||||||||||
Client ID | Azure Auth Type is AAD User. | Enter the Client ID of the application created in the Azure Active Directory portal -https://portal.azure.com.
Default value: N/A Example: hewr42ap-32jd-pd95-ms38-b342bnasdh80 | ||||||||||||||||||||||
Client Secret | Azure Auth Type is AAD User. | Enter the Client ID of the application created in the Azure Active Directory portal - https://portal.azure.com.
Default value: N/A Example: jhweufv!befDIufwb*dsfwujfnb$KH&_ | ||||||||||||||||||||||
Tenant ID | Azure Auth Type is AAD User. | Enter the Tenant ID of the application created in the Azure Active Directory portal - https://portal.azure.com.
Default value: N/A Example: b342bnas-he38-apms-dhjd-pd95wr423280 | ||||||||||||||||||||||
Advanced Properties | Other parameters that you want to specify as URL properties. See the following resources for a list of parameters that can be specified in this field set.
In case of Snowflake, DO NOT configure parameters that are already present in the Account Properties field set. For example:
This field set consists of the following fields:
| |||||||||||||||||||||||
URL Properties | String/Integer | The account parameter's name and its corresponding value. Click + to add more rows. Add each URL property-value pair in a separate row. Specify the name of the parameter in the URL Property Name field and its value in the URL Property Value field. |
Click Validate after entering the required details to ensure that all fields have been filled accurately. Click Apply to save the settings.
Note | ||
---|---|---|
| ||
If you have used Pipeline parameters or expressions to define values for the account fields above, the account validation (done by clicking the Validate button) is not supported. However, the Snaps that use this account may connect to the endpoint successfully depending on the accuracy and validity of the expressions and parameters used. |
...
Supported CDW | Certified JDBC JAR File |
---|---|
Azure Synapse | mssql-jdbc-8.4.1.jre8.jar |
BigQuery | SimbaJDBCDriverforGoogleBigQuery42_1.2.19.1023.zip |
Databricks Lakehouse Platform (DLP) | SimbaSparkJDBC42-2.6.21.1021.jar |
Redshift | redshift-jdbc42-1.2.43.1067.jar |
Snowflake | snowflake-jdbc-3.12.16.jar |
...