In this article
Table of Contents | ||||||
---|---|---|---|---|---|---|
|
...
You can use this account type to connect <Snap Pack name> Databricks Snaps with data sources that use <Account Type Name> accountthe Databricks account with Google Cloud Storage as source.
Prerequisites
Valid client ID.
Valid tenant URL.
...
...
Limitations
Known Issues
Supports only reading JSON files.
...A valid Databricks account.
Certified JDBC JAR File: databricks-jdbc-2.6.25-1.jar
Limitations and Known Issues
None.
Account Settings
<Insert image here>
Appsplus panel macro confluence macro | ||
---|---|---|
| ||
**Delete Before Publishing** Choose from the following sentences to document specific field types. Drop-down lists/Option Buttons (radio buttons): You must list the LoV and describe them if their meaning isn't apparent. In this case, format the LoV in italics, regular font for the LoV's description. In either case, list the LoVs as a bullet list.
Check boxes:
Text Fields
Numeric Text Fields
Notes in field descriptions
|
Info |
---|
|
...
Field Name
...
Field Type
...
Field Dependency
...
Description
Default Value:
Example:
Label*
Default Value: ELT Database Account
Example: ELT RS Account
...
String
...
None.
Specify a unique label for the account.
...
Field set
Specify advanced parameters that you want to include in the request.
This field set consists of the following fields:
Field 1
Field 2
Field 3
Field 1*
Default Value:
Example:
...
String
...
Debug mode checkbox is not selected.
Field 2
Default Value:
Example:
...
String
...
None.
Troubleshooting
...
Error
...
Reason
...
Resolution
...
Account validation failed.
...
The Pipeline ended before the batch could complete execution due to a connection error.
...
Verify that the Refresh token field is configured to handle the inputs properly. If you are not sure when the input data is available, configure this field as zero to keep the connection always open.
...
Related Links
...
Endpoint Doc Link 1
...
Endpoint Doc Link 2
...
Endpoint Doc Link 3
...
Info |
---|
|
Field Name | Field Type | Field Dependency | Description | ||
---|---|---|---|---|---|
Label* Default Value: N/A | String | None. | Specify a unique label for the account. | ||
Account Properties* | Use this field set to configure the information required to establish a JDBC connection with the account. | ||||
Download JDBC Driver Automatically Default Value: Not Selected | Checkbox | None. | Select this checkbox to allow the Snap account to download the certified JDBC Driver for DLP. The following fields are disabled when this checkbox is selected.
To use a JDBC Driver of your choice, clear this checkbox, upload (to SLDB), and choose the required JAR files in the JDBC JAR(s) and/or ZIP(s): JDBC Driver field. Use of Custom JDBC JAR version You can use a different JAR file version outside of the recommended listed JAR file versions. Spark JDBC and Databricks JDBC If you do not select this checkbox and use an older JDBC JAR file (older than version 2.6.25), ensure that you use:
| ||
JDBC Driver Class Default Value: com.databricks.client.jdbc.Driver | String | None. | Specify the JDBC driver class to use. | ||
JDBC JARs | Use this field set to define list of JDBC JAR files to be loaded. | ||||
JDBC Driver | String | None. | Specify or upload the JDBC driver to use.
| ||
JDBC URL* Default Value: N/A | String | None. | Enter the JDBC driver connection string that you want to use in the syntax provided below, for connecting to your DLP instance. See Microsoft's JDBC and ODBC drivers and configuration parameters for more information. jdbc:spark://dbc-ede87531-a2ce.cloud.databricks.com:443/default;transportMode=http;ssl=1;httpPath= Avoid passing Password inside the JDBC URL If you specify the password inside the JDBC URL, it is saved as it is and is not encrypted. We recommend passing your password using the Password field provided, instead, to ensure that your password is encrypted. | ||
Use Token Based Authentication Default value: Selected | Checkbox | None. | Select this checkbox to use token-based authentication for connecting to the target database (DLP) instance. Activates the Token field. | ||
Token* Default value: N/A | String | Appears when Use Token Based Authentication checkbox is selected. | Enter the token value for accessing the target database/folder path. | ||
Database name* Default Value: N/A | String | None. | Enter the name of the database to use by default. This database is used if you do not specify one in the Databricks Select or Databricks Insert Snaps. | ||
Source/Target Location* Default Value: None | Dropdown | None. | Select the target data warehouse. If you want to load the queries from ADLS Gen2 as source, then the selected datawarehouse would serve as a target and vice versa. Following are the options available:
This activates the following fields:
| ||
GCS Bucket Default Value: N/A | String | Appears when Google Cloud Storage is selected for Source/Target Location. | Specify the GCS Bucket to use for staging data to be used for loading to the target table. | ||
GCS Folder Default Value: N/A | String | Appears when Google Cloud Storage is selected for Source/Target Location. | Specify the relative path to a folder in the GCS Bucket. This is used as a root folder for staging data. | ||
GCS Authorization type Default Value: Service Account | String | Appears when Google Cloud Storage is selected for Source/Target Location. | Select the authentication type to load data. By default the authentication type is Service Account. | ||
Service Account Email* Default Value: N/A | String/Expression | Appears when Google Cloud Storage is selected for Source/Target Location. | Specify the service account email allowed to connect to the BigQuery database. This will be used as the default username when retrieving connections. The Email must be valid in order to set up the data source | ||
Service Account Key File Path* Default Value: N/A | String/Expression | Appears when Google Cloud Storage is selected for Source/Target Location. | Specify the path to Key file used to authenticate the service account email address with the BigQuery database. | ||
Advanced Properties | Other parameters that you want to specify to configure the account. | ||||
URL properties | Use this field set to define the account parameter's name and its corresponding value. Click + to add the parameters and the corresponding values. Add each URL property-value pair in a separate row. It consists of the following fields:
| ||||
URL property name Default Value: N/A | N/A | None. | Specify the name of the parameter for the URL property. | ||
URL property value Default Value: N/A | N/A | None. | Specify the value for the URL property parameter. | ||
Batch size* Default Value: N/A | Integer | None. | Specify the number of Snowflake queries that you want to execute at a time.
| ||
Fetch size* Default Value: 100 | Integer | None. | Specify the number of rows a query must fetch for each execution. Large values could cause the server to run out of memory. | ||
Min pool size* Default Value: 3 | Integer | None. | Specify the minimum number of idle connections that you want the pool to maintain at a time. | ||
Max pool size* Default Value: 15 | Integer | None. | Specify the maximum number of connections that you want the pool to maintain at a time. | ||
Max life time* Default Value: 60 | Integer | None. | Specify the maximum lifetime of a connection in the pool, in seconds.
Minimum value: 0 | ||
Idle Timeout* Default Value: 5 | Integer | None. | Specify the maximum amount of time in seconds that a connection is allowed to sit idle in the pool. 0 indicates that idle connections are never removed from the pool. Minimum value: 0 | ||
Checkout timeout* Default Value: 10000 | Integer | None. | Specify the maximum time in milliseconds you want the system to wait for a connection to become available when the pool is exhausted. If you provide 0, the Snap waits infinitely until the connection is available. Therefore, we recommend you not to specify 0 for Checkout Timeout. Minimum value: 0 |
Snap Pack History
Expand | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|
|
...