Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

In this article

Table of Contents
minLevel1
maxLevel2
absoluteUrltrue

...

You can use this account type to connect <Snap Pack name> Snaps with data sources that use <Account Type Name> Generic Hive Database account.

Prerequisites

  • Valid client ID.

  • Valid tenant URL.

  • ...

  • ...

Limitations

Known Issues

  • Supports only reading JSON files.

  • ....

Account Settings

<Insert image here>

Appsplus panel macro confluence macro
data{"features":["title","icon","rounded","collapsible","border"],"title":"Documenting Fields Based On Data Type/UI Element","titleColor":"#000000","titleColorBG":"#ff5c30","titleSize":"14","titleBold":true,"titleItalic":false,"titleUnderline":false,"lozenge":"Hot stuff","lozengeColor":"#172B4D","lozengeColorBG":"#fff0b3","height":200,"panelPadding":12,"panelColor":"#172B4D","panelColorBG":"#FFFFFF","borderColor":"#ff5c30","borderRadius":3,"borderStyle":"solid","borderWidth":1,"icon":"editor/info","iconPrimary":"#FFFFFF","iconSecondary":"#0052CC","newMacro":false}

**Delete Before Publishing**

Choose from the following sentences to document specific field types.

Drop-down lists/Option Buttons (radio buttons):

You must list the LoV and describe them if their meaning isn't apparent. In this case, format the LoV in italics, regular font for the LoV's description. In either case, list the LoVs as a bullet list.

  • <State what the should do in this field>. The available options are: <bullet list of LoVs>
    Specify the Salesforce API to be used. The available options are:...

  • Select the <category> that you want to use. Available options are...
    * Option 1<italicized>. <third person singular form of the verb>
    * Option 2<italicized>. <third person singular form of the verb>
    Select the API that you want to use. Available options are:
    Bulk API. Sends the Snap execution request details as a bulk API call.
    REST API. ...
    OR
    Select one of the three following modes in which the Snap executes:
    * Validate & Execute. Performs limited execution of the Snap and generates a data preview during Pipeline validation, then performs full execution of the Snap (unlimited records) during Pipeline runtime.
    * Execute only. Performs full execution of the Snap during Pipeline execution without generating preview data.
    * Disabled. Disables the Snap and all Snaps downstream from it.

Check boxes:

  • If selected, <Snap behavior>.
    If selected, an empty file is written when the incoming document has no data.

  • If selected, <behavior>. If not selected/Otherwise, <behavior>
    Use "If not selected" if the first sentence is long.
    If selected, the Snap uses the file path value as is. Otherwise, the Snap uses the file path value in the URL.
    If selected, an empty file is written when the incoming document has empty data. If there is no incoming document at the input view of the Snap, no file is written regardless of the value of the property.

  • Select to <action>
    Use this if the behavior is binary. Either this or that, where the converse behavior is apparent/obvious.
    Select to execute the Pipeline during validation.

Text Fields

  • Describe what the user shall specify in this field. Additional details, as applicable, in a separate sentence. Include caveats such as the field being conditionally mandatory, limitations, etc.
    Enter the name for new account.
    Specify the account ID to use to log in to the endpoint.
    Required if IAM Role is selected.
    Do not use this field if you are using batch processing.

Numeric Text Fields

  • Describe what the field represents/contains. Additional details, as applicable, in a separate sentence. Include caveats such as the field being conditionally mandatory, limitations, etc. Include special values that impact the field's behavior as a bullet list.
    The number of records in a batch.
    The number of seconds for which you want the Snap to wait between retries.
    The number of seconds for which the Snap waits between retries.
    Use the following special values:
    * 0: Disables batching.
    * 1: Includes all documents in a single request.

Notes in field descriptions

  • Confluence’s new editor does not allow nesting of most macros inside another macro, especially the Note/Alert/Warning/Info (Panel) macros inside a table macro and Excerpt macros inside Expand or Panel Macro+ macros. So, as a workaround use the Footnotes approach as mentioned below:

    • Assign numbers at the Note locations in the form of follow through phrases like See Note 2 below this table. or such.

    • Add your Notes---an appropriate Note/Alert/Warning/Info (Panel) macro---immediately below the macro (for example, Table macro) beginning the content with the corresponding number assigned.

Info
  • Asterisk ( * ): Indicates a mandatory field.

  • Suggestion icon ( (blue star) ): Indicates a list that is dynamically populated based on the configuration.

  • Expression icon ( (blue star) ): Indicates the value is an expression (if enabled) or a static value (if disabled). Learn more about Using Expressions in SnapLogic.

  • Add icon ( (blue star) ): Indicates that you can add fields in the fieldset.

  • Remove icon ( (blue star) ): Indicates that you can remove fields from the fieldset.

...

Field Name

...

Field Type

...

Field Dependency

...

Description

Default Value:

Example:

Label*

Default ValueELT Database Account
ExampleELT RS Account

...

String

...

None.

Specify a unique label for the account.

...

Field set

Specify advanced parameters that you want to include in the request.

This field set consists of the following fields:

  • Field 1

  • Field 2

  • Field 3

Field 1*

Default Value
Example

...

String

...

Debug mode checkbox is not selected.

Field 2

Default Value
Example

...

String

...

None.

Troubleshooting

...

Error

...

Reason

...

Resolution

...

Account validation failed.

...

The Pipeline ended before the batch could complete execution due to a connection error.

...

Verify that the Refresh token field is configured to handle the inputs properly. If you are not sure when the input data is available, configure this field as zero to keep the connection always open.

...

...

Endpoint Doc Link 1

...

Endpoint Doc Link 2

...

None.

Limitations

  • The Hive Snap Pack does not validate with Apache Hive JDBC v1.2.1 jars or earlier because of a defect in Hive. HDP 2.6.3 and HDP 2.6.1 run on Apache Hive JDBC v1.2.1 jars.

  • To validate Snaps that must work with HDP 2.6.3 and HDP 2.6.1, use JDBC v2.0.0 jars.

Known Issues

  • "Method not supported" error while validating Apache Hive JDBC v1.2.1 or earlier

Account Settings

...

Info
  • Asterisk ( * ): Indicates a mandatory field.

  • Suggestion icon ( (blue star) ): Indicates a list that is dynamically populated based on the configuration.

  • Expression icon ( (blue star) ): Indicates the value is an expression (if enabled) or a static value (if disabled). Learn more about Using Expressions in SnapLogic.

  • Add icon ( (blue star) ): Indicates that you can add fields in the fieldset.

  • Remove icon ( (blue star) ): Indicates that you can remove fields from the fieldset.

Field Name

Field Type

Description

Label*

Default Value: N/A
Example: Snapuser 

String

Specify a unique label for the account.

Account properties

Username

Default Value: N/A
Example: Snapuser 

String

Specify the username that is allowed to connect to the database. Username will be used as the default username when retrieving connections. The username must be valid in order to set up the data source.

Password

Default Value: N/A
Example: Sn@pUser.3

String

Specify the password used to connect to the data source. Password will be used as the default when retrieving connections. The password must be valid in order to set up the data source.

JDBC URL

Default Value: N/A
Example: jdbc:hive://hostname/dbname:sasl.qop=auth-int

String/Expression

Specify the URL of the JDBC database.

JDBC JARs

Use this filedset to specify the JDBC JAR files to be loaded. A different driver binary for a driver must have a different name, the same name can not be reused for a different driver. If this property is left blank, a default JDBC driver will be loaded.

Enter the following JDBC jars to configure the Generic Hive Database account for the concerned cluster.

For HDP

  • Hive-jdbc-2.0.0.2.3.5.0-81-standalone.jar

  • Zookeeper-3.4.6.jar (Use this for setting up Hive with Zookeeper)

For CDH

  • hive_metastore.jar

  • hive_service.jar

  • hiveJDBC4.jar

  • libfb303-0.9.0.jar

  • libthrift-0.9.0.jar

  • TCLIServiceClient.jar

  • Zookeeper-3.3.6.jar (Use this for setting up Hive with Zookeeper)

  • The JDBC driver can be uploaded through Designer or Manager and it is stored on a per-project basis. That is, only users with access to that project will see JDBC drivers uploaded. To provide access to all users of your org, place the driver in the /shared project.

  • See Advanced Configurations: Configuring Hive with Kerberos section below for a list of JAR files to be uploaded when configuring Hive with Kerberos.

JDBC Driver Class

Default Value: N/A
Example: jdbc:hive://hostname/dbname:sasl.qop=auth-int

String

Specify the JDBC Driver class name. 

For HDP Clusters

Enter the following value: org.apache.hive.jdbc.HiveDriver

For CDH Clusters

Enter the following value: com.cloudera.hive.jdbc4.HS2Driver

Advanced properties

Auto commit

Default Value: Selected

Checkbox

Select this checkbox to commit a batch immediately after the batch executes. So, only the current executing batch will be rolled back if the Snap fails. If you deselect, then a transaction is started for the Snap run and committed upon run success. The transaction is rolled back if the Snap fails.

For a DB Execute Snap, assume that a stream of documents enter the input view of the Snap and the SQL statement property has JSON paths in the WHERE clause. If the number of documents are large, the Snap executes in more than one batches rather than executing one per each document. Each batch would contain a certain number of WHERE clause values. If Auto commit is turned on, a failure would only roll back the records in the current batch. If Auto commit is turned off, the entire operation would be rolled back. For a single execute statement (with no input view), the setting has no practical effect.

Batch size*

Default Value: 50
Example: 10

Integer

Specify the number of statements to execute at a time. Using a large batch size could use up the JDBC placeholder limit of 2100.

Fetch size*

Default Value: 100
Example: 100

Integer

Specify the number of rows to fetch at a time when executing a query. Large values could cause the server to run out of memory.

Max pool size*

Default Value: 50
Example: 10

Integer

Specify the maximum number of idle connections a pool will maintain at a time.

Max idle time*

Default Value: 30
Example: 30

Integer

Specify the minutes a connection can exist in the pool before it is destroyed.

Idle connection Test period*

Default Value: 5

Integer

Specify the number of minutes for a connection to remain idle before a test query is run. This helps keep database connections from timing out.

Checkout timeout*

Default Value10000
Example10000

Integer

Specify the number of milliseconds to wait for a connection to be available in the pool. Zero waits forever. After set time, then an exception will be thrown and the pipeline will fail.

Url properties

Use this fieldset to specify properties to use in JDBC Url. These properties will need to be configured when setting up SSL connection. See Advanced Configurations: Configuring Hive with SSL section below for details.

Url property name

Default Value: N/A
ExamplemaxAllowedPacket 

String

Specify a name for the URL property to be used by the account.

Url property value

Default ValueN/A
Example1000

String

Specify a value for the URL property name.

Hadoop properties

Authentication method

Default ValueNone

Dropdown list

Select the Authentication method to use when connecting to the Hadoop service.  

  • None: Allows connection even without the Username and Password

  • Kerberos: Allows connection with Kerberos details such as Client Principal, Keytab file, and Service principal

  • User ID: Allows connection with Username only

  • User ID and Password: Allows connection with Username and Password

Use Zookeeper 

Default ValueDeselected

Checkbox

Select if Zookeeper be used to locate the Hadoop service instead of a specific hostname. If the checkbox is selected, use Zookeeper to resolve the location of the database instead of using the hostname field in the standard block.

Zookeeper Versions

When using Zookeeper in combination with a Hive account, add the Zookeeper JAR package file on the Groundplex associated with that Hive account. The version of Zookeeper on the Groundplex should be the same as the version your Hive account uses.

For HDP users, in addition to the zookeeper.jar package, you might also require the curator-client-X.X.X.jar and curator-framework-X.X.X.jar package files on the Groundplex.

Zookeeper URL


Default Value: N/A
Examplehostname1:port,hostname2:port/;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2

String

Specify the URL of the Zookeeper service. Zookeeper URL formats are different for CDH and HDP.

  • For HDP:

    • Format: hostname1:port,hostname2:port/;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2

    • Example: na77sl-ihdc-ux02011.clouddev.snaplogic.com:2181,na77sl-ihdc-ux02012.clouddev.snaplogic.com:2181,na77sl-ihdc-ux02013.clouddev.snaplogic.com:2181/;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2

  • For CDH:

    • Format: zk=hostname1:port,hostname2:port/hiveserver2

    • Example: Zk = jdbc:hive2://cdhclusterqa-1-1.cloudev.snaplogic.com:2181.cdhclusterqa-1-2.clouddev.snaplogic.com:2181.cdhclusterqa-1-3.clouddev.smaplogic.com:2181/hiveserver2

This is NOT the URL for the Hadoop service being sought.

Hive properties

JDBC Subprotocol*

Default ValueHive
ExampleImpala

Dropdown list

Specify the JDBC Subprotocol to be used. This is required when the Authentication method is KerberosAvailable options are:

  • Hive

  • Impala

Kerberos properties

Use this fieldset to configure information required for the Kerberos authentication. These properties must be configured if you select Kerberos in the Authentication method property.

Client Principal

Default Value: N/A
Example

String

Specify the principal used to authenticate to Kerberos KDC (Kerberos Key Distribution Center - Network service used by the clients and servers for authentication). 

Keytab File

Default Value: N/A
Example:  

String

Specify the Keytab file (file used to store encryption keys) used to authenticate to Kerberos KDC.

Service Principal

Default Value: N/A
Example:  hive/host@REALM or impala/host@REALM

String

Specify the principal used by an instance of a service.

Examples: 

  • If you are connecting to a specific server: hive/host@REALM or impala/host@REALM

  • If you are connecting(more common for the Snap) to any compliant host (see Use Zookeeper property's description) in which case the principal is: 'hive/_HOST@REALM' or 'impala/_HOST@REALM'. 

...