Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

Nodes (Min/Rec)

Minimum: 1

Recommended: 2 or more nodes

SnapLogic Project and Enterprise platform package nodes can be configured in the following sizes:

  • 2 vCPU and 8 GB RAM 

  • 4 vCPU and 16 GB RAM 

  • 8 vCPU and 32 GB RAM

  • 16 vCPU and 64 GB RAM

We recommend two nodes for high availability. For requirements about clustering nodes, refer to Node Cluster.

All nodes within a Snaplex must be of the same size.

RAM (Min)

Minimum: 8 GB

Depending on the size, number, and design of Pipelines, more RAM is required to maintain an acceptable level of performance.

CPU (Min)

Minimum: 2 core

All Snaps execute in parallel in their own threads: the more cores that are available to the Snaplex, the more performant the system.

DISK Disk (Min/Rec)

Minimum: 40 GB

Recommended: 100 GB

Local disk space is required for logging and for any Snap that uses the local disk for temporary storage (for example, Sort and Join Snaps). For details, refer to Temporary Folder.

SnapLogic does not have restrictions on the disk size of your Groundplex nodes.

...

To communicate with the SnapLogic Integration CloudControl Plane, Groundplexes use a combination of HTTP/HTTPS requests and WebSockets communication over the TLS (SSL) tunnel. In order for For this combination to operate effectively, you must configure the firewall to allow the following network communication requirements:

Feature

Required

Consequence

HTTP outbound Port 443

Yes

Does not function

HTTP HEAD

Desired

Without HEAD support, a full GET requires more time and bandwidth

Websockets WebSockets (WSS protocol)

Yes

Does not function

JCC node: 8090 (HTTP), 8081 (HTTPS)

Feedmaster: 8090 (HTTP), 8084 (HTTPS), 8089 (Message queue)

Yes

Unable to reach Snaplex neighbor - https://hostname:8081

Needs to be available for communication between the nodes in a Snaplex.

  • The nodes of a Snaplex need to communicate among themselves, so it is important that they each node can each resolve each other's hostnameshost names. This requirement is required crucial when you are making local calls into the Snaplex nodes for the execution of Pipelines rather than going instead of initiating it through the SnapLogic Platform. The Pipelines are load-balanced by SnapLogic with Tasks passed to the target node.

  • Communication between the customer-managed Groundplex and the SnapLogic-managed S3 bucket is over HTTPS with TLS enforced by default. The AWS-provided S3 URL also uses an HTTPS connection with TLS enforced by default. If direct access from the Groundplex to the SnapLogic AWS S3 bucket is blocked, then the connection to the AWS S3 bucket communication falls back to a connection via through the SnapLogic Control Plane that still uses TLS 1.2.

...

In the SnapLogic Platform, the Snaps actually communicate to and from the application endpoints. The protocols and ports required for this communication are mostly determined by the endpoints themselves, and not by SnapLogic. Cloud and SaaS applications commonly communicate using HTTPS, although older applications and non-cloud or SaaS applications might have their own requirements. 

For example, the following table shows some of these requirements:

RedShift

Application

Protocol

Default Port

Netezza

TCP

5480

Oracle

TCP

54391521

NetezzaRedShift

TCP

54805439

Salesforce

HTTPS

443

Oracle

TCP

1521

Each of these application connections may might allow the use of a proxy for the network connection, but it is a configuration option of the application’s connection—not one applied by SnapLogic.

...

  • 8084—The FeedMaster's HTTPS port. Requests for the pipelines are Pipelines are sent here as well as in addition to some internal requests from the other Groundplex nodes.

  • 8089—The FeedMaster's embedded ActiveMQ broker SSL port. The other Groundplex nodes connect to this port to send /and receive messages.

The machine hosting the FeedMaster nodes needs to have those ports open on the local firewall, and the other Groundplex nodes need to allow outbound requests to the FeedMaster nodes on those ports.

Groundplex Name and Associated Nodes

...

Every Snaplex requires a name, for example, ground-dev or ground-prod. In the SnapLogic Designer, you can choose which the Snaplex where Pipelines are executed. The Self-managed Snaplex configuration has an Environment variable associated with it, for example, dev or prod. When you configure the nodes for your Groundplex, you must set the jcc.environment to the Environment value that you have configured for the Snaplex.

The hostname host name of the system used by a Groundplex can not have an underscore (_) in its name as per DNS standards. Avoid special characters as well.

After the Snaplex service is started on a node, the service connects to the SnapLogic Cloud service. Runtime logs from the Snaplex are written to the /opt/snaplogic/run/log (or c:\opt\snaplogic\run\log) directory. The Dashboard shows the nodes that are currently connected nodes for each Snaplex.

Understanding the Distribution of Data Processing across Snaplex Nodes

When a Pipeline or Task is executed, the work is assigned to one of the JCC nodes in the Snaplex. Depending on a number of variables, the distribution of work across JCC nodes is determined by the number of threads in use, the amount of available memory available, and the average system load.

...

Starting multiple nodes with the JCC service pointing to the same Snaplex configuration automatically forms a cluster of nodes, as long as if you follow these requirements for nodes in a Snaplex:

...

We recommend that you set up JCC nodes in a Snaplex within in the same network and data center. Communication between JCC nodes in the same Snaplex is required for the following reasons:

  • The Pipeline Execute Snap communicates directly with neighboring JCC nodes in a Snaplex to start child Pipeline executions and send documents between parent and child Pipelines.

  • The data displayed in Data Preview is written to and read from neighboring JCC nodes in the Snaplex.

  • The requests and responses made in Ultra Pipelines are exchanged between a FeedMaster node and all of the JCC nodes in a  Snaplexa Snaplex.

  • A Ground Triggered Task (invoked from a Groundplex) can be executed on a neighboring JCC node due to load balancing—in because of load-balancing, in which case, the Pipeline prepare request and the bodies of the request and response are transferred between nodes.

Therefore, any Any extra latency or network hops between neighboring JCC nodes can introduce performance and reliability problems.

...