In this article
Table of Contents | ||
---|---|---|
|
Overview
SnapLogic supports Kubernetes orchestration on your Groundplex instances. You can deploy Snaplex nodes in your Kubernetes environment by setting up a Helm Chart that defines the node configuration for discoverability in the Kubernetes environment. This article explains how you can deploy and configure SnapLogic Snaplex nodes in a Kubernetes environment and contains an attached Helm Chart that you can use.
The autoscaling solution described in the article Deploying a Groundplex in Kubernetes with Elastic Scaling is no longer available. You can configure your Groundplex to use autoscaling through Kubernetes-based metrics. Consult your CSM to learn more.
Workflow
...
Prerequisites
You must be an Org-level administrator.
You are familiar with the SnapLogic Snaplex installation process.
Downloading the Configuration File from SnapLogic Manager
Open an existing Snaplex in the Org:
Navigate to the target Snaplex in Manager.
Click on the Snaplex name to display the Update Snaplex dialog.
Alternatively, if none exists, Create a Snaplex.
On the Downloads tab, click to copy the Configuration link. Paste this link into your Helm Chart.
download the Configuration file using the option or use the API
api/1/rest/public/snaplex/config/<path_to_groundplex>
to download theslpropz
file.Click Cancel to exit the dialog.
Since the configuration link has an expiration, to ensure that the Kubernetes pods continue to run:
Delete all query string parameters from the Configuration Link URL.
In the following example URL, delete everything from the question mark to the end:Use the API to retrieve the information needed to configure a node on a Groundplex.
plexGET https://{controlplane_path}/api/1/rest/public/snaplex/config/{plex_path}?{query_parameters}
.
An example would be similar to the path provided below:https://elastic.snaplogic.com/api/1/rest/
PlatformQApublic/snaplex/config/
Ground_Triggered?expires=1613086219&user_id=testuser22@snaplogic.com&_sl_authproxy_key=1BN...snaplogic/shared/
MyGround
Use the path retrieved from the API in the
snaplogic_config_link
parameter of the Helm chart. Learn more at Retrieve config info for a Groundplex and Download the slpropz configuration file
Set the parameter
Deployingsnaplogic_secret
in the Helm chart YAML file to the name of the Kubernetes secret you create, as described in theDeploy the Helm Chart section.
Info |
---|
For a Zero Trust Kubernetes Installation, open up port TCP 443 and websockets connections for the sites mentioned below: |
Running the Snaplex with Org Credentials
You can associate Org admin credentials with the SnapLogic secret created when enabling enhanced encryption. Doing so This makes it easier to share the Snaplex service with the users in your Org. Both the credentials for the SnapLogic Org admin your organization's users. The SnapLogic Org admin credentials and the enhanced encryption secret are in JSON format as key/value pairs.
Generate a key and encode it for each value.
NOTE: We recommend that you use base64 Base64 to encode the values.To create the SnapLogic secret:
Create the YAML file with the following two keys: username and password.
Example YAML File
Code Block apiVersion: v1 kind: Secret metadata: name: mysecret type: Opaque data: username: <base64<Base64 username> password: <base64<Base64 password>
TIP: Run the following command to encode your username/password into the text of the secret:
$ echo -n "snaplogic_username_or_password" | base64Base64
IMPORTANT: If your password includes any of the following characters, you must escape the character with a backslash (
\
) in the string that you pass to the encoder:\
(backslash)$
(dollar sign)'
(apostrophe or single-quote)`
(backtick)"
(double-quotes)&
(ampersand)|
(pipe symbol);
(semicolon)!
(exclamation mark)
For example, if your password ismypa$$word
, pass the stringmypa\$\$word
to the Base64 encoder.Run the following command:
$ kubectl apply -f snaplogic_secret.yaml
(Optional) If Enhanced Encryption is enabled for your Org, create the Enhanced Encryption secret by running the following commands:
$ kubectl create secret generic enhanced-encryption-secret --from-file=keystore_jks --from-file=keystore_pass
$ kubectl apply -f enhanced_encryption_secret.yaml
After the secret is created, delete the YAML file , which because it is no longer needed.
Info |
---|
See the Kubernetes documentation regarding the management of the secret. |
You can now deploy the Helm Chart.
...
Deploy the Helm
...
chart
About the Helm Chart
The helm chart defines the values for your Groundplex nodes in your Kubernetes environment.
You can download the Helm Chart package, which contains the following:
values.yaml
- : This file is the helm chart.templates folder
- : Boilerplate charts based on parameters.Chart.yaml
- : This file contains metadata about the artifact repository.
...
The following list describes each field parameter in the values.yaml file:
Field | Description | Guidelines |
---|---|---|
Regular nodes count |
...
Specifies the number of JCC |
...
deployed to run the SnapLogic container | If HPA is enabled, then the minReplicas and maxReplicas counts are used from |
FeedMaster nodes count |
...
Specifies the number of FeedMaster |
...
Pods to deploy. | N/A |
Docker image of SnapLogic Snaplex image |
...
Specifies the repository where the image resides and the tag indicating the version of the image. |
...
Although you can specify a version of your Snaplex, we recommend that you enter |
...
the latest version for the most recently released SnapLogic build. | N/A |
SnapLogic configuration link |
...
Specifies the link to the SnapLogic JCC configuration file (also known as .slpropz). | N/A | |
SnapLogic Org admin credential |
...
Specifies the secret (an encoded username and password) to authenticate the deployment. | N/A | |
Enhanced encryption secret |
...
Specifies the secret (an additional encoded username and password) to authenticate the deployment, available only to the user |
...
. | N/A |
CPU and memory limits for the nodes |
...
Specifies the upper limits and requests for the CPU and memory resources. |
...
You can set these values for only the upper limits |
...
. The lower limits are system-defined and cannot be modified. | Limits: The maximum amount of resources (CPU and memory) that a container can consume. Requests: The minimum amount of resources that a container needs to function properly. This ensures that the container is always scheduled on a node that can fulfill its minimum requirements. |
Default file ulimit and process ulimit |
...
Specifies the number of possible open file descriptors and processes |
...
Probes. Monitors the SnapLogic app.
...
for the JCC containers. | The value should be greater than the number of slots configured for the node (Maximum Slots under the Snaplex's node properties). If not set, then the node defaults will be used. (/etc/security/limits.conf). | |
Probes | Monitors the SnapLogic application. | |
Autoscaling | Sets autoscaling properties. Enabled to false by default. Contact |
...
SnapLogic CSM for recommendations for this setting. | The HorizontalPodAutoScaler is enabled by setting this property to true. Contact SnapLogic CSM for recommendations for this setting. | |
minReplicas maxReplicas | minReplicas defines the minimum number of Pods that must be running. maxReplicas defines the maximum number of Pods that can be scheduled on the node(s) | The general guideline is to start with 1:2 or 1:3 Pods per node. |
targetAvgCPUUtilization | This is the average CPU utilization across all Pods. HPA will scale up or scale down Pods to maintain this average. | N/A |
targetAvgMemoryUtilization | This parameter specifies the average memory utilization (as a percentage of the requested memory) that the HPA should maintain across all the replicas of a particular deployment or stateful set. | N/A |
scaleDownStabilizationWindowSeconds | This parameter controls the amount of time the HPA waits (like a cool-down period) before scaling down the number of pods after a decreasein resource utilization. | N/A |
Termination Grace Period Seconds |
...
The time to shut down a node gracefully before it gets terminated. | N/A | |
IPv6 |
...
Enabled to |
...
False by default. You can enable IPv6 for your connections if your infrastructure supports it. | N/A |
Info |
---|
The regular nodes count and feed master nodes count are similar to the minReplicas and maxReplicas fields, which refer to the number of Pods needed to run the application. These are used to configure the pod count statically rather than using autoscaling, as in the case of minReplicas and maxReplicas. The YAML files set the default name values for Groundplex nodes. |
Example Helm Chart
Code Block |
---|
# Default values for snaplogic-snaplex. # This is a YAML-formatted file. # Declare variables to be passed into your templates. # Regular nodes count jccCount: 1 # Feedmaster nodes count feedmasterCount: 1 # Docker image of SnapLogic snaplex image: repository: snaplogic/snaplex tag: latest # SnapLogic configuration link snaplogic_config_link: # SnapLogic Org admin credential #snaplogic_secret: # Enhanced encryption secret #enhanced_encrypt_secret: # CPU and memory limits/requests for the nodes limits: memory: 8Gi cpu: 2000m requests: memory: 8Gi cpu: 2000m # Default file ulimit and process ulimit sl_file_ulimit: 8192 sl_process_ulimit: 4096 # Enable/disable startup, liveness and readiness probes probes: enabled: true # JCC HPA autoscaling: enabled: false minReplicas: maxReplicas: # Average count of Snaplex queued pipelines (e.g. targetPlexQueueSize: 5), leave empty to disable # To enable this metric, Prometheus and Prometheus-Adapter are required to install. targetPlexQueueSize: # Average CPU utilization (e.g. targetAvgCPUUtilization: 50 means 50%), leave empty to disable. # To enable this metric, Kubernetes Metrics Server is required to install. targetAvgCPUUtilization: # Average memory utilization (e.g. targetAvgMemoryUtilization: 50 means 50%), leave empty to disable. # To enable this metric, Kubernetes Metrics Server is required to install. targetAvgMemoryUtilization: # window to consider waiting while scaling up. default is 0s if empty. scaleUpStabilizationWindowSeconds: # window to consider waiting while scaling down. default is 300s if empty. scaleDownStabilizationWindowSeconds: # grace period seconds after JCC termination signal before force shutdown, default is 30s if empty. terminationGracePeriodSeconds: 900 # Enable IPv6 service for DNS routing to pods enableIPv6: false |
...
Configure the following parameters in the Helm Chart and name the file values.yaml.
In the Helm Chart console, run the following command:
$ helm install snaplogic <helm_chart_folder>
Where <helm_chart_folder> is the Helm Chart zip file, which you can download from this document.Run the
helm list
command to determine the status of the deployment.
The following sample output shows a successful deployment:Note If you deploy the Snaplex JCC nodes but resources in the Kubernetes environment are not available to fulfill the Helm Chart, then your deployment goes into a pending state until sufficient resources are available.
The The following sample output shows the pending status of resources:
Once After you deploy your Helm Chart, you can deploy a load balancer.
...
In the values.yaml file of the Helm Chart, set the value
enableIPv6: true
.Set the global property for the
jcc.k8s_subdomain_service
in this format --<Helm Release Name>-snaplogic-snaplex-ipv6
.
...
Deploy a
...
load balancer
Note |
---|
Use base64 Base64 to encrypt the values. |
To add load balancers to your JCC and FeedMaster nodes:
In the Helm console, run the
helm list
command to list the services.In SnapLogic Manager, navigate to the target Project folder, then click the target Snaplex; the Update Snaplex dialog appears.
On the Settings tab of the Update Snaplex dialog, enter the corresponding values in the following fields:
Load balancer. Enter the protocol and port number of the Snaplex JCC node. See Refer to PORT(s) associated with
snaplogic-snaplex-regular
.Ultra load balancer. Enter the the protocol and port number of the FeedMaster node. See Refer to PORT(s) associated with
snaplogic-snaplex-feed
.
Review the information, then click Update.
Once After your Snaplex and FeedMaster nodes are deployed, you can start designing and running Pipelines and Tasks.
Installing utilities
For Kubernetes-based deployments, users have to build images/containers that install the utilities in appropriate locations. Those images can use the official Snaplogic image as the base image. When deploying the Snaplex to Kubernetes, the users would then use that image/container for deployment and have necessary dependencies and utilities in place.
Disk sizing guidelines
By default, Kubernetes pods use the disk space of the node they run on, called ephemeral disk
. If ephemeral disk
runs low, the Kubernetes pod taking the most disk space on the node will be evicted (e.g. restarted), and that disk space will be freed up. Kubernetes pods do not retain ephemeral disk space across restarts, so each time a pod restarts, its filesystem will be essentially cleared.
The amount of ephemeral disk
space a Kubernetes worker node needs is dependent on the workloads running on that node, e.g., the number of ground plexes on K8s, the number of other pods, etc.
If more disk space is needed, a Persistent Volume can be used. In cloud environments like AWS, this is often EBS storage. Persistent volumes can be mounted to pods, and they retain data across restarts.
Best Practices
Avoid running processes on in the same pod container as the JCC node, so that the JCC can have has the maximum amount of memory available on that pod, as requested.
Do not overwrite the
global . properties
file options unless you are working with your CSM to customize your Groundplex.Request resources upfront. Do this by setting The requests determine the minimum required resources, while limits set the maximum resources a Container can consume. Setting them to the same amount ensures stability and exact resource usage. To do this, set the pod’s request and limit to the same value, as shown in the image below:
FAQ
Expand | ||
---|---|---|
| ||
You can split the license volume across multiple Nodes and Pods if you do not exceed the allocated CPU and memory (for a 4-container configuration). |
Expand | ||
---|---|---|
| ||
If a user is licensed for four 8*32 nodes, then they should deploy not more than four Kubernetes containers, each assigned no more than 8vCPU and 32GB RAM. |
Downloads
Download and extract the following files, using the values.yaml
file as the basis for your Helm Chart.
Attachments | ||||
---|---|---|---|---|
|