Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

We need two Snaps—one to load the data from the CSV file in an S3 location and another to read the data loaded in the target table and generate some insights.

We configure the Databricks - Bulk Load Snap to load employee data from the CSV into a new table, company_employees. We configure this Snap’s account to connect to the AWS S3 service using Source Location Credentials to read the CSV file.

...

Here is how we do it:

  • Select the Drop and create table as the Load action.

  • Define the schema for the new table in the Target Table Columns fieldset.

  • Choose the source data type and indicate that the file contains a valid header.

  • Specify the file names (with relative paths, here) to load the data from.

  • As our CSV file in the S3 location is not encrypted, we leave the corresponding fields blank.

...

Upon validation, the Snap retrieves and displays the data from the company_employees table that matches the WHERE condition specified.

...

Download this Pipeline

Downloads

Info
  1. Download and import the Pipeline into SnapLogic.

  2. Configure Snap accounts as applicable.

  3. Provide Pipeline parameters as applicable.

...