On this Page
Table of Contents | ||||
---|---|---|---|---|
|
Snap Type: | Write | ||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Description: |
Warning | ||
---|---|---|
| ||
Ensure that this property is not selected when choosing the Upload exising files on Google Cloud Storage option in Upload type property, else an exception will be thrown and the Snap will not execute. |
Name of the Google Cloud Storage bucket to be used for the operation. This is a suggestible field and will list all the buckets within the given account.
Example: project1234
Default value: [None]
This is a drop-down menu consisting of two options: Upload incoming documents and Upload existing files from Google Cloud Storage. It specifies the data source to the Snap, that is incoming files or existing files in Google Cloud Storage bucket have to be uploaded. To upload data from existing files in Google Cloud Storage bucket select the option Upload existing files from Google Cloud Storage.
Default value: Upload existing files from Google Cloud Storage.
Info |
---|
Based on the selection here, only one of the succeeding sections has to be configured. The succeeding sections being Properties for uploading incoming documents and Properties for uploading existing files from Google Cloud Storage. |
For selecting the file preferred format of the temporary file. This is a drop-down list that has three options: CSV, JSON, and AVRO.
Default value: CSV
Warning | ||
---|---|---|
| ||
The file format should be selected based on the data type they support, for example if the incoming document contains arrays or lists then selecting CSV in the File format property will throw an execption and the Snap will not execute. In order to avoid this, AVRO or JSON must be selected. Similarly, AVRO file format should not be used for Date Time data types. If the incoming documents do not contain all the table columns in the same order as the destination table then do not use CSV. |
The name of the temporary file that will be created on the Google Cloud Storage bucket, if not provided then a system generated file name will be used.
Default value: [None]
Specifies whether the temporary file created for the load operation has to be retained or deleted after the Snap's execution. By default the temporary file will be deleted.
Default value: Selected
This is a drop-down list that has three options: CSV, JSON, and AVRO.
Default value: CSV
The file's location in the Google Cloud Storage bucket.
Example: gs://gcs_existingbucket/exisitng_file.csv.
Default value: [None]
Specifies that the CSV file contains headers, using this option will enable the Snap in differentiating between the headers and records.
Default value: Not selected
Specifies the delimiter for the CSV file. This is needed only for CSV file types.
Example: | (pipe)
Default value: , (comma)
Info | ||
---|---|---|
| ||
All custom delimiters supported by BigQuery are supported as well. |
Similar operation as Preserve temp file, the files from which the data is loaded to the destination table will be deleted after Snap's execution if this option is enabled.
Default value: Not selected
Executes the Snap during a pipeline Save operation so that the output view can produce the preview data.
Default value: Not selectedThis Snap performs a bulk load operation into a Google BigQuery database, it . Depending upon the Snap's configuration, it does so either by using data from incoming documents or by using existing files in the Google Cloud Storage bucket depending upon the Snap's configuration. The Snap supports all three file types supported by Google BigQuery - CSV, JSON, and AVRO.
When using incoming documents:
In the case where data from incoming documents is being loaded, the data is first uploaded to a temporary file on Google Cloud Storage and from that temporary file the data is loaded into the destination table. The user can choose to either retain or delete this temporary file after the Snap terminates.
When using existing files from Google Cloud Storage:
In the case where existing files from Google Cloud Storage are being used, the data is loaded directly from the specified files into the destination table, there are no temporary files created for this operation. However, the user can choose to either retain or delete these existing files after the Snap terminates.
ETL Transformations & Data Flow
The Google BigQuery Bulk Load (Cloud Storage) Snap performs a bulk load of the input records into the specified database. If the data is being loaded from incoming documents, it is sent to a temporary file in the cloud storage and from there to the destination table. The temporary file is retained after Snap's execution however . However, the user can choose to delete it if they so wish and configure the Snap accordingly.
Input & Output
- Input: Any Snap that can pass a document output view, such as Structure or JSON Generator. Pipleline parameters can also be passed, only for bucket ID, project ID, table ID, and so on. In cases where existing files from the Google Cloud Storage bucket are being used, then input data is taken from the source files.
- Output: The output is in document view format. It displays the statistics for the completed bulk load operation.
Modes
- Ultra pipelinesPipelines: May work in Works in Ultra Pipelines.Spark (Deprecated) mode: Not supported in Spark mode.
Note |
---|
The Consider the following points must be considered when using this Snap:
|
- When using incoming documents with arrays/lists, CSV cannot be selected as the file format for the temporary file since that will cause an error. JSON or AVRO must be selected in such a case.
- When using incoming documents with date time data types, AVRO cannot be selected as the file format for the temporary file since that will cause an error. CSV or JSON must be selected in such a case.
- When uploading incoming documents via CSV file or uploading existing CSV file from Google Cloud Storage, ensure that the CSV file contains all the destination table columns in the same order as per the table. Otherwise, use the workarounds to handle CSV file that does not contain all the table columns in the same order in the Troubleshooting section.
- When uploading from existing documents on Google Cloud Storage, enabling the Create table if not present property will throw an exception error. In order to avoid this it should be ensured that the destination table exists.
Account & Access
This Snap uses account references created on the Accounts page of SnapLogic Manager to handle access to this endpoint. See Google BigQuery Account for information on setting up this type of account.
Views
Input | This Snap has exactly at most one document input view. |
---|---|
Output | This Snap has exactly one document output view. |
Error | This Snap has at most one document error view and produces zero or more documents in the view. |
Mismatch in the order of columns/number of columns between incoming document/existing Google Cloud Storage files in CSV format:
- Incoming documents: Use JSON/AVRO format for the temp file in the Cloud Storage Bulk Load Snap.
- Existing files from Google Cloud Storage: Read the file using File Reader Snap followed by the CSV Parser Snap and then use Google BigQuery Bulk Load (Cloud Storage) Snap selecting Upload type as Upload incoming documents and File format as JSON or AVRO.
Settings
Label
Required. Project ID of the project containing the table, this is a suggestible field and can be populated based on the Account settings.
Example: project1234
Default value: [None]
Required. Dataset ID of the dataset containing the table, this is a suggestible field and all the datasets in the specified project will be populated.
Example: dataset1234
Default value: [None]
Required. Name of the target table into which the records are to be loaded, this is a suggestible field and all the tables in the datasets will be listed.
Example: table1234
Default value: [None]
Note | ||
---|---|---|
| ||
So, ensure that you include the time zone in all the datetime values that you load into Google BigQuery tables using this Snap. For example: "2020-08-29T18:38:07.370 America/Los_Angeles", “2020-09-11T10:05:14.000-07:00", “2020-09-11T17:05:14.000Z” |
Prerequisites:
Write access to the Google BigQuery Account and Read & Write access from/to the Google Cloud Storage account is required.
Limitations and Known Issues
When using incoming documents with arrays/lists, CSV cannot be selected as the file format for the temporary file. This will cause an error. Select JSON or AVRO instead.
When using incoming documents with date time data types, AVRO cannot be selected as the file format for the temporary file. This will cause an error. Select CSV or JSON.
When uploading incoming documents via a CSV file or uploading an existing CSV file from Google Cloud Storage, make sure the CSV file contains all the destination table columns in the same order as per the table. Otherwise, use the workarounds in the Troubleshooting section to handle a CSV file that does not contain all the table columns in the same order.
When uploading from existing documents on Google Cloud Storage, enabling the Create table if not the present check box throws an exception error. To avoid this, ensure the destination table exists.
Known Issue
Copying data by creating a table with the same name in Google BigQuery immediately after deleting it, might not insert the rows as expected. This behavior is due to the way the tables are cached and the internal table ID is propagated throughout the system.
Workaround
We recommend you avoid rewriting in Google BigQuery and suggest the following workarounds. You can choose to use them individually or in unison, to suit your requirement.
Truncate the existing table instead of deleting it.
Add some randomly generated prefix to the table name each time you create the table (new or unique table name every time).
Configurations:
Account & Access
This Snap uses account references created on the Accounts page of SnapLogic Manager to handle access to this endpoint. See Google BigQuery Account for information on setting up this type of account.
Views
Input | This Snap has exactly at most one document input view. |
---|---|
Output | This Snap has exactly one document output view. |
Error | This Snap has at most one document error view and produces zero or more documents in the view. |
Troubleshooting:
Mismatch in the order of columns/number of columns between incoming document/existing Google Cloud Storage files in CSV format:
Incoming documents: Use JSON/AVRO format for the temp file in the Cloud Storage Bulk Load Snap.
Existing files from Google Cloud Storage: Read the file using File Reader Snap followed by the CSV Parser Snap, then use BigQuery Bulk Load (Cloud Storage) Snap and selec Upload type as Upload incoming documents and File format as JSON or AVRO.
Snap Settings
Label*
Insert excerpt | ||||||
---|---|---|---|---|---|---|
|
Project ID*
Specify the Project ID of the project containing the table. This is a suggestible field and can be populated based on the Account settings.
Default value: None
Example: project1234
Dataset ID
Specify the dataset ID of the dataset containing the table. This is a suggestible field and all the datasets in the specified project will be populated.
Default value: None
Example: dataset1234
Table ID*
Specify the name of the target table into which the records are to be loaded. This is a suggestible field and all the tables in the datasets will be listed.
Default value: None
Example: table1234
Create table if not present
Multiexcerpt include macro | ||||
---|---|---|---|---|
|
Default value: Not selected
Warning | ||
---|---|---|
| ||
Ensure that you do not select Create table if not present check box if you choose the Upload existing files on Google Cloud Storage option for Upload type, else an exception is thrown and the Snap does not execute. |
Bucket name
Specify the name of the Google Cloud Storage bucket to be used for the operation. This is a suggestible field and will list all the buckets within the given account.
Default value: None
Example: project1234
Upload type
Choose an upload type for uploading your documents. The available options are:
- Upload incoming documents
- Upload existing files from Google Cloud Storage: To upload data from existing files in Google Cloud Storage bucket select the option Upload existing files from Google Cloud Storage.
These options specify the data source to the Snap. Incoming files or existing files in Google Cloud Storage bucket have to be uploaded.
Default value: Upload existing files from Google Cloud Storage
Example: Upload incoming documents
Info |
---|
Based on the option you choose, only one of the succeeding sections has to be configured. The succeeding sections being Properties for uploading incoming documents and Properties for uploading existing files from Google Cloud Storage. |
Properties for uploading incoming documents
This sub-section has to be configured if the option Upload incoming documents was selected in the Upload type property.
File format
Choose the file preferred format of the temporary file. The available options are: CSV, JSON, and AVRO.
Default value: CSV
Example: JSON
Warning | ||
---|---|---|
| ||
The file format should be selected based on the data type they support, for example if the incoming document contains arrays or lists then selecting CSV in the File format property will throw an execption and the Snap will not execute. To avoid this, AVRO or JSON must be selected. Similarly, AVRO file format should not be used for Date Time data types. If the incoming documents do not contain all the table columns in the same order as the destination table then do not use CSV. |
Temp file name
The name of the temporary file that is created on the Google Cloud Storage bucket. If a filename is not provided, then a system generated file name is used.
Note |
---|
When using a temporary file to load data, use a unique temporary file for each of the BigQuery Bulk Load Snaps in the Pipeline. Else, the snaps may fail randomly with |
Default value: [None]
Preserve temp file
Specify whether the temporary file created for the load operation has to be retained or deleted after the Snap's execution. By default the temporary file is deleted.
Default value: Selected
Batching
Select if you want the Snap to process the input records in batches.
Batch Size
The number of records batched per request. If the input has 10,000 records and the batch size is set to 100, the total number of requests would be 100.
Batch Timeout (milliseconds)
Time in milliseconds to elapse following which the batch if not empty will be processed even though it might be lesser than the given batch size.
Properties for uploading existing files from Google Cloud Storage
This sub-section has to be configured if the option "Upload existing files from Google Cloud Storage" was selected in the Upload type property.
File paths
Multiple files can be selected based on the need. When the pipeline is executed, the output data will have as many records listed. Based on the number of files added, the Snap will group them into categories (CSV with header & delimiter, CSV with delimiter but without header, JSON, and AVRO). This distinction is maintained in the output preview as well (shown distinctly according to File type).
File format
This is a drop-down list that has three options: CSV, JSON, and AVRO.
Default value: CSV
File path
The file's location in the Google Cloud Storage bucket.
Example: gs://gcs_existingbucket/exisitng_file.csv.
Default value: [None]
CSV file contains headers
Specifies that the CSV file contains headers. Use this option to enable the Snap to differentiate between the headers and records.
Default value: Not selected
CSV delimiter
Specify the delimiter for the CSV file. This is needed only for CSV file types.
Default value: , (comma)
Example: | (pipe)
Info | ||
---|---|---|
| ||
All custom delimiters supported by BigQuery are supported as well. |
Delete files upon exit
Similar operation as Preserve temp file. If this option is enabled, the files from which the data is loaded to the destination table are deleted after Snap's execution.
Default value: Not selected
Multiexcerpt include macro | ||||
---|---|---|---|---|
|
Multiexcerpt include macro | ||||
---|---|---|---|---|
|
Note | |||||||||
---|---|---|---|---|---|---|---|---|---|
| |||||||||
Google BigQuery tables support columns with NUMERIC data type to allow storing big decimal numbers (up to 38 digits with nine decimal places). But Snaps in Google BigQuery Snap Pack that load data into tables cannot create numeric columns. When the Create table if not present check box is selected, the Snaps create the required table schema, but map big decimals to a FLOAT64 column. So, to store the data into numeric columns using these Snaps, we recommend the following actions:
The Google API converts this string into a number with full precision and saves it in the numeric column. Example:
|
Examples
Basic Use Case - 1 (Upload incoming documents)
In this example, incoming documents are used for the bulk load. It will cover two scenarios:
Temporary file is preserved.
Temporary file is deleted.
The following pipeline is executed:
The CSV Generator Snap provides the input data to the Snap, a . A preview of the sample data from the incoming document is shown below:
Scenario - 1 - Temporary File Preserved
The following image is the configuration of the Google BigQuery Bulk Load (Cloud Storage) Snap:
The Preserve temp file checkbox is selected and the temprary file is specified as tempfile.csv. Upon execution, the records from the incoming CSV file will be are loaded into the temporary file present in the gcs_cloud_1 bucket and from . From there it is loaded into the destination table csvtablenew in the babynames dataset within the case16370 project.
The following is a preview of the data output from the Google BigQuery Bulk Load (Cloud Storage) Snap:
Scenario - 2 - Temporary File Deleted
The following image is the configuration of the Google BigQuery Bulk Load (Cloud Storage) Snap:
The Preserve temp file checkbox is not selected and the temprary file is specified as tempfile.csv. Upon execution, the records from the incoming CSV file will be are loaded into the temporary file present in the gcs_cloud_1 bucket and from . From there it is loaded into the destination table csvtablenew in the babynames dataset within the case16370 project. The temporary file is deleted.
The following is a preview of the data output from the Google BigQuery Bulk Load (Cloud Storage) Snap:
Basic Use Case - 2 (Upload Existing Files)
In this example, incoming documents are used for the bulk load. It will cover two scenarios:
Cloud Storage file is preserved.
Cloud Storage file is deleted
The following pipeline is executed:
.
Scenario - 1 - Cloud Storage File Preserved
The following image is the configuration of the Google BigQuery Bulk Load (Cloud Storage) Snap:
An AVRO file is chosen as the input file and the Delete files upon exit checkbox is not selected. Upon the Snap's execution, the data from the input file will be is loaded into the table named avrotable in the babynames dataset within the case16370 project. The AVRO file is not deleted.
The following is a preview of the Snap's output:
Scenario - 2 - Cloud Storage File Deleted
The following image is the configuration of the Google BigQuery Bulk Load (Cloud Storage) Snap:
P
The following is a preview of the Snap's output:
Exported The exported pipeline is available in Downloads section below.
Typical Snap Configurations
The key configuration of the Snap lies in how the values are passed to the Snap, this . This can be done in the following ways:
Without expressions
The values are passed to the Snap directly.With expressions
Using pipeline parameters
The values are passed as
pipeline parameterspipeline parameters. The parameters to be applied have to be selected by enabling the corresponding checkbox under the Capture column.
Advanced Use Case
The following Use Case describes a pipeline, Pipeline with a broader business logic involving multiple ETL transformations, that . It shows how typically in an enterprise environment, Bulk Load functionality can typically be used.
This pipeline Pipeline reads social data and writes it into an existing file in the Google Cloud Storage. Bulk load of this social data is then performed from the existing file in the Google Cloud Storage to the destination table.
Extract: Social data matching a certain criteria is extracted by the Query Snap.
Transform: The extracted data from the Query Snap is transformed by the CSV Formatter Snap.
Load: The incoming data from the CSV Formatter Snap is written into a CSV file in Cloud Storage by the File Writer Snap.
Extract & Load: The
GoogleBigQuery Bulk Load (Cloud Storage) Snap extracts the data from the existing CSV file in Cloud Storage and performs bulk load of this extracted data into the destination table.
Extract: The Google BigQuery Execute Snap extracts the data inserted into the destination table by the
GoogleBigQuery Bulk Load (Cloud Storage) Snap.
Exported The exported pipeline is available in the Downloads section below.
Downloads
Multiexcerpt include macro | ||||
---|---|---|---|---|
|
Attachments | ||||
---|---|---|---|---|
|
Insert excerpt | ||||||
---|---|---|---|---|---|---|
|