In this article

Overview

Salesforce Read is a Read-type Snap that provides the ability to retrieve all records for all fields of a Salesforce object from Salesforce.

Overview of settings

Prerequisites

None.

Support for Ultra Pipelines

Works in Ultra Task Pipelines.

Limitations

  • When using Primary Key (PK) Chunking mode in this Snap, the output document schemas in preview mode and execution mode may differ. For more information, refer to the note in PK Chunking.
  • When you run more than 10 Salesforce Snaps simultaneously, the following error is displayed
    Cannot get input stream from next records URL" or "INVALID_QUERY_LOCATOR"
    We recommend that you not use more than 10 Salesforce Snaps simultaneously because this might lead to the opening of more than 10 query cursors in Salesforce. Refer to this Salesforce Knowledge Article for more information.

Known Issues

None.

Snap Views

Type FormatNumber of ViewsExamples of Upstream and Downstream SnapsDescription
InputDocument
Min: 0
Max: 1

Mapper

Copy

This Snap has at most one input view.
OutputDocument
Min: 1
Max: 2
Mapper

The snap allows you to add an optional second output view that exposes the schema of the target object as the output document.

ErrorDocument

The error view contains error, reason, resolution and stack trace. For more information, see Handling Errors with an Error Pipeline

Snap Settings

Field

Field Type

Description

Label*


Default Value: Salesforce Read
Example
: Salesforce Read

String

Specify the name for the Snap. You can modify this to be more specific, especially if you have more than one of the same Snap in your Pipeline.

Salesforce API*


Default ValueREST API 
ExampleBulk API
 

Dropdown list

Choose the Salesforce API mode to use during the pipeline execution. The available options are:

  • REST API
  • Bulk API

We recommend you to set the Salesforce API mode to Bulk API if the table size of the Salesforce object referred to in the Object Type field is large (that is, 10,000 records or more) to prevent the time-out or connection error.


Service Version*

Dropdown list

Batch Size*


Default Value2000 
Example
2000

Dropdown list

Specify the size of the records to process in each batch for downloading large query results. Each batch read requires an API call against Salesforce to retrieve the set of records.

  1. REST API: The batch sizes 200 and 2000 are valid in REST API.
    According to the SFDC REST API documentation:
    "There is no guarantee that the requested batch size is the actual batch size. Changes are made as necessary to maximize performance."
    In fact, if the size of each record in the query result is relatively large (for example, a few hundred output fields), SFDC may respond in a batch size of 200, And if the size of each record in the query result is relatively small, SFDC may respond in a batch size of 2000 regardless if the Batch size is set to 200 or 2000.

  2. Bulk API: If you select the Use PK chunking if supported checkbox, the value from Batch Size is used as the PK chunking batch size. Refer to PK Chunking for more details.

    The status of the submitted bulk job and its batches can be monitored by logging in to your Salesforce.com account and going to Setup. From Setup, enter Bulk Data Load Jobs in the Quick Find box, then select Bulk Data Load Jobs. 


    According to the SFDC Bulk API documentation (Number of attempts to query):
    15 attempts at 10 minutes each to process the batch. There is also a 2-minute limit on the time to process the query. If more than 15 attempts are made for the query, an error message of 'Tried more than fifteen times' is returned. If the query takes more than 2 minutes to process, a QUERY_TIMEOUT error is returned.
    Therefore, if the query is expected to take a long time to execute (for example, if the target SObject has more than several hundred thousand records and the number of output fields in a record is more than a few hundred), it is highly recommended to set this property to 100,000 or 250,000 to avoid timeout errors.

The Snap validates using the Bulk API, even if you configure the Snap to use Bulk API with Primary Key (PK) Chunking.


Use PK chunking if supported


Default ValueDeselected

Expression/Checkbox

Appears when you select Bulk API from the Salesforce API dropdown.

Select this checkbox to use PK chunking for the bulk API request if the object supports PK chunking.

If you select this checkbox, the value from Batch Size is used as the PK chunking batch size. Please note that the chunk size applies to the number of records in the queried table rather than the number of records in the query result. For example, if the queried table has 10 million records and 250,000 is selected for Batch Size, 40 batches are created to execute the bulk query job. One additional control batch is also created and it does not process any records. Learn more about PK Chunking.

You can enable expressions for this field in order to use input values from the pipeline parameters.

Pipeline parameters are strings, and any non-empty string is treated as true in the SnapLogic Expression Language. If you create expressions for Boolean properties that use pipeline parameters, keep this in mind. For example, _IncludeDeleted does not work correctly when the value is false. Instead, use _IncludeDeleted == 'true'


Object Type*


Default ValueAccount
Example
Account

String/Expression/Suggestion

Specify the name of the Salesforce object or select one from the suggested list.

The Snap does not support Net Zero Cloud Salesforce objects


Output Fields

Use this field set to enter a list of field names for the SOQL SELECT statement. If left empty, the Snap selects all fields.

Output Fields

Default Value: N/A
Example
:

  • Id
  • Name
  • ShippingAddress
String/Expression/SuggestionEnter output field for SOQL statement.

Output Field Limit

Default Value:  N/A

String/Expression

Specify the number of output fields to return from the Salesforce object.


Output Field Offset


Default Value:  N/A

String/Expression

Defines a starting field index for the output fields. This is where the result set should start. 

If you enter an offset value that is greater than 1, the first field "ID" is always returned, but the following fields start from the offset position. This is because the first field is the only unique identifier.


Where Clause
 

Default Value: N/A
Example: Id > '001i0000007FVjpAAG'

String/Expression

Enter the WHERE clause for the SOQL SELECT statement. Do not include the word WHERE.

  • Do not use quotes for field names in the WHERE clause. Using quotes results in an error.
  • Use only single quotes for values in the WHERE clause as using double quotes results in an error.
  • The above rules do not apply when you are using SnapLogic expressions, you can use quotes for the field names and values as applicable. 


Order By Clause


Default Value: N/A
Example
:LastName, FirstName

String/Expression
Enter the ORDER BY clause that you want to use with your SOQL SELECT Query.


PK Chunking does not support the ORDER BY clause.


Limit Clause


Default Value: N/A
Example
: 2000

String/Expression
Enter the LIMIT clause that you want to use with your SOQL SELECT Query.


PK Chunking does not support the LIMIT clause.



Polling Interval*


Default Value:  5
Example
:  5

String

Define the polling interval in seconds for the Bulk API read execution. At each polling interval, the Snap checks the status of the Bulk API read batch processing.
Maximum value:  60 


Polling Timeout*


Default Value: 3000
Example
: 300

String

This property allows you to define the polling timeout in seconds for the Bulk API read batch execution. If the timeout occurs while waiting for the completion of the read batch execution, the Snap displays an exception.


Process Date/time


Default ValueDeselected

Checkbox

All date/time fields from Salesforce.com  are retrieved as string type.

Select this checkbox to allow the Snap to convert date/time fields to corresponding date/time types by accessing the metadata of the given SObject.

  • Salesforce datetime is converted to Joda DateTime,
  • Salesforce date to Joda LocalDate, 
  • Salesforce time to Joda LocalTime.

Deselect this checkbox to allow the Snap to send these date/time fields without any conversion.


Include Deleted Records

Default ValueDeselected

Expression/Checkbox

Select this checkbox to allow the Snap to include deleted records in the query. This feature is supported in REST API version 29.0 or later and Bulk API version 39.0 or later.

You can enable expressions for this field in order to use input values from the pipeline parameters.

Pipeline parameters are strings, and any non-empty string is treated as true in the SnapLogic Expression Language. If you create expressions for Boolean properties that use pipeline parameters, keep this in mind. For example, _IncludeDeleted does not work correctly when the value is false. Instead, use _IncludeDeleted == 'true'



Pass Through


Default ValueSelected

Checkbox

Select this checkbox to pass the input document to the output view under the key 'original'. This is applicable for REST and Bulk APIs.


Ignore Empty Results


Default ValueSelected

 

Checkbox

Select this checkbox to ignore empty results; no document will be written to the output view when the operation does not produce any result. If this property is not selected and Pass Through is selected, the input document will be passed through to the output view.

If you do not select both the checkboxes, Ignore Empty Result and Pass Through, the Snap writes an empty output as below:

[{}]


This property does not apply when you run the Snap in the Bulk API mode.

When you run the Snap in Bulk API mode, if the Salesforce result is empty, the Snap does not pass the input document to the output view even if the Pass Through checkbox is selected.


Bulk Content Type


Default Value:  XML
ExampleJSON

 

Dropdown list

Select the content type for Bulk API: JSON or XML.  

The numeric type field values will be read as numbers in JSON content type, and as strings in XML content type, in the output documents.

JSON content type for Bulk API is available in Salesforce API version 36.0 or higher.

In REST API, the number-type field values will always be read as numbers. If the Bulk API has been selected along with 100,000/ 250,000 as batch size value, the content-type will always be CSV regardless of the value set in this property.


Number Of Retries


Default Value: 1
Example
: 3

String/Expression

Specify the maximum number of retry attempts in case of a network failure.

  • If the value is larger than zero in PK-Chunking, the Snap overrides user settings and sets Use temp files in PK-Chunking property to true.

  • In Bulk API and PK-Chunking modes, the Snap downloads each batch of data into a temporary local file. When the download is complete, the Snap writes the output documents parsed from the temporary file to avoid the possibility of duplicated output documents in the downstream Pipeline. Ensure that the local drive has sufficient free disk space in the node where the Pipeline executes. All temporary local files are deleted when they are no longer needed.

Minimum value: 0


Retry Interval (seconds)


Default Value: 1
Example
: 3

String/Expression

Specify the minimum number of seconds for which the Snap must wait before attempting recovery from a network failure.

Minimum value: 0


Match Data Type


Default Value: Not selected

 

Checkbox

Select this checkbox to match the data types of the Bulk API results with the data types of the REST API. This property applies only when the content type is XML for Bulk API (it does not apply to JSON). If the Bulk content type is XML, Salesforce.com returns all values as strings. If Match data type is selected, the Snap attempts to convert string values to the corresponding data types if the original data type is one of the following: boolean, integer, double, currency, or percent. 

This property is ignored in REST APIs or when the Bulk Content Type is JSON. For Bulk API, Salesforce.com does not return any value for null.

This Snap does not honor the selection of the Match Data Type checkbox if the Use PK chunking if supported checkbox is also selected.


Advanced Properties  

Use this field set to define additional advanced properties that you want to add to the Snap's settings. Additional advanced properties are not required by default, and the field-set represents an empty table property. Click  to add an advanced property. 

This field set contains the following fields:

  • Properties
  • Values

Properties

Dropdown list

You can use one or more of the following properties in this field:

PropertyDescription

Use temp files in PK-Chunking


Default Valuefalse
Exampletrue

Enter true if Groundplex is located where the network connection to SFDC is relatively slow. This will have the Snap in PK-Chunking download CSV files into compressed temporary local files in order to prevent SFDC to close input streams prematurely. Ensure that the Groundplex node has sufficient free disk space. Compressed local files are less than 10% of the total CSV data size.

Setting this property to true may also be necessary if any of the downstream Snaps are relatively slow in processing the document stream. By default, the Snap in PK-Chunking downloads, parses and writes to the output view in streaming mode without using large memory or temporary files. In a normal network condition, this streaming mode is efficient and scalable. However, it has been observed that SFDC tends to close HTTP input streams prematurely if client apps pause reading the input stream momentarily to perform other processing.

Number of threads in PK-Chunking


Default Value8
Example: 2

In PK-Chunking, the Snap invokes multiple threads to download CSV files in parallel with one CSV file per each thread. The maximum number of threads is 8 by default, but a lower number may be more efficient if the network connection to SFDC is relatively slow or downstream Snaps are slow in processing documents.

Validate record count downloaded in Bulk API


Default Valuefalse
Exampletrue

This is a boolean property that applies to Bulk API where the batch size 10,000 or less (not including PK Chunking). If set to true, the Snap checks if the number of records downloaded is the same as the one processed in Salesforce.com.

If it is enabled (set to true), it is recommended that the Error View is enabled to allow the Snap to show record-count mismatch.


Disable automatic vertical slicing


Default Valuefalse
Exampletrue

The Salesforce SELECT query does not support wildcards in field lists, and its query length cannot exceed 20,000 characters. When Output fields, Output field limit, and Output field offset properties are empty, the Snap formats the entire list of column names in the SELECT statement. If the query exceeds 20,000 characters, the Snap divides the query into two or more queries, executes them separately, and merges the query results automatically. This Snap feature is called Vertical Slicing.

The Vertical Slicing feature may fail if columns used in the WHERE clause or ORDER BY clause change their values during query executions (for example, if the column is SystemModStamp). Since the query length threshold used in the Snap is smaller than 20,000 characters, disabling the automatic vertical slicing feature may resolve the issue if the actual query length is between the internal threshold and 20,000 characters.


ValuesString

The value that you want to associate with the property selected in the corresponding Properties field. The default values for the expected properties are:

PropertyValue

Use temp files in PK-Chunking

false
Number of threads in PK-Chunking8
Validate record count downloaded in Bulk APIfalse
Disable automatic vertical slicingfalse


Snap Execution


Default Value: Validate & Execute
Example
: Execute only

Dropdown list


PK Chunking

PK chunking splits bulk queries on very large tables into chunks based on the record IDs, or primary keys, of the queried records. Each chunk is processed as a separate batch that counts toward your daily batch limit. PK chunking is supported for the following objects: Account, Campaign, CampaignMember, Case, Contact, Lead, LoginHistory, Opportunity, Task, User, and custom objects. 

PK chunking works by adding record ID boundaries to the query with a WHERE clause, limiting the query results to a smaller chunk of the total results. The remaining results are fetched with additional queries that contain successive boundaries. The number of records within the ID boundaries of each chunk is referred to as the chunk size. The first query retrieves records between a specified starting ID and the starting ID plus the chunk size, the next query retrieves the next chunk of records, and so on. Since Salesforce.com appends a WHERE clause to the query in the PK Chunking mode, if SOQL Query has LIMIT clause in it, the Snap will submit a regular bulk query job without PK Chunking.

Learn more about the Salesforce PK Chunking Header.

If you select the Use PK chunking if supported checkbox, the value from Batch Size is used as the PK chunking batch size. Please note that the chunk size applies to the number of records in the queried table rather than the number of records in the query result. For example, if the queried table has 10 million records and 250,000 is selected for Batch Size, 40 batches are created to execute the bulk query job. One additional control batch is also created and it does not process any records. The status of the submitted bulk job and its batches can be monitored by logging in to your Salesforce.com account and going to Setup. From Setup, enter Bulk Data Load Jobs in the Quick Find box, then select Bulk Data Load Jobs.

PK Chunking requires Service Version 28.0 or later, cannot be used with Order By Clause or Limit Clause, and isn’t available for all Salesforce objects. Therefore, if Bulk API and Use PK chunking if supported are selected but other settings in the Snap don’t support PK chunking, the Snap will submit a regular bulk query job without PK chunking and display a warning that PK chunking was not used.

If your account doesn’t support PK chunking and you’re using the Bulk API, deselect the Use PK chunking if supported checkbox.


The output document schemas in preview mode  (validation) and execution mode for this Snap may differ when using Primary Key (PK) Chunking mode. This is because the Snap intentionally generates the output preview using regular Bulk API instead of PK Chunking, to reduce the costs involved in the PK Chunking operation.
Salesforce recommends the use of PK Chunking if the target Salesforce object (table) is relatively large (for example, more than few 100,000 records).

Examples


Pipeline: Salesforce.com Data to a File: This Pipeline reads data using a Salesforce read and writes it to a file. 

Reading Records from an Object

The following Salesforce Read Snap shows how the Snap is configured and how the object records are read. The Snap reads records from the Account object, and retrieves values for the fields, Idname, & type where the type field value is Analyst:  

Successful execution of the Snap gives the following preview:

 

Reading Records in Bulk 

The Salesforce Read Snap reads the records from the Standard Object, Account, and retrieves values for the 10 output fields (Output field limit) starting from the 3rd field (Output field offset). Additionally, we are passing the values dynamically for the Access token and the Instance URL fields in the Account settings of the Snap by defining the respective values in the pipeline parameters.  

1. The Salesforce Read Pipeline. 

2. The Key and Value parameters are assigned using the Edit Pipeline property on the designer. 

For this Pipeline, define the two Pipeline parameters:

  • access_Token
  • instance_URL

3. The Salesforce Read Snap reads the records from the Standard object, Account, to the extent of 10 output fields starting from the 3rd record(by defining the properties- Output field limit and Output field offset with the values 10 and 3 respectively).

4. Create a dynamic account and toggle (enable) the expressions for Access Token and Instance URL properties in order to pass the values dynamically.

Set Access token to _access_token and Instance URL to _instance_url. Note that the values are to be passed manually and are not suggestible. 

 

5. Successful execution of the Pipeline displays the below output preview:

Using Second Output View

This example Pipeline demonstrates how you can add an optional second output view that exposes the schema of the target object as the output document. For this, we configure the Pipeline using the Mapper and Salesforce Read Snaps.


First, we configure the Mapper Snap as follows. Upon validating the Snap, the target schema is populated in the Mapper Snap. Once this is available, we define the target path variables from the target schema.

Upon validation, the following output is generated in the Snap's preview.

Next, we configure the Salesforce Read Snap to read the specified records.

Upon validation, we can see the following outputs in both the output views of the Snap.

Output0 (default)

Output 1