Bulk Ingestion
With the Data Cloud Ingestion API, you can upsert or delete large data sets. Prepare a CSV file for the data you want to upload, create a job, upload job data, and let Salesforce take care of the rest.
The Ingestion API upserts or deletes data in Data Cloud using jobs. A bulk job typically goes through following stages:
- Create a job to specify the object type of the data being processed and the operation that’s performed on the data. The operations supported are upserting bulk data or deleting bulk data.
- After the job is created upload data in CSV format to the job.
- To signal the data is ready to be processed, close the job. You can choose to abort the job if necessary.
- Monitor the progress of the job and act on any failed records.
- Delete a job.
Lists the field names for the object that you're processing in the first row in the CSV file. Each subsequent row corresponds to a record in your Data Cloud data lake. All the records in a CSV file must be for the same object. You specify this object when you first create the job.
- Include all required fields when you create a record.
- Each field-name header in the file must be the same as the Datasource Object's field names. Results only include columns that are a match.
- Updating records works as a full replace. Patch semantics aren’t supported.
- Files must be in UTF-8 format. Upload data must not exceed 150 MB.
- CSV files are expected to be formatted according to RFC 4180, Common Format, and MIME Type for CSV Files.
- Only supports comma field delimiters.
- Empty field values are set to null.
The Ingestion API supports ISO 8601 UTC with Zulu format.
-
Use the yyyy-MM-dd format to specify date values. For example, in 2021-07-05
- yyyy is the four-digit year
- MM is the two-digit month
- dd is the two-digit day
-
Use the yyyy-MM-ddTHH:mm
.SSSZ format to specify dateTime fields. For example, 2021-07-05T09:31:44.457Z - yyyy is the four-digit year
- MM is the two-digit month
- dd is the two-digit day
- 'T' is a separator indicating that time-of-day follows
- HH is the two-digit hour
- mm is the two-digit minute
- ss is the two-digit seconds
- SSS is the optional three-digit milliseconds (000=999)
- 'Z' is the reference UTC timezone
-
Creates a job that represents the bulk operation for loading data into the Data Cloud org.
-
Uploads CSV data for the job to process. A job can consist of up to 100 data files. After all of the data has been uploaded, close the job to indicate the data is ready to be processed.
-
Close a job by updating the job state to UploadComplete. When you close a job, the uploaded data is enqueued for processing. To abort a job, update the job state to Aborted. If you abort a job, the job doesn’t get queued or processed. Any data uploaded as part of the aborted job is deleted.
-
To delete, a job must have a state of UploadComplete, JobComplete, Aborted, or Failed. When you delete a job, the stored job data is also deleted and its metadata is removed.
-
Retrieves detailed information about the specified job.
-
Retrieves all jobs in Data Cloud.
See Also