Ingest Data into Data Cloud
When do you use Bulk and Streaming Ingestion?
Consider the data source and the problem you’re trying to solve when choosing the ingestion mode.
- Moving large amounts of data on a daily, weekly, or monthly schedule.
- Legacy systems where you can only export data in off-peak hours.
- A new Data Cloud org that you want to backfill with 30, 60, 90, or X days of data.
- Source system built on modern streaming architectures.
- Change data capture events.
- Consuming data from webhooks.
Example
Imagine you’re an integrator working for Northern Trail Outfitters (NTO). You need to extract the runner profiles and activity logs from NTO’s Track My Run mobile app and load them into Data Cloud. The marketer you’re working with has indicated that he needs the last 90 days of historical data and wants all new and updated data on a go-forward basis.
You query the last 90 days from your data warehouse as requested and you chunk up the data into 150 MB CSV files. You then load the data into Data Cloud via the bulk ingestion, understanding that the bulk ingestion is an appropriate solution for one-time operations. Now that the initial dataset is loaded, you decide it’s best to synchronize future changes with an upsert operation via streaming ingestion. These updates can be forwarded to Data Cloud in micro-batches as soon as the data becomes available.