Create a Databricks File Federation Data Stream (Beta)

Create a data stream to start the flow of data from your Databricks source to Data Cloud .

User Permissions Needed
To create a data stream:Data Cloud Admin

Before you begin:

  • Make sure that the Databricks File Federation connection is set up.
  • Review IP addresses to make sure that the Databricks connection has the necessary access.
  1. In Data Cloud, on the Data Streams tab, click New.

  2. Select Databricks File Federation as the source of the data stream, and click Next.

  3. From the menus, select the connection and database to use.

  4. Under Schemas, select the schema to use.

  5. From the Available Objects, select the objects that you want to include.

    You can't choose objects listed under In Use Objects.

  6. Click Next.

  7. Enter the object name and object API name. See Data Lake Object Naming Standards.

  8. Select the category to specify the type of data to ingest. See Category.

  9. Select a primary key to uniquely identify the incoming records. See Primary Key.

  10. Select the source fields you want to ingest and click Next.

    Fields with compatible data types are listed under Supported Fields. It's recommended to select all supported fields.

    Unsupported Fields lists the fields with data types that aren't supported in Data Cloud.

  11. Select a data space. If you don't have multiple data spaces provisioned, select Default. For more information, see Data Spaces.

  12. Click Deploy.

This table lists the mapping of Databricks data types with Data Cloud data types.

Databricks Data TypeIs Auto-mapped to Data Cloud Data Type
INT, SMALLINT, TINYINT, BIGINTNumber
FLOAT, DOUBLE. DECIMAL(p,s)Number
STRINGText
TIMESTAMP with TimezoneDateTime
DATEDate
BOOLEANBoolean
VARIANT, OBJECT, ARRAY, GEOGRAPHY, GEOMETRYNot Suported
BINARY, VARBINARYNot Supported
STRUCTNot Supported