USER MANUALS

Databricks

When Virtual DataPort runs on Windows, this feature requires the Microsoft Visual C++ 2013 Redistributable Package. If you are unsure if it is installed, download it and install it anyway. The installer of this package will detect if it is already installed.

There are two ways of configuring a Databricks data source to perform bulk data loads:

  1. Recommended way:

    1. Follow the steps of the section Bulk Data Load on a Distributed Object Storage like HDFS, S3 or ADLS.

    2. In the tab Read & Write of the data source, in Server time zone, select the time zone of the DataBricks server.

    3. In the same tab, select the format of the data: Parquet or Delta.

  2. Or, using the Databricks API and DBFS (explained in section Using the Databricks API and DBFS for Bulk Loads of the appendix).

Add feedback