Databricks has built-in keyword bindings for all of the data formats natively supported by Apache Spark. Databricks uses Delta Lake as the default protocol for reading and writing data and tables, whereas Apache Spark uses Parquet.
These articles provide an overview of many of the options and configurations available when you query data on Databricks.
The following data formats have built-in keyword configurations in Apache Spark DataFrames and SQL:
Databricks also provides a custom keyword for loading MLflow experiments.
Some data formats require additional configuration or special considerations for use:
Databricks recommends loading images as
Hive tables are natively supported by Apache Spark, but require configuration on Databricks.
Databricks can directly read compressed files in many file formats. You can also unzip compressed files on Databricks if necessary.
LZO requires a codec installation.