The upload data UI allows you to upload CSV or TSV files to create or overwrite a managed Delta Lake table.
You can create managed Delta tables in Unity Catalog or in the Hive Metastore.
To learn how to load files from cloud storage such as Azure Data Lake Storage Gen2, AWS S3, or Google Cloud Storage, check out the tutorial on COPY INTO.
You must have access to a running compute resource and permissions to create tables in a target schema.
Workspace admins can disable the upload data UI.
You can use the UI to create a Delta table by importing small CSV or TSV files from your local machine.
The upload UI supports uploading up to 10 files at a time.
The total size of uploaded files must be under 100 megabytes.
The file must be a CSV or TSV and have the extension “.csv” or “.tsv”.
Compressed files such as
tarfiles are not supported.
Click New > File upload.
Alternatively, you can go to the Add data UI and select Upload data.
Click the file browser button or drag and drop files directly on the drop zone.
Imported files are uploaded to a secure internal location within your account which is garbage collected daily.
You can upload data to the staging area without connecting to compute resources, but you must select an active compute resource to preview and configure your table.
You can preview 50 rows of your data when you configure the options for the uploaded table. Click the grid or list buttons under the file name to switch the presentation of your data.
Databricks stores data files for managed tables in the locations configured for the containing schema. You need proper permissions to create a table in a schema.
Select the desired schema in which to create a table by doing the following:
(For Unity Catalog-enabled workspaces only) You can select a catalog or the legacy
Select a schema.
(Optional) Edit the table name.
You can use the dropdown to select Overwrite existing table or Create new table. Operations that attempt to create new tables with name conflicts display an error message.
To create the table, click Create at the bottom of the page.
Format options depend on the file format you upload. Common format options appear in the header bar, while less commonly used options are available on the Advanced attributes dialog.
For CSV, the following options are available.
First row contains the header (enabled by default): This option specifies whether the CSV/TSV file contains a header.
Column delimiter: The separator character between columns. Only a single character is allowed, and backslash is not supported. This defaults to comma for CSV files.
Automatically detect column types (enabled by default): Automatically detect column types from file content. You can edit types in the preview table. If this is set to false, all column types are inferred as
Rows span multiple lines (disabled by default): Whether a column’s value can span multiple lines in the file.
The data preview updates automatically when you edit format options.
When you upload multiple files, the following rules apply:
The schema for the uploaded data is the result of merging all detected schemas. This cannot be disabled.
Header settings apply to all files. Make sure headers are consistently absent or present in all uploaded files to avoid data loss.
Uploaded files combine by appending all data as rows in the target table. Joining or merging records during file upload is not supported.
You can edit column names and types.
To edit types, click the icon with the type.
To edit the column name, click the input box at the top of the column.
Column names do not support commas, backslashes, or unicode characters (such as emojis).
Column data types are inferred by default for CSV files. You can interpret all columns as
STRING type by disabling Advanced attributes > Automatically detect column types.
Schema inference does a best effort detection of column types. Changing column types can lead to some values being cast to
NULLif the value cannot be cast correctly to the target data type. Casting
TIMESTAMPcolumns is not supported. Databricks recommends that you create a table first and then transform these columns using SQL functions afterwards.
To support table column names with special characters, the upload data UI leverages Column Mapping.
To add comments to columns, create the table and navigate to Data Explorer where you can add comments.
The upload data UI supports the following data types. For more information about individual data types see SQL data types.
8-byte signed integer numbers.
Values comprising values of fields year, month, and day, without a time-zone.
8-byte double-precision floating point numbers.
Character string values.
Values comprising values of fields year, month, day, hour, minute, and second, with the session local timezone.