Delta Live Tables properties reference
This article provides a reference for Delta Live Tables JSON setting specification and table properties in Databricks. For more details on using these various properties and configurations, see the following articles:
Delta Live Tables pipeline configurations
Fields |
---|
id Type: A globally unique identifier for this pipeline. The identifier is assigned by the system and cannot be changed. |
name Type: A user-friendly name for this pipeline. The name can be used to identify pipeline jobs in the UI. |
storage Type: A location on DBFS or cloud storage where output data and metadata required for pipeline execution are stored. Tables and metadata are stored in subdirectories of this location. When the The |
configuration Type: An optional list of settings to add to the Spark configuration of the cluster that will run the pipeline. These settings are read by the Delta Live Tables runtime and available to pipeline queries through the Spark configuration. Elements must be formatted as |
libraries Type: An array of notebooks containing the pipeline code and required artifacts. |
clusters Type: An array of specifications for the clusters to run the pipeline. If this is not specified, pipelines will automatically select a default cluster configuration for the pipeline. |
development Type: A flag indicating whether to run the pipeline in
The default value is |
continuous Type: A flag indicating whether to run the pipeline continuously. The default value is |
target Type: The name of a database for persisting pipeline output
data. Configuring the |
channel Type: The version of the Delta Live Tables runtime to use. The supported values are:
The |
edition Type The Delta Live Tables product edition to run the pipeline. This setting allows you to choose the best product edition based on the requirements of your pipeline:
The |
photon Type: A flag indicating whether to use Photon runtime to run the pipeline. Photon is the Databricks high performance Spark engine. Photon enabled pipelines are billed at a different rate than non-Photon pipelines. The |
pipelines.maxFlowRetryAttempts Type: The maximum number of attempts to retry a flow before failing a pipeline update when a retryable failure occurs. The default value is two. By default, when a retryable failure occurs, the Delta Live Tables runtime attempts to run the flow three times including the original attempt. |
pipelines.numUpdateRetryAttempts Type: The maximum number of attempts to retry an update before failing the update when a retryable failure occurs. The retry is run as a full update. The default is five. This parameter applies only to triggered updates run in production mode. There is no retry when your pipeline runs in development mode. |
Delta Live Tables table properties
In addition to the table properties supported by Delta Lake, you can set the following table properties.
Table properties |
---|
pipelines.autoOptimize.managed Default: Enables or disables automatic scheduled optimization of this table. |
pipelines.autoOptimize.zOrderCols Default: None An optional string containing a comma-separated list of column names to z-order this table by.
For example, |
pipelines.reset.allowed Default: Controls whether a full-refresh is allowed for this table. |
CDC table properties
The following table properties are added to control the behavior of tombstone management for DELETE
events when using CDC:
Table properties |
---|
pipelines.cdc.tombstoneGCThresholdInSeconds Default: 5 minutes Set this value to match the highest expected interval between out-of-order data. |
pipelines.cdc.tombstoneGCFrequencyInSeconds Default: 60 seconds Controls how frequently tombstones are checked for cleanup. |
Pipelines trigger interval
You can specify a pipeline trigger interval for the entire Delta Live Tables pipeline or as part of a dataset declaration. See Pipelines trigger interval.
pipelines.trigger.interval |
---|
The default is based on flow type:
The value is a number plus the time unit. The following are the valid time units:
You can use the singular or plural unit when defining the value, for example:
|