Configure settings for Databricks jobs
This article provides details on configuring Databricks Jobs and individual job tasks in the Jobs UI. To learn about using the Databricks CLI to edit job settings, see Jobs CLI. To learn about using the Jobs API, see the Jobs API.
Some configuration options are available on the job, and other options are available on individual tasks. For example, the maximum concurrent runs can be set only on the job, while parameters must be defined for each task.
Edit a job
To change the configuration for a job:
Click
Workflows in the sidebar.
In the Name column, click the job name.
The side panel displays the Job details. You can change the trigger for the job, compute configuration, notifications, maximum number of concurrent runs, and add or change tags. If job access control is enabled, you can also edit job permissions.
Add tags to a job
To add labels or key:value attributes to your job, you can add tags when you edit the job. You can use tags to filter jobs in the Jobs list; for example, you can use a department
tag to filter all jobs that belong to a specific department.
Note
Because job tags are not designed to store sensitive information such as personally identifiable information or passwords, Databricks recommends using tags for non-sensitive values only.
Tags also propagate to job clusters created when a job is run, allowing you to use tags with your existing cluster monitoring.
To add or edit tags, click + Tag in the Job details side panel. You can add the tag as a key and value, or a label. To add a label, enter the label in the Key field and leave the Value field empty.
Control access to a job
Job access control enables job owners and administrators to grant fine-grained permissions on their jobs. Job owners can choose which other users or groups can view the results of the job. Owners can also choose who can manage their job runs (Run now and Cancel run permissions).
See Jobs access control for details.
Configure maximum concurrent runs
Click Edit concurrent runs to set the maximum number of parallel runs for this job. Databricks skips the run if the job has already reached its maximum number of active runs when attempting to start a new run. Set this value higher than the default of 1 to perform multiple runs of the same job concurrently. This is useful, for example, if you trigger your job on a frequent schedule and want to allow consecutive runs to overlap with each other, or you want to trigger multiple runs that differ by their input parameters.
Configure a timeout for a job
To configure a maximum completion time for a job, click Timeout. If the job does not complete in this time, Databricks sets its status to “Timed Out”.
Edit a task
To set task configuration options:
Click
Workflows in the sidebar.
In the Name column, click the job name.
Click the Tasks tab and select the task to edit.
Define task dependencies
You can define the order of execution of tasks in a job using the Depends on dropdown menu. You can set this field to one or more tasks in the job.

Note
Depends on is not visible if the job consists of only a single task.
Configuring task dependencies creates a Directed Acyclic Graph (DAG) of task execution, a common way of representing execution order in job schedulers. For example, consider the following job consisting of four tasks:

Task 1 is the root task and does not depend on any other task.
Task 2 and Task 3 depend on Task 1 completing first.
Finally, Task 4 depends on Task 2 and Task 3 completing successfully.
Databricks runs upstream tasks before running downstream tasks, running as many of them in parallel as possible. The following diagram illustrates the order of processing for these tasks:

Configure a cluster for a task
To configure the cluster where a task runs, click the Cluster dropdown menu. You can edit a shared job cluster, but you cannot delete a shared cluster if it is still used by other tasks.
To learn more about selecting and configuring clusters to run tasks, see Use Databricks compute with your jobs.
Configure dependent libraries
Dependent libraries will be installed on the cluster before the task runs. You must set all task dependencies to ensure they are installed before the run starts. Follow the recommendations in Manage library dependencies for specifying dependencies.
Configure a timeout for a task
To configure a maximum completion time for a task, click Timeout. If the task does not complete in this time, Databricks sets its status to “Timed Out”.
Configure a retry policy
To configure a policy that determines when and how many times failed runs are retried, click + Add next to Retries. The retry interval is calculated in milliseconds between the start of the failed run and the subsequent retry run.
Note
If you configure both Timeout and Retries, the timeout applies to each retry.