All the recently terminated clusters are displayed in the ‘Terminated Clusters’ section in the cluster list page. The section displays clusters terminated in the last 7 days with a max cap of 70 such clusters. The bottom of the section also displays up to 30 jobs clusters that are recently terminated by jobs scheduler.
You can manually terminate a cluster both from the cluster list and details page.
You can also set auto-termination for a cluster. During cluster creation, you can specify an inactivity period in minutes after which you want the cluster to terminate. If the difference between the current time and the last command run on the cluster is more than the inactivity period specified, then Databricks will automatically terminate that cluster.
A cluster is inactive if all commands on the cluster have finished executing, including Spark Jobs, Structured Streaming, and JDBC. This does not include commands run by SSH-ing into the cluster and running bash commands.
You may opt out of auto-termination by unselecting the checkbox above, or by specifying an inactivity period of “0” in the API.
Auto-termination is best supported in the latest Spark versions. Older spark versions have known limitations which may result in inaccurate reporting of cluster activity. For example, clusters running JDBC, R, or streaming commands may report a stale activity time which will lead to premature cluster termination. Users are strongly recommended to upgrade to the most recent spark version to benefit from bug fixes and improvements to the autotermination feature.
Clusters do not report activity resulting from the use of DStreams. This means that an autoterminating cluster may be terminated while it is running DStreams. Please turn off autotermination for clusters running DStreams or consider using Structured Streaming.
Databricks records information whenever a cluster is terminated.
This page lists common termination reasons and describes potential steps for remediation.
Cloud Provider Limit¶
Databricks launches a cluster by requesting resources on behalf of your cloud account. Sometimes, these requests fail because they would exceed your cloud account’s resource limits. In AWS, common error codes include:
AWS limits the number of running instances for each node type. Possible solutions include:
- Requesting a cluster with fewer nodes.
- Requesting a cluster with a different node type.
- Ask AWS support to increase instance limits.
The cluster creation request exceeded the EBS volume limit. AWS has two types of volume limits: a limit on the total number of EBS volumes, and a limit on the total storage size of EBS volumes. Potential remediation steps:
- Requesting a cluster with fewer nodes.
- Check which of the two limits was exceeded. (AWS trusted advisor shows service limits for free). If the request exceeded the total number of EBS volumes, try reducing the requested number of volumes per node. If the request exceeded the total EBS storage size, try reducing the requested storage size and/or the number of EBS volumes.
- Ask AWS support to increase EBS volume limits.
- AWS limits the rate of API requests made for an AWS account. Please wait a while before retrying the request.
Cloud Provider Shutdown¶
The spark driver is a single point of failure because it holds all cluster state. If the instance hosting the driver node is shut down, Databricks must terminate the cluster. In AWS, common error codes include:
- Instance was terminated by a direct request to AWS which did not originate from Databricks. Please contact your AWS administrator for more details.
- AWS could not satisfy the instance request. Please wait a while and retry the request. Contact AWS support if the problem persists.
- Instance was terminated by AWS because the current spot price has exceeded the maximum bid made for this instance. To avoid this issue, use an on-demand instance for the driver, choose a different availability zone, or specify a higher spot bid price.
For other shutdown-related error codes, refer to AWS docs
Cloud Provider Launch Failures¶
In AWS, common error codes include:
Databricks was not authorized to launch the requested instances. Possible reasons include:
- Your AWS administrator invalidated Databricks’ AWS access key or IAM role used to launch instances.
- You are trying to launch a cluster using an IAM role that Databricks does not have permission to use. Please contact your AWS administrator who set up the IAM role. For more information, see Secure Access to S3 Buckets using IAM Roles
- Unsupported with message “EBS-optimized instances are not supported for your requested configuration”
- This error indicates that the selected instance type is not available in the selected availability zone (AZ). It does not actually have anything to do with EBS-optimization being enabled. To remediate, you can choose a different instance type or AZ.
- AuthFailure.ServiceLinkedRoleCreationNotPermitted - The provided credentials do not have permission to create the service-linked role for EC2 Spot Instances
- This error indicates that the Databricks administrator needs to update the credentials used to launch instances in your account. Instructions and the updated policy can be found here.
See here for a complete list of AWS error codes.
This means that Databricks was able to launch the cluster, but lost connection to the instance hosting the Spark driver.
This may be caused by incorrect networking configuration (ex: changing security group settings for Databricks workers), or may be a transient AWS networking issue.