clusters
Creates, updates, deletes, gets or lists a clusters resource.
Overview
| Name | clusters |
| Type | Resource |
| Id | databricks_workspace.compute.clusters |
Fields
The following fields are returned by SELECT queries:
- get
- list
| Name | Datatype | Description |
|---|---|---|
cluster_id | string | Canonical identifier for the cluster. This id is retained during cluster restarts and resizes, while each new cluster has a globally unique id. |
driver_instance_pool_id | string | The optional ID of the instance pool for the driver of the cluster belongs. The pool cluster uses the instance pool with id (instance_pool_id) if the driver pool is not assigned. |
driver_node_type_id | string | The node type of the Spark driver. Note that this field is optional; if unset, the driver node type will be set as the same value as `node_type_id` defined above. This field, along with node_type_id, should not be set if virtual_cluster_size is set. If both driver_node_type_id, node_type_id, and virtual_cluster_size are specified, driver_node_type_id and node_type_id take precedence. |
instance_pool_id | string | The optional ID of the instance pool to which the cluster belongs. |
node_type_id | string | This field encodes, through a single value, the resources available to each of the Spark nodes in this cluster. For example, the Spark nodes can be provisioned and optimized for memory or compute intensive workloads. A list of available node types can be retrieved by using the :method:clusters/listNodeTypes API call. |
policy_id | string | The ID of the cluster policy used to create the cluster if applicable. |
spark_context_id | integer | A canonical SparkContext identifier. This value *does* change when the Spark driver restarts. The pair `(cluster_id, spark_context_id)` is a globally unique identifier over all Spark contexts. |
cluster_name | string | Cluster name requested by the user. This doesn't have to be unique. If not specified at creation, the cluster name will be an empty string. For job clusters, the cluster name is automatically set based on the job and job run IDs. |
creator_user_name | string | Creator user name. The field won't be included in the response if the user has already been deleted. |
single_user_name | string | Single user name if data_security_mode is `SINGLE_USER` |
autoscale | object | Parameters needed in order to automatically scale clusters up and down based on load. Note: autoscaling works best with DB runtime versions 3.0 or later. |
autotermination_minutes | integer | Automatically terminates the cluster after it is inactive for this time in minutes. If not set, this cluster will not be automatically terminated. If specified, the threshold must be between 10 and 10000 minutes. Users can also set this value to 0 to explicitly disable automatic termination. |
aws_attributes | object | Attributes related to clusters running on Amazon Web Services. If not specified at cluster creation, a set of default values will be used. |
azure_attributes | object | Attributes related to clusters running on Microsoft Azure. If not specified at cluster creation, a set of default values will be used. |
cluster_cores | number | Number of CPU cores available for this cluster. Note that this can be fractional, e.g. 7.5 cores, since certain node types are configured to share cores between Spark nodes on the same instance. |
cluster_log_conf | object | The configuration for delivering spark logs to a long-term storage destination. Three kinds of destinations (DBFS, S3 and Unity Catalog volumes) are supported. Only one destination can be specified for one cluster. If the conf is given, the logs will be delivered to the destination every `5 mins`. The destination of driver logs is `$destination/$clusterId/driver`, while the destination of executor logs is `$destination/$clusterId/executor`. |
cluster_log_status | object | Cluster log delivery status. |
cluster_memory_mb | integer | Total amount of cluster memory, in megabytes |
cluster_source | string | Determines whether the cluster was created by a user through the UI, created by the Databricks Jobs Scheduler, or through an API request. (API, JOB, MODELS, PIPELINE, PIPELINE_MAINTENANCE, SQL, UI) |
custom_tags | object | Additional tags for cluster resources. Databricks will tag all cluster resources (e.g., AWS instances and EBS volumes) with these tags in addition to `default_tags`. Notes: - Currently, Databricks allows at most 45 custom tags - Clusters can only reuse cloud resources if the resources' tags are a subset of the cluster tags |
data_security_mode | string | Data security mode decides what data governance model to use when accessing data from a cluster.<br /><br />The following modes can only be used when `kind = CLASSIC_PREVIEW`. * `DATA_SECURITY_MODE_AUTO`:<br />Databricks will choose the most appropriate access mode depending on your compute configuration.<br />* `DATA_SECURITY_MODE_STANDARD`: Alias for `USER_ISOLATION`. * `DATA_SECURITY_MODE_DEDICATED`:<br />Alias for `SINGLE_USER`.<br /><br />The following modes can be used regardless of `kind`. * `NONE`: No security isolation for<br />multiple users sharing the cluster. Data governance features are not available in this mode. *<br />`SINGLE_USER`: A secure cluster that can only be exclusively used by a single user specified in<br />`single_user_name`. Most programming languages, cluster features and data governance features<br />are available in this mode. * `USER_ISOLATION`: A secure cluster that can be shared by multiple<br />users. Cluster users are fully isolated so that they cannot see each other's data and<br />credentials. Most data governance features are supported in this mode. But programming languages<br />and cluster features might be limited.<br /><br />The following modes are deprecated starting with Databricks Runtime 15.0 and will be removed for<br />future Databricks Runtime versions:<br /><br />* `LEGACY_TABLE_ACL`: This mode is for users migrating from legacy Table ACL clusters. *<br />`LEGACY_PASSTHROUGH`: This mode is for users migrating from legacy Passthrough on high<br />concurrency clusters. * `LEGACY_SINGLE_USER`: This mode is for users migrating from legacy<br />Passthrough on standard clusters. * `LEGACY_SINGLE_USER_STANDARD`: This mode provides a way that<br />doesn’t have UC nor passthrough enabled. (DATA_SECURITY_MODE_AUTO, DATA_SECURITY_MODE_DEDICATED, DATA_SECURITY_MODE_STANDARD, LEGACY_PASSTHROUGH, LEGACY_SINGLE_USER, LEGACY_SINGLE_USER_STANDARD, LEGACY_TABLE_ACL, NONE, SINGLE_USER, USER_ISOLATION) |
default_tags | object | Tags that are added by Databricks regardless of any `custom_tags`, including: - Vendor: Databricks - Creator: <username_of_creator> - ClusterName: <name_of_cluster> - ClusterId: <id_of_cluster> - Name: <Databricks internal use> |
docker_image | object | Custom docker image BYOC |
driver | object | Node on which the Spark driver resides. The driver node contains the Spark master and the Databricks application that manages the per-notebook Spark REPLs. |
driver_node_type_flexibility | object | Flexible node type configuration for the driver node. |
enable_elastic_disk | boolean | Autoscaling Local Storage: when enabled, this cluster will dynamically acquire additional disk space when its Spark workers are running low on disk space. |
enable_local_disk_encryption | boolean | Whether to enable LUKS on cluster VMs' local disks |
executors | array | Nodes on which the Spark executors reside. |
gcp_attributes | object | Attributes related to clusters running on Google Cloud Platform. If not specified at cluster creation, a set of default values will be used. |
init_scripts | array | The configuration for storing init scripts. Any number of destinations can be specified. The scripts are executed sequentially in the order provided. If `cluster_log_conf` is specified, init script logs are sent to `<destination>/<cluster-ID>/init_scripts`. |
is_single_node | boolean | This field can only be used when `kind = CLASSIC_PREVIEW`. When set to true, Databricks will automatically set single node related `custom_tags`, `spark_conf`, and `num_workers` |
jdbc_port | integer | Port on which Spark JDBC server is listening, in the driver nod. No service will be listeningon on this port in executor nodes. |
kind | string | The kind of compute described by this compute specification.<br /><br />Depending on `kind`, different validations and default values will be applied.<br /><br />Clusters with `kind = CLASSIC_PREVIEW` support the following fields, whereas clusters with no<br />specified `kind` do not. * [is_single_node](/api/workspace/clusters/create#is_single_node) *<br />[use_ml_runtime](/api/workspace/clusters/create#use_ml_runtime) *<br />[data_security_mode](/api/workspace/clusters/create#data_security_mode) set to<br />`DATA_SECURITY_MODE_AUTO`, `DATA_SECURITY_MODE_DEDICATED`, or `DATA_SECURITY_MODE_STANDARD`<br /><br />By using the [simple form], your clusters are automatically using `kind = CLASSIC_PREVIEW`.<br /><br />[simple form]: https://docs.databricks.com/compute/simple-form.html (CLASSIC_PREVIEW) |
last_restarted_time | integer | the timestamp that the cluster was started/restarted |
last_state_loss_time | integer | Time when the cluster driver last lost its state (due to a restart or driver failure). |
num_workers | integer | Number of worker nodes that this cluster should have. A cluster has one Spark Driver and `num_workers` Executors for a total of `num_workers` + 1 Spark nodes. Note: When reading the properties of a cluster, this field reflects the desired number of workers rather than the actual current number of workers. For instance, if a cluster is resized from 5 to 10 workers, this field will immediately be updated to reflect the target size of 10 workers, whereas the workers listed in `spark_info` will gradually increase from 5 to 10 as the new nodes are provisioned. |
remote_disk_throughput | integer | If set, what the configurable throughput (in Mb/s) for the remote disk is. Currently only supported for GCP HYPERDISK_BALANCED disks. |
runtime_engine | string | Determines the cluster's runtime engine, either standard or Photon. This field is not compatible with legacy `spark_version` values that contain `-photon-`. Remove `-photon-` from the `spark_version` and set `runtime_engine` to `PHOTON`. If left unspecified, the runtime engine defaults to standard unless the spark_version contains -photon-, in which case Photon will be used. (NULL, PHOTON, STANDARD) |
spark_conf | object | An object containing a set of optional, user-specified Spark configuration key-value pairs. Users can also pass in a string of extra JVM options to the driver and the executors via `spark.driver.extraJavaOptions` and `spark.executor.extraJavaOptions` respectively. |
spark_env_vars | object | An object containing a set of optional, user-specified environment variable key-value pairs. Please note that key-value pair of the form (X,Y) will be exported as is (i.e., `export X='Y'`) while launching the driver and workers. In order to specify an additional set of `SPARK_DAEMON_JAVA_OPTS`, we recommend appending them to `$SPARK_DAEMON_JAVA_OPTS` as shown in the example below. This ensures that all default databricks managed environmental variables are included as well. Example Spark environment variables: `{"SPARK_WORKER_MEMORY": "28000m", "SPARK_LOCAL_DIRS": "/local_disk0"}` or `{"SPARK_DAEMON_JAVA_OPTS": "$SPARK_DAEMON_JAVA_OPTS -Dspark.shuffle.service.enabled=true"}` |
spark_version | string | The Spark version of the cluster, e.g. `3.3.x-scala2.11`. A list of available Spark versions can be retrieved by using the :method:clusters/sparkVersions API call. |
spec | object | The spec contains a snapshot of the latest user specified settings that were used to create/edit the cluster. Note: not included in the response of the ListClusters API. |
ssh_public_keys | array | SSH public key contents that will be added to each Spark node in this cluster. The corresponding private keys can be used to login with the user name `ubuntu` on port `2200`. Up to 10 keys can be specified. |
start_time | integer | Time (in epoch milliseconds) when the cluster creation request was received (when the cluster entered a `PENDING` state). |
state | string | Current state of the cluster. (ERROR, PENDING, RESIZING, RESTARTING, RUNNING, TERMINATED, TERMINATING, UNKNOWN) |
state_message | string | A message associated with the most recent state transition (e.g., the reason why the cluster entered a `TERMINATED` state). |
terminated_time | integer | Time (in epoch milliseconds) when the cluster was terminated, if applicable. |
termination_reason | object | Information about why the cluster was terminated. This field only appears when the cluster is in a `TERMINATING` or `TERMINATED` state. |
total_initial_remote_disk_size | integer | If set, what the total initial volume size (in GB) of the remote disks should be. Currently only supported for GCP HYPERDISK_BALANCED disks. |
use_ml_runtime | boolean | This field can only be used when `kind = CLASSIC_PREVIEW`. `effective_spark_version` is determined by `spark_version` (DBR release), this field `use_ml_runtime`, and whether `node_type_id` is gpu node or not. |
worker_node_type_flexibility | object | Flexible node type configuration for worker nodes. |
workload_type | object | Cluster Attributes showing for clusters workload types. |
| Name | Datatype | Description |
|---|---|---|
cluster_id | string | Canonical identifier for the cluster. This id is retained during cluster restarts and resizes, while each new cluster has a globally unique id. |
driver_instance_pool_id | string | The optional ID of the instance pool for the driver of the cluster belongs. The pool cluster uses the instance pool with id (instance_pool_id) if the driver pool is not assigned. |
driver_node_type_id | string | The node type of the Spark driver. Note that this field is optional; if unset, the driver node type will be set as the same value as `node_type_id` defined above. This field, along with node_type_id, should not be set if virtual_cluster_size is set. If both driver_node_type_id, node_type_id, and virtual_cluster_size are specified, driver_node_type_id and node_type_id take precedence. |
instance_pool_id | string | The optional ID of the instance pool to which the cluster belongs. |
node_type_id | string | This field encodes, through a single value, the resources available to each of the Spark nodes in this cluster. For example, the Spark nodes can be provisioned and optimized for memory or compute intensive workloads. A list of available node types can be retrieved by using the :method:clusters/listNodeTypes API call. |
policy_id | string | The ID of the cluster policy used to create the cluster if applicable. |
spark_context_id | integer | A canonical SparkContext identifier. This value *does* change when the Spark driver restarts. The pair `(cluster_id, spark_context_id)` is a globally unique identifier over all Spark contexts. |
cluster_name | string | Cluster name requested by the user. This doesn't have to be unique. If not specified at creation, the cluster name will be an empty string. For job clusters, the cluster name is automatically set based on the job and job run IDs. |
creator_user_name | string | Creator user name. The field won't be included in the response if the user has already been deleted. |
single_user_name | string | Single user name if data_security_mode is `SINGLE_USER` |
autoscale | object | Parameters needed in order to automatically scale clusters up and down based on load. Note: autoscaling works best with DB runtime versions 3.0 or later. |
autotermination_minutes | integer | Automatically terminates the cluster after it is inactive for this time in minutes. If not set, this cluster will not be automatically terminated. If specified, the threshold must be between 10 and 10000 minutes. Users can also set this value to 0 to explicitly disable automatic termination. |
aws_attributes | object | Attributes related to clusters running on Amazon Web Services. If not specified at cluster creation, a set of default values will be used. |
azure_attributes | object | Attributes related to clusters running on Microsoft Azure. If not specified at cluster creation, a set of default values will be used. |
cluster_cores | number | Number of CPU cores available for this cluster. Note that this can be fractional, e.g. 7.5 cores, since certain node types are configured to share cores between Spark nodes on the same instance. |
cluster_log_conf | object | The configuration for delivering spark logs to a long-term storage destination. Three kinds of destinations (DBFS, S3 and Unity Catalog volumes) are supported. Only one destination can be specified for one cluster. If the conf is given, the logs will be delivered to the destination every `5 mins`. The destination of driver logs is `$destination/$clusterId/driver`, while the destination of executor logs is `$destination/$clusterId/executor`. |
cluster_log_status | object | Cluster log delivery status. |
cluster_memory_mb | integer | Total amount of cluster memory, in megabytes |
cluster_source | string | Determines whether the cluster was created by a user through the UI, created by the Databricks Jobs Scheduler, or through an API request. (API, JOB, MODELS, PIPELINE, PIPELINE_MAINTENANCE, SQL, UI) |
custom_tags | object | Additional tags for cluster resources. Databricks will tag all cluster resources (e.g., AWS instances and EBS volumes) with these tags in addition to `default_tags`. Notes: - Currently, Databricks allows at most 45 custom tags - Clusters can only reuse cloud resources if the resources' tags are a subset of the cluster tags |
data_security_mode | string | Data security mode decides what data governance model to use when accessing data from a cluster.<br /><br />The following modes can only be used when `kind = CLASSIC_PREVIEW`. * `DATA_SECURITY_MODE_AUTO`:<br />Databricks will choose the most appropriate access mode depending on your compute configuration.<br />* `DATA_SECURITY_MODE_STANDARD`: Alias for `USER_ISOLATION`. * `DATA_SECURITY_MODE_DEDICATED`:<br />Alias for `SINGLE_USER`.<br /><br />The following modes can be used regardless of `kind`. * `NONE`: No security isolation for<br />multiple users sharing the cluster. Data governance features are not available in this mode. *<br />`SINGLE_USER`: A secure cluster that can only be exclusively used by a single user specified in<br />`single_user_name`. Most programming languages, cluster features and data governance features<br />are available in this mode. * `USER_ISOLATION`: A secure cluster that can be shared by multiple<br />users. Cluster users are fully isolated so that they cannot see each other's data and<br />credentials. Most data governance features are supported in this mode. But programming languages<br />and cluster features might be limited.<br /><br />The following modes are deprecated starting with Databricks Runtime 15.0 and will be removed for<br />future Databricks Runtime versions:<br /><br />* `LEGACY_TABLE_ACL`: This mode is for users migrating from legacy Table ACL clusters. *<br />`LEGACY_PASSTHROUGH`: This mode is for users migrating from legacy Passthrough on high<br />concurrency clusters. * `LEGACY_SINGLE_USER`: This mode is for users migrating from legacy<br />Passthrough on standard clusters. * `LEGACY_SINGLE_USER_STANDARD`: This mode provides a way that<br />doesn’t have UC nor passthrough enabled. (DATA_SECURITY_MODE_AUTO, DATA_SECURITY_MODE_DEDICATED, DATA_SECURITY_MODE_STANDARD, LEGACY_PASSTHROUGH, LEGACY_SINGLE_USER, LEGACY_SINGLE_USER_STANDARD, LEGACY_TABLE_ACL, NONE, SINGLE_USER, USER_ISOLATION) |
default_tags | object | Tags that are added by Databricks regardless of any `custom_tags`, including: - Vendor: Databricks - Creator: <username_of_creator> - ClusterName: <name_of_cluster> - ClusterId: <id_of_cluster> - Name: <Databricks internal use> |
docker_image | object | Custom docker image BYOC |
driver | object | Node on which the Spark driver resides. The driver node contains the Spark master and the Databricks application that manages the per-notebook Spark REPLs. |
driver_node_type_flexibility | object | Flexible node type configuration for the driver node. |
enable_elastic_disk | boolean | Autoscaling Local Storage: when enabled, this cluster will dynamically acquire additional disk space when its Spark workers are running low on disk space. |
enable_local_disk_encryption | boolean | Whether to enable LUKS on cluster VMs' local disks |
executors | array | Nodes on which the Spark executors reside. |
gcp_attributes | object | Attributes related to clusters running on Google Cloud Platform. If not specified at cluster creation, a set of default values will be used. |
init_scripts | array | The configuration for storing init scripts. Any number of destinations can be specified. The scripts are executed sequentially in the order provided. If `cluster_log_conf` is specified, init script logs are sent to `<destination>/<cluster-ID>/init_scripts`. |
is_single_node | boolean | This field can only be used when `kind = CLASSIC_PREVIEW`. When set to true, Databricks will automatically set single node related `custom_tags`, `spark_conf`, and `num_workers` |
jdbc_port | integer | Port on which Spark JDBC server is listening, in the driver nod. No service will be listeningon on this port in executor nodes. |
kind | string | The kind of compute described by this compute specification.<br /><br />Depending on `kind`, different validations and default values will be applied.<br /><br />Clusters with `kind = CLASSIC_PREVIEW` support the following fields, whereas clusters with no<br />specified `kind` do not. * [is_single_node](/api/workspace/clusters/create#is_single_node) *<br />[use_ml_runtime](/api/workspace/clusters/create#use_ml_runtime) *<br />[data_security_mode](/api/workspace/clusters/create#data_security_mode) set to<br />`DATA_SECURITY_MODE_AUTO`, `DATA_SECURITY_MODE_DEDICATED`, or `DATA_SECURITY_MODE_STANDARD`<br /><br />By using the [simple form], your clusters are automatically using `kind = CLASSIC_PREVIEW`.<br /><br />[simple form]: https://docs.databricks.com/compute/simple-form.html (CLASSIC_PREVIEW) |
last_restarted_time | integer | the timestamp that the cluster was started/restarted |
last_state_loss_time | integer | Time when the cluster driver last lost its state (due to a restart or driver failure). |
num_workers | integer | Number of worker nodes that this cluster should have. A cluster has one Spark Driver and `num_workers` Executors for a total of `num_workers` + 1 Spark nodes. Note: When reading the properties of a cluster, this field reflects the desired number of workers rather than the actual current number of workers. For instance, if a cluster is resized from 5 to 10 workers, this field will immediately be updated to reflect the target size of 10 workers, whereas the workers listed in `spark_info` will gradually increase from 5 to 10 as the new nodes are provisioned. |
remote_disk_throughput | integer | If set, what the configurable throughput (in Mb/s) for the remote disk is. Currently only supported for GCP HYPERDISK_BALANCED disks. |
runtime_engine | string | Determines the cluster's runtime engine, either standard or Photon. This field is not compatible with legacy `spark_version` values that contain `-photon-`. Remove `-photon-` from the `spark_version` and set `runtime_engine` to `PHOTON`. If left unspecified, the runtime engine defaults to standard unless the spark_version contains -photon-, in which case Photon will be used. (NULL, PHOTON, STANDARD) |
spark_conf | object | An object containing a set of optional, user-specified Spark configuration key-value pairs. Users can also pass in a string of extra JVM options to the driver and the executors via `spark.driver.extraJavaOptions` and `spark.executor.extraJavaOptions` respectively. |
spark_env_vars | object | An object containing a set of optional, user-specified environment variable key-value pairs. Please note that key-value pair of the form (X,Y) will be exported as is (i.e., `export X='Y'`) while launching the driver and workers. In order to specify an additional set of `SPARK_DAEMON_JAVA_OPTS`, we recommend appending them to `$SPARK_DAEMON_JAVA_OPTS` as shown in the example below. This ensures that all default databricks managed environmental variables are included as well. Example Spark environment variables: `{"SPARK_WORKER_MEMORY": "28000m", "SPARK_LOCAL_DIRS": "/local_disk0"}` or `{"SPARK_DAEMON_JAVA_OPTS": "$SPARK_DAEMON_JAVA_OPTS -Dspark.shuffle.service.enabled=true"}` |
spark_version | string | The Spark version of the cluster, e.g. `3.3.x-scala2.11`. A list of available Spark versions can be retrieved by using the :method:clusters/sparkVersions API call. |
spec | object | The spec contains a snapshot of the latest user specified settings that were used to create/edit the cluster. Note: not included in the response of the ListClusters API. |
ssh_public_keys | array | SSH public key contents that will be added to each Spark node in this cluster. The corresponding private keys can be used to login with the user name `ubuntu` on port `2200`. Up to 10 keys can be specified. |
start_time | integer | Time (in epoch milliseconds) when the cluster creation request was received (when the cluster entered a `PENDING` state). |
state | string | Current state of the cluster. (ERROR, PENDING, RESIZING, RESTARTING, RUNNING, TERMINATED, TERMINATING, UNKNOWN) |
state_message | string | A message associated with the most recent state transition (e.g., the reason why the cluster entered a `TERMINATED` state). |
terminated_time | integer | Time (in epoch milliseconds) when the cluster was terminated, if applicable. |
termination_reason | object | Information about why the cluster was terminated. This field only appears when the cluster is in a `TERMINATING` or `TERMINATED` state. |
total_initial_remote_disk_size | integer | If set, what the total initial volume size (in GB) of the remote disks should be. Currently only supported for GCP HYPERDISK_BALANCED disks. |
use_ml_runtime | boolean | This field can only be used when `kind = CLASSIC_PREVIEW`. `effective_spark_version` is determined by `spark_version` (DBR release), this field `use_ml_runtime`, and whether `node_type_id` is gpu node or not. |
worker_node_type_flexibility | object | Flexible node type configuration for worker nodes. |
workload_type | object | Cluster Attributes showing for clusters workload types. |
Methods
The following methods are available for this resource:
| Name | Accessible by | Required Params | Optional Params | Description |
|---|---|---|---|---|
get | select | cluster_id, deployment_name | Retrieves the information for a cluster given its identifier. Clusters can be described while they are | |
list | select | deployment_name | filter_by, page_size, page_token, sort_by | Return information about all pinned and active clusters, and all clusters terminated within the last |
create | insert | deployment_name, spark_version | Creates a new Spark cluster. This method will acquire new instances from the cloud provider if | |
change_owner | exec | deployment_name, cluster_id, owner_username | Change the owner of the cluster. You must be an admin and the cluster must be terminated to perform | |
delete | exec | deployment_name, cluster_id | Terminates the Spark cluster with the specified ID. The cluster is removed asynchronously. Once the | |
edit | exec | deployment_name, cluster_id, spark_version | Updates the configuration of a cluster to match the provided attributes and size. A cluster can be | |
events | exec | deployment_name, cluster_id | Retrieves a list of events about the activity of a cluster. This API is paginated. If there are more | |
permanent_delete | exec | deployment_name, cluster_id | Permanently deletes a Spark cluster. This cluster is terminated and resources are asynchronously | |
pin | exec | deployment_name, cluster_id | Pinning a cluster ensures that the cluster will always be returned by the ListClusters API. Pinning a | |
resize | exec | deployment_name, cluster_id | Resizes a cluster to have a desired number of workers. This will fail unless the cluster is in a | |
restart | exec | deployment_name, cluster_id | Restarts a Spark cluster with the supplied ID. If the cluster is not currently in a RUNNING state, | |
start | exec | deployment_name, cluster_id | Starts a terminated Spark cluster with the supplied ID. This works similar to createCluster except: | |
unpin | exec | deployment_name, cluster_id | Unpinning a cluster will allow the cluster to eventually be removed from the ListClusters API. | |
update | exec | deployment_name, cluster_id, update_mask | Updates the configuration of a cluster to match the partial set of attributes and size. Denote which |
Parameters
Parameters can be passed in the WHERE clause of a query. Check the Methods section to see which parameters are required or optional for each operation.
| Name | Datatype | Description |
|---|---|---|
cluster_id | string | The cluster about which to retrieve information. |
deployment_name | string | The Databricks Workspace Deployment Name (default: dbc-abcd0123-a1bc) |
filter_by | object | Filters to apply to the list of clusters. |
page_size | integer | Use this field to specify the maximum number of results to be returned by the server. The server may further constrain the maximum number of results returned in a single page. |
page_token | string | Use next_page_token or prev_page_token returned from the previous request to list the next or previous page of clusters respectively. |
sort_by | object | Sort the list of clusters by a specific criteria. |
SELECT examples
- get
- list
Retrieves the information for a cluster given its identifier. Clusters can be described while they are
SELECT
cluster_id,
driver_instance_pool_id,
driver_node_type_id,
instance_pool_id,
node_type_id,
policy_id,
spark_context_id,
cluster_name,
creator_user_name,
single_user_name,
autoscale,
autotermination_minutes,
aws_attributes,
azure_attributes,
cluster_cores,
cluster_log_conf,
cluster_log_status,
cluster_memory_mb,
cluster_source,
custom_tags,
data_security_mode,
default_tags,
docker_image,
driver,
driver_node_type_flexibility,
enable_elastic_disk,
enable_local_disk_encryption,
executors,
gcp_attributes,
init_scripts,
is_single_node,
jdbc_port,
kind,
last_restarted_time,
last_state_loss_time,
num_workers,
remote_disk_throughput,
runtime_engine,
spark_conf,
spark_env_vars,
spark_version,
spec,
ssh_public_keys,
start_time,
state,
state_message,
terminated_time,
termination_reason,
total_initial_remote_disk_size,
use_ml_runtime,
worker_node_type_flexibility,
workload_type
FROM databricks_workspace.compute.clusters
WHERE cluster_id = '{{ cluster_id }}' -- required
AND deployment_name = '{{ deployment_name }}' -- required
;
Return information about all pinned and active clusters, and all clusters terminated within the last
SELECT
cluster_id,
driver_instance_pool_id,
driver_node_type_id,
instance_pool_id,
node_type_id,
policy_id,
spark_context_id,
cluster_name,
creator_user_name,
single_user_name,
autoscale,
autotermination_minutes,
aws_attributes,
azure_attributes,
cluster_cores,
cluster_log_conf,
cluster_log_status,
cluster_memory_mb,
cluster_source,
custom_tags,
data_security_mode,
default_tags,
docker_image,
driver,
driver_node_type_flexibility,
enable_elastic_disk,
enable_local_disk_encryption,
executors,
gcp_attributes,
init_scripts,
is_single_node,
jdbc_port,
kind,
last_restarted_time,
last_state_loss_time,
num_workers,
remote_disk_throughput,
runtime_engine,
spark_conf,
spark_env_vars,
spark_version,
spec,
ssh_public_keys,
start_time,
state,
state_message,
terminated_time,
termination_reason,
total_initial_remote_disk_size,
use_ml_runtime,
worker_node_type_flexibility,
workload_type
FROM databricks_workspace.compute.clusters
WHERE deployment_name = '{{ deployment_name }}' -- required
AND filter_by = '{{ filter_by }}'
AND page_size = '{{ page_size }}'
AND page_token = '{{ page_token }}'
AND sort_by = '{{ sort_by }}'
;
INSERT examples
- create
- Manifest
Creates a new Spark cluster. This method will acquire new instances from the cloud provider if
INSERT INTO databricks_workspace.compute.clusters (
spark_version,
apply_policy_default_values,
autoscale,
autotermination_minutes,
aws_attributes,
azure_attributes,
clone_from,
cluster_log_conf,
cluster_name,
custom_tags,
data_security_mode,
docker_image,
driver_instance_pool_id,
driver_node_type_flexibility,
driver_node_type_id,
enable_elastic_disk,
enable_local_disk_encryption,
gcp_attributes,
init_scripts,
instance_pool_id,
is_single_node,
kind,
node_type_id,
num_workers,
policy_id,
remote_disk_throughput,
runtime_engine,
single_user_name,
spark_conf,
spark_env_vars,
ssh_public_keys,
total_initial_remote_disk_size,
use_ml_runtime,
worker_node_type_flexibility,
workload_type,
deployment_name
)
SELECT
'{{ spark_version }}' /* required */,
{{ apply_policy_default_values }},
'{{ autoscale }}',
{{ autotermination_minutes }},
'{{ aws_attributes }}',
'{{ azure_attributes }}',
'{{ clone_from }}',
'{{ cluster_log_conf }}',
'{{ cluster_name }}',
'{{ custom_tags }}',
'{{ data_security_mode }}',
'{{ docker_image }}',
'{{ driver_instance_pool_id }}',
'{{ driver_node_type_flexibility }}',
'{{ driver_node_type_id }}',
{{ enable_elastic_disk }},
{{ enable_local_disk_encryption }},
'{{ gcp_attributes }}',
'{{ init_scripts }}',
'{{ instance_pool_id }}',
{{ is_single_node }},
'{{ kind }}',
'{{ node_type_id }}',
{{ num_workers }},
'{{ policy_id }}',
{{ remote_disk_throughput }},
'{{ runtime_engine }}',
'{{ single_user_name }}',
'{{ spark_conf }}',
'{{ spark_env_vars }}',
'{{ ssh_public_keys }}',
{{ total_initial_remote_disk_size }},
{{ use_ml_runtime }},
'{{ worker_node_type_flexibility }}',
'{{ workload_type }}',
'{{ deployment_name }}'
RETURNING
cluster_id,
driver_instance_pool_id,
driver_node_type_id,
instance_pool_id,
node_type_id,
policy_id,
spark_context_id,
cluster_name,
creator_user_name,
single_user_name,
autoscale,
autotermination_minutes,
aws_attributes,
azure_attributes,
cluster_cores,
cluster_log_conf,
cluster_log_status,
cluster_memory_mb,
cluster_source,
custom_tags,
data_security_mode,
default_tags,
docker_image,
driver,
driver_node_type_flexibility,
enable_elastic_disk,
enable_local_disk_encryption,
executors,
gcp_attributes,
init_scripts,
is_single_node,
jdbc_port,
kind,
last_restarted_time,
last_state_loss_time,
num_workers,
remote_disk_throughput,
runtime_engine,
spark_conf,
spark_env_vars,
spark_version,
spec,
ssh_public_keys,
start_time,
state,
state_message,
terminated_time,
termination_reason,
total_initial_remote_disk_size,
use_ml_runtime,
worker_node_type_flexibility,
workload_type
;
# Description fields are for documentation purposes
- name: clusters
props:
- name: deployment_name
value: "{{ deployment_name }}"
description: Required parameter for the clusters resource.
- name: spark_version
value: "{{ spark_version }}"
description: |
The Spark version of the cluster, e.g. `3.3.x-scala2.11`. A list of available Spark versions can be retrieved by using the :method:clusters/sparkVersions API call.
- name: apply_policy_default_values
value: {{ apply_policy_default_values }}
description: |
When set to true, fixed and default values from the policy will be used for fields that are omitted. When set to false, only fixed values from the policy will be applied.
- name: autoscale
description: |
Parameters needed in order to automatically scale clusters up and down based on load. Note: autoscaling works best with DB runtime versions 3.0 or later.
value:
max_workers: {{ max_workers }}
min_workers: {{ min_workers }}
- name: autotermination_minutes
value: {{ autotermination_minutes }}
description: |
Automatically terminates the cluster after it is inactive for this time in minutes. If not set, this cluster will not be automatically terminated. If specified, the threshold must be between 10 and 10000 minutes. Users can also set this value to 0 to explicitly disable automatic termination.
- name: aws_attributes
description: |
Attributes related to clusters running on Amazon Web Services. If not specified at cluster creation, a set of default values will be used.
value:
availability: "{{ availability }}"
ebs_volume_count: {{ ebs_volume_count }}
ebs_volume_iops: {{ ebs_volume_iops }}
ebs_volume_size: {{ ebs_volume_size }}
ebs_volume_throughput: {{ ebs_volume_throughput }}
ebs_volume_type: "{{ ebs_volume_type }}"
first_on_demand: {{ first_on_demand }}
instance_profile_arn: "{{ instance_profile_arn }}"
spot_bid_price_percent: {{ spot_bid_price_percent }}
zone_id: "{{ zone_id }}"
- name: azure_attributes
description: |
Attributes related to clusters running on Microsoft Azure. If not specified at cluster creation, a set of default values will be used.
value:
availability: "{{ availability }}"
first_on_demand: {{ first_on_demand }}
log_analytics_info:
log_analytics_primary_key: "{{ log_analytics_primary_key }}"
log_analytics_workspace_id: "{{ log_analytics_workspace_id }}"
spot_bid_max_price: {{ spot_bid_max_price }}
- name: clone_from
description: |
When specified, this clones libraries from a source cluster during the creation of a new cluster.
value:
source_cluster_id: "{{ source_cluster_id }}"
- name: cluster_log_conf
description: |
The configuration for delivering spark logs to a long-term storage destination. Three kinds of destinations (DBFS, S3 and Unity Catalog volumes) are supported. Only one destination can be specified for one cluster. If the conf is given, the logs will be delivered to the destination every `5 mins`. The destination of driver logs is `$destination/$clusterId/driver`, while the destination of executor logs is `$destination/$clusterId/executor`.
value:
dbfs:
destination: "{{ destination }}"
s3:
destination: "{{ destination }}"
canned_acl: "{{ canned_acl }}"
enable_encryption: {{ enable_encryption }}
encryption_type: "{{ encryption_type }}"
endpoint: "{{ endpoint }}"
kms_key: "{{ kms_key }}"
region: "{{ region }}"
volumes:
destination: "{{ destination }}"
- name: cluster_name
value: "{{ cluster_name }}"
description: |
Cluster name requested by the user. This doesn't have to be unique. If not specified at creation, the cluster name will be an empty string. For job clusters, the cluster name is automatically set based on the job and job run IDs.
- name: custom_tags
value: "{{ custom_tags }}"
description: |
Additional tags for cluster resources. Databricks will tag all cluster resources (e.g., AWS instances and EBS volumes) with these tags in addition to `default_tags`. Notes: - Currently, Databricks allows at most 45 custom tags - Clusters can only reuse cloud resources if the resources' tags are a subset of the cluster tags
- name: data_security_mode
value: "{{ data_security_mode }}"
description: |
Data security mode decides what data governance model to use when accessing data from a cluster.
The following modes can only be used when `kind = CLASSIC_PREVIEW`. * `DATA_SECURITY_MODE_AUTO`:
Databricks will choose the most appropriate access mode depending on your compute configuration.
* `DATA_SECURITY_MODE_STANDARD`: Alias for `USER_ISOLATION`. * `DATA_SECURITY_MODE_DEDICATED`:
Alias for `SINGLE_USER`.
The following modes can be used regardless of `kind`. * `NONE`: No security isolation for
multiple users sharing the cluster. Data governance features are not available in this mode. *
`SINGLE_USER`: A secure cluster that can only be exclusively used by a single user specified in
`single_user_name`. Most programming languages, cluster features and data governance features
are available in this mode. * `USER_ISOLATION`: A secure cluster that can be shared by multiple
users. Cluster users are fully isolated so that they cannot see each other's data and
credentials. Most data governance features are supported in this mode. But programming languages
and cluster features might be limited.
The following modes are deprecated starting with Databricks Runtime 15.0 and will be removed for
future Databricks Runtime versions:
* `LEGACY_TABLE_ACL`: This mode is for users migrating from legacy Table ACL clusters. *
`LEGACY_PASSTHROUGH`: This mode is for users migrating from legacy Passthrough on high
concurrency clusters. * `LEGACY_SINGLE_USER`: This mode is for users migrating from legacy
Passthrough on standard clusters. * `LEGACY_SINGLE_USER_STANDARD`: This mode provides a way that
doesn’t have UC nor passthrough enabled.
- name: docker_image
description: |
Custom docker image BYOC
value:
basic_auth:
password: "{{ password }}"
username: "{{ username }}"
url: "{{ url }}"
- name: driver_instance_pool_id
value: "{{ driver_instance_pool_id }}"
description: |
The optional ID of the instance pool for the driver of the cluster belongs. The pool cluster uses the instance pool with id (instance_pool_id) if the driver pool is not assigned.
- name: driver_node_type_flexibility
description: |
Flexible node type configuration for the driver node.
value:
alternate_node_type_ids:
- "{{ alternate_node_type_ids }}"
- name: driver_node_type_id
value: "{{ driver_node_type_id }}"
description: |
The node type of the Spark driver. Note that this field is optional; if unset, the driver node type will be set as the same value as `node_type_id` defined above. This field, along with node_type_id, should not be set if virtual_cluster_size is set. If both driver_node_type_id, node_type_id, and virtual_cluster_size are specified, driver_node_type_id and node_type_id take precedence.
- name: enable_elastic_disk
value: {{ enable_elastic_disk }}
description: |
Autoscaling Local Storage: when enabled, this cluster will dynamically acquire additional disk space when its Spark workers are running low on disk space.
- name: enable_local_disk_encryption
value: {{ enable_local_disk_encryption }}
description: |
Whether to enable LUKS on cluster VMs' local disks
- name: gcp_attributes
description: |
Attributes related to clusters running on Google Cloud Platform. If not specified at cluster creation, a set of default values will be used.
value:
availability: "{{ availability }}"
boot_disk_size: {{ boot_disk_size }}
first_on_demand: {{ first_on_demand }}
google_service_account: "{{ google_service_account }}"
local_ssd_count: {{ local_ssd_count }}
use_preemptible_executors: {{ use_preemptible_executors }}
zone_id: "{{ zone_id }}"
- name: init_scripts
description: |
The configuration for storing init scripts. Any number of destinations can be specified. The scripts are executed sequentially in the order provided. If `cluster_log_conf` is specified, init script logs are sent to `<destination>/<cluster-ID>/init_scripts`.
value:
- abfss:
destination: "{{ destination }}"
dbfs:
destination: "{{ destination }}"
file:
destination: "{{ destination }}"
gcs:
destination: "{{ destination }}"
s3:
destination: "{{ destination }}"
canned_acl: "{{ canned_acl }}"
enable_encryption: {{ enable_encryption }}
encryption_type: "{{ encryption_type }}"
endpoint: "{{ endpoint }}"
kms_key: "{{ kms_key }}"
region: "{{ region }}"
volumes:
destination: "{{ destination }}"
workspace:
destination: "{{ destination }}"
- name: instance_pool_id
value: "{{ instance_pool_id }}"
description: |
The optional ID of the instance pool to which the cluster belongs.
- name: is_single_node
value: {{ is_single_node }}
description: |
This field can only be used when `kind = CLASSIC_PREVIEW`. When set to true, Databricks will automatically set single node related `custom_tags`, `spark_conf`, and `num_workers`
- name: kind
value: "{{ kind }}"
description: |
The kind of compute described by this compute specification.
Depending on `kind`, different validations and default values will be applied.
Clusters with `kind = CLASSIC_PREVIEW` support the following fields, whereas clusters with no
specified `kind` do not. * [is_single_node](/api/workspace/clusters/create#is_single_node) *
[use_ml_runtime](/api/workspace/clusters/create#use_ml_runtime) *
[data_security_mode](/api/workspace/clusters/create#data_security_mode) set to
`DATA_SECURITY_MODE_AUTO`, `DATA_SECURITY_MODE_DEDICATED`, or `DATA_SECURITY_MODE_STANDARD`
By using the [simple form], your clusters are automatically using `kind = CLASSIC_PREVIEW`.
[simple form]: https://docs.databricks.com/compute/simple-form.html
- name: node_type_id
value: "{{ node_type_id }}"
description: |
This field encodes, through a single value, the resources available to each of the Spark nodes in this cluster. For example, the Spark nodes can be provisioned and optimized for memory or compute intensive workloads. A list of available node types can be retrieved by using the :method:clusters/listNodeTypes API call.
- name: num_workers
value: {{ num_workers }}
description: |
Number of worker nodes that this cluster should have. A cluster has one Spark Driver and `num_workers` Executors for a total of `num_workers` + 1 Spark nodes. Note: When reading the properties of a cluster, this field reflects the desired number of workers rather than the actual current number of workers. For instance, if a cluster is resized from 5 to 10 workers, this field will immediately be updated to reflect the target size of 10 workers, whereas the workers listed in `spark_info` will gradually increase from 5 to 10 as the new nodes are provisioned.
- name: policy_id
value: "{{ policy_id }}"
description: |
The ID of the cluster policy used to create the cluster if applicable.
- name: remote_disk_throughput
value: {{ remote_disk_throughput }}
description: |
If set, what the configurable throughput (in Mb/s) for the remote disk is. Currently only supported for GCP HYPERDISK_BALANCED disks.
- name: runtime_engine
value: "{{ runtime_engine }}"
description: |
Determines the cluster's runtime engine, either standard or Photon. This field is not compatible with legacy `spark_version` values that contain `-photon-`. Remove `-photon-` from the `spark_version` and set `runtime_engine` to `PHOTON`. If left unspecified, the runtime engine defaults to standard unless the spark_version contains -photon-, in which case Photon will be used.
- name: single_user_name
value: "{{ single_user_name }}"
description: |
Single user name if data_security_mode is `SINGLE_USER`
- name: spark_conf
value: "{{ spark_conf }}"
description: |
An object containing a set of optional, user-specified Spark configuration key-value pairs. Users can also pass in a string of extra JVM options to the driver and the executors via `spark.driver.extraJavaOptions` and `spark.executor.extraJavaOptions` respectively.
- name: spark_env_vars
value: "{{ spark_env_vars }}"
description: |
An object containing a set of optional, user-specified environment variable key-value pairs. Please note that key-value pair of the form (X,Y) will be exported as is (i.e., `export X='Y'`) while launching the driver and workers. In order to specify an additional set of `SPARK_DAEMON_JAVA_OPTS`, we recommend appending them to `$SPARK_DAEMON_JAVA_OPTS` as shown in the example below. This ensures that all default databricks managed environmental variables are included as well. Example Spark environment variables: `{"SPARK_WORKER_MEMORY": "28000m", "SPARK_LOCAL_DIRS": "/local_disk0"}` or `{"SPARK_DAEMON_JAVA_OPTS": "$SPARK_DAEMON_JAVA_OPTS -Dspark.shuffle.service.enabled=true"}`
- name: ssh_public_keys
value:
- "{{ ssh_public_keys }}"
description: |
SSH public key contents that will be added to each Spark node in this cluster. The corresponding private keys can be used to login with the user name `ubuntu` on port `2200`. Up to 10 keys can be specified.
- name: total_initial_remote_disk_size
value: {{ total_initial_remote_disk_size }}
description: |
If set, what the total initial volume size (in GB) of the remote disks should be. Currently only supported for GCP HYPERDISK_BALANCED disks.
- name: use_ml_runtime
value: {{ use_ml_runtime }}
description: |
This field can only be used when `kind = CLASSIC_PREVIEW`. `effective_spark_version` is determined by `spark_version` (DBR release), this field `use_ml_runtime`, and whether `node_type_id` is gpu node or not.
- name: worker_node_type_flexibility
description: |
Flexible node type configuration for worker nodes.
value:
alternate_node_type_ids:
- "{{ alternate_node_type_ids }}"
- name: workload_type
description: |
Cluster Attributes showing for clusters workload types.
value:
clients:
jobs: {{ jobs }}
notebooks: {{ notebooks }}
Lifecycle Methods
- change_owner
- delete
- edit
- events
- permanent_delete
- pin
- resize
- restart
- start
- unpin
- update
Change the owner of the cluster. You must be an admin and the cluster must be terminated to perform
EXEC databricks_workspace.compute.clusters.change_owner
@deployment_name='{{ deployment_name }}' --required
@@json=
'{
"cluster_id": "{{ cluster_id }}",
"owner_username": "{{ owner_username }}"
}'
;
Terminates the Spark cluster with the specified ID. The cluster is removed asynchronously. Once the
EXEC databricks_workspace.compute.clusters.delete
@deployment_name='{{ deployment_name }}' --required
@@json=
'{
"cluster_id": "{{ cluster_id }}"
}'
;
Updates the configuration of a cluster to match the provided attributes and size. A cluster can be
EXEC databricks_workspace.compute.clusters.edit
@deployment_name='{{ deployment_name }}' --required
@@json=
'{
"cluster_id": "{{ cluster_id }}",
"spark_version": "{{ spark_version }}",
"apply_policy_default_values": {{ apply_policy_default_values }},
"autoscale": "{{ autoscale }}",
"autotermination_minutes": {{ autotermination_minutes }},
"aws_attributes": "{{ aws_attributes }}",
"azure_attributes": "{{ azure_attributes }}",
"cluster_log_conf": "{{ cluster_log_conf }}",
"cluster_name": "{{ cluster_name }}",
"custom_tags": "{{ custom_tags }}",
"data_security_mode": "{{ data_security_mode }}",
"docker_image": "{{ docker_image }}",
"driver_instance_pool_id": "{{ driver_instance_pool_id }}",
"driver_node_type_flexibility": "{{ driver_node_type_flexibility }}",
"driver_node_type_id": "{{ driver_node_type_id }}",
"enable_elastic_disk": {{ enable_elastic_disk }},
"enable_local_disk_encryption": {{ enable_local_disk_encryption }},
"gcp_attributes": "{{ gcp_attributes }}",
"init_scripts": "{{ init_scripts }}",
"instance_pool_id": "{{ instance_pool_id }}",
"is_single_node": {{ is_single_node }},
"kind": "{{ kind }}",
"node_type_id": "{{ node_type_id }}",
"num_workers": {{ num_workers }},
"policy_id": "{{ policy_id }}",
"remote_disk_throughput": {{ remote_disk_throughput }},
"runtime_engine": "{{ runtime_engine }}",
"single_user_name": "{{ single_user_name }}",
"spark_conf": "{{ spark_conf }}",
"spark_env_vars": "{{ spark_env_vars }}",
"ssh_public_keys": "{{ ssh_public_keys }}",
"total_initial_remote_disk_size": {{ total_initial_remote_disk_size }},
"use_ml_runtime": {{ use_ml_runtime }},
"worker_node_type_flexibility": "{{ worker_node_type_flexibility }}",
"workload_type": "{{ workload_type }}"
}'
;
Retrieves a list of events about the activity of a cluster. This API is paginated. If there are more
EXEC databricks_workspace.compute.clusters.events
@deployment_name='{{ deployment_name }}' --required
@@json=
'{
"cluster_id": "{{ cluster_id }}",
"end_time": {{ end_time }},
"event_types": "{{ event_types }}",
"limit": {{ limit }},
"offset": {{ offset }},
"order": "{{ order }}",
"page_size": {{ page_size }},
"page_token": "{{ page_token }}",
"start_time": {{ start_time }}
}'
;
Permanently deletes a Spark cluster. This cluster is terminated and resources are asynchronously
EXEC databricks_workspace.compute.clusters.permanent_delete
@deployment_name='{{ deployment_name }}' --required
@@json=
'{
"cluster_id": "{{ cluster_id }}"
}'
;
Pinning a cluster ensures that the cluster will always be returned by the ListClusters API. Pinning a
EXEC databricks_workspace.compute.clusters.pin
@deployment_name='{{ deployment_name }}' --required
@@json=
'{
"cluster_id": "{{ cluster_id }}"
}'
;
Resizes a cluster to have a desired number of workers. This will fail unless the cluster is in a
EXEC databricks_workspace.compute.clusters.resize
@deployment_name='{{ deployment_name }}' --required
@@json=
'{
"cluster_id": "{{ cluster_id }}",
"autoscale": "{{ autoscale }}",
"num_workers": {{ num_workers }}
}'
;
Restarts a Spark cluster with the supplied ID. If the cluster is not currently in a RUNNING state,
EXEC databricks_workspace.compute.clusters.restart
@deployment_name='{{ deployment_name }}' --required
@@json=
'{
"cluster_id": "{{ cluster_id }}",
"restart_user": "{{ restart_user }}"
}'
;
Starts a terminated Spark cluster with the supplied ID. This works similar to createCluster except:
EXEC databricks_workspace.compute.clusters.start
@deployment_name='{{ deployment_name }}' --required
@@json=
'{
"cluster_id": "{{ cluster_id }}"
}'
;
Unpinning a cluster will allow the cluster to eventually be removed from the ListClusters API.
EXEC databricks_workspace.compute.clusters.unpin
@deployment_name='{{ deployment_name }}' --required
@@json=
'{
"cluster_id": "{{ cluster_id }}"
}'
;
Updates the configuration of a cluster to match the partial set of attributes and size. Denote which
EXEC databricks_workspace.compute.clusters.update
@deployment_name='{{ deployment_name }}' --required
@@json=
'{
"cluster_id": "{{ cluster_id }}",
"update_mask": "{{ update_mask }}",
"cluster": "{{ cluster }}"
}'
;