Skip to content
This repository has been archived by the owner on Jun 13, 2023. It is now read-only.

Latest commit

 

History

History
49 lines (41 loc) · 5.63 KB

NewTaskCluster.md

File metadata and controls

49 lines (41 loc) · 5.63 KB

NewTaskCluster

If new_cluster, a description of a cluster that is created only for this task.

Properties

Name Type Description Notes
num_workers int If num_workers, number of worker nodes that this cluster must have. A cluster has one Spark driver and num_workers executors for a total of num_workers + 1 Spark nodes. When reading the properties of a cluster, this field reflects the desired number of workers rather than the actual current number of workers. For example, if a cluster is resized from 5 to 10 workers, this field immediately updates to reflect the target size of 10 workers, whereas the workers listed in `spark_info` gradually increase from 5 to 10 as the new nodes are provisioned. [optional]
autoscale AutoScale [optional]
spark_version str The Spark version of the cluster. A list of available Spark versions can be retrieved by using the Runtime versions API call.
spark_conf Dict[str, object] An arbitrary object where the object key is a configuration propery name and the value is a configuration property value. [optional]
aws_attributes AwsAttributes [optional]
node_type_id str This field encodes, through a single value, the resources available to each of the Spark nodes in this cluster. For example, the Spark nodes can be provisioned and optimized for memory or compute intensive workloads A list of available node types can be retrieved by using the List node types API call. [optional]
driver_node_type_id str The node type of the Spark driver. This field is optional; if unset, the driver node type is set as the same value as `node_type_id` defined above. [optional]
ssh_public_keys List[str] Set to empty array. Cluster SSH is not supported. [optional]
custom_tags Dict[str, str] An object with key value pairs. The key length must be between 1 and 127 UTF-8 characters, inclusive. The value length must be less than or equal to 255 UTF-8 characters. [optional]
cluster_log_conf ClusterLogConf [optional]
init_scripts List[InitScriptInfo] The configuration for storing init scripts. Any number of scripts can be specified. The scripts are executed sequentially in the order provided. If `cluster_log_conf` is specified, init script logs are sent to `<destination>/<cluster-id>/init_scripts`. [optional]
spark_env_vars Dict[str, object] An arbitrary object where the object key is an environment variable name and the value is an environment variable value. [optional]
enable_elastic_disk bool Autoscaling Local Storage: when enabled, this cluster dynamically acquires additional disk space when its Spark workers are running low on disk space. Refer to Autoscaling local storage for details. [optional]
driver_instance_pool_id str The optional ID of the instance pool to use for the driver node. You must also specify `instance_pool_id`. Refer to Instance Pools API for details. [optional]
instance_pool_id str The optional ID of the instance pool to use for cluster nodes. If `driver_instance_pool_id` is present, `instance_pool_id` is used for worker nodes only. Otherwise, it is used for both the driver node and worker nodes. Refer to Instance Pools API for details. [optional]
policy_id str A cluster policy ID. Either `node_type_id` or `instance_pool_id` must be specified in the cluster policy if they are not specified in this job cluster object. [optional]
enable_local_disk_encryption bool Determines whether encryption of disks locally attached to the cluster is enabled. [optional]
docker_image DockerImage [optional]
runtime_engine str The type of runtime engine to use. If not specified, the runtime engine type is inferred based on the `spark_version` value. Allowed values include: * `PHOTON`: Use the Photon runtime engine type. * `STANDARD`: Use the standard runtime engine type. This field is optional. [optional]
gcp_attributes GcpAttributes [optional]
azure_attributes AzureAttributes [optional]

Example

from databricks_jobs.models.new_task_cluster import NewTaskCluster

# TODO update the JSON string below
json = "{}"
# create an instance of NewTaskCluster from a JSON string
new_task_cluster_instance = NewTaskCluster.from_json(json)
# print the JSON string representation of the object
print NewTaskCluster.to_json()

# convert the object into a dict
new_task_cluster_dict = new_task_cluster_instance.to_dict()
# create an instance of NewTaskCluster from a dict
new_task_cluster_form_dict = new_task_cluster.from_dict(new_task_cluster_dict)

[Back to Model list] [Back to API list] [Back to README]