Reference documentation and code samples for the Google Cloud Dataproc V1 Client class ClusterConfig.
The cluster config.
Generated from protobuf message google.cloud.dataproc.v1.ClusterConfig
Namespace
Google \ Cloud \ Dataproc \ V1Methods
__construct
Constructor.
data
array
Optional. Data for populating the Message object.
↳ config_bucket
string
Optional. A Cloud Storage bucket used to stage job dependencies, config files, and job driver console output. If you do not specify a staging bucket, Cloud Dataproc will determine a Cloud Storage location (US, ASIA, or EU) for your cluster's staging bucket according to the Compute Engine zone where your cluster is deployed, and then create and manage this project-level, per-location bucket (see Dataproc staging and temp buckets
). This field requires a Cloud Storage bucket name, not a gs://...
URI to a Cloud Storage bucket.
↳ temp_bucket
string
Optional. A Cloud Storage bucket used to store ephemeral cluster and jobs data, such as Spark and MapReduce history files. If you do not specify a temp bucket, Dataproc will determine a Cloud Storage location (US, ASIA, or EU) for your cluster's temp bucket according to the Compute Engine zone where your cluster is deployed, and then create and manage this project-level, per-location bucket. The default bucket has a TTL of 90 days, but you can use any TTL (or none) if you specify a bucket (see Dataproc staging and temp buckets
). This field requires a Cloud Storage bucket name, not a gs://...
URI to a Cloud Storage bucket.
↳ gce_cluster_config
GceClusterConfig
Optional. The shared Compute Engine config settings for all instances in a cluster.
↳ master_config
↳ worker_config
InstanceGroupConfig
Optional. The Compute Engine config settings for the cluster's worker instances.
↳ secondary_worker_config
InstanceGroupConfig
Optional. The Compute Engine config settings for a cluster's secondary worker instances
↳ software_config
↳ initialization_actions
array< NodeInitializationAction
>
Optional. Commands to execute on each node after config is completed. By default, executables are run on master and all worker nodes. You can test a node's role
metadata to run an executable on a master or worker node, as shown below using curl
(you can also use wget
): ROLE=$(curl -H Metadata-Flavor:Google http://metadata/computeMetadata/v1/instance/attributes/dataproc-role
) if [[ "${ROLE}" == 'Master' ]]; then ... master specific actions ... else ... worker specific actions ... fi
↳ encryption_config
↳ autoscaling_config
AutoscalingConfig
Optional. Autoscaling config for the policy associated with the cluster. Cluster does not autoscale if this field is unset.
↳ security_config
↳ lifecycle_config
↳ endpoint_config
↳ metastore_config
↳ dataproc_metric_config
↳ auxiliary_node_groups
getConfigBucket
Optional. A Cloud Storage bucket used to stage job dependencies, config files, and job driver console output.
If you do not specify a staging bucket, Cloud
Dataproc will determine a Cloud Storage location (US,
ASIA, or EU) for your cluster's staging bucket according to the
Compute Engine zone where your cluster is deployed, and then create
and manage this project-level, per-location bucket (see Dataproc staging and temp
buckets
). This field requires a Cloud Storage bucket name, not a gs://...
URI to
a Cloud Storage bucket.
string
setConfigBucket
Optional. A Cloud Storage bucket used to stage job dependencies, config files, and job driver console output.
If you do not specify a staging bucket, Cloud
Dataproc will determine a Cloud Storage location (US,
ASIA, or EU) for your cluster's staging bucket according to the
Compute Engine zone where your cluster is deployed, and then create
and manage this project-level, per-location bucket (see Dataproc staging and temp
buckets
). This field requires a Cloud Storage bucket name, not a gs://...
URI to
a Cloud Storage bucket.
var
string
$this
getTempBucket
Optional. A Cloud Storage bucket used to store ephemeral cluster and jobs data, such as Spark and MapReduce history files. If you do not specify a temp bucket, Dataproc will determine a Cloud Storage location (US, ASIA, or EU) for your cluster's temp bucket according to the Compute Engine zone where your cluster is deployed, and then create and manage this project-level, per-location bucket. The default bucket has a TTL of 90 days, but you can use any TTL (or none) if you specify a bucket (see Dataproc staging and temp buckets ).
This field requires a Cloud Storage bucket name, not a gs://...
URI to
a Cloud Storage bucket.
string
setTempBucket
Optional. A Cloud Storage bucket used to store ephemeral cluster and jobs data, such as Spark and MapReduce history files. If you do not specify a temp bucket, Dataproc will determine a Cloud Storage location (US, ASIA, or EU) for your cluster's temp bucket according to the Compute Engine zone where your cluster is deployed, and then create and manage this project-level, per-location bucket. The default bucket has a TTL of 90 days, but you can use any TTL (or none) if you specify a bucket (see Dataproc staging and temp buckets ).
This field requires a Cloud Storage bucket name, not a gs://...
URI to
a Cloud Storage bucket.
var
string
$this
getGceClusterConfig
Optional. The shared Compute Engine config settings for all instances in a cluster.
hasGceClusterConfig
clearGceClusterConfig
setGceClusterConfig
Optional. The shared Compute Engine config settings for all instances in a cluster.
$this
getMasterConfig
Optional. The Compute Engine config settings for the cluster's master instance.
hasMasterConfig
clearMasterConfig
setMasterConfig
Optional. The Compute Engine config settings for the cluster's master instance.
$this
getWorkerConfig
Optional. The Compute Engine config settings for the cluster's worker instances.
hasWorkerConfig
clearWorkerConfig
setWorkerConfig
Optional. The Compute Engine config settings for the cluster's worker instances.
$this
getSecondaryWorkerConfig
Optional. The Compute Engine config settings for a cluster's secondary worker instances
hasSecondaryWorkerConfig
clearSecondaryWorkerConfig
setSecondaryWorkerConfig
Optional. The Compute Engine config settings for a cluster's secondary worker instances
$this
getSoftwareConfig
Optional. The config settings for cluster software.
hasSoftwareConfig
clearSoftwareConfig
setSoftwareConfig
Optional. The config settings for cluster software.
$this
getInitializationActions
Optional. Commands to execute on each node after config is completed. By default, executables are run on master and all worker nodes.
You can test a node's role
metadata to run an executable on
a master or worker node, as shown below using curl
(you can also use wget
):
ROLE=$(curl -H Metadata-Flavor:Google http://metadata/computeMetadata/v1/instance/attributes/dataproc-role
)
if [[ "${ROLE}" == 'Master' ]]; then
... master specific actions ...
else
... worker specific actions ...
fi
setInitializationActions
Optional. Commands to execute on each node after config is completed. By default, executables are run on master and all worker nodes.
You can test a node's role
metadata to run an executable on
a master or worker node, as shown below using curl
(you can also use wget
):
ROLE=$(curl -H Metadata-Flavor:Google http://metadata/computeMetadata/v1/instance/attributes/dataproc-role
)
if [[ "${ROLE}" == 'Master' ]]; then
... master specific actions ...
else
... worker specific actions ...
fi
$this
getEncryptionConfig
Optional. Encryption settings for the cluster.
hasEncryptionConfig
clearEncryptionConfig
setEncryptionConfig
Optional. Encryption settings for the cluster.
$this
getAutoscalingConfig
Optional. Autoscaling config for the policy associated with the cluster.
Cluster does not autoscale if this field is unset.
hasAutoscalingConfig
clearAutoscalingConfig
setAutoscalingConfig
Optional. Autoscaling config for the policy associated with the cluster.
Cluster does not autoscale if this field is unset.
$this
getSecurityConfig
Optional. Security settings for the cluster.
hasSecurityConfig
clearSecurityConfig
setSecurityConfig
Optional. Security settings for the cluster.
$this
getLifecycleConfig
Optional. Lifecycle setting for the cluster.
hasLifecycleConfig
clearLifecycleConfig
setLifecycleConfig
Optional. Lifecycle setting for the cluster.
$this
getEndpointConfig
Optional. Port/endpoint configuration for this cluster
hasEndpointConfig
clearEndpointConfig
setEndpointConfig
Optional. Port/endpoint configuration for this cluster
$this
getMetastoreConfig
Optional. Metastore configuration.
hasMetastoreConfig
clearMetastoreConfig
setMetastoreConfig
Optional. Metastore configuration.
$this
getDataprocMetricConfig
Optional. The config for Dataproc metrics.
hasDataprocMetricConfig
clearDataprocMetricConfig
setDataprocMetricConfig
Optional. The config for Dataproc metrics.
$this
getAuxiliaryNodeGroups
Optional. The node group settings.
setAuxiliaryNodeGroups
Optional. The node group settings.
$this