Stay organized with collectionsSave and categorize content based on your preferences.
Single node clusters are Dataproc clusters with only one node. This single
node acts as the master and worker for your
Dataproc cluster. While single
node clusters only have one node, most Dataproc concepts and features
still apply, except thoselisted below.
There are a number of situations where single node Dataproc clusters can
be useful, including:
Trying out new versions of Spark and Hadoop or other open source components
Building proof-of-concept (PoC) demonstrations
Lightweight data science
Small-scale non-critical data processing
Education related to the Spark and Hadoop ecosystem
Single node cluster semantics
The following semantics apply to single node Dataproc clusters:
Single node clusters are configured the same as multi node Dataproc
clusters, and include services such as HDFS and YARN.
Single node clusters show 0 workers since the single node acts as
both master and worker.
Single node clusters are given hostnames that follow the patternclustername-m.
You can use this hostname to SSH into or connect to aweb UIon the node.
Single node clusters cannot be upgraded to multi node clusters. Once created,
single node clusters are restricted to one node. Similarly, multi node
clusters cannot be scaled down to single node clusters.
Limitations
Single node clusters are not recommended for large-scale parallel data
processing. If you exceed the resources on a single node cluster, a multi node
Dataproc cluster is recommended.
Single node clusters are not available withhigh-availabilitysince there is only one node in the cluster.
You can create a single node Dataproc cluster using thegcloudcommand-line tool. To create a single node cluster, pass the--single-nodeflag to thegcloud dataproc clusters createcommand.
gcloud dataproc clusters createcluster-name\
--region=region\
--single-node \
... other args
REST API
You can create a single node cluster through theDataproc REST APIusing aclusters.createrequest. When making this request, you must:
Add the property"dataproc:dataproc.allow.zero.workers":"true"to theSoftwareConfigof the cluster request.
You can create a single node cluster by selecting "Single Node
(1 master, 0 workers)" on the Cluster type section of
the Set up cluster panel on the DataprocCreate a clusterpage.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Hard to understand","hardToUnderstand","thumb-down"],["Incorrect information or sample code","incorrectInformationOrSampleCode","thumb-down"],["Missing the information/samples I need","missingTheInformationSamplesINeed","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2025-09-04 UTC."],[[["\u003cp\u003eSingle node Dataproc clusters utilize a single node as both the master and worker, simplifying cluster management for certain use cases.\u003c/p\u003e\n"],["\u003cp\u003eThese clusters are useful for tasks like testing new Spark/Hadoop versions, creating proof-of-concept demos, lightweight data science, small-scale data processing, and educational purposes.\u003c/p\u003e\n"],["\u003cp\u003eWhile most Dataproc features apply, single node clusters lack high-availability and do not support preemptible VMs or scaling to multi-node setups.\u003c/p\u003e\n"],["\u003cp\u003eSingle node clusters can be created using the \u003ccode\u003egcloud\u003c/code\u003e command-line tool with the \u003ccode\u003e--single-node\u003c/code\u003e flag, the Dataproc REST API with specific configurations, or the Google Cloud console by choosing the "Single Node" cluster type.\u003c/p\u003e\n"],["\u003cp\u003eSingle node clusters are not suitable for large-scale data processing, as they are limited by the resources of a single node.\u003c/p\u003e\n"]]],[],null,["Single node clusters are Dataproc clusters with only one node. This single\nnode acts as the master and worker for your\nDataproc cluster. While single\nnode clusters only have one node, most Dataproc concepts and features\nstill apply, except those [listed below](#limitations).\n\nThere are a number of situations where single node Dataproc clusters can\nbe useful, including:\n\n- Trying out new versions of Spark and Hadoop or other open source components\n- Building proof-of-concept (PoC) demonstrations\n- Lightweight data science\n- Small-scale non-critical data processing\n- Education related to the Spark and Hadoop ecosystem\n\nSingle node cluster semantics\n\nThe following semantics apply to single node Dataproc clusters:\n\n- Single node clusters are configured the same as multi node Dataproc clusters, and include services such as HDFS and YARN.\n- Single node clusters report as master nodes for [initialization actions](/dataproc/docs/concepts/configuring-clusters/init-actions).\n- Single node clusters show 0 workers since the single node acts as both master and worker.\n- Single node clusters are given hostnames that follow the pattern `clustername-m`. You can use this hostname to SSH into or connect to a [web UI](/dataproc/docs/concepts/accessing/cluster-web-interfaces) on the node.\n- Single node clusters cannot be upgraded to multi node clusters. Once created, single node clusters are restricted to one node. Similarly, multi node clusters cannot be scaled down to single node clusters.\n\nLimitations\n\n- Single node clusters are not recommended for large-scale parallel data\n processing. If you exceed the resources on a single node cluster, a multi node\n Dataproc cluster is recommended.\n\n- Single node clusters are not available with\n [high-availability](/dataproc/docs/concepts/configuring-clusters/high-availability)\n since there is only one node in the cluster.\n\n- Single node clusters cannot use [preemptible VMs](/dataproc/docs/concepts/compute/preemptible-vms).\n\nCreate a single node cluster \n\ngcloud command\n\n\nYou can create a single node Dataproc cluster using the `gcloud`\ncommand-line tool. To create a single node cluster, pass the\n`--single-node` flag to the\n[`gcloud dataproc clusters create`](/sdk/gcloud/reference/dataproc/clusters/create)\ncommand. \n\n```\ngcloud dataproc clusters create cluster-name \\\n --region=region \\\n --single-node \\\n ... other args\n```\n\n\u003cbr /\u003e\n\nREST API\n\n\nYou can create a single node cluster through the\n[Dataproc REST API](/dataproc/docs/reference/rest) using a\n[clusters.create](/dataproc/docs/reference/rest/v1/projects.regions.clusters/create)\nrequest. When making this request, you must:\n\n1. Add the property `\"dataproc:dataproc.allow.zero.workers\":\"true\"` to the [SoftwareConfig](/dataproc/docs/reference/rest/v1/ClusterConfig#SoftwareConfig) of the cluster request.\n2. Don't submit values for `workerConfig` and `secondaryWorkerConfig` (see [ClusterConfig](/dataproc/docs/reference/rest/v1/ClusterConfig)).\n\n| To examine and construct the JSON body of a Dataproc API clusters create request, open the Dataproc [Create a cluster](https://console.cloud.google.com/dataproc/clustersAdd) page, fill in the applicable fields, then click the **Equivalent REST** button at the bottom of the left panel to view the POST request with the completed JSON request body.\n\n\u003cbr /\u003e\n\nConsole\n\n\nYou can create a single node cluster by selecting \"Single Node\n(1 master, 0 workers)\" on the Cluster type section of\nthe Set up cluster panel on the Dataproc\n[Create a cluster](https://console.cloud.google.com/dataproc/clustersAdd)\npage."]]