단일 노드 클러스터는 노드가 하나뿐인 Dataproc 클러스터입니다. 이 단일 노드는 Dataproc 클러스터의 마스터 및 작업자의 역할을 합니다. 단일 노드 클러스터에는 노드가 하나뿐이지만 아래 나와 있는 사항을 제외하고 Dataproc의 개념과 기능이 대부분 그대로 적용됩니다.
단일 노드 Dataproc 클러스터를 사용할 수 있는 상황은 다음과 같이 다양합니다.
새로운 버전의 Spark 및 Hadoop 또는 다른 오픈소스의 구성요소 사용
개념 증명(PoC) 데모 구축
간단한 데이터 과학
소규모의 중요하지 않은 데이터 처리
Spark 및 Hadoop 생태계 관련 교육
단일 노드 클러스터의 의미 체계
단일 노드 Dataproc 클러스터에는 다음과 같은 시맨틱스가 적용됩니다.
단일 노드 클러스터는 멀티 노드 Dataproc 클러스터와 동일하게 구성되며 HDFS 및 YARN 같은 서비스가 포함됩니다.
[[["이해하기 쉬움","easyToUnderstand","thumb-up"],["문제가 해결됨","solvedMyProblem","thumb-up"],["기타","otherUp","thumb-up"]],[["이해하기 어려움","hardToUnderstand","thumb-down"],["잘못된 정보 또는 샘플 코드","incorrectInformationOrSampleCode","thumb-down"],["필요한 정보/샘플이 없음","missingTheInformationSamplesINeed","thumb-down"],["번역 문제","translationIssue","thumb-down"],["기타","otherDown","thumb-down"]],["최종 업데이트: 2025-08-27(UTC)"],[[["\u003cp\u003eSingle node Dataproc clusters utilize a single node as both the master and worker, simplifying cluster management for certain use cases.\u003c/p\u003e\n"],["\u003cp\u003eThese clusters are useful for tasks like testing new Spark/Hadoop versions, creating proof-of-concept demos, lightweight data science, small-scale data processing, and educational purposes.\u003c/p\u003e\n"],["\u003cp\u003eWhile most Dataproc features apply, single node clusters lack high-availability and do not support preemptible VMs or scaling to multi-node setups.\u003c/p\u003e\n"],["\u003cp\u003eSingle node clusters can be created using the \u003ccode\u003egcloud\u003c/code\u003e command-line tool with the \u003ccode\u003e--single-node\u003c/code\u003e flag, the Dataproc REST API with specific configurations, or the Google Cloud console by choosing the "Single Node" cluster type.\u003c/p\u003e\n"],["\u003cp\u003eSingle node clusters are not suitable for large-scale data processing, as they are limited by the resources of a single node.\u003c/p\u003e\n"]]],[],null,["Single node clusters are Dataproc clusters with only one node. This single\nnode acts as the master and worker for your\nDataproc cluster. While single\nnode clusters only have one node, most Dataproc concepts and features\nstill apply, except those [listed below](#limitations).\n\nThere are a number of situations where single node Dataproc clusters can\nbe useful, including:\n\n- Trying out new versions of Spark and Hadoop or other open source components\n- Building proof-of-concept (PoC) demonstrations\n- Lightweight data science\n- Small-scale non-critical data processing\n- Education related to the Spark and Hadoop ecosystem\n\nSingle node cluster semantics\n\nThe following semantics apply to single node Dataproc clusters:\n\n- Single node clusters are configured the same as multi node Dataproc clusters, and include services such as HDFS and YARN.\n- Single node clusters report as master nodes for [initialization actions](/dataproc/docs/concepts/configuring-clusters/init-actions).\n- Single node clusters show 0 workers since the single node acts as both master and worker.\n- Single node clusters are given hostnames that follow the pattern `clustername-m`. You can use this hostname to SSH into or connect to a [web UI](/dataproc/docs/concepts/accessing/cluster-web-interfaces) on the node.\n- Single node clusters cannot be upgraded to multi node clusters. Once created, single node clusters are restricted to one node. Similarly, multi node clusters cannot be scaled down to single node clusters.\n\nLimitations\n\n- Single node clusters are not recommended for large-scale parallel data\n processing. If you exceed the resources on a single node cluster, a multi node\n Dataproc cluster is recommended.\n\n- Single node clusters are not available with\n [high-availability](/dataproc/docs/concepts/configuring-clusters/high-availability)\n since there is only one node in the cluster.\n\n- Single node clusters cannot use [preemptible VMs](/dataproc/docs/concepts/compute/preemptible-vms).\n\nCreate a single node cluster \n\ngcloud command\n\n\nYou can create a single node Dataproc cluster using the `gcloud`\ncommand-line tool. To create a single node cluster, pass the\n`--single-node` flag to the\n[`gcloud dataproc clusters create`](/sdk/gcloud/reference/dataproc/clusters/create)\ncommand. \n\n```\ngcloud dataproc clusters create cluster-name \\\n --region=region \\\n --single-node \\\n ... other args\n```\n\n\u003cbr /\u003e\n\nREST API\n\n\nYou can create a single node cluster through the\n[Dataproc REST API](/dataproc/docs/reference/rest) using a\n[clusters.create](/dataproc/docs/reference/rest/v1/projects.regions.clusters/create)\nrequest. When making this request, you must:\n\n1. Add the property `\"dataproc:dataproc.allow.zero.workers\":\"true\"` to the [SoftwareConfig](/dataproc/docs/reference/rest/v1/ClusterConfig#SoftwareConfig) of the cluster request.\n2. Don't submit values for `workerConfig` and `secondaryWorkerConfig` (see [ClusterConfig](/dataproc/docs/reference/rest/v1/ClusterConfig)).\n\n| To examine and construct the JSON body of a Dataproc API clusters create request, open the Dataproc [Create a cluster](https://console.cloud.google.com/dataproc/clustersAdd) page, fill in the applicable fields, then click the **Equivalent REST** button at the bottom of the left panel to view the POST request with the completed JSON request body.\n\n\u003cbr /\u003e\n\nConsole\n\n\nYou can create a single node cluster by selecting \"Single Node\n(1 master, 0 workers)\" on the Cluster type section of\nthe Set up cluster panel on the Dataproc\n[Create a cluster](https://console.cloud.google.com/dataproc/clustersAdd)\npage."]]