Stay organized with collectionsSave and categorize content based on your preferences.
This page explains the administration capabilities that the
Cloud Data Fusion Studio provides to manage configurations.
Cloud Data Fusion follows a hierarchy, where each instance can have
multiplenamespaces. From the Cloud Data Fusion Studio, administrators
can centrally manage all namespaces, or manage them individually.
Cloud Data Fusion Studio provides the following system and namespace
controls.
System administration
ClickingSystem adminin the Cloud Data Fusion Studio displays the
following tabs:
TheManagementtab: view the health status of various
Cloud Data Fusion services. You can also view logs for each of the
services.
TheConfigurationtab: create, view, and edit the following controls:
System compute profiles.Compute profilesindicate which
provisioner to use when creating a cluster for pipeline execution and
applying the associated configurations.
Provisionersare responsible for creating, initializing, and destroying the
cloud environment that pipelines run in. Each provisioner exposes a set of
configurations that are used to control what type of cluster is created
and deleted. Different provisioners create different types of clusters.
Each compute profile has a scope: system or user. You can use a system compute
profile for any namespaces under it. User compute profiles exist within a
namespace, and only pipelines in that namespace can use the user compute
profiles.
On the system administratorConfigurationstab, you can create a system
compute profile that's applied to all the namespaces. Cloud Data Fusion
assigns a default compute profile.
When you create a compute profile, you select the provisioner, which the profile
uses to create and configure the cloud runtime details.
System preferences
Preferences are predefined configurations that apply at various levels within
Cloud Data Fusion, including the system itself, namespaces, applications
(which contain pipelines), and individual programs within pipelines. Preferences
provide a way to set default values for commonly used configurations. These
defaults can be inherited by pipelines and programs at lower levels, reducing
repetitive configuration tasks. For more information, seeManage macros,
preferences, and runtime arguments.
HTTP call action
The HTTP call action on theSystem adminpage lets you interact with
Cloud Data Fusion's own API, or potentially other Google Cloud service
APIs, directly from the Cloud Data Fusion Studio interface. However, for
building data processing pipelines with external data sources, instead use the
HTTP plugin and its HTTP call Executor within your pipelines for a more
comprehensive solution. It differs slightly from the HTTP call action, but the
underlying concepts are alike.
Configurations and use cases
The HTTP call action is primarily used for administrative tasks or
configuration purposes within Cloud Data Fusion. It lets you interact
with the Cloud Data Fusion API or other Google Cloud services
that expose an HTTP API, directly from the Cloud Data Fusion
Studio.
Configurations
You can define the following details for an HTTP call:
URL: the target endpoint of the web service you want to
call.
Method: the HTTP method to use, such asGET,POST, orPUT.
Optional:Headers: any custom headers required for the
request.
Optional:Body: data to be sent in the request body, such as
forPOSTandPUTcalls.
You can then execute the defined HTTP call and view the response from the
web service within the Cloud Data Fusion Studio.
Use cases
Test Cloud Data Fusion API calls.You can use the HTTP call
action to test or explore Cloud Data Fusion API functionalities
directly from the web interface. This action can be helpful for
understanding API behavior or troubleshooting potential issues.
Manage namespaces (advanced).While there's a dedicated UI for
namespace management, the HTTP call action can be used for advanced tasks
by directly making calls to the Cloud Data Fusion API for namespace
creation, deletion, or configuration.
Interacting with other Google Cloud services (limited).If other Google Cloud services you use have a publicly documented
HTTP API, you can use the HTTP call action to interact with those
services, though this is an uncommon use case.
Important considerations
Security: be cautious when using the HTTP call action,
especially with sensitive data or Cloud Data Fusion API calls that
could impact your environment. Be sure that you understand the
implications of each API call before executing it.
Limitations: the HTTP call action is primarily for
administrative tasks and testing purposes. It's not designed for building
complex data processing pipelines that involve data manipulation within
Cloud Data Fusion.
Alternative for pipelines: for integrating external data sources
or services into your data pipelines, use the HTTP plugin and its
associated HTTP call executor within your pipeline definitions. This
provides a more robust and controlled way to manage HTTP interactions
within your data processing workflows.
Namespace administration
ClickingNamespace adminin the Cloud Data Fusion Studio lets you
manage the configurations for the specific namespace. For each namespace, you
can define the following aspects:
Compute profiles: the profiles set up inNamespace adminareusercompute profiles. Only pipelines in that namespace can use these user
compute profiles. For more information, seeManage compute profiles.
Preferences: preferences defined at namespace level are applicable to
the namespace, applications (which contain pipelines), and individual
programs within pipelines. For more information, seeManage macros,
preferences, and runtime arguments.
Connections: Cloud Data Fusion lets you reuse connections to
sources and sinks in data pipelines. You can add connections in the
Namespace Admin page. For more information, seeCreate and manage connections.
Drivers: some plugins in Cloud Data Fusion require a JDBC driver to
be added to the namespace. For example, before you can run a pipeline with a
MySQL batch source plugin, you must add the supported MySQL driver to the
namespace. You can upload or remove JDBC drivers to a namespace from the
Namespace Admin page, or directly from the Hub. For more information, seePlugin drivers.
Source Control Management: to efficiently manage the development process
of deployed pipelines, Source Control Management lets you connect a
namespace with the repository of your source control system. For more
information, seeManage pipelines using Source Control Management.
Service account: to control access to Google Cloud resources,
namespaces in Cloud Data Fusion use theCloud Data Fusion API Service Agentby default.
For better data isolation, you can associate a customized
Identity and Access Management (IAM) service account (known as a Per Namespace
Service Account) with each namespace. The customized IAM service
account, which can be different for different namespaces, lets you control
access to Google Cloud resources between namespaces for pipeline
design-time operations in Cloud Data Fusion, such as pipeline preview,
Wrangler, and pipeline validation. For more information, seeAccess control with namespace service account.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Hard to understand","hardToUnderstand","thumb-down"],["Incorrect information or sample code","incorrectInformationOrSampleCode","thumb-down"],["Missing the information/samples I need","missingTheInformationSamplesINeed","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2025-09-04 UTC."],[[["\u003cp\u003eCloud Data Fusion Studio enables administrators to manage configurations across multiple namespaces, either centrally or individually, offering both system and namespace-specific controls.\u003c/p\u003e\n"],["\u003cp\u003eSystem administrators can manage compute profiles, define system-wide preferences, and view the health status of Cloud Data Fusion services through the Management and Configuration tabs.\u003c/p\u003e\n"],["\u003cp\u003eThe HTTP call action allows for direct interaction with Cloud Data Fusion's API and other Google Cloud service APIs for administrative tasks, but the HTTP plugin is recommended for building pipelines with external data sources.\u003c/p\u003e\n"],["\u003cp\u003eNamespace administrators can configure user compute profiles, preferences, connections, drivers, source control management, and service accounts, each specific to the namespace.\u003c/p\u003e\n"],["\u003cp\u003eCompute profiles determine which provisioner is used to create the cloud environment for pipelines, with system profiles applying to all namespaces and user profiles confined to a single namespace.\u003c/p\u003e\n"]]],[],null,["# Manage Studio administration\n\nThis page explains the administration capabilities that the Cloud Data Fusion Studio provides to manage configurations. Cloud Data Fusion follows a hierarchy, where each instance can have multiple [namespaces](/data-fusion/docs/how-to/control-access-in-namespace). From the Cloud Data Fusion Studio, administrators can centrally manage all namespaces, or manage them individually. Cloud Data Fusion Studio provides the following system and namespace controls.\n\n\u003cbr /\u003e\n\nSystem administration\n---------------------\n\nClicking **System admin** in the Cloud Data Fusion Studio displays the\nfollowing tabs:\n\n- The **Management** tab: view the health status of various Cloud Data Fusion services. You can also view logs for each of the services.\n- The **Configuration** tab: create, view, and edit the following controls:\n\n - **Namespaces** . For more information, see [Access control with a namespace service account](/data-fusion/docs/how-to/control-access-in-namespace).\n - **System compute profiles** . [Compute profiles](/data-fusion/docs/concepts/overview#concepts) indicate which provisioner to use when creating a cluster for pipeline execution and applying the associated configurations.\n\n*Provisioners* are responsible for creating, initializing, and destroying the\ncloud environment that pipelines run in. Each provisioner exposes a set of\nconfigurations that are used to control what type of cluster is created\nand deleted. Different provisioners create different types of clusters.\n| **Note:** The Dataproc provisioner is recommended.\n\nEach compute profile has a scope: system or user. You can use a system compute\nprofile for any namespaces under it. User compute profiles exist within a\nnamespace, and only pipelines in that namespace can use the user compute\nprofiles.\n\nOn the system administrator **Configurations** tab, you can create a system\ncompute profile that's applied to all the namespaces. Cloud Data Fusion\nassigns a default compute profile.\n\nWhen you create a compute profile, you select the provisioner, which the profile\nuses to create and configure the cloud runtime details.\n| **Note:** In general, provisioner configurations can be overridden at runtime. For more information, see [Manage compute profiles](/data-fusion/docs/how-to/manage-compute-profiles).\n\nSystem preferences\n------------------\n\nPreferences are predefined configurations that apply at various levels within\nCloud Data Fusion, including the system itself, namespaces, applications\n(which contain pipelines), and individual programs within pipelines. Preferences\nprovide a way to set default values for commonly used configurations. These\ndefaults can be inherited by pipelines and programs at lower levels, reducing\nrepetitive configuration tasks. For more information, see [Manage macros,\npreferences, and runtime arguments](/data-fusion/docs/how-to/manage-macros-prefs-and-runtime-args).\n\n### HTTP call action\n\nThe HTTP call action on the **System admin** page lets you interact with\nCloud Data Fusion's own API, or potentially other Google Cloud service\nAPIs, directly from the Cloud Data Fusion Studio interface. However, for\nbuilding data processing pipelines with external data sources, instead use the\nHTTP plugin and its HTTP call Executor within your pipelines for a more\ncomprehensive solution. It differs slightly from the HTTP call action, but the\nunderlying concepts are alike. \n\n#### Configurations and use cases\n\nThe HTTP call action is primarily used for administrative tasks or\nconfiguration purposes within Cloud Data Fusion. It lets you interact\nwith the Cloud Data Fusion API or other Google Cloud services\nthat expose an HTTP API, directly from the Cloud Data Fusion\nStudio.\n\n##### Configurations\n\nYou can define the following details for an HTTP call:\n\n- **URL**: the target endpoint of the web service you want to call.\n- **Method** : the HTTP method to use, such as `GET`, `POST`, or `PUT`.\n- Optional: **Headers**: any custom headers required for the request.\n- Optional: **Body** : data to be sent in the request body, such as for `POST` and `PUT` calls.\n\nYou can then execute the defined HTTP call and view the response from the\nweb service within the Cloud Data Fusion Studio.\n\n##### Use cases\n\n- **Test Cloud Data Fusion API calls.** You can use the HTTP call action to test or explore Cloud Data Fusion API functionalities directly from the web interface. This action can be helpful for understanding API behavior or troubleshooting potential issues.\n- **Manage namespaces (advanced).** While there's a dedicated UI for namespace management, the HTTP call action can be used for advanced tasks by directly making calls to the Cloud Data Fusion API for namespace creation, deletion, or configuration.\n- **Interacting with other Google Cloud services (limited).** If other Google Cloud services you use have a publicly documented HTTP API, you can use the HTTP call action to interact with those services, though this is an uncommon use case.\n\n##### Important considerations\n\n- **Security**: be cautious when using the HTTP call action, especially with sensitive data or Cloud Data Fusion API calls that could impact your environment. Be sure that you understand the implications of each API call before executing it.\n- **Limitations**: the HTTP call action is primarily for administrative tasks and testing purposes. It's not designed for building complex data processing pipelines that involve data manipulation within Cloud Data Fusion.\n- **Alternative for pipelines**: for integrating external data sources or services into your data pipelines, use the HTTP plugin and its associated HTTP call executor within your pipeline definitions. This provides a more robust and controlled way to manage HTTP interactions within your data processing workflows.\n\nNamespace administration\n------------------------\n\nClicking **Namespace admin** in the Cloud Data Fusion Studio lets you\nmanage the configurations for the specific namespace. For each namespace, you\ncan define the following aspects:\n\n- **Compute profiles** : the profiles set up in **Namespace admin** are *user* compute profiles. Only pipelines in that namespace can use these user compute profiles. For more information, see [Manage compute profiles](/data-fusion/docs/how-to/manage-compute-profiles).\n- **Preferences** : preferences defined at namespace level are applicable to the namespace, applications (which contain pipelines), and individual programs within pipelines. For more information, see [Manage macros,\n preferences, and runtime arguments](/data-fusion/docs/how-to/manage-macros-prefs-and-runtime-args).\n- **Connections** : Cloud Data Fusion lets you reuse connections to sources and sinks in data pipelines. You can add connections in the Namespace Admin page. For more information, see [Create and manage connections](/data-fusion/docs/how-to/managing-connections).\n- **Drivers** : some plugins in Cloud Data Fusion require a JDBC driver to be added to the namespace. For example, before you can run a pipeline with a MySQL batch source plugin, you must add the supported MySQL driver to the namespace. You can upload or remove JDBC drivers to a namespace from the Namespace Admin page, or directly from the Hub. For more information, see [Plugin drivers](/data-fusion/docs/how-to/source-control-management).\n- **Source Control Management** : to efficiently manage the development process of deployed pipelines, Source Control Management lets you connect a namespace with the repository of your source control system. For more information, see [Manage pipelines using Source Control Management](/data-fusion/docs/how-to/source-control-management).\n- **Service account** : to control access to Google Cloud resources, namespaces in Cloud Data Fusion use the [Cloud Data Fusion API Service Agent](/iam/docs/understanding-roles#datafusion.serviceAgent) by default.\n\nFor better data isolation, you can associate a customized\nIdentity and Access Management (IAM) service account (known as a Per Namespace\nService Account) with each namespace. The customized IAM service\naccount, which can be different for different namespaces, lets you control\naccess to Google Cloud resources between namespaces for pipeline\ndesign-time operations in Cloud Data Fusion, such as pipeline preview,\nWrangler, and pipeline validation. For more information, see\n[Access control with namespace service account](/data-fusion/docs/how-to/control-access-in-namespace).\n\nWhat's next\n-----------\n\n- Learn more about [compute profiles](/data-fusion/docs/how-to/manage-compute-profiles).\n- Learn more about [macros, preferences, and runtime arguments](/data-fusion/docs/how-to/manage-macros-prefs-and-runtime-args)."]]