All Products
Search
Document Center

DataWorks:Configure global Spark parameters

Last Updated:Mar 26, 2026

Global Spark parameters let you set workspace-level Spark properties for EMR tasks across multiple DataWorks modules, and optionally enforce those properties to take precedence over per-module settings.

How it works

DataWorks provides two ways to configure Spark parameters for scheduling nodes:

MethodScopeWhere to configure
Global Spark parameters (this topic)Workspace levelManagement Center > Cluster Management > SPARK Parameters
Per-module Spark parametersIndividual nodeScheduling Configurations panel on the node editing page (Data Development / Data Studio only)

Conflict resolution: When the same Spark parameter is configured in both the DataWorks Management Center and the E-MapReduce (EMR) console, the Management Center configuration takes precedence for tasks submitted from DataWorks.

Priority override: When you enable Global Configuration Has Priority, global Spark properties override any per-module configuration.

Limitations

  • Only the following roles can configure global Spark parameters:

    • An Alibaba Cloud account

    • A Resource Access Management (RAM) user or RAM role with the AliyunDataWorksFullAccess permission

    • A RAM user with the Workspace Administrator role

  • Global Spark parameters apply only to the following node types:

  • Global Spark parameters can only be set for these modules: Data Development (Data Studio), Data Quality, Data Analysis, and Operation Center.

Prerequisites

Before you begin, ensure that you have:

Configure global Spark parameters

  1. Log on to the DataWorks console. In the top navigation bar, select the target region.

  2. In the left-side navigation pane, choose More > Management Center. Select the target workspace from the drop-down list, then click Go to Management Center.

  3. In the left-side navigation pane of the SettingCenter page, click Cluster Management.

  4. Find the target EMR cluster and click SPARK Parameters.

    SPARK Parameters page

  5. Click Edit SPARK Parameters in the upper-right corner.

    Note

    Global Spark parameters apply at the workspace level. Confirm the target workspace before making changes.

  6. Configure the following parameters for each module:

    Note

    To enable Ranger access control for Spark in DataWorks, add spark.hadoop.fs.oss.authorization.method with the value ranger as a Spark property to ensure that Ranger access control is enabled.

    ParameterDescription
    Spark propertyThe Spark properties (Spark Property Name and Spark Property Value) applied when a module runs EMR tasks. For valid property names and values, see Spark Configurations and Spark Configurations on Kubernetes.
    Global Configuration Has PriorityIf selected, global Spark properties override per-module configurations. In this case, tasks are run based on the globally configured Spark properties.
  7. Save the configuration.

What's next