This topic describes the FAQ about Data Analytics.

What can I do if DataWorks does not run a node as scheduled and the operational logs indicate that the node is waiting for slot or gateway resources?

DataWorks offers you a certain scheduling limit for free. If the number of concurrent nodes reaches the limit, DataWorks cannot run more nodes until it completes a running node.

We recommend that you specify different scheduling time for nodes to maximize the utilization of your computing resources. If you need to extend the scheduling limit, that includes the following content for assessment:
  1. The business scenarios
  2. The expected maximum number of nodes to be concurrently run during peak hours
  3. The reason why you cannot change the scheduling time of the nodes to avoid high concurrency
Note If you encounter this issue, purchase an exclusive resource group to run nodes. For more information, see DataWorks exclusive resources.
To guarantee smooth running of Data Integration nodes, you must purchase both of the following exclusive resource groups:
  • Exclusive resource group for scheduling: Nodes that are run on such a resource group do not need to wait for gateway resources.
  • Exclusive resource group for Data Integration: Sync nodes that are run on such a resource group can use the resource group exclusively.

How do I add fields to a table in the production environment?

If you are using an Alibaba Cloud account, add fields to the table on the Workspace Tables tab of the DataStudio page and then commit the table to the production environment.

If you are a Resource Access Management (RAM) user, apply for the Administrator Expert or Space Administrator role, add fields to the table on the Workspace Tables tab of the DataStudio page, and then commit the table to the production environment.

How do I use a MaxCompute table in DataWorks?

Currently, DataWorks does not provide any graphical user interfaces (GUIs) for you to upload a MaxCompute table. To use a MaxCompute table in DataWorks, follow these steps:
  1. On the DataStudio page of DataWorks, create a file resource with the same name as the MaxCompute table and upload the file. In this example, upload the userlog3.txt file.
    Note Do not select Upload to MaxCompute.
  2. Add the MaxCompute table on the MaxCompute client. In this example, run the following command on the MaxCompute client:
    add table userlog3 -f;
  3. To use the MaxCompute table in DataWorks, directly specify the name of the uploaded file.

How do I grant access to the DataWorks console only to users on the internal network?

If you want to grant access to the DataWorks console only to users on the internal network of the enterprise, follow these steps: Log on to the RAM console and configure a security policy to only allow access from the public IP addresses mapping the private IP addresses of the enterprise.

For more information, see Set security policies for RAM users.

Can a Python resource call another Python resource?

A Python resource can call another Python resource in the same workspace.

What causes the failure to find the workspace when I call an API operation?

Symptom: When you call the createManualDagTest operation, an error message is reported indicating that the workspace does not exist.

Analysis: In versions earlier than DataWorks V2.0, the workspace name used to call an API operation must be in lowercase. If the workspace name that you specify to call an API operation contains uppercase letters, the code converts the uppercase letters to lowercase letters. As a result, the target workspace cannot be found.

In DataWorks V2.0 and later versions, the code does not interchange uppercase letters and lowercase letters in workspace names.

This issue does not occur if the workspace name is in lowercase or the DataWorks version is V2.0 or later.

How do I configure scheduling properties for an ODPS Spark node?

After you create an ODPS Spark node, click Properties in the right-side navigation pane of the node configuration tab. In the Properties pane that appears, configure scheduling properties of the node in the Schedule section.

Then, set other required parameters of the ODPS Spark node on the node configuration tab. DataWorks runs the node based on your configuration.