Data Lake Analytics nodes are supported in DataWorks. You can create a Data Lake Analytics
node in the DataWorks console to build an online extract, transform, and load (ETL)
process.
Background information
Data Lake Analytics nodes are used to connect to Data Lake Analytics (DLA), an interactive
query and analytics service that is provided by Alibaba Cloud. For more information,
see DLA documentation.
Procedure
- Go to the DataStudio page.
- Log on to the DataWorks console.
- In the left-side navigation pane, click Workspaces.
- In the top navigation bar, select the region where the target workspace resides. Find
the target workspace and click Data Analytics in the Actions column.
- On the Data Development tab, move the pointer over
and choose .Alternatively, you can click the target workflow, right-click UserDefined, and then choose .
- In the New node dialog box, set the Node name and Destination folder parameters.
Note The node name must be 1 to 128 characters in length and can contain letters, digits,
underscores (_), and periods (.). It is not case-sensitive.
- Click Submit.
- Configure the Data Lake Analytics node.
On the configuration tab of the Data Lake Analytics node, select a connection from
the Select data source drop-down list. If you cannot find the required connection
in the drop-down list, click
New data source to add a connection on the
Data Source page. For more information, see
Supported data stores and plug-ins.
- Write the SQL statements of the node.
After you select a connection, write SQL statements based on the syntax that is supported
by DLA. You can write data manipulation language (DML) or data definition language
(DDL) statements in the code editor.
- Click
in the toolbar.
- Click
in the toolbar to run the SQL statements you have saved.
If you need to change the resource group on which you test the Data Lake Analytics
node on the
DataStudio page, click

in the toolbar and select your desired exclusive resource group.
Note To access a data store in a virtual private cloud (VPC), a node must be run on an
exclusive resource group for scheduling. In this example, the data store is in a VPC.
You must select an exclusive resource group for scheduling that is connected to the
target DLA data store.
- Click the Scheduling configuration tab in the right-side navigation pane. On the Scheduling configuration tab, set the
scheduling properties for the node. For more information, see Basic attributes.
You must select an exclusive resource group for scheduling that is connected to the
target DLA data store to periodically run the node.
- Save and commit the node.
Notice You must set the Rerun attribute and Dependent upstream node parameters on the Scheduling configuration tab before you can commit the node.
- Click
in the toolbar to save the node.
- Click
in the toolbar to commit the node.
- In the Submit New Version dialog box, enter your comments in the Change description field.
- Click OK.
In a workspace in standard mode, you must click
Publish in the upper-right corner after you commit the AnalyticDB for MySQL node. For more
information, see
Deploy a node.
- Test the node. For more information, see Auto triggered nodes.