Data Lake Analytics nodes are supported in DataWorks. You can create a Data Lake Analytics node in the DataWorks console to build an online extract, transform, and load (ETL) process.
- Go to the DataStudio page.
- Log on to the DataWorks console.
- In the left-side navigation pane, click Workspaces.
- In the top navigation bar, select the region where the target workspace resides. Find the target workspace and click Data Analytics in the Actions column.
- On the Data Development tab, move the pointer over and choose .Alternatively, you can click the target workflow, right-click UserDefined, and then choose .
- In the New node dialog box, set the Node name and Destination folder parameters.Note The node name must be 1 to 128 characters in length and can contain letters, digits, underscores (_), and periods (.). It is not case-sensitive.
- Click Submit.
- Configure the Data Lake Analytics node.
If you need to change the resource group on which you test the Data Lake Analytics node on the DataStudio page, click in the toolbar and select your desired exclusive resource group.Note To access a data store in a virtual private cloud (VPC), a node must be run on an exclusive resource group for scheduling. In this example, the data store is in a VPC. You must select an exclusive resource group for scheduling that is connected to the target DLA data store.
- Select a connection.On the configuration tab of the Data Lake Analytics node, select a connection from the Select data source drop-down list. If you cannot find the required connection in the drop-down list, click New data source to add a connection on the Data Source page. For more information, see Supported data stores and plug-ins.
- Write the SQL statements of the node.After you select a connection, write SQL statements based on the syntax that is supported by DLA. You can write data manipulation language (DML) or data definition language (DDL) statements in the code editor.
- Click in the toolbar.
- Click in the toolbar to run the SQL statements you have saved.
- Select a connection.
- Click the Scheduling configuration tab in the right-side navigation pane. On the Scheduling configuration tab, set the
scheduling properties for the node. For more information, see Basic attributes.You must select an exclusive resource group for scheduling that is connected to the target DLA data store to periodically run the node.
- Save and commit the node.Notice You must set the Rerun attribute and Dependent upstream node parameters on the Scheduling configuration tab before you can commit the node.
In a workspace in standard mode, you must click Publish in the upper-right corner after you commit the AnalyticDB for MySQL node. For more information, see Deploy a node.
- Click in the toolbar to save the node.
- Click in the toolbar to commit the node.
- In the Submit New Version dialog box, enter your comments in the Change description field.
- Click OK.
- Test the node. For more information, see Auto triggered nodes.