This topic describes how to import tasks that are exported from open source engines into DataWorks.

Procedure

  1. Go to the Open Source engine import page.
    1. Log on to the DataWorks console.
    2. In the left-side navigation pane, click Workspaces.
    3. After you select the region where the required workspace resides, find the desired workspace and click Data Analytics in the Actions column.
    4. Click the Icon icon in the upper-left corner and choose All Products > Other > Migration Assistant.
    5. In the left-side navigation pane, choose Cloud tasks > Scheduling Engine Import.
  2. Create an import task.
    1. On the Import Tasks page, click Create Import Task in the upper-right corner.
    2. In the Create Import Task dialog box, configure the parameters.
      Create Import Task
      Parameter Description
      Name The name of the import task. The name can contain uppercase and lowercase letters, digits, underscores (_), and periods (.).
      Engine type The type of the open source engine. Valid values: Apache Airflow, Azkaban, Cai Yun Jian, Data Go, Oozie, Standard format, and E-MapReduce (EMR).
      Upload From The source of the package that you want to import. Valid values: Local and OSS.
      • If you select Local for this parameter, perform the following steps to upload a package from your on-premises machine:
        1. Click Upload File.
        2. In the dialog box that appears, select the package you want to upload and click Open.
        3. Click Check.
        4. When the message The resource package has passed the check appears, verify that the file format and content are correct.
        Note The size of the package that you upload cannot exceed 30 MB. If the size of the package exceeds 30 MB, select OSS for this parameter.
      • If you select OSS for this parameter, enter the endpoint of an OSS object in the OSS Endpoint field. Then, click Check and Preview in sequence to check and preview the package that you want to upload.
      Remarks The description of the import task.
    3. Click OK. The Edit import task page appears.
  3. Edit the import task.
    1. On the Edit import task page, specify Import objects.
      Periodic tasks is selected for Import objects by default. If you want to import data objects of another type, select the required value from the Import objects drop-down list. Import objects
    2. Optional:Click Advanced Settings. In the Advanced Settings dialog box, configure the mappings between the compute engine instances and the node types. Then, click OK.
      If multiple compute engine instances are bound to the destination workspace, you must complete the settings in the Advanced Settings dialog box. You can configure the mappings between compute engine instances and nodes of the Shell, Hive, and Sqoop types.
    3. On the Edit import task page, click start import in the upper-right corner.
  4. View the import report.
    1. In the Import progress dialog box, view the task import progress.
    2. After the task import is complete, click Return to import task list.
    3. Find the task on the Import Tasks page and click View Import Report in the Actions column. On the page that appears, view the task information in the Basic Information, Import results, Details, and Import Settings sections.