All Products
Search
Document Center

Data Management:Create a workspace and import resources

Last Updated:Mar 28, 2026

Set up a Notebook workspace in DMS to query and analyze data with Spark or Python. This guide walks you through creating a workspace, configuring storage, adding a compute resource, and accessing your data.

Prerequisites

Before you begin, ensure that you have:

  • An Object Storage Service (OSS) bucket in the Singapore region with the Standard storage class. If you don't have one, create a bucket first.

  • An AnalyticDB for MySQL cluster registered in DMS (required only if you plan to use Spark images). If the cluster isn't registered, see Register an Alibaba Cloud database instance.

  • A virtual private cloud (VPC) with a configured vSwitch and security group.

Step 1: Create a workspace

  1. Log on to the DMS console V5.0.

  2. Click the 2023-01-28_15-57-17.png icon in the upper-left corner and choose All Features > Data Development > Notebook.

    Note

    If you are not using DMS in simple mode, choose Data Development > Notebook in the top navigation bar.

  3. Click Create Workspace. In the Create Workspace dialog box, set Workspace Name and Region, then click OK.

    Note
    • The workspace name can contain letters, digits, and underscores (_).

    • You can select only the Singapore region for the workspace.

  4. In the workspace list, click Go to Workspace in the Actions column.

    Note

    Only the workspace creator can access a workspace by default. To enable collaborative development, grant development permissions to the users who need access.

Step 2: Add workspace members (optional)

Skip this step if you're working alone.

If multiple users need access, assign each user a role. Users must be added to DMS before you can assign roles. For details, see Manage users.

image

Step 3: Configure an OSS bucket for code storage

  1. In your workspace, click Storage Management on the image tab.

  2. Click the image icon next to Code Storage Space.

  3. In the Select OSS Directory dialog box, select a bucket. The bucket must be in the same region as your workspace and use the Standard storage class.

  4. Click OK.

Step 4: Add a compute resource

Add and start a compute resource before using Notebook to query and analyze data.

Resources are automatically released when all notebook kernels are exited and the maximum idle period elapses. Size your resource based on your workload to avoid unnecessary costs.
  1. On the image tab, click Resource Configuration.

  2. Click Add Resource and configure the following parameters.

    ParameterDescription
    Resource NameA name for the resource.
    Resource IntroductionA description of the resource.
    ImageThe runtime environment. Options: Spark 3.5+Python 3.9, Spark 3.3+Python 3.9, Python 3.9. Choose a Spark image if you need distributed processing; choose Python 3.9 for lightweight data exploration.
    AnalyticDB InstanceThe AnalyticDB for MySQL cluster to use. Required when Image is set to Spark 3.3 or Spark 3.5.
    AnalyticDB Resource GroupThe resource group within the AnalyticDB for MySQL cluster.
    Executor SpecThe resource specifications for each Spark executor. For reference, see the Type column in Spark application configuration parameters.
    Executor CountThe number of Spark executors. During public preview, the maximum is 6 executors per notebook. Contact DMS technical support if you need more.
    Driver SpecificationsThe resource specifications for the Spark driver. Valid values: General_XSmall_v1 (2 CPU cores, 8 GB memory), General_Small_v1 (4 CPU cores, 16 GB memory), General_Medium_v1 (8 CPU cores, 32 GB memory), General_Large_v1 (16 CPU cores, 64 GB memory). For lightweight workloads, General_XSmall_v1 is sufficient. For large Spark jobs, consider General_Medium_v1 or above.
    NotebookQuantityThe compute specifications for a Python 3.9 notebook. Appears only when Image is set to Python 3.9. Same valid values as Driver Specifications.
    VPC IDThe VPC in which the resource runs.
    Zone IDThe zone of the VPC.
    vSwitch IDThe vSwitch in the VPC.
    Security Group IDThe security group for the resource.
  3. Click Save.

  4. Find the resource in the list, click Start in the Action column, and then click OK. The resource enters the Running state after about 1 minute.

Step 5: Configure an OSS bucket for user data (optional)

Skip this step if you only need to access data already in DMS.

To read data from external OSS buckets in your notebook, configure one or more OSS paths in the User Storage Space area.

  1. In your workspace, click Storage Management on the image tab.

  2. In the User Storage Space area, configure an OSS path.

    Note

    Mount paths must start with /mnt/.

    image

  3. Click the image icon to save the path.

Step 6: View data

  1. In your workspace, click the image tab to open SQL Console.

  2. Use SQL Console to query data and explore table definitions:

    • Query data: Enter SQL statements directly, or use Copilot to generate them. SQL syntax follows the logical data warehouse standard, which supports MySQL syntax across multiple databases (such as AnalyticDB for MySQL and ApsaraDB RDS for MySQL). DMS automatically converts and optimizes your statements. When you use Copilot, it generates SQL based on database, table, and column metadata. If the generated results are inaccurate, edit the knowledge that Copilot uses to improve future accuracy. You can also like accurate results to further improve quality over time.

    • View table usage notes: Click a database to expand it, then double-click a table name to open the table details page. View or edit the table description on the Usage Notes tab.

Manage resources

On the Resource Configuration page, you can manage resources at any time. For more information about how to navigate to this page, see Step 4: Add a compute resource.

OperationDescription
StopManually stop a running resource.
EditEdit the resource configuration. The resource must be stopped before you can edit it.
StartManually start a stopped resource.
Auto-releaseResources are automatically released when all notebook kernels exit and the maximum idle period elapses.
View Spark jobsClick SparkUI in the Action column to open the History Server page, then click an application ID to view its Spark jobs. SparkUI is available only when the default resource is used and the resource contains a Spark image.

What's next

Use Notebook to query and analyze data