All Products
Search
Document Center

Platform For AI:DLC billing

Last Updated:Mar 02, 2026

This topic describes the billable items, billing methods, and pricing for Deep Learning Containers (DLC).

Important

The pricing in this topic is for reference only. Actual prices are shown on your bills.

Billable items

The following figures show the billable items:

fd0d7ba16722bbf09f49d5a33f9bdcf9

Billing methods

Select a billing method based on your business needs:

Billing method

Billable item

Billing entity

Billing rule

Stop billing

Pay-as-you-go

Public resources

Runtime of DLC jobs (duration of public resource usage)

Charged based on the runtime of DLC jobs that use public resources.

  • The DLC job completes.

  • The DLC job is stopped.

Subscription

AI computing resources (general computing resources and Lingjun resources)

For more information, see Billing of AI computing resources.

For more information, see Billing of AI computing resources.

Not applicable

Public resources

Pay-as-you-go

With pay-as-you-go, you are charged based on the runtime of DLC jobs that use public resources.

Resource type

Billing formula

Unit price

Billing duration

Scaling

Notes

General computing public resources

Bill amount = Number of nodes × (Unit price / 60) × Runtime (minutes)

Prices vary by region. To view pricing, go to the DLC create job page. In the Resource Information section, set Resource Source to Public Resources and select a specification. For more information, see Create a job using the console.image

Based on DLC job runtime.

Not applicable

None

AI computing resources

Subscription

AI computing resources include general computing resources and Lingjun resources. With subscription, you purchase AI computing resources upfront and use the resource quota to submit DLC jobs. For more information, see Billing of AI computing resources.

Billing examples

Important

The following billing examples are for reference only. Actual fees are shown on the console or buy page.

Public resources example

  • Scenario:

    You create a training job in the China (Shanghai) region using a general computing public resource with the ecs.g6.2xlarge specification. The job uses one node and runs for 1 minute and 15 seconds.

  • Calculation:

    Bill amount = 1 × 0.6 / 60 × 1.25 = 0.0125 USD

AI computing resources example

For billing examples of AI computing resources, see Billing of AI computing resources.

Appendix: List of public resource specifications

The following table lists some general computing public resource specifications supported by DLC. For a complete list, see the Resource Information section on the DLC create job page. For more information, see Create a job using the console. Available specifications vary by region. The specifications displayed on the console prevail.

Resource type

Specification

GPU type

ecs.g6.xlarge

4 vCPUs + 16 GB memory

None

ecs.c6.large

2 vCPUs + 4 GB memory

None

ecs.g6.large

2 vCPUs + 8 GB memory

None

ecs.g6.2xlarge

8 vCPUs + 32 GB memory

None

ecs.g6.4xlarge

16 vCPUs + 64 GB memory

None

ecs.g6.8xlarge

32 vCPUs + 128 GB memory

None

ecs.r7.large

2 vCPUs + 16 GB memory

None

ecs.r7.xlarge

4 vCPUs + 32 GB memory

None

ecs.r7.2xlarge

8 vCPUs + 64 GB memory

None

ecs.r7.4xlarge

16 vCPUs + 128 GB memory

None

ecs.r7.6xlarge

24 vCPUs + 192 GB memory

None

ecs.r7.8xlarge

32 vCPUs + 256 GB memory

None

ecs.r7.16xlarge

64 vCPUs + 512 GB memory

None

ecs.g5.xlarge

4 vCPUs + 16 GB memory

None

ecs.g7.xlarge

4 vCPUs + 16 GB memory

None

ecs.g7.2xlarge

8 vCPUs + 32 GB memory

None

ecs.g5.2xlarge

8 vCPUs + 32 GB memory

None

ecs.g6.3xlarge

12 vCPUs + 48 GB memory

None

ecs.g7.3xlarge

12 vCPUs + 48 GB memory

None

ecs.g7.4xlarge

16 vCPUs + 64 GB memory

None

ecs.r7.3xlarge

12 vCPUs + 96 GB memory

None

ecs.c6e.8xlarge

32 vCPUs + 64 GB memory

None

ecs.g6.6xlarge

24 vCPUs + 96 GB memory

None

ecs.g7.6xlarge

24 vCPUs + 96 GB memory

None

ecs.g5.4xlarge

16 vCPUs + 64 GB memory

None

ecs.hfc6.8xlarge

32 vCPUs + 64 GB memory

None

ecs.g7.8xlarge

32 vCPUs + 128 GB memory

None

ecs.hfc6.10xlarge

40 vCPUs + 96 GB memory

None

ecs.g6.13xlarge

52 vCPUs + 192 GB memory

None

ecs.g5.8xlarge

32 vCPUs + 128 GB memory

None

ecs.hfc6.16xlarge

64 vCPUs + 128 GB memory

None

ecs.g7.16xlarge

64 vCPUs + 256 GB memory

None

ecs.hfc6.20xlarge

80 vCPUs + 192 GB memory

None

ecs.g6.26xlarge

104 vCPUs + 384 GB memory

None

ecs.g5.16xlarge

64 vCPUs + 256 GB memory

None

ecs.r5.8xlarge

32 vCPUs + 256 GB memory

None

ecs.re6.13xlarge

52 vCPUs + 768 GB memory

None

ecs.re6.26xlarge

104 vCPUs + 1,536 GB memory

None

ecs.re6.52xlarge

208 vCPUs + 3,072 GB memory

None

ecs.g7.32xlarge

128 vCPUs + 512 GB memory

None

ecs.gn7i-c8g1.2xlarge

8 vCPUs + 30 GB memory

1 × NVIDIA A10

ecs.gn6v-c8g1.2xlarge

8 vCPUs + 32 GB memory

1 × NVIDIA V100

ecs.gn6e-c12g1.24xlarge

96 vCPUs + 736 GB memory

8 × NVIDIA V100

ecs.gn6v-c8g1.16xlarge

64 vCPUs + 256 GB memory

8 × NVIDIA V100

ecs.gn6v-c10g1.20xlarge

82 vCPUs + 336 GB memory

8 × NVIDIA V100

ecs.gn6e-c12g1.12xlarge

48 vCPUs + 338 GB memory

4 × NVIDIA V100

ecs.gn6v-c8g1.8xlarge

32 vCPUs + 128 GB memory

4 × NVIDIA V100

ecs.gn6i-c24g1.24xlarge

96 vCPUs + 372 GB memory

4 × NVIDIA T4

ecs.gn5-c8g1.4xlarge

16 vCPUs + 120 GB memory

2 × NVIDIA P100

ecs.gn7i-c32g1.16xlarge

64 vCPUs + 376 GB memory

2 × NVIDIA A10

ecs.gn6i-c24g1.12xlarge

48 vCPUs + 186 GB memory

2 × NVIDIA T4

ecs.gn6e-c12g1.3xlarge

12 vCPUs + 92 GB memory

1 × NVIDIA V100

ecs.gn5-c4g1.xlarge

4 vCPUs + 30 GB memory

1 × NVIDIA P100

ecs.gn5-c8g1.2xlarge

8 vCPUs + 60 GB memory

1 × NVIDIA P100

ecs.gn5-c28g1.7xlarge

28 vCPUs + 112 GB memory

1 × NVIDIA P100

ecs.gn6i-c4g1.xlarge

4 vCPUs + 15 GB memory

1 × NVIDIA T4

ecs.gn6i-c8g1.2xlarge

8 vCPUs + 31 GB memory

1 × NVIDIA T4

ecs.gn6i-c16g1.4xlarge

16 vCPUs + 62 GB memory

1 × NVIDIA T4

ecs.gn6i-c24g1.6xlarge

24 vCPUs + 93 GB memory

1 × NVIDIA T4

ecs.gn7i-c32g1.8xlarge

32 vCPUs + 188 GB memory

1 × NVIDIA A10

ecs.gn7e-c16g1.4xlarge

16 vCPUs + 125 GB memory

1 × GU50

ecs.gn7-c12g1.3xlarge

12 vCPUs + 95 GB memory

1 × GU50

ecs.gn7i-c16g1.4xlarge

16 vCPUs + 60 GB memory

1 × NVIDIA A10

ecs.gn7-c13g1.26xlarge

104 vCPUs + 760 GB memory

8 × GU50

ecs.ebmgn7e.32xlarge

128 vCPUs + 1,024 GB memory

8 × GU50

ecs.gn7i-c32g1.32xlarge

128 vCPUs + 752 GB memory

4 × NVIDIA A10

ecs.gn7-c13g1.13xlarge

52 vCPUs + 380 GB memory

4 × GU50

ecs.gn7s-c32g1.32xlarge

128 vCPUs + 1,000 GB memory

4 × NVIDIA A30

ecs.gn7s-c56g1.14xlarge

56 vCPUs + 440 GB memory

1 × NVIDIA A30

ecs.gn7s-c48g1.12xlarge

48 vCPUs + 380 GB memory

1 × NVIDIA A30

ecs.gn7s-c16g1.4xlarge

16 vCPUs + 120 GB memory

1 × NVIDIA A30

ecs.gn7s-c8g1.2xlarge

8 vCPUs + 60 GB memory

1 × NVIDIA A30

ecs.gn7s-c32g1.8xlarge

32 vCPUs + 250 GB memory

1 × NVIDIA A30