All Products
Search
Document Center

Platform For AI:Release notes

Last Updated:May 22, 2025

2025

March

Date

Feature

Description

Reference

2025-03-28

DLC supports mounting of OSS using ossfs

DLC supports mounting of OSS using ossfs. This delivers the best performance for reading and writing OSS for computing-intensive tasks such as autonomous driving (usually sequential and random read, sequential append write).

Use cloud storage for a DLC training job

2025-03-27

Node status upgraded

The status of computing power nodes is optimized. The scheduling status will no longer add status codes. This improves user experience.

Node

2025-03-19

DLC supports custom roles for Ray jobs

DLC supports user-defined worker roles for Ray jobs to implement mixed running of heterogeneous resources.

Submit training jobs

2025-03-19

Resource quota supports node-level scaling

Resource quota supports node-level scaling. This facilitates the management, update, and transfer of computing power among quotas.

Manage resource quotas

2025-03-07

PAI training service launched in US (Silicon Valley)

Deep Learning Containers (DLC) and AI resource quotas are available in the US (Silicon Valley) region. You can submit training jobs that use resource quotas and pay-as-you-go public resources.

Regions and zones

February

Date

Feature

Description

Reference

2025-02-28

DLC supports read/write permission on storage service mount configurations

PAI-DLC allows you to configure read/write permission when you mount a storage instance, such as OSS, NAS, and CPFS. This allows you to manage the permissions of the storage instance in a fine-grained manner.

Use cloud storage for a DLC training job

2025-02-21

AI scheduling engine v2.0 implements multi-level job preemption

The quota-based scheduling engine of PAI uses job type classification (such as training, inference, development, and priority) and dynamic priority evaluation algorithms to trigger the preemption mechanism. This ensures that high-priority jobs can be quickly executed. In addition, combined with the AIMaster preemptive rollback technology, interrupted jobs automatically save the intermediate state and enter the queue. After resources are released, they are preferentially resumed for efficient scheduling in resource-intensive scenarios.

Preemption policies

2025-02-20

EAS releases DeepSeek web search to help building enterprise-level AI assistant

EAS releases a DeepSeek + web search solution based on PAI-RAG, to help enterprise customers build multi-scenario AI assistants.

The PAI-RAG scenario-based deployment on EAS is fully upgraded. It supports flexible configuration of the web search capability. The feature uses the general search API of Alibaba Cloud to easily access and use real-time data to obtain more accurate and comprehensive search results. PAI-RAG supports flexible selection of LLMs. You can easily deploy dedicated AI assistants with DeepSeek and web search. In addition, PAI-RAG provides comprehensive ecological capabilities to support one-click deployment to enterprise WeChat, WeChat public account, DingTalk group chat and other platforms.

2025-02-10

DLC supports multiple connections (nconnect) when mounting file storage services

DLC supports multiple connections (nconnect) when mounting file storage services NAS or CPFS. This allows fine-grained control over the number of mount connections, optimizes the concurrent access performance of multiple nodes, and ensures the stability of large-scale training jobs.

Use cloud storage for a DLC training job

2025-02-07

EAS releases distributed inference

With the advent of ultra-large-scale MoE models such as Qwen-max and Deepseek, it is difficult for a single device to handle their huge parameter sizes. EAS offers a multi-machine distributed inference solution that overcomes hardware limitations and efficiently supports the deployment and operation of models with large parameter sizes. EAS supports multiple parallelism methods, such as pipeline parallelism, tensor parallelism, and data parallelism. It is also compatible with high-performance inference engine frameworks, such as BladeLLM, vLLM, and SGLang.

Multi-machine distributed inference

January

Date

Feature

Description

Reference

2025-01-21

DLC supports training timeout alert

DLC allows you to configure training timeout alerts. You can customize timeout alert rules for training jobs in environment preparation, queuing, and running states. When a rule is triggered, an alert notification is sent to help you monitor the training process for exceptions.

Notification rule

2025-01-21

DLC supports training status notifications

DLC allows you to subscribe to training status notifications and add status events such as queuing, bidding, environment preparation, and running. This allows you to track the training process and improve the message notification capability of the training service.

Notification rule

2025-01-20

DLC supports direct mounting of storage services when submitting jobs

When you use DLC to submit training jobs, you can directly select different storage instances. Currently, OSS, general-purpose NAS, extreme NAS, general-purpose CPFS, and intelligent computing CPFS are supported. This reduces the use threshold and is convenient for users.

Submit training jobs

2025-01-20

Ray on DLC supports using idle resources

DLC supports submitting Ray jobs using idle resources, facilitating customers to run multiple types of jobs on a single set of resources. This enables resource sharing between jobs and improves overall resource utilization.

Use idle resources

2025-01-20

Artlab launches industry-tailored tools

ArtLab launches industry-tailored tools. In the first phase, applications such as realistic e-commerce product rendering (for home appliances and furniture), corporate style poster generation, and creative footwear design have been released. More applications will be released in the future.

2025-01-20

ArtLab launches AIGC application module

PAI-ArtLab launches an AIGC application module to support online applications that encapsulate ComfyUI workflows to generate image based on text or image. This reduces the threshold for users to use AIGC tools and reduces user costs through the serverless service mode.

1. Out-of-the-box: No environment configuration required, you can start of the application with one click.

2. The platform has built-in enterprise-level AIGC applications, such as enterprise-style poster generator and enterprise event avatars generator.

3. Serverless mode: You are charged only for GPU inference, significantly reducing costs.

2025-01-20

Model Gallery supports model inference acceleration

The model inference acceleration of PAI-Model Gallery can match supported inference acceleration capabilities (vllm and BladeLLM) based on machine specifications.

2025-01-16

EAS upgrades BladeLLM high-performance deployment service

PAI-EAS supports scenario-based deployment of BladeLLM to achieve faster response time and higher throughput for LLM inference.

BladeLLM is an inference engine developed by PAI. It provides efficient runtime, high-performance operator implementation, and extreme hybrid quantization. PAI-EAS fully integrates BladeLLM to launch the LLM high-performance inference service. It supports the deployment of preset models and custom models as well as advanced options such as model parallelism and speculative sampling. This provides customers with efficient LLM deployment solutions.

Get started with BladeLLM

2025-01-02

Model Gallery launched in China (Hong Kong) and other regions

PAI-Model Gallery integrates pre-trained models in LLM, CV, NLP, and speech fields to provide one-stop and zero-code model training, model compression, model evaluation, and model deployment capabilities. It is newly launched in the following regions: China (Hong Kong), Japan (Tokyo), Indonesia (Jakarta), Germany (Frankfurt), and US (Virginia).

2024

December

Date

Feature

Description

Reference

2024-12-23

DLC pay-as-you-go bills distinguish job types

DLC training jobs support system tags (key:acs:pai:payType) that distinguish pay-as-you-go jobs from preempted jobs. Customers can quickly identify and filter pay-as-you-go jobs and check the consumption and discount.

View billing details

2024-12-16

Machine Learning Designer supports grouping LLM data preprocessing components

Machine Learning Designer supports grouping multiple serial data preprocessing nodes (DLC) for execution, avoiding repeated data writes to disk and the time consumption of starting and stopping distributed tasks, thereby improving execution efficiency and supporting automatic intelligent aggregation.

Grouping of LLM data processing components

2024-12-16

DLC launchs preemptible jobs that use general computing resources

PAI supports preemptible jobs based on general-purpose computing resources to provide customers with more cost-effective AI computing power.

2024-12-10

PAI training service launched in Germany (Frankfurt)

Deep Learning Containers (DLC) and AI resource quotas are available in the Germany (Frankfurt) region. You can submit training jobs that use resource quotas and pay-as-you-go public resources.

2024-12-09

DLC job status upgrades to v2.0

Based on the resource quota, the Queuing and PreAllocation states are merged into Queuing. This provides clear and simple task status information for easy use and understanding.

2024-12-06

DLC sanity check supports custom items

The DLC sanity check feature supports more than 15 items, such as computing performance checks, node communication checks, computing and communication cross-checks, and model simulation checks. This further improves its troubleshooting capabilities in case of computing power and network failures. The check items are open to users to choose based on their business requirements to achieve self-management and control.

Sanity check

November

Date

Feature

Description

Reference

2024-11-20

DSW supports dynamic mounting of OSS datasets

1. Allows dynamic mounting or unmounting of OSS datasets without restarting the instance. 2. Provides an easy-to-use SDK that allows users to mount or unmount datasets through simple configuration or a single line of code. 3. Supports dynamically mounting of datasets from AI assets (PAI public datasets or custom datasets) or directly mounting OSS storage paths.

Mount datasets or OSS paths

2024-11-20

DSW instances support custom access configurations

With the rapid development of AIGC, various WebUI frameworks and application development frameworks have become the mainstream choice for developers. As a one-stop AI development platform, PAI-DSW provides the custom service access configuration feature. During application development, developers can share services with collaborative developers for testing and verification in a secure manner at any time.

Access service over Internet

October

Date

Feature

Description

Reference

2024-10-17

AI general computing resource group supports L20 in international regions

AI general computing resource group of PAI supports L20 (gn8is series) in international regions

2024-10-12

DLC job status upgrades to v1.0

Computing power type includes resource quotas, bidding resources, and public computing power. Business mode includes subscription, bidding, and pay-as-you-go. At job and instance levels, the EnvPreparing and Bidding statuses are added. The Created, Queuing, and PreAllocation statuses are simplified. This provides clearer and simpler job status information for ease of use and understanding.

2024-10-11

ComfyUI serverless available in PAI-ArtLab

ComfyUI serverless is available in the ArtLab toolbox. You can use ComfyUI for text-to-image or image-to-image generation. The serverless mode reduces costs. You are charged only for model inference.

ArtLab

2024-10-10

QuickStart supports DPO and CPT for LLM

PAI Quick Start-Model Gallery provides more complete LLM training capability. On top of the original Supervised Fine-Tuning (SFT), Model Gallery now supports Direct preference optimization (DPO) and Continued Pre-training (CPT).

QuickStart

September

Date

Feature

Description

Reference

2024-09-29

DSW integrates Tongyi Lingma

The AI coding assistant Tongyi Lingma (Personal Edition) is integrated into DSW, offering features such as line or method level code generation, natural language to code, unit test generation, comment generation, code explanation, AI coding chat, and troubleshooting. Users can use the feature without installation and login, experiencing efficient and graceful coding.

Using Tongyi Lingma for development

2024-10-08

PAI Training Service is available in the China (Hong Kong) and Indonesia (Jakarta) regions

Deep Learning Containers (DLC) and AI resource quotas are available in the China (Hong Kong) and Indonesia (Jakarta) regions. You can submit training jobs that use resource quotas and pay-as-you-go public resources.

August

Date

Feature

Description

Reference

2024-11-11

Judge model feature officially released

The judge model service of PAI uses a fine-tuned LLM based on Qwen2 as a judge to score responses from evaluated models. This service is suitable for open-ended and complex scenarios. Main advantages: 1. Accuracy: The judge model can classify subjective questions into scenarios such as open-ended discussions, creative writing, code generation, and role-playing. It then develop tailored criteria for each scenario, significantly enhancing evaluation accuracy. 2. Efficiency: Without the need for manual data labeling, the judge model can independently analyze and evaluate LLMs based on questions and model answers, greatly boosting evaluation efficiency. 3. Ease of use: PAI offers various usage methods, such as task creation in the console, API calls, and SDK calls. This allows for both quick trials and flexible integration for developers. 4. Cost-effectiveness: The judge model provides performance evaluation at a competitive price. Its performance is comparable to that of ChatGPT-4 in Chinese language scenarios.

Judge model overview

2024-09-03

DSW supports NotebookLab (Lightweight Edition)

1. Lightweight coding is supported in Notebook. You can use a browser to code, without the need to start other resources in advance. 2.Notebooks are managed as assets and are decoupled from instance resources, making it easier to archive or share Notebooks as documents or code samples.

Notebook Lab

2024-08-26

EAS supports LLM Intelligent Router to improve LLM inference efficiency

When customers deploy LLM services on EAS, they can enable the LLM Intelligent Router feature. LLM Intelligent Router can evenly allocate the computing power and video memory of backend inference instances and improve the resource usage of clusters.

Use LLM Intelligent Router to improve inference efficiency

2024-08-26

DLC jobs with general computing resources support CPU affinity

The general computing resources of PAI DLC supports CPU core binding to improve job performance.

2024-08-15

EAS supports dedicated gateway

EAS supports the dedicated gateway feature to meet inference requirements for security isolation and access control. This reduces network risks in high-concurrency and high-throughput business scenarios. Dedicated gateways allow you to configure whitelists for access over virtual private clouds (VPCs) and the Internet and implement fine-grained management. Dedicated gateways can also ensure the stability of connections between services. You can use PrivateLink to connect to the VPC of your enterprise. You can also implement independent control over access to the Internet.

Use a dedicated gateway

2024-08-15

PAI workspace supports custom roles

Workspace is a key concept in Platform for AI (PAI). Workspaces allow your organization or team to manage computing resources, user permissions, and AI assets in a centralized manner to achieve seamless collaboration at every stage of AI development. In specific scenarios, the current preset roles of a workspace cannot meet the management requirements of customers. For example, you cannot assign a RAM role the permissions to use DSW without assigning the role the permissions on DLC. PAI provides the feature that allows you to customize roles and related permissions.

Manage members of a workspace

2024-08-05

Discontinuation of earlier versions of PAI-PyTorch algorithm components

Earlier versions of the PyTorch algorithm components, including PyTorch100 and PyTorch 131-based components, are officially discontinued in all clusters of Platform for AI (PAI) on August 30, 2024 because of a system upgrade. If you have a PyTorch job that is submitted to MaxCompute by using the pai -name pytorch100/pytorch131 PAI command, we recommend that you migrate the jobs before the discontinuation date. We recommend that you use Deep Learning Containers (DLC) to submit a PyTorch job. For more information, see Submit training jobs. Starting August 31, 2024, existing jobs that use earlier versions of PyTorch algorithm components are no longer guaranteed by the Platform for AI Service Level Agreement. For more information, see Alibaba Cloud product terms of service.

If you have questions or require technical support, contact us in your dedicated DingTalk group or submit a ticket.

Thank you for your cooperation.

July

Date

Feature

Description

Reference

2024-07-03

EAS supports GPU sharing

When you deploy a model in EAS, you can split and use the computing power based on the ratio of GPU computing power and the memory size. This helps reduce resource costs and improve resource utilization. On the deployment page, you can schedule instances based on GPU memory and computing power. This allows multiple instances to share a single GPU.

EAS overview

2024-07-03

EAS supports instance health check

The health check feature of EAS ensures high service availability. By performing fast fault detection and automatic recovery, this feature can facilitate enterprise-level inference service deployment. The Kubernetes health check mechanism can automatically detect and recover failed containers. This ensures that traffic is allocated only to healthy instances.

EAS overview

June

Date

Feature

Description

Reference

2024-07-01

QuickStart supports LLM evaluation

PAI-QuickStart provides the LLM evaluation feature. You can evaluate the comprehensive capabilities of a model based on authoritative public datasets, such as CMMLU, C- Eval, or MMLU, or custom datasets to compare the performance of multiple models and determine whether the model capabilities are suitable for your business scenarios.

Model evaluation

2024-06-19

General computing resources of PAI support mounting CPFS for Lingjun (invitational preview)

PAI leverages Alibaba Cloud storage services to provide cost-effective solution for storage and computing in large language model scenarios. You can mount CPFS for Lingjun for a PAI training that runs on general computing resources.

2024-06-12

Machine Learning Designer is available in the China (Ulanqab) region

Machine Learning Designer is available in the China (Ulanqab) region. You can deploy your services in this region in the PAI console.

2024-06-11

Machine Learning Designer provides the Notebook component

Machine Learning Designer provides the Notebook component, which can be connected to DSW instances. The component allows you to directly use notebooks to write, debug, and run code in a pipeline.

Notebook

May

Date

Feature

Description

Reference

2024-07-01

QuickStart supports fine-tuning LLMs by using QLoRA, LoRA, and full parameter

PAI allows you to fine-tune an LLM by using QLoRA, LoRA, and full-parameter fine-tuning in QuickStart. You can select a training method based on your business requirements to reduce costs.

Fine-tune, evaluate, and deploy a Qwen2.5 model

2024-06-07

DSW supports instance RAM role configuration

After you associate the default role of PAI with an instance, you do not need to configure an AccessKey pair in the following scenarios in DSW:

  • Submit a training task to the current workspace by using the PAI SDK.

  • Submit a training task to the current workspace by using the DLC SDK.

  • Submit a task to MaxCompute projects on which the instance owner has execution permissions by using the ODPS SDK.

  • Access data in the bucket that is the default storage path of the current workspace by using the OSS SDK.

  • Use the Tongyi Lingma service in a web integrated development environment (IDE).

If you use a custom RAM role, DSW uses the temporary access credentials of the role to access specific Alibaba Cloud services, such as OSS or RDS. This ensures secure communication between the DSW instance and other Alibaba Cloud services.

Associate a RAM role with a DSW instance

April

Date

Feature

Description

Reference

2024-04-29

EAS supports serverless deployment of AI painting services

EAS provides the serverless deployment feature for model services that have intermittent or unpredictable traffic patterns. If you deploy an AI painting service by using the serverless method, you are charged only for GPU consumption.

Quickly deploy Stable Diffusion for text-to-image generation in EAS

March

Date

Feature

Description

Reference

2024-03-25

DSW supports integrated AI and big data development

Customers can submit data analysis or preprocessing tasks to MaxCompute or E-MapReduce by using multiple methods such as Python in DSW. The processed data can be used in model trainings that are performed on on-premises GPU devices or in Deep Learning Containers (DLC).

Connect a DSW instance to an EMR cluster

2024-03-25

The file transfer station feature is available in DSW

DSW provides the file transfer station feature that can accelerate the upload process when you need to upload large files such as large models from your on-premises computer to a DSW instance. After you upload the large file, you can use the uploaded large file in multiple DSW instances in your RAM account.

File transfer station

2024-03-15

PAI-Lingjun Intelligent Computing Service available in the Singapore region

PAI-Lingjun Intelligent Computing Service is a next-generation intelligent computing service independently developed by Alibaba Cloud to provide optimized heterogeneous cluster instances. Trained based on a large number of AI applications, PAI-Lingjun Intelligent Computing Service is proven to provide high performance, efficiency, and resource utilization. PAI-Lingjun Intelligent Computing Service can meet the requirements of various industries, such as autonomous driving, basic scientific research, new drug research and development, finance, and meta-universe. The service provides affordable and accessible intelligent computing power to facilitate technological innovation and industrial upgrade.

PAI-Lingjun Intelligent Computing Service is available in the Singapore region. You can activate the service in the console.

2024-06-06

Discontinuation of GPU-related server and algorithm components in Machine Learning Designer

The warranty for V100 and P100 server clusters in which services are running is expired. The algorithm components that are related to TensorFlow(GPU), MXNet, and PyTorch in Machine Learning Designer of Platform for AI (PAI) are discontinued on March 1, 2024. You can continue using the related algorithm components of the cloud-native version and submit training jobs to Deep Learning Containers (DLC) in PAI. We recommend that you use the Python-based components in Machine Learning Designer to run DLC jobs. The Python-based components can work in the same manner as the discontinued components. The existing tasks of the discontinued algorithm components are not covered by the SLA starting from June 1, 2024. The V100 and P100 server clusters are discontinued on June 30, 2024.

Python Script

February

Date

Feature

Description

Reference

2024-02-28

Machine Learning Designer provides data preprocessing operators for LLM and common templates

High-quality data preprocessing is an important step in LLM application. Machine Learning Designer of Platform for AI (PAI) provides commonly used high-performance operators for data preprocessing, such as deduplication, standardization, and sensitive information masking. You can use large-scale distributed computing capabilities based on MaxCompute when you preprocess data for LLMs to improve efficiency and improve the reliability and performance of LLMs.

Component reference: Data processing for foundation model

2024-02-04

EAS serverless deployment is in invitational preview

EAS provides the serverless deployment feature for model services that have intermittent or unpredictable traffic patterns. If you deploy a model service in EAS by using the serverless deployment feature, you are charged only when GPU computing occurs. For example, if you deploy an AI painting model, you are charged based on the actual painting duration.

EAS overview

January

Date

Feature

Description

Reference

2024-02-04

QuickStart Released on the International Site (alibabacloud.com)

The QuickStart feature is now available in the Singapore region.

2024-02-04

EAS simple deployment released

EAS provides simplified deployment methods for common deployment scenarios, including ModelScope model deployment, Hugging Face model deployment, Triton deployment, TFServing deployment, LLM deployment, and SD web application deployment. In these scenarios, you need to only provide the storage directory of the model to start services and applications in a few clicks.

2024-02-01

Quick deployment of AI video generation applications by using EAS

Users can use Elastic Algorithm Service (EAS) to deploy web applications for AI video generation based on ComfyUI and Stable Video Diffusion models. EAS can help you quickly implement AI-powered text-to-video or image-to-video generation in industries such as live streaming and short video platforms, gaming and Internet entertainment, and animation production.

Deploy an AI video generation application in EAS

2023

December

Date

Feature

Description

Reference

2023-12-13

Machine Learning Designer is available in Indonesia (Jakarta)

Machine Learning Designer is available in Indonesia (Jakarta). You can select the region Indonesia (Jakarta) on the Platform for AI (PAI) console.

2023-12-06

DSW supports SSH logon

Allows customers to access DSW instances by using the machines in their VPC or from the on-premises development environment in a more convenient manner and facilitates development and training in DSW.

SSH direct connection

November

Date

Feature

Description

Reference

2023-11-20

PAI releases the automated machine learning platform AutoML

PAI releases AutoML, an enhanced machine learning service of PAI. AutoML integrates various algorithms and distributed computing resources supported by PAI and supports multiple access methods. You can use AutoML to automatically find the optimal hyperparameter values and improve the model tuning efficiency.

How AutoML works

October

Date

Feature

Description

Reference

2023-10-27

Subscription AI Training is available on the International site (alibabacloud.com)

AI Training of Platform for AI (PAI) is available on the international site (alibabacloud.com) in the following regions: China (Beijing), China (Shanghai), China (Hangzhou), China (Shenzhen), and Singapore. AI Training supports the subscription billing method. You can access AI Training in the PAI console.

September

Date

Feature

Description

Reference

2023-09-28

EAS supports push-button deployment of the Tongyi Qianwen model

You can use PAI-EAS to deploy web UI applications that are based on the open source Tongyi Qianwen model, and use the web UI and API operations to perform model inference.

Qwen-7B is a 7 billion-parameter model of the Tongyi Qianwen series developed by Alibaba Cloud. Qwen-7B is a large language model based on Transformer, and is trained on ultra-large-scale pre-trained data. The pre-trained data covers a wide range of data types, including a large number of texts, professional books, and code. In addition, Qwen-7B-Chat, an LLM AI assistant is developed based on Qwen-7B.

Quickly deploy Qwen in EAS.

2023-09-18

DLC supports monitoring metric subscription and alert

PAI-DLC offers detailed monitoring metrics for job resource conditions, enabling users to configure flexible alert rules for distributed training jobs.

Training monitoring and alerting

2023-09-18

EasyCKPT high-performance CKPT released

PAI-EasyCKPT, designed for PyTorch model training, features a near-zero overhead saving mechanism and ensures accurate model saving and recovery throughout the training process. Compatible with Megatron and DeepSpeed, it requires minimal code changes for use.

Use EasyCkpt to save and resume foundation model trainings

August

Date

Feature

Description

Reference

2023-09-04

Support for deploying and fine-tuning Stable Diffusion models

  • Deploy a Stable Diffusion model.

  • Fine-tune a Stable Diffusion model.

  • Quickly start and deploy a Stable Diffusion web UI.

  • Deploy a LoRA SD model by using Kohya_ss.

AIGC

July

June

May

Date

Feature

Description

Reference

2023-05-20

Support for PAI SDK for Python

PAI SDK for Python provides easy-to-use HighLevel APIs for machine learning engineers to easily train and deploy models on PAI and complete end-to-end machine learning processes.

Install and configure PAI SDK for Python

April

Date

Feature

Description

Reference

2023-04-19

Support for elastic resource pool in Elastic Algorithm Service (EAS)

EAS supports auto scaling of service resources. If node resources in the dedicated resource group are insufficient during a service scale-out, new instances of the service are created in the pay-as-you-go public resource group and billed based on public resource group rules. During a service scale-in, service instances that reside in the public resource group are released first.

Elastic resource pool

2023-04-04

Support for the new version of Elastic Algorithm Service (EAS) page for quick service deployment

EAS supports the following deployment methods: Deploy Service by Using Image, Deploy Web App by Using Image, and Deploy Service by Using Model and Processor. You can deploy AI services or applications to EAS in just a few clicks.

EAS overview

March

Date

Feature

Description

Reference

2023-03-23

Upgraded model management feature

Models trained by Machine Learning Designer can be registered on the model management page. You can change the approval status of model versions to trigger model-related events, such as automated messaging to DingTalk groups through chatbots, or calling the specified HTTP or HTTPS service.

Model version approval status and events

February

Date

Feature

Description

Reference

2023-02-13

Support for preemptible resource instances in Elastic Algorithm Service (EAS)

When you use a public resource group to deploy a service to EAS, you can specify preemptible instances for the service to reduce costs.

Specify preemptible instances

2023-02-06

Support for multiple instance types for Elastic Algorithm Service (EAS)

When you deploy services in EAS, you can specify multiple instance specifications in the configuration file. The system then prepares resources based on the instance types you specified in the configuration file. This method reduces the chances of insufficient resources in scenarios where only a single instance type is specified.

Specify multiple instance types

January

Date

Feature

Description

Reference

2023-01-13

Support for push-button pipeline deployment as an EAS online service

Machine Learning Designer supports push-button pipeline deployment. You can deploy a batch data-processing pipeline that implements data pre-processing, feature engineering, and model prediction to Elastic Algorithm Service (EAS) as an online service after packaging the pipeline as a model.

Deploy a pipeline as an online service

2022

December

Feature

Description

Date

Region

Reference

Support for computing resources based on Yitian 710 in Elastic Algorithm Service (EAS)

EAS supports computing resources based on Yitian 710 processors. These resources can help you reduce the costs of model deployment, model inference, as well as improve efficiency.

2022-12-8

All regions

EAS overview

New algorithm components

Machine Learning Designer provides a variety of new algorithm components, including Prophet, MTable Expander, MTable Assembler, and Time Window SQL. You can find and use the components in the left-side directory tree of components on the Machine Learning Designer platform.

2022-12-5

All regions

Component reference: Overview of all components

Support for custom templates

You can create a custom template based on a pipeline that runs successfully in Machine Learning Designer. You can use this template to quickly build similar pipelines to improve efficiency.

2022-12-1

All regions

Create a pipeline from a custom template

December

Date

Feature

Description

Applicable customers

Reference

Support for O&M on EAS nodes

You can perform O&M operations on the nodes in resource groups. The operations include viewing node information, stopping and restarting node scheduling, and removing instances in nodes.

2022-11-30

All regions

EAS overview

Support for querying the updates of Data Science Workshop (DSW) instances

You can view the changes to the states of a DSW instance throughout the lifecycle of the instance.

You can view the details of a DSW instance and change its configurations.

2022-11-18

All regions

Create a DSW instance

September

Feature

Description

Date

Region

Reference

Support for the service grouping and asynchronous inference features in Elastic Algorithm Service (EAS)

When you create an EAS service, you can specify the service group to which the EAS service belongs. The service group has a unified ingress. The ingress allocates traffic to each EAS service based on the traffic allocation policy. You can also specify the traffic allocation ratio of each service in a service group to ensure high resource utilization.

PAI provides the queue service and asynchronous inference features. These features allow you to consume inference services by distributing requests, subscribing to requests and pushing inference results, or periodically querying inference results.

2022-09-30

All regions

November

Date

Feature

Description

Applicable customers

Reference

New algorithm components

Machine Learning Designer provides a variety of new training and prediction components, including XGBoost, Density-Based Spatial Clustering of Applications with Noise (DBSCAN), Gaussian Mixture Mode (GMM), Ridge Regression, and Lasso Regression. You can find and use the components in the left-side directory tree of components on the Machine Learning Designer platform.

2022-08-02

All regions

July

Feature

Description

Date

Region

Reference

Python Script V2 added

The Python Script V2 component is added to Machine Learning Designer of PAI. You can use the component to develop custom algorithms. You can also use the component together with the pre-set algorithms of PAI to support more scenarios.

2022-07-15

  • China (Hong Kong)

  • Singapore

  • India (Mumbai) Closed Down

  • US (Silicon Valley)

  • US (Virginia)

  • Germany (Frankfurt)

Python Script

Designer officially launched in US (Virginia)

Designer is officially launched in US (Virginia). You can select the corresponding region in the PAI console and create a workspace to use Designer-related features.

2022-07-05

US (Virginia)

None

Automatic stress testing available

The automatic stress testing feature is available for EAS. You can use EAS-benchmark, a distributed stress testing tool, to create stress testing tasks for prediction services that are deployed in EAS.

2022-07-04

  • China (Hong Kong)

  • Singapore

  • Indonesia (Jakarta)

  • India (Mumbai) Closed Down

  • US (Silicon Valley)

  • US (Virginia)

  • Germany (Frankfurt)

N/A

October

Date

Feature

Description

Applicable customers

Reference

Visualized analytical reports supported

Analytical reports can be visualized by using Tensorboard. Visual deep learning components of Machine Learning Designer allow you to use the Tensorboard dashboard to view visualized analytical reports. On the dashboard, you can view visualized feature importance evaluation, correlation analysis, and scatter charts.

2022-06-22

  • China (Hong Kong)

  • Singapore

  • India (Mumbai) Closed Down

  • US (Virginia)

  • Germany (Frankfurt)

Use TensorBoard to visualize analytical reports

Designer officially launched in China (Hong Kong)

Designer officially launched in China (Hong Kong), offering hundreds of PAI self-developed machine learning algorithms and dozens of industry templates. You can use them on the PAI console as needed.

2022-06-20

China (Hong Kong)

None

May

Feature

Description

Publish date

Publish region

Reference

Available in Singapore and US (Silicon Valley)

Machine Learning Designer is available in the Singapore and US (Silicon Valley) regions. Machine Learning Designer provides hundreds of self-developed machine learning algorithms and dozens of industry templates. You can use them as needed in the PAI console.

2022-05-10

  • Singapore

  • US (Silicon Valley)

None

April

Feature

Description

Date

Region

Reference

Fully-managed Flink resources supported

Fully-managed Flink resources can be purchased and associated with workspaces. Then, you can use multiple components or use the PyAlink Script component alone to build pipelines for large-scale distributed training of models.

2022-04-30

Germany (Frankfurt)

Flink resource quotas

New anomaly detection, recommendation, data source, and custom algorithm components added

Components including PyAlink Script, Read CSV File, IForest Outlier, LOF Outlier, One-Class SVM Outlier, and Swing Recommendation are added to Machine Learning Designer. The PyAlink Script component allows you to call hundreds of algorithms that are under the Alink framework.

2022-04-16

Germany (Frankfurt)

September

Date

Feature

Description

Applicable customers

Reference

Designer officially launched in Germany (Frankfurt)

Designer officially launched in Germany (Frankfurt), offering hundreds of PAI self-developed machine learning algorithms and dozens of industry templates. You can use them on the PAI console as needed.

2022-03-30

Germany (Frankfurt)

None

TensorFlow 2.7 supported by PAI-Blade

TensorFlow 2.7 is supported by PAI-Blade. You can select a version based on your needs.

2022-03-27

All regions

None

DSW is officially launched in five regions including Singapore

You can create DSW instances and use DSW features to build and train models in these regions.

2022-03-21

  • Singapore

  • Malaysia (Kuala Lumpur)

  • Indonesia (Jakarta)

  • Germany (Frankfurt)

None

CronHPA feature supported, and gRPC and WebSocket protocols supported for image deployment and release

The CronHPA feature is available in EAS. This feature allows you to perform scheduled auto-scaling on service instances. In addition, EAS supports the deployment of services by using the open source TensorFlow Serving system or the Triton software.

2022-03-21

  • China (Hong Kong)

  • Singapore

  • Malaysia (Kuala Lumpur)

  • Indonesia (Jakarta)

  • India (Mumbai) Closed Down

  • US (Silicon Valley)

  • Germany (Frankfurt)

Scheduled scaling