This topic describes the terms related to management, AI development, and modules in Machine Learning Platform for AI (PAI).
Terms related to management
Term | Description |
workspace | Workspace is a key concept in PAI. Workspaces can help enterprises and teams manage computing resources and control permissions in a unified manner. Workspaces also provide AI developers with development tools and capabilities to allow different teams to collaborate throughout the entire lifecycle of AI development. PAI runs on top of DataWorks. Therefore, workspaces in PAI are mapped to workspaces in DataWorks. Default workspace: The default workspace contains commonly used pay-as-you-go resources. You must activate the default workspace before you can use these resources. The default workspace can help first-time users quickly get started with model development without the need to understand concepts such as resource groups. |
Deep Learning Containers (DLC) | DLC is a cloud-native platform for basic AI computing jobs. DLC provides an elastic, stable, easy-to-use, and high-performance environment for model training. DLC provides multiple algorithm frameworks, allows you to run a large number of deep learning jobs in a distributed manner, and supports custom algorithm frameworks. DLC supports the following types of clusters:
|
resource group |
|
member | Members are Alibaba Cloud accounts or Resource Access Management (RAM) users that join workspaces. Members in the same workspace can assume different roles to collaborate throughout the AI development pipeline. Only the owner and administrators of a workspace can modify the members in the workspace. |
role | Roles are mappings between members and permissions. You can use the system roles or create custom roles. The system roles include:
|
dependency | PAI relies on other Alibaba Cloud services. To use the features provided by PAI, you must first activate these Alibaba Cloud services and complete RAM authorization by using an Alibaba Cloud account or the resource administrator role. The Alibaba Cloud services that you need to activate include Object Storage Service (OSS), Apsara File Storage NAS (NAS), Log Service, Container Registry, and API Gateway. |
Terms related to AI development
Term | Description |
dataset | Datasets are used in labeling, model training, and model evaluation. You can create datasets to use structured data or unstructured data or mount directories in datastores such as OSS, NAS, and MaxCompute. In addition, you can centrally manage the storage, versions, and schemas of datasets in PAI. |
pipeline | Pipelines are directed acyclic graphs (DAGs) that consist of upstream components and downstream components connected based on logical scheduling. You can submit multiple runs for a pipeline to generate PipelineRuns. |
PipelineDraft | PipelineDrafts are configurable pipeline objects on the canvas of Designer. You can submit runs for PipelineDrafts to generate PipelineRuns. |
component | Components are the smallest configurable units in pipelines and PipelineDrafts, and the smallest executable units in PipelineRuns. Components consist of the following types:
|
node | Nodes represent components that are dragged and dropped to the canvas to form a pipeline. |
snapshot | Each time the system runs a single node in a PipelineDraft, multiple nodes in the PipelineDraft, or the entire PipelineDraft, the system creates a snapshot for the configurations of the PipelineDraft. The snapshot includes the node configurations, runtime parameters, and execution mode. Snapshots can be used in PipelineDraft versioning and configuration rollbacks. |
PipelineRun | A PipelineRun represents a single run of a pipeline. After you can use Designer to submit a run for a PipelineDraft or use the SDK to submit a run for a pipeline, a PipelineRun is generated. |
job | Jobs use different types of computing resources. You can create multiple types of jobs, such as DLC jobs and MC jobs. Compared with run and PipelineRun, job is an underlying concept. |
run | A run represents a single execution of a pipeline. Run is equivalent to the same concept in MLflow. All runs must belong to an experiment. You can use runs to track the training jobs that you submitted in PAI. You can also use the MLflow client to submit runs from an on-premises machine. |
model | Models are generated by training jobs based on datasets, algorithms, and code. You can use models to make predictions. |
Processor | A processor is a package of online prediction logic, including the logic for loading models and handling requests. In most cases, processors are deployed together with model files to provision services. Processors consist of the following types:
|
service | You can deploy model files together with the online prediction logic to provision services. EAS allows you to create, update, start, stop, scale out, and scale in services. |
image | Machine Learning Platform for AI allows you to manage the following Docker images as AI assets:
You can use images in pipelines to create custom components to complete specific jobs. Images can also provide runtime environments for DSW instances and training jobs. |
instance | Instances are the smallest units for provisioning computing resources. Instances consist of the following types:
|
Terms related to modules
Term | Description |
iTAG | iTAG is a dataset labeling tool that is developed with black box models to help improve the quality and efficiency of dataset labeling. |
Designer | Designer is intended for pipeline design in the AI sector. Designer provides a variety of built-in machine learning algorithm components. You can drag and drop these components to train models without coding. |
Data Science Workshop (DSW) | DSW is an integrated development environment (IDE) intended for interactive AI development in the cloud. DSW consists of JupyterLab, VS Code, and Terminal. You can use images to deploy DSW instances that use NAS as the storage. |
Deep Learning Containers (DLC) | You can submit training jobs to computing resource groups, such as DLC clusters, in the current workspace. After you submit training jobs, you can view the details about the jobs in the Jobs module of the PAI console. |
Elastic Algorithm Service (EAS) | EAS allows you to deploy complex models as services on a large scale with a few clicks. EAS supports real-time scaling and provides a sophisticated monitoring and maintenance system. |
AI Asset Management | This module allows you to manage key AI assets, including datasets, models, and source code repositories. |
Scenario-based solution | A collection of solutions provided by PAI to help you resolve issues in vertical markets. |
AI Competency Center |
|