Why Alibaba Cloud
Alibaba Cloud provides a full-stack solution for generative AI development, which includes a collection of compute-optimized ECS instances with high-performance GPUs, a powerful AI computing platform, a one-stop AI development platform, and acceleration features for AI model training and inference. This solution helps you build and optimize foundation models (FMs) based on your business needs, so you can create new and intelligent customer experiences, and drive business transformation with innovations in generative AI.
-
Powered by Compute-Optimized GPUs
Train and run large-scale FMs containing billions of parameters with high performance, scalability, and cost effectiveness on Alibaba Cloud ECS instances and NVIDIA GPUs
-
A Diverse Selection of Pre-Built FMs
Choose FMs for your business from a wide range of pre-built models in PAI-EAS, including Alibaba Cloud's Tongyi Qianwen (Qwen), Stable Diffusion, Llama 2, and more from Hugging Face
-
Streamlined FM Integration
Integrate and deploy FMs in easy steps with Machine Learning Platform for AI and ECS, and tap into the capabilities and services of Alibaba Cloud to speed up innovation
-
Model Training / Inference Optimization
Apply end-to-end performance optimization to accelerate dataset processing, model training, and model inference, and enhance it with a GPU-based AI acceleration solution
2023 Alibaba Cloud Global Summit: Powering Up Future Business with Generative AI
At the 2023 Alibaba Cloud Global Summit, Mr. Dongliang Guo, VP of Product and Solution, discussed the significance of the cloud computing industry when driving AI transformation, introduced Alibaba Cloud’s latest innovations (including Cloud-Native AI Suite 2.0 and the 8th Generation of ECS), and demonstrated the exciting capabilities of multi-modality content generation jointly supported by Alibaba Cloud and key partners.
Learn MoreLearn More About Generative AI on Alibaba Cloud
1. GPUs for Model Training and Inference
-
Model Training: gn7 series of ECS instances power large-scale training tasks with high-performance NVIDIA GPUs
Model Inference: gn6 series of ECS instances provide a cost-effective choice for model inference tasks
Learn More >
2. AI Acceleration
-
You can leverage GPU Accelerator AIACC to speed up AI training tasks by up to 70% and inference tasks by 2-3 times according to Stanford DAWN Deep Learning Benchmark.
Learn More >
3. Platform for AI
-
PAI provides an end-to-end optimization solution to including PAI-iTAG for data labeling, PAI-DSW for model building, PAI-DLC for model training, and PAI-EAS for model inference and deployment.
Learn More >
A Wide Selection of Open-Source FMs
Tongyi Qianwen (Qwen)
Alibaba Cloud has released two open-source Tongyi Qianwen models: Qwen, the large language model (including Qwen-7B and Qwen-14B), and Qwen-Chat, the chat model (including Qwen-7B-Chat and Qwen-14B-Chat). Qwen-14B and Qwen-7B outperform other baseline models of a similar size on a series of benchmark datasets (including MMLU, C-Eval, GSM8K, MATH, HumanEval, MBPP, BBH, etc.) that evaluate the models' capabilities on natural language understanding, mathematical problem solving, coding, etc. Please refer to OpenCompass Large Language Model Leaderboard or download Qwen Technical Report for details.
You can use PAI-EAS to deploy web UI applications based on the open source Qwen models, and use the web UI and API operations to perform model inference.
Llama 2
Llama is an open-source LLM. Compared to its predecessor, Llama 1, Llama 2 has more training data and a longer context length, making it more versatile and accurate in generating natural language at scale. You can run Llama 2 on PAI-EAS in WebUI with just a few clicks or use API with Python Wrapper code to set up Llama 2 service.
Stable Diffusion
Stable Diffusion is an open-source and popular cross-modal generation model that leverages advanced deep learning algorithms and techniques to generate visually compelling images based on text descriptions. You can deploy the Stable Diffusion models and create a text-to-image generation service with a few clicks in PAI-EAS.
ChatGLM
ChatGLM is an open-source LLM based on the General Language Model (GLM) architecture. ChatGLM-6B is trained on 1 trillion tokens and uses supervised fine-tuning (SFT), self-feedback system, and human feedback reinforcement learning (RLHF) to better align with human preferences. You can quickly deploy the ChatGLM application in PAI-EAS.
Deploy and Run Llama 2 Now on PAI-EAS
Alibaba Cloud Keeps Driving Innovation in Generative AI