All Products
Search
Document Center

Container Compute Service:Overview of Knative

Last Updated:Mar 26, 2026

Knative is an open source serverless framework built on Kubernetes. Container Compute Service (ACS) Knative is fully compatible with open source Knative and Kubernetes APIs, and adds managed hosting, predictive scaling, and deeper integration with Alibaba Cloud services.

Knative addresses three common pain points in Kubernetes-based development:

  • Deployment complexity: Knative abstracts away pods, Services, Deployments, and Ingress configuration. A single Knative Service definition handles all of it.

  • Idle resource cost: Knative scales pod counts to zero when no traffic is received and scales back out automatically when requests arrive.

  • Vendor lock-in: Because Knative runs on any Kubernetes cluster, workloads are portable across cloud providers and on-premises environments.

Key concepts

Knative consists of three components. Each can be used independently or together, so you can adopt Knative incrementally based on your needs.

  • Knative Serving handles serverless workload management: deployment, version management, canary releases, and request-based auto scaling. When no traffic is received, Knative Serving scales pods to zero.

  • Knative Eventing provides event management: connect to external event sources, register and subscribe to events, and filter events before routing them to Services or functions. It decouples event producers from event consumers.

  • Knative Functions lets you create, build, and deploy Knative Services without deep knowledge of Kubernetes, containers, or Knative internals. Deploy stateless, event-driven functions as Knative Services directly to a cluster.

How request-based auto scaling works

Knative Serving injects a queue-proxy sidecar container into each pod to collect concurrency and requests per second (RPS) metrics. The autoscaler uses these metrics—rather than CPU or memory—to adjust pod count, which more directly reflects actual service load.

When traffic drops to zero, the autoscaler switches from Serve mode (direct routing) to Proxy mode. Incoming requests are held at the Proxy while new pods start. Once the pods are Ready, the Proxy forwards the queued requests and switches back to Serve mode.

To achieve the same request-based auto scaling without Knative, you would need to create a Deployment and a Service, configure an Ingress, and configure Horizontal Pod Autoscaler (HPA) separately. A Knative Service handles all of this in a single YAML file.

How version management and canary releases work

Every Knative Service has two underlying objects:

  • Configuration: represents the current desired state of the Service. Each update to the Service creates a new, uniquely identified revision.

  • Route: maps traffic to one or more revisions with configurable proportions.

This model supports rolling back to a previous revision and splitting traffic across revisions for canary releases or A/B tests. For example, you can route 70% of traffic to Revision V1 and 30% to Revision V2.

image

Why ACS Knative

ACS Knative is fully compatible with open source Knative and Kubernetes APIs, and extends it in several areas.

image

Capability as a Service

ACS Knative lets you deploy applications with a few clicks, without purchasing resources to build a system. A console with visual operations simplifies the use of Kubernetes clusters and Knative.

Managed components and Ingress

In ACS clusters, Knative Serving and Knative Eventing are hosted and managed by ACS with high availability at no extra charge. ACS also provides four Ingress options:

Ingress Description
Application Load Balancer (ALB) Fully managed, O&M-free; backed by the Alibaba Cloud ALB service with built-in auto scaling
Microservices Engine (MSE) Next-generation Ingress complying with Kubernetes Ingress standards; supports all features of traditional traffic gateways
Service Mesh (ASM) Istio-compatible; centralized traffic management for microservices with mesh observability and secure inter-service communication
Kourier Open source, lightweight Ingress based on the Envoy architecture; compatible with the open source Knative version

ALB, MSE, and ASM Ingress controllers are created and hosted by ACS. Kourier is compatible with the open source version.

Reduced cold start time

Open source Knative scales pods to zero during off-peak hours to save costs, but the next request incurs a cold start. ACS Knative offers two ways to reduce this latency:

  • Reserved instances: Reserve a low-specification burstable instance to keep a pod warm. This balances cost savings with startup time.

  • Predictive scaling with Advanced Horizontal Pod Autoscaler (AHPA): AHPA analyzes historical metrics to predict demand and pre-provisions resources before traffic peaks. Use this when your application workload fluctuates on a regular schedule.

For a full comparison, see Comparison between Alibaba Cloud Knative and open source Knative.

Ecosystem integration

ACS Knative integrates with Alibaba Cloud observability and application services without requiring you to provision or manage additional servers:

Use cases

Use case Description
Web service hosting Deploy web services with simplified configuration. The revision mechanism gives each version a unique identifier for rollback and traffic splitting. Use Route-based traffic proportions to run canary releases and A/B tests.
Serverless applications Let developers focus on business logic rather than IaaS resources. ACS Knative scales out based on request concurrency and scales pods to zero when traffic stops, reducing idle costs.
AI inference services In GPU-accelerated scenarios, teams can focus on AI model development and deployment rather than infrastructure maintenance. ACS Knative scales inference services based on actual workloads and can run on cloud nodes, in data centers, or on edge devices.
Event-driven workloads Connect to external event sources using Knative Eventing. For example, configure an event source to receive IoT sensor data and trigger downstream processing logic such as data storage, real-time analysis, and alerting. Supported sources include Apache Kafka, GitHub, MNS, and Apache RocketMQ.

Get started

Prerequisites: Deploy Knative in your ACS cluster before proceeding.

Set up components and Ingress

Configure auto scaling

Manage versions and traffic

Access Knative Services

The default domain for a Knative Service follows the format {route}.{namespace}.{default-example.com}.

Advanced features

Observe and monitor

  • View the Knative dashboard — connect to Managed Service for Prometheus to view response latency and request concurrency metrics.

Billing

ACS Knative itself is free. You are charged only for the cloud resources created when you use it, such as SLB instances and NAT gateways. For details, see Billing rules.

FAQ

See Knative FAQ for common issues. Frequently asked questions include:

Contact us

For questions or feedback about Knative, join DingTalk group 23302777.

What's next