All Products
Search
Document Center

Realtime Compute for Apache Flink:Feature overview

Last Updated:Sep 30, 2025

Data development

Category

Feature

Description

Reference

Development

Developing SQL drafts

When you write code for an SQL draft in the Realtime Compute for Apache Flink console, you can use various built-in connectors provided by the service, user-defined functions (UDFs), and custom connectors.

Developing a PyFlink program for Python deployment

You can develop your Python program in an on-premises environment, and deploy it and start the deployment in the development console of Realtime Compute for Apache Flink.

Develop a Python API draft

Developing and packaging a Flink program for JAR deployment

Realtime Compute for Apache Flink DataStream provides flexible programming models and APIs, which allow you to specify data conversions, perform operations, and configure operators to meet requirements for complex business logic and data processing.

Develop a JAR draft

Scripts

A script can contain the CALL command and DDL, DQL, and DML statements. You can use scripts to create and manage catalogs and tables, perform data queries, and manage Apache Paimon tables. You can use the Explain statement in a script to view execution plans and troubleshoot issues.

Scripts

Function management

You can use a UDF in developing an SQL draft.

Manage UDFs

Draft debugging

You can enable the debugging feature to simulate deployment running, check outputs, and verify the business logic of SELECT and INSERT statements. This feature improves the development efficiency and reduces the risks of poor data quality.

Debug a draft

Data access

Connectors

Realtime Compute for Apache Flink provides a variety of built-in connectors to read data from upstream data stores, write data to downstream data stores, and synchronize data. Realtime Compute for Apache Flink also allows you to upload and use custom connectors.

Data formats

Realtime Compute for Apache Flink supports multiple data formats that can be used with connectors. A data format defines how to map binary data onto table columns.

Data formats

Metadata Management

Built-in catalogs

After you create a catalog, you can directly read metadata in the development console of Realtime Compute for Apache Flink without the need to manually register tables. This improves data development efficiency and data accuracy.

Custom catalogs

If built-in catalogs cannot meet your business requirements, you can use custom catalogs.

Manage custom catalogs

O&M

Category

Feature

Description

Reference

Deployment management

Deployment creation

Deployment effectively isolates development from production. Creating a deployment does not affect existing, running jobs. A job is officially submitted after the deployment is started or restarted.

Create a deployment

Deployment configuration

You must configure a deployment before starting it.

Configure a deployment

Dynamic parameter update

Dynamic update of the parameter configuration of a Flink deployment can make the parameter configuration take effect more quickly. This helps reduce the service interruption time caused by deployment startup and cancellation and facilitates dynamic scaling of TaskManagers and checkpoint-based troubleshooting.

Dynamically update the parameter configuration for dynamic scaling

Start a deployment

After you develop and deploy a draft, you must start the deployment on the Deployments page to run the deployment.

Start a deployment

Job logs

If an exception occurs at deployment startup or at runtime, you can view the logs for troubleshooting.

State management

The checkpoints and savepoints of a deployment are collectively called its state set, which is used to manage the lifecycle of the deployment.

State management

Deployment diagnostics and optimization

The intelligent deployment diagnostics feature is used to monitor the health status of a deployment, analyze and diagnose exception logs, exceptions, and risks of a deployment, and provide optimization suggestions based on the diagnostic results. Flink deployments support two automatic tuning modes to allocate resources: Autopilot and scheduled tuning.

Deployment diagnostics and optimization

Data lineage

You can use the data lineage feature to track data in a deployment. This way, you can efficiently manage and optimize data streams, quickly identify issues, and evaluate the impact of the issues.

View data lineage

Monitoring and alerting

Realtime Compute for Apache Flink allows you to use CloudMonitor (a free monitoring service) or Application Real-Time Monitoring Service (ARMS) to implement deployment monitoring and alerting. You can configure metric alert rules or subscribe to event-triggered alerts, detecting and handling exceptions at the earliest opportunity.

Deployment monitoring and alerting

Workflows

Workflow management

A workflow is a directed acyclic graph (DAG) that you can create by dragging tasks and establishing associations among tasks. If you want to run tasks at specific points in time, you can create a workflow and configure tasks and scheduling policies in the workflow.

Manage workflows

Workflow instance management

After you create a workflow, you can perform operations on existing workflow instances and task instances.

Manage workflow instances and task instances

Resources

Resources

Realtime Compute for Apache Flink creates directories in the Object Storage Service (OSS) bucket that is associated with your Realtime Compute for Apache Flink workspace when you purchase the workspace. The directories are used to store data files (checkpoints, savepoints, logs, and JAR packages, etc.) that are required for running a deployment or generated at run time.

Artifacts

Queue management

Queue management

You can isolate and manage resources through queue management.

Manage queues

Security service

Category

Feature

Description

Reference

Security center

Hive Kerberos

You can register a Hive cluster that supports Kerberos authentication in the console to allow Realtime Compute for Apache Flink deployments to access the Hive cluster.

Register a Hive cluster that supports Kerberos authentication

Key hosting

Realtime Compute for Apache Flink allows you to configure variables and keys to prevent security risks that are caused by the use of information such as plaintext AccessKey pairs and passwords. You can reuse variables and keys to avoid repeatedly writing the same code or values and simplify configuration management. You can reference a variable or a key in various scenarios, such as the development of SQL, JAR, or Python drafts, log output configuration, and UI-based parameter configuration.

Manage variables

Permission management

Member management

To allow multiple users to collaborate in a namespace in the development console of Realtime Compute for Apache Flink, such as performing SQL development and O&M, you can add members to the namespace and assign roles that have predefined permissions to the members.

Manage Members

Role management

You can create custom roles to grant permissions based on your business requirements. This improves the flexibility of permission management and ensures security.

Role management

Workspace management

Category

Feature

Description

Reference

Namespace management

Namespace management

A namespace is a basic unit for managing deployments in Realtime Compute for Apache Flink. Configurations, deployments, and permissions of each namespace are separately managed. You can create multiple namespaces and assign separate resources and permissions to each namespace. This way, the resources and permissions of multiple tenants are isolated among namespaces.

Manage namespaces

Resource tag management

A tag consists of a tag key and a tag value. Tags are used to identify cloud resources. Tags allow you to categorize, search for, and aggregate cloud resources that have the same characteristics from different dimensions. This helps you manage cloud resources in an efficient manner.

Manage resource tags