This topic describes how to synchronize data from a user-created MySQL database to a Message Queue for Apache Kafka instance by using Data Transmission Service (DTS). This allows you to extend message processing capabilities.

Prerequisites

  • The version of the user-created MySQL database is 5.1, 5.5, 5.6, 5.7, or 8.0.
  • The version of the destination Kafka instance is 0.10 or 1.0.
  • In the destination Kafka instance, a topic is created to receive the synchronized data. For more information, see Create a topic.

Background information

Message Queue for Apache Kafka is a distributed, high-throughput, and scalable message queue service provided by Alibaba Cloud. It provides fully managed services for the open-source Apache Kafka to solve the long-standing shortcomings of open-source products. Message Queue for Apache Kafka allows you to focus on business development without the need to consider deployment and O&M. Message Queue for Apache Kafka is widely used in big data fields such as log collection, monitoring data aggregation, streaming data processing, and online and offline analysis. It has become indispensable for the big data ecosystem.

Precautions

  • DTS uses read and write resources of the source and destination databases during initial full data synchronization. This may increase the database load. If the database performance is unfavorable, the specification is low, or the data volume is large, database services may become unavailable. For example, DTS occupies a large amount of read and write resources in the following cases: a large number of slow SQL queries are performed on the source database, the tables have no primary keys, or a deadlock occurs in the destination database. Before synchronizing data, you must evaluate the performance of the source and destination databases. We recommend that you synchronize data during off-peak hours. For example, you can synchronize data when the CPU usage of the source and destination databases is less than 30%.
  • The source database must have PRIMARY KEY or UNIQUE constraints and all fields must be unique. Otherwise, duplicate data may exist in the destination database.

Limits

  • You can select only tables as the objects to be synchronized.
  • DTS does not synchronize the data in a renamed table to the destination Kafka cluster. This applies if the new table name is not included in the objects to be synchronized. To synchronize the data in a renamed table to the destination Kafka cluster, you must modify the objects to be synchronized. For more information, see Add an object to a data synchronization task.

Data format

The data that is synchronized to the Kafka cluster is stored in the Avro format. For more information, see DTS Avro schema.

After the data is synchronized to the Kafka cluster, you must parse the data based on the Avro schema.

Billing

For more information, see Pricing.

Preparations

Create an account for a user-created MySQL database and configure binary logging

Procedure

  1. Purchase a data synchronization instance. For more information, see Purchase procedure.
    Note On the buy page, set Source Instance to MySQL, set Target Instance to Kafka, and set Synchronization Topology to One-Way Synchronization.
  2. Log on to the DTS console.
  3. In the left-side navigation pane, click Data Synchronization.
  4. At the top of the Synchronization Tasks page, select the region where the destination instance resides.
    Select a region
  5. Find the data synchronization instance and click Configure Synchronization Channel in the Actions column.
  6. Configure the source and destination instances.
    Configure the source and destination instances
    Section Parameter Description
    N/A Synchronization Task Name DTS automatically generates a task name. We recommend that you specify an informative name for easy identification. You do not need to use a unique task name.
    Source Instance Details Instance Type Select RDS Instance, User-Created Database in ECS Instance, or User-Created Database Connected over Express Connect, VPN Gateway, or Smart Access Gateway based on where the source database is deployed.

    This topic uses User-Created Database in ECS Instance as an example to describe how to configure a data synchronization task. You can also follow the procedure to configure data synchronization tasks for other types of user-created MySQL databases.

    Instance Region The region of the source instance. The region is the same as the source region that you selected when you purchased the data synchronization instance. You cannot change the value of this parameter.
    ECS Instance ID Select the ID of the ECS instance that is connected to the user-created MySQL database.
    Database Type The value of this parameter is set to MySQL and cannot be changed.
    Port Number Enter the service port number of the user-created MySQL database.
    Database Account Enter the account of the user-created MySQL database. The account must have the REPLICATION SLAVE permission, the REPLICATION CLIENT permission, the SHOW VIEW permission, and the permission to perform SELECT operations on the required objects.
    Database Password Enter the password for the database account.
    Destination Instance Details Instance Type Select User-Created Database Connected over Express Connect, VPN Gateway, or Smart Access Gateway.
    Note You cannot select Message Queue for Apache Kafka as the instance type. You can use Message Queue for Apache Kafka as a user-created Kafka database to configure data synchronization.
    Instance Region The region of the destination instance. The region is the same as the destination region that you selected when you purchased the data synchronization instance. You cannot change the value of this parameter.
    Peer VPC Select the ID of the VPC to which the destination Message Queue for Apache Kafka instance belongs. To obtain the VPC ID, you can log on to the Message Queue for Apache Kafka console and navigate to the Instance Details page of the Message Queue for Apache Kafka instance. In the Basic Information section, you can view the VPC ID.
    Database Type Select Kafka.
    IP Address Enter an IP address from the Default Endpoint parameter of the Message Queue for Apache Kafka instance.
    Note To obtain an IP address, you can log on to the Message Queue for Apache Kafka console and navigate to the Instance Details page of the Message Queue for Apache Kafka instance. In the Basic Information section, you can obtain an IP address from the Default Endpoint parameter.
    Port Number Enter the service port number of the Message Queue for Apache Kafka instance. The default port number is 9092.
    Database Account Enter the username that is used to log on to the Message Queue for Apache Kafka instance.
    Note If the instance type of the Message Queue for Apache Kafka instance is VPC Instance, you do not need to specify the database account or database password.
    Database Password Enter the password for the username.
    Topic
    1. Click Get Topic List.
    2. Select a topic name from the drop-down list.
    Kafka Version Select a version based on the version of the Message Queue for Apache Kafka instance.
    Encryption Select Non-encrypted or SCRAM-SHA-256 based on your business and security requirements.
  7. In the lower-right corner of the page, click Set Whitelist and Next.
  8. Select the objects to be synchronized.
    Parameter Description
    Objects to be synchronized Select tables from the Available section and click the Right arrow icon to move the tables to the Selected section.
    Note DTS automatically maps the table names to the topic name that you select when you configure the source and destination instances. If you want to change the topic, you can move the pointer over the table and click Edit. You must specify a topic that exists in the Message Queue for Apache Kafka instance.
  9. In the lower-right corner of the page, click Next.
  10. Configure initial synchronization.
    Message Queue for Apache Kafka: Configure initial synchronization
    Parameter Description
    Initial Synchronization Select both Initial Schema Synchronization and Initial Full Data Synchronization. Before synchronizing incremental data, DTS synchronizes the schemas and historical data of the required objects from the source database to the destination database.
    Filter options Ignore DDL in incremental synchronization phase is selected by default. In this case, DTS does not synchronize DDL operations that are performed on the source database during incremental data synchronization.
  11. In the lower-right corner of the page, click Precheck.
    Note
    • Before you can start the data synchronization task, a precheck is performed. You can start the data synchronization task only after the task passes the precheck.
    • If the task fails to pass the precheck, click the Info icon icon next to each failed item to view details. Troubleshoot the issues based on the causes and run the precheck again.
  12. Close the Precheck dialog box after the following message is displayed: The precheck is passed.
  13. Wait until the initial synchronization is complete and the data synchronization task is in the Synchronizing state.
    On the Synchronization Tasks page, view the status of the data synchronization task.View the status of a data synchronization task.