Try 'sink.batch.size' = '100000'
[email protected] <[email protected]> 于2021年5月28日周五 下午4:05写道:
> 我使用FlinkCDC读取Mysql数据插入Doris,数据总数插入500条左右后会停止写入。
> 我尝试使用同样的方法写入Mysql,数据是正常的,排除Flink的问题。
> 希望得到大佬们帮助,谢谢!
>
> Translation:
> I use flinkcdc to read MySQL data and insert it into Doris. After about
> 500 pieces of data are inserted, writing will stop.
> I try to use the same method to write mysql, the data is normal, and the
> problem of Flink is excluded.
> Hope to get your help, thank you!
>
> doris table:
> CREATE TABLE `ods_business_order` (
> `id` bigint(20) NOT NULL COMMENT '主键',
> ...
> )
> UNIQUE KEY(`id`)
> DISTRIBUTED BY HASH(`id`) BUCKETS 10
> PROPERTIES("replication_num" = "1");
>
>
> flinksql sink table:
> CREATE TABLE `doris_business_order` (
> `id` bigint,
> ...
> ) WITH (
> 'connector' = 'doris',
> 'fenodes' = 'XXX',
> 'table.identifier' = 'stage_order.ods_business_order',
> 'username' = 'XXX',
> 'password' = 'XXX',
> 'sink.batch.size' = '1'
> );
>
>
> flinksql source table:
> CREATE TABLE `mysql_business_order` (
> `id` bigint ,
> ...
> ) WITH (
> 'connector' = 'mysql-cdc',
> 'hostname' = 'XXX',
> 'port' = '3306',
> 'username' = 'XXX',
> 'password' = 'XXX',
> 'database-name' = 'mcs_order100000054',
> 'table-name' = 'business_order2021[0-9]*'
> );
>
>
>
>
> [email protected]
>