[ https://issues.apache.org/jira/browse/HUDI-3693?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
chacha.tang reassigned HUDI-3693: --------------------------------- Assignee: chacha.tang > error will be occured when writing data into hudi using flink job more than > once in same flink session > ------------------------------------------------------------------------------------------------------ > > Key: HUDI-3693 > URL: https://issues.apache.org/jira/browse/HUDI-3693 > Project: Apache Hudi > Issue Type: Bug > Components: flink > Reporter: chacha.tang > Assignee: chacha.tang > Priority: Minor > > 环境: > hudi: 0.10.1 > flink:1.13.2 > flink cluster setup in standalone mode, and use the below flink sql to launch > the flink job > #create the cow sink table > CREATE TABLE t1( > uuid VARCHAR(20) PRIMARY KEY NOT ENFORCED, > name VARCHAR(10), > age INT, > ts TIMESTAMP(3), > `partition` VARCHAR(20) > ) > PARTITIONED BY (`partition`) > WITH ( > 'connector' = 'hudi', > 'path' = '/user/hive/warehouse/hudi.db/t1', > 'write.tasks' = '1', > 'table.type' = 'COPY_ON_WRITE' > ); > #write data first > INSERT INTO t1 VALUES ('id1','Danny',20,TIMESTAMP '1970-01-01 > 00:00:01','par1') > #write data in twice > INSERT INTO t1 VALUES ('id1','Danny',20,TIMESTAMP '1970-01-01 > 00:00:01','par1') > the first time the flink job is finished successfully, but it failed in > twice, and the below exception occured > org.apache.hudi.common.fs.HoodieWrapperFileSystem cannot be cast to > org.apache.hudi.common.fs.HoodieWrapperFileSystem > I search issues, and found issue#3885 mentioned same problem, But the root > casuse is different.My problem is involed by hadoop fs cache. The > HoodieWrapperFileSystem is cached in fs cache when it create by first flink > job in flink taskmanger.When I launched the second flink job to write data, > it used cached HoodieWrapperFileSystem and whose childFlinkClassload is > different with this flink job.So the exception occured. -- This message was sent by Atlassian Jira (v8.20.1#820001)