You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hudi.apache.org by "chacha.tang (Jira)" <ji...@apache.org> on 2022/03/23 10:09:00 UTC

[jira] [Assigned] (HUDI-3693) error will be occured when writing data into hudi using flink job more than once in same flink session

     [ https://issues.apache.org/jira/browse/HUDI-3693?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

chacha.tang reassigned HUDI-3693:
---------------------------------

    Assignee: chacha.tang

> error will be occured when writing data into hudi using flink job more than once in same flink session
> ------------------------------------------------------------------------------------------------------
>
>                 Key: HUDI-3693
>                 URL: https://issues.apache.org/jira/browse/HUDI-3693
>             Project: Apache Hudi
>          Issue Type: Bug
>          Components: flink
>            Reporter: chacha.tang
>            Assignee: chacha.tang
>            Priority: Minor
>
> 环境:
> hudi: 0.10.1
> flink:1.13.2
> flink cluster setup in standalone mode, and use the below flink sql to launch the flink job
> #create the cow sink table
> CREATE TABLE t1(
>   uuid VARCHAR(20) PRIMARY KEY NOT ENFORCED,
>   name VARCHAR(10),
>   age INT,
>   ts TIMESTAMP(3),
>   `partition` VARCHAR(20)
> )
> PARTITIONED BY (`partition`)
> WITH (
>   'connector' = 'hudi',
>   'path' = '/user/hive/warehouse/hudi.db/t1',
>   'write.tasks' = '1',
>   'table.type' = 'COPY_ON_WRITE'
> );
> #write data first
> INSERT INTO t1 VALUES ('id1','Danny',20,TIMESTAMP '1970-01-01 00:00:01','par1')
> #write data  in twice
> INSERT INTO t1 VALUES ('id1','Danny',20,TIMESTAMP '1970-01-01 00:00:01','par1')
> the first time the flink job is finished successfully, but it failed in twice, and the below exception occured 
> org.apache.hudi.common.fs.HoodieWrapperFileSystem cannot be cast to org.apache.hudi.common.fs.HoodieWrapperFileSystem
> I search issues, and found issue#3885 mentioned same problem, But the root casuse is different.My problem is involed by hadoop fs cache. The HoodieWrapperFileSystem is cached in fs cache when it create by first flink job in flink taskmanger.When I launched the second flink job to write data, it used cached HoodieWrapperFileSystem and whose childFlinkClassload is different with this flink job.So the exception occured.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)