JIAN WANG created FLINK-18592:
---------------------------------

             Summary: StreamingFileSink fails due to truncating HDFS file 
failure
                 Key: FLINK-18592
                 URL: https://issues.apache.org/jira/browse/FLINK-18592
             Project: Flink
          Issue Type: Bug
    Affects Versions: 1.10.1
            Reporter: JIAN WANG


I meet the issue on flink-1.10.1. I use flink on YARN(3.0.0-cdh6.3.2) with 
StreamingFileSink. 

code part like this:

*public* *static* <IN> StreamingFileSink<IN> build(String dir, 
BucketAssigner<IN, String> assigner, String prefix)

{                  return StreamingFileSink.forRowFormat(new Path(dir), new 
SimpleStringEncoder<IN>())                          
.withRollingPolicy(DefaultRollingPolicy._builder_()                             
                                               
.withRolloverInterval(TimeUnit.HOURS.toMillis(2))                               
                                              
.withInactivityInterval(TimeUnit.MINUTES.toMillis(10))                          
                                          .withMaxPartSize(1024L * 1024L * 
1024L * 50) // Max 50GB                                                         
  .build())                                  .withBucketAssigner(assigner)      
                                                                                
                     
.withOutputFileConfig(OutputFileConfig._builder_().withPartPrefix(prefix).build())
                       .build();     }

 

The error is 

java.io.IOException: Problem while truncating file: 
hdfs:///business_log/hashtag/2020-06-25/.hashtag-122-37.inprogress.8e65f69c-b5ba-4466-a844-ccc0a5a93de2

Due to this issue, it can not restart from the latest checkpoint and savepoint.

 

Currently, my workaround is that we keep latest 3 checkpoint, and if it fails, 
I manually restart from penult checkpoint.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to