Hi,
Which version of Flink is this? It cannot recover because it expects more data
to have been written than is there, which seems to indicate that flushing did
not work correctly.
Best,
Aljoscha
> On 19. Dec 2017, at 00:40, xiatao123 wrote:
>
> Hi Das,
> Have you got your .pending issue re
Hi Das,
Have you got your .pending issue resolved? I am running into the same
issue where the parquet files are all in pending status.
Please help to share your solutions.
Thanks,
Tao
--
Sent from: http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/
Hi Stefan ,
My bad , I'm really sorry. I have copied wrong exception stack , during the
recovery after error I'm seeing below exception
Caused by:
org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.HadoopIllegalArgumentException):
Cannot truncate to a larger file size. Current size: 3113238
Hi,
I think there are two different things mixed up in your analysis. The stack
trace that you provided is caused by a failing checkpoint - in writing, not in
reading. It seems to fail from a Timeout of your HDFS connection. This close
method has also nothing to do with the close method in the