Hi, the exception says "org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException): Failed to CREATE_FILE /y=2017/m=01/d=10/H=18/M=12/_part-0-0.in-progress for DFSClient_NONMAPREDUCE_1062142735_3". I would assume that your output format tries to create a file that already exists. Maybe you need to check the code to generate filenames.
Best, Fabian 2017-01-11 3:13 GMT+01:00 Biswajit Das <biswajit...@gmail.com>: > Hello , > > I have to create a custom Parquet writer with rolling sink , I'm seeing > error like this , I'm expecting every partition should write in a new file > ?? Any tips ? > > --------------- > 18:12:12.551 [flink-akka.actor.default-dispatcher-5] DEBUG > akka.event.EventStream - shutting down: StandardOutLogger started > Exception in thread "main" > org.apache.flink.runtime.client.JobExecutionException: > Job execution failed. > at org.apache.flink.runtime.jobmanager.JobManager$$ > anonfun$handleMessage$1$$anonfun$applyOrElse$8.apply$ > mcV$sp(JobManager.scala:822) > at org.apache.flink.runtime.jobmanager.JobManager$$ > anonfun$handleMessage$1$$anonfun$applyOrElse$8.apply(JobManager.scala:768) > at org.apache.flink.runtime.jobmanager.JobManager$$ > anonfun$handleMessage$1$$anonfun$applyOrElse$8.apply(JobManager.scala:768) > at scala.concurrent.impl.Future$PromiseCompletingRunnable. > liftedTree1$1(Future.scala:24) > at scala.concurrent.impl.Future$PromiseCompletingRunnable.run( > Future.scala:24) > at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:41) > at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec( > AbstractDispatcher.scala:401) > at scala.concurrent.forkjoin.ForkJoinTask.doExec( > ForkJoinTask.java:260) > at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue. > pollAndExecAll(ForkJoinPool.java:1253) > at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue. > runTask(ForkJoinPool.java:1346) > at scala.concurrent.forkjoin.ForkJoinPool.runWorker( > ForkJoinPool.java:1979) > at scala.concurrent.forkjoin.ForkJoinWorkerThread.run( > ForkJoinWorkerThread.java:107) > Caused by: java.lang.RuntimeException: Could not forward element to next > operator > at org.apache.flink.streaming.runtime.tasks.OperatorChain$ > CopyingChainingOutput.collect(OperatorChain.java:376) > at org.apache.flink.streaming.runtime.tasks.OperatorChain$ > CopyingChainingOutput.collect(OperatorChain.java:358) > at org.apache.flink.streaming.runtime.tasks.OperatorChain$ > BroadcastingOutputCollector.collect(OperatorChain.java:399) > at org.apache.flink.streaming.runtime.tasks.OperatorChain$ > BroadcastingOutputCollector.collect(OperatorChain.java:381) > at org.apache.flink.streaming.api.operators.AbstractStreamOperator$ > CountingOutput.collect(AbstractStreamOperator.java:346) > at org.apache.flink.streaming.api.operators.AbstractStreamOperator$ > CountingOutput.collect(AbstractStreamOperator.java:329) > at org.apache.flink.streaming.api.operators.StreamSource$ > NonTimestampContext.collect(StreamSource.java:161) > at org.apache.flink.streaming.connectors.kafka.internals. > AbstractFetcher.emitRecord(AbstractFetcher.java:225) > at org.apache.flink.streaming.connectors.kafka.internals. > SimpleConsumerThread.run(SimpleConsumerThread.java:379) > Caused by: org.apache.hadoop.ipc.RemoteException(org.apache. > hadoop.hdfs.protocol.AlreadyBeingCreatedException): Failed to CREATE_FILE > /y=2017/m=01/d=10/H=18/M=12/_part-0-0.in-progress for > DFSClient_NONMAPREDUCE_1062142735_3 >