This could be related to https://issues.apache.org/jira/browse/FLINK-5063 where some issues related to the cleanup of checkpointing files were fixed.
-Max On Mon, Nov 21, 2016 at 10:05 PM, static-max <flasha...@googlemail.com> wrote: > Update: I deleted the /flink/recovery folder on HDFS and even then I get the > same Exception after the next checkpoint. > > 2016-11-21 21:51 GMT+01:00 static-max <flasha...@googlemail.com>: >> >> Hi Stephan, >> >> it's not a problem, but makes finding other errors on my NameNode >> complicated as I have this error message every minute. >> Can't we just delete the directory recursively? >> >> Regards, >> Max >> >> 2016-10-11 17:59 GMT+02:00 Stephan Ewen <se...@apache.org>: >>> >>> Hi! >>> >>> I think to some extend this is expected. There is some cleanup code that >>> deletes files and then issues parent directory remove requests. It relies >>> on the fact that the parent directory is only removed if it is empty (after >>> the last file was deleted). >>> >>> Is this a problem right now, or just a confusing behavior? >>> >>> Greetings, >>> Stephan >>> >>> >>> On Tue, Oct 11, 2016 at 5:25 PM, static-max <flasha...@googlemail.com> >>> wrote: >>>> >>>> Hi, >>>> >>>> I get many (multiple times per minute) errors in my Namenode HDFS >>>> logfile: >>>> >>>> 2016-10-11 17:17:07,596 INFO ipc.Server >>>> (Server.java:logException(2401)) - IPC Server handler 295 on 8020, call >>>> org.apache.hadoop.hdfs.protocol.ClientProtocol.delete from datanode1:34872 >>>> Call#2361 Retry#0 >>>> org.apache.hadoop.fs.PathIsNotEmptyDirectoryException: `/flink/recovery >>>> is non empty': Directory is not empty >>>> at >>>> org.apache.hadoop.hdfs.server.namenode.FSDirDeleteOp.delete(FSDirDeleteOp.java:89) >>>> at >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.delete(FSNamesystem.java:3829) >>>> at >>>> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.delete(NameNodeRpcServer.java:1071) >>>> at >>>> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.delete(ClientNamenodeProtocolServerSideTranslatorPB.java:619) >>>> at >>>> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) >>>> at >>>> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:640) >>>> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982) >>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2313) >>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2309) >>>> at java.security.AccessController.doPrivileged(Native Method) >>>> at javax.security.auth.Subject.doAs(Subject.java:422) >>>> at >>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1724) >>>> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2307) >>>> >>>> That is the directory I configured for Jobmanager HA. I deleted it >>>> before starting the YARN session but that did not help. The folder gets >>>> created by Flink without problems. >>>> >>>> I'm using latest Flink Master (Commit: 6731ec1) and build it for Hadoop >>>> 2.7.3. >>>> >>>> Any idea is highly appreciated. Thanks a lot! >>> >>> >> >