Update: I deleted the /flink/recovery folder on HDFS and even then I get the same Exception after the next checkpoint.
2016-11-21 21:51 GMT+01:00 static-max <flasha...@googlemail.com>: > Hi Stephan, > > it's not a problem, but makes finding other errors on my NameNode > complicated as I have this error message every minute. > Can't we just delete the directory recursively? > > Regards, > Max > > 2016-10-11 17:59 GMT+02:00 Stephan Ewen <se...@apache.org>: > >> Hi! >> >> I think to some extend this is expected. There is some cleanup code that >> deletes files and then issues parent directory remove requests. It relies >> on the fact that the parent directory is only removed if it is empty (after >> the last file was deleted). >> >> Is this a problem right now, or just a confusing behavior? >> >> Greetings, >> Stephan >> >> >> On Tue, Oct 11, 2016 at 5:25 PM, static-max <flasha...@googlemail.com> >> wrote: >> >>> Hi, >>> >>> I get many (multiple times per minute) errors in my Namenode HDFS >>> logfile: >>> >>> 2016-10-11 17:17:07,596 INFO ipc.Server (Server.java:logException(2401)) >>> - IPC Server handler 295 on 8020, call >>> org.apache.hadoop.hdfs.protocol.ClientProtocol.delete >>> from datanode1:34872 Call#2361 Retry#0 >>> org.apache.hadoop.fs.PathIsNotEmptyDirectoryException: `/flink/recovery >>> is non empty': Directory is not empty >>> at org.apache.hadoop.hdfs.server.namenode.FSDirDeleteOp.delete( >>> FSDirDeleteOp.java:89) >>> at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.delete(F >>> SNamesystem.java:3829) >>> at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.del >>> ete(NameNodeRpcServer.java:1071) >>> at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServ >>> erSideTranslatorPB.delete(ClientNamenodeProtocolServerSideTr >>> anslatorPB.java:619) >>> at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocol >>> Protos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNam >>> enodeProtocolProtos.java) >>> at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcIn >>> voker.call(ProtobufRpcEngine.java:640) >>> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982) >>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2313) >>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2309) >>> at java.security.AccessController.doPrivileged(Native Method) >>> at javax.security.auth.Subject.doAs(Subject.java:422) >>> at org.apache.hadoop.security.UserGroupInformation.doAs(UserGro >>> upInformation.java:1724) >>> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2307) >>> >>> That is the directory I configured for Jobmanager HA. I deleted it >>> before starting the YARN session but that did not help. The folder gets >>> created by Flink without problems. >>> >>> I'm using latest Flink Master (Commit: 6731ec1) and build it for Hadoop >>> 2.7.3. >>> >>> Any idea is highly appreciated. Thanks a lot! >>> >> >> >