Update: I deleted the /flink/recovery folder on HDFS and even then I get
the same Exception after the next checkpoint.

2016-11-21 21:51 GMT+01:00 static-max <flasha...@googlemail.com>:

> Hi Stephan,
>
> it's not a problem, but makes finding other errors on my NameNode
> complicated as I have this error message every minute.
> Can't we just delete the directory recursively?
>
> Regards,
> Max
>
> 2016-10-11 17:59 GMT+02:00 Stephan Ewen <se...@apache.org>:
>
>> Hi!
>>
>> I think to some extend this is expected. There is some cleanup code that
>> deletes files and then  issues parent directory remove requests. It relies
>> on the fact that the parent directory is only removed if it is empty (after
>> the last file was deleted).
>>
>> Is this a problem right now, or just a confusing behavior?
>>
>> Greetings,
>> Stephan
>>
>>
>> On Tue, Oct 11, 2016 at 5:25 PM, static-max <flasha...@googlemail.com>
>> wrote:
>>
>>> Hi,
>>>
>>> I get many (multiple times per minute) errors in my Namenode HDFS
>>> logfile:
>>>
>>> 2016-10-11 17:17:07,596 INFO  ipc.Server (Server.java:logException(2401))
>>> - IPC Server handler 295 on 8020, call 
>>> org.apache.hadoop.hdfs.protocol.ClientProtocol.delete
>>> from datanode1:34872 Call#2361 Retry#0
>>> org.apache.hadoop.fs.PathIsNotEmptyDirectoryException: `/flink/recovery
>>> is non empty': Directory is not empty
>>>         at org.apache.hadoop.hdfs.server.namenode.FSDirDeleteOp.delete(
>>> FSDirDeleteOp.java:89)
>>>         at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.delete(F
>>> SNamesystem.java:3829)
>>>         at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.del
>>> ete(NameNodeRpcServer.java:1071)
>>>         at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServ
>>> erSideTranslatorPB.delete(ClientNamenodeProtocolServerSideTr
>>> anslatorPB.java:619)
>>>         at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocol
>>> Protos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNam
>>> enodeProtocolProtos.java)
>>>         at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcIn
>>> voker.call(ProtobufRpcEngine.java:640)
>>>         at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982)
>>>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2313)
>>>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2309)
>>>         at java.security.AccessController.doPrivileged(Native Method)
>>>         at javax.security.auth.Subject.doAs(Subject.java:422)
>>>         at org.apache.hadoop.security.UserGroupInformation.doAs(UserGro
>>> upInformation.java:1724)
>>>         at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2307)
>>>
>>> That is the directory I configured for Jobmanager HA. I deleted it
>>> before starting the YARN session but that did not help. The folder gets
>>> created by Flink without problems.
>>>
>>> I'm using latest Flink Master (Commit: 6731ec1) and build it for Hadoop
>>> 2.7.3.
>>>
>>> Any idea is highly appreciated. Thanks a lot!
>>>
>>
>>
>

Reply via email to