Tao Jie created HDFS-13769: ------------------------------ Summary: Namenode gets stuck when deleting large dir in trash Key: HDFS-13769 URL: https://issues.apache.org/jira/browse/HDFS-13769 Project: Hadoop HDFS Issue Type: Improvement Affects Versions: 3.1.0, 2.8.2 Reporter: Tao Jie Assignee: Tao Jie
Similar to the situation discussed in HDFS-13671, Namenode gets stuck for a long time when deleting trash dir with a large mount of data. We found log in namenode: {quote} 2018-06-08 20:00:59,042 INFO namenode.FSNamesystem (FSNamesystemLock.java:writeUnlock(252)) - FSNamesystem write lock held for 23018 ms via java.lang.Thread.getStackTrace(Thread.java:1552) org.apache.hadoop.util.StringUtils.getStackTrace(StringUtils.java:1033) org.apache.hadoop.hdfs.server.namenode.FSNamesystemLock.writeUnlock(FSNamesystemLock.java:254) org.apache.hadoop.hdfs.server.namenode.FSNamesystem.writeUnlock(FSNamesystem.java:1567) org.apache.hadoop.hdfs.server.namenode.FSNamesystem.delete(FSNamesystem.java:2820) org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.delete(NameNodeRpcServer.java:1047) {quote} One simple solution is to avoid deleting large data in one delete RPC call. We implement a trashPolicy that divide the delete operation into several delete RPCs, and each single deletion would not delete too many files. Any thought? [~linyiqun] -- This message was sent by Atlassian JIRA (v7.6.3#76005) --------------------------------------------------------------------- To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org