[
https://issues.apache.org/jira/browse/HDFS-17741?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
liuguanghua resolved HDFS-17741.
--------------------------------
Assignee: liuguanghua
Resolution: Invalid
> The old blocks should be added into MarkedDeleteQueue when create a file with
> overwrite
> ---------------------------------------------------------------------------------------
>
> Key: HDFS-17741
> URL: https://issues.apache.org/jira/browse/HDFS-17741
> Project: Hadoop HDFS
> Issue Type: Bug
> Reporter: liuguanghua
> Assignee: liuguanghua
> Priority: Major
> Labels: pull-request-available
>
> When create a file with overwrite, the old blocks will not delete immediately
> because these blocks not add into MarkedDeleteQueue.
> This will have two consequences
> (1)The block will delete when datanode block report comes.
> (2) If the block is in the PendingReconstructionBlocks, it will be
> timeouted after 300s.
>
--
This message was sent by Atlassian Jira
(v8.20.10#820010)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]