J.Andreina created HDFS-6753:
--------------------------------

             Summary: When one the Disk is full and all the volumes configured 
are unhealthy , then Datanode is not considering it as failure and datanode 
process is not shutting down .
                 Key: HDFS-6753
                 URL: https://issues.apache.org/jira/browse/HDFS-6753
             Project: Hadoop HDFS
          Issue Type: Bug
            Reporter: J.Andreina


Env Details :
=============
Cluster has 3 Datanode
Cluster installed with "Rex" user
dfs.datanode.failed.volumes.tolerated  = 3
dfs.blockreport.intervalMsec                  = 18000
dfs.datanode.directoryscan.interval     = 120
DN_XX1.XX1.XX1.XX1 data dir                         = 
/mnt/tmp_Datanode,/home/REX/data/dfs1/data,/home/REX/data/dfs2/data,/opt/REX/dfs/data
 
 
/home/REX/data/dfs1/data,/home/REX/data/dfs2/data,/opt/REX/dfs/data - 
permission is denied ( hence DN considered the volume as failed )
 
Expected behavior is observed when disk is not full:
========================================
 
Step 1: Change the permissions of /mnt/tmp_Datanode to root
 
Step 2: Perform write operations ( DN detects that all Volume configured is 
failed and gets shutdown )
 
Scenario 1: 
===========
 
Step 1 : Make /mnt/tmp_Datanode disk full and change the permissions to root
Step 2 : Perform client write operations ( disk full exception is thrown , but 
Datanode is not getting shutdown ,  eventhough all the volume configured has 
failed)
 
{noformat}
 
2014-07-21 14:10:52,814 ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: 
XX1.XX1.XX1.XX1:50010:DataXceiver error processing WRITE_BLOCK operation  src: 
/XX2.XX2.XX2.XX2:10106 dst: /XX1.XX1.XX1.XX1:50010
 
org.apache.hadoop.util.DiskChecker$DiskOutOfSpaceException: Out of space: The 
volume with the most available space (=4096 B) is less than the block size 
(=134217728 B).
 
at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.RoundRobinVolumeChoosingPolicy.chooseVolume(RoundRobinVolumeChoosingPolicy.java:60)
 
{noformat}
 
Observations :
==============
1. Write operations does not shutdown Datanode , eventhough all the volume 
configured is failed ( When one of the disk is full and for all the disk 
permission is denied)
 
2. Directory scannning fails , still DN is not getting shutdown
 
 
 
{noformat}
 
2014-07-21 14:13:00,180 WARN 
org.apache.hadoop.hdfs.server.datanode.DirectoryScanner: Exception occured 
while compiling report: 
 
java.io.IOException: Invalid directory or I/O error occurred for dir: 
/mnt/tmp_Datanode/current/BP-1384489961-XX2.XX2.XX2.XX2-845784615183/current/finalized
 
at org.apache.hadoop.fs.FileUtil.listFiles(FileUtil.java:1164)
 
at 
org.apache.hadoop.hdfs.server.datanode.DirectoryScanner$ReportCompiler.compileReport(DirectoryScanner.java:596)
 
{noformat}



--
This message was sent by Atlassian JIRA
(v6.2#6252)

Reply via email to