Hi vikas,

Please include your exclude file in hdfs-site.xml under the property 
dfs.hosts.exclude.
please make note that this file must be included at the namenode startup and 
then u hv to edit it for the datanodes you want to exclude. Secondly, after 
editin it for the ip:port of the retiring datanode, use the command bin/hadoop 
dfsadmin -refreshNodes. 

also, Decommission does not happen momentarily since it requires replication 
of potentially a large number of blocks and we do not want the cluster 
to be overwhelmed with just this one job. The decommission progress can 
be monitored on the name-node Web UI. Until all blocks are replicated 
the node will be in "Decommission In Progress" state. When decommission 
is done the state will change to "Decommissioned". The nodes can be 
removed whenever decommission is finished.



The decommission process can be terminated at any time by editing the 
configuration or the exclude files and repeating the -refreshNodes command.


hope it helps.

*------------------------*

Cheers !!!

Siddharth Tiwari
TCS world wide Data warehouse and Analytic Team - Americas

Have a refreshing day !!!


Date: Mon, 19 Sep 2011 12:50:13 +0530
Subject: Decommission of datanode(Urgent)
From: vikas.srivast...@one97.net
To: user@hive.apache.org; ayonsi...@yahoo.com; viral.baja...@gmail.com
CC: nitin2.ku...@one97.net; abhinav.mis...@one97.net

Hey folks !!

i tried to decommission datanode from hadoop cluster... steps i followed

1: add  this in core site
 <property>
    <name>dfs.hosts.exclude</name>
    <value>/home/hadoop/excludes</value>
    <final>true</final>
  </property>


1: add  this in mapred-site
  <property>
    <name>mapred.hosts.exclude</name>

    <value>/home/hadoop/excludes</value>
    <final>true</final>
  </property>
3:create a excludes file and add ip:port in that


exp: 10.0.3.31:50010

4: run cmd

hadoop dfsadmin -refreshNodes


5: After that my live nodes became 0 and all nodes became dead.. i checked 
namenode logs where i found these error msgs


2011-09-19 12:33:47,695 INFO org.apache.hadoop.ipc.Server: IPC Server handler 
24 on 9000, call sendHeartbeat(DatanodeRegistration(10.0.3.16:50010, 
storageID=DS-1703098060-10.0.3.16-50010-1298269611944, infoPort=50075, 
ipcPort=50020), 2012206694400, 1650194042865, 271003275264, 0, 1) from 
10.0.3.16:38587: error: 
org.apache.hadoop.hdfs.server.protocol.DisallowedDatanodeException: Datanode 
denied communication with namenode: 10.0.3.16:50010

org.apache.hadoop.hdfs.server.protocol.DisallowedDatanodeException: Datanode 
denied communication with namenode: 10.0.3.16:50010
        at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.handleHeartbeat(FSNamesystem.java:2235)

        at 
org.apache.hadoop.hdfs.server.namenode.NameNode.sendHeartbeat(NameNode.java:704)
        at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)

        at java.lang.reflect.Method.invoke(Method.java:597)
        at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
        at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959)
        at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955)

        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:396)
        at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953)
2011-09-19 12:33:47,701 INFO org.apache.hadoop.ipc.Server: IPC Server handler 7 
on 9000, call sendHeartbeat(DatanodeRegistration(10.0.5.36:50010, 
storageID=DS-809855347-10.0.5.36-50010-1316252293924, infoPort=50075, 
ipcPort=50020), 1938687860736, 1390486994944, 457712619520, 0, 1) from 
10.0.5.36:58924: error: 
org.apache.hadoop.hdfs.server.protocol.DisallowedDatanodeException: Datanode 
denied communication with namenode: 10.0.5.36:50010

org.apache.hadoop.hdfs.server.protocol.DisallowedDatanodeException: Datanode 
denied communication with namenode: 10.0.5.36:50010
        at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.handleHeartbeat(FSNamesystem.java:2235)

        at 
org.apache.hadoop.hdfs.server.namenode.NameNode.sendHeartbeat(NameNode.java:704)
        at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)

        at java.lang.reflect.Method.invoke(Method.java:597)
        at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
        at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959)
        at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955)

        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:396)
        at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953)





please suggest any help would be appreciated!!!!!!!!!!


-- 
With Regards

Vikas Srivastava

DWH & Analytics TeamMob:+91 9560885900
One97 | Let's get talking !
                                          

Reply via email to