[ https://issues.apache.org/jira/browse/HDFS-2076?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Tsz Wo Nicholas Sze resolved HDFS-2076. --------------------------------------- Resolution: Cannot Reproduce Resolving as "Cannot Reproduce". Please feel free to reopen it if it is not the case. > ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(1 > ----------------------------------------------------------------------------- > > Key: HDFS-2076 > URL: https://issues.apache.org/jira/browse/HDFS-2076 > Project: Hadoop HDFS > Issue Type: Bug > Components: datanode > Affects Versions: 0.20.2 > Environment: hadoop -hdfs > Reporter: chakali ranga swamy > > see sir > datanode log socket and datasteam problem unable to upload text file to DFS i > deleted tmp folders dfs and mapred again i formated "hadoop namenode -format" > start-all.sh done then > dfs folder contains: > data node ,name node,secondarynamenode > mapred: empty > about space:----------------- > linux-8ysi:/etc/hadoop/hadoop-0.20.2 # df -h > Filesystem Size Used Avail Use% Mounted on > /dev/sda5 25G 16G 7.4G 69% / > udev 987M 212K 986M 1% /dev > /dev/sda7 42G 5.5G 34G 14% /home > ------------------------------------------- > http://localhost:50070/dfshealth.jsp------------------ > NameNode 'localhost:54310' > Started: Wed Jun 15 04:13:14 IST 2011 > Version: 0.20.2, r911707 > Compiled: Fri Feb 19 08:07:34 UTC 2010 by chrisdo > Upgrades: There are no upgrades in progress. > Browse the filesystem > Namenode Logs > Cluster Summary > 10 files and directories, 0 blocks = 10 total. Heap Size is 15.5 MB / 966.69 > MB (1%) > Configured Capacity : 24.61 GB > DFS Used : 24 KB > Non DFS Used : 17.23 GB > DFS Remaining : 7.38 GB > DFS Used% : 0 % > DFS Remaining% : 29.99 % > Live Nodes : 1 > Dead Nodes : 0 > NameNode Storage: > Storage Directory Type State > /tmp/Testinghadoop/dfs/name IMAGE_AND_EDITS Active > Hadoop, 2011. > ---------------------------------------- > core-site.xml > --------------------------------- > <?xml version="1.0"?> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> > <!-- Put site-specific property overrides in this file. --> > <configuration> > <property> > <name>hadoop.tmp.dir</name> > <value>/tmp/Testinghadoop/</value> > <description>A base for other temporary directories.</description> > </property> > <property> > <name>fs.default.name</name> > <value>hdfs://localhost:54310</value> > <description>The name of the default file system. A URI whose > scheme and authority determine the FileSystem implementation. The > uri's scheme determines the config property (fs.SCHEME.impl) naming > the FileSystem implementation class. The uri's authority is used to > determine the host, port, etc. for a filesystem.</description> > </property> > </configuration> > ------------------------------------------------ > hdfs-site.xml > ---------------------------------- > <?xml version="1.0"?> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> > <!-- Put site-specific property overrides in this file. --> > <configuration> > <property> > <name>dfs.permissions</name> > <value>true</value> > <description> > If "true", enable permission checking in HDFS. > If "false", permission checking is turned off, > but all other behavior is unchanged. > Switching from one parameter value to the other does not change the mode, > owner, or group of files or directories. > </description> > </property> > <property> > <name>dfs.replication</name> > <value>1</value> > <description>Default block replication. > The actual number of replications can be specified when the file is created. > The default is used if replication is not specified in create time. > </description> > </property> > </configuration> > --------------------------------------- > mapred-site.xml > ---------------------------------- > <?xml version="1.0"?> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> > <!-- Put site-specific property overrides in this file. --> > <configuration> > <property> > <name>mapred.job.tracker</name> > <value>localhost:54311</value> > <description>The host and port that the MapReduce job tracker runs > at. If "local", then jobs are run in-process as a single map > and reduce task. > </description> > </property> > </configuration> > ---------------------------------------------------------------------------------- > please give suggetions about this error: > ------------------------------------------------------------------------------------------------------------------ > linux-8ysi:/etc/hadoop/hadoop-0.20.2/conf # hadoop fsck / > RUN_JAVA > /usr/java/jre1.6.0_25/bin/java > .Status: HEALTHY > Total size: 0 B > Total dirs: 7 > Total files: 1 (Files currently being written: 1) > Total blocks (validated): 0 > Minimally replicated blocks: 0 > Over-replicated blocks: 0 > Under-replicated blocks: 0 > Mis-replicated blocks: 0 > Default replication factor: 1 > Average block replication: 0.0 > Corrupt blocks: 0 > Missing replicas: 0 > Number of data-nodes: 1 > Number of racks: 1 > The filesystem under path '/' is HEALTHY > linux-8ysi:/etc/hadoop/hadoop-0.20.2/conf # hadoop dfsadmin -report > RUN_JAVA > /usr/java/jre1.6.0_25/bin/java > Configured Capacity: 26425618432 (24.61 GB) > Present Capacity: 7923564544 (7.38 GB) > DFS Remaining: 7923539968 (7.38 GB) > DFS Used: 24576 (24 KB) > DFS Used%: 0% > Under replicated blocks: 0 > Blocks with corrupt replicas: 0 > Missing blocks: 0 > ------------------------------------------------- > Datanodes available: 1 (1 total, 0 dead) > Name: 127.0.0.1:50010 > Decommission Status : Normal > Configured Capacity: 26425618432 (24.61 GB) > DFS Used: 24576 (24 KB) > Non DFS Used: 18502053888 (17.23 GB) > DFS Remaining: 7923539968(7.38 GB) > DFS Used%: 0% > DFS Remaining%: 29.98% > Last contact: Wed Jun 15 05:54:00 IST 2011 > i got this error: > ---------------------------- > linux-8ysi:/etc/hadoop/hadoop-0.20.2 # hadoop dfs -put spo.txt In > RUN_JAVA > /usr/java/jre1.6.0_25/bin/java > 11/06/15 04:50:18 WARN hdfs.DFSClient: DataStreamer Exception: > org.apache.hadoop.ipc.RemoteException: java.io.IOException: File > /user/root/In/spo.txt could only be replicated to 0 nodes, instead of 1 > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1271) > at org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:422) > at sun.reflect.GeneratedMethodAccessor5.invoke(Unknown Source) > at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) > at java.lang.reflect.Method.invoke(Unknown Source) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Unknown Source) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953) > at org.apache.hadoop.ipc.Client.call(Client.java:740) > at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220) > at $Proxy0.addBlock(Unknown Source) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source) > at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) > at java.lang.reflect.Method.invoke(Unknown Source) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59) > at $Proxy0.addBlock(Unknown Source) > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.locateFollowingBlock(DFSClient.java:2937) > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2819) > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2000(DFSClient.java:2102) > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2288) > 11/06/15 04:50:18 WARN hdfs.DFSClient: Error Recovery for block null bad > datanode[0] nodes == null > 11/06/15 04:50:18 WARN hdfs.DFSClient: Could not get block locations. Source > file "/user/root/In/spo.txt" - Aborting... > put: java.io.IOException: File /user/root/In/spo.txt could only be replicated > to 0 nodes, instead of 1 > 11/06/15 04:50:18 ERROR hdfs.DFSClient: Exception closing file > /user/root/In/spo.txt : org.apache.hadoop.ipc.RemoteException: > java.io.IOException: File /user/root/In/spo.txt could only be replicated to 0 > nodes, instead of 1 > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1271) > at org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:422) > at sun.reflect.GeneratedMethodAccessor5.invoke(Unknown Source) > at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) > at java.lang.reflect.Method.invoke(Unknown Source) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Unknown Source) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953) > org.apache.hadoop.ipc.RemoteException: java.io.IOException: File > /user/root/In/spo.txt could only be replicated to 0 nodes, instead of 1 > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1271) > at org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:422) > at sun.reflect.GeneratedMethodAccessor5.invoke(Unknown Source) > at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) > at java.lang.reflect.Method.invoke(Unknown Source) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Unknown Source) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953) > at org.apache.hadoop.ipc.Client.call(Client.java:740) > at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220) > at $Proxy0.addBlock(Unknown Source) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source) > at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) > at java.lang.reflect.Method.invoke(Unknown Source) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59) > at $Proxy0.addBlock(Unknown Source) > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.locateFollowingBlock(DFSClient.java:2937) > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2819) > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2000(DFSClient.java:2102) > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2288) > regards > Ranga Swamy > 8904524975 -- This message was sent by Atlassian JIRA (v6.2#6252)