Hadoop-Hdfs-21-Build - Build # 151 - Still Failing

2011-05-20 Thread Apache Jenkins Server
See https://builds.apache.org/hudson/job/Hadoop-Hdfs-21-Build/151/

###
## LAST 60 LINES OF THE CONSOLE 
###
[...truncated 895 lines...]
clover.setup:

clover.info:
 [echo] 
 [echo]  Clover not found. Code coverage reports disabled.
 [echo]   

clover:

ivy-download:
  [get] Getting: 
http://repo2.maven.org/maven2/org/apache/ivy/ivy/2.1.0/ivy-2.1.0.jar
  [get] To: 
/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-21-Build/trunk/ivy/ivy-2.1.0.jar
  [get] Error getting 
http://repo2.maven.org/maven2/org/apache/ivy/ivy/2.1.0/ivy-2.1.0.jar to 
/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-21-Build/trunk/ivy/ivy-2.1.0.jar

BUILD FAILED
java.net.ConnectException: Connection timed out
at java.net.PlainSocketImpl.socketConnect(Native Method)
at java.net.PlainSocketImpl.doConnect(PlainSocketImpl.java:333)
at java.net.PlainSocketImpl.connectToAddress(PlainSocketImpl.java:195)
at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:182)
at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:366)
at java.net.Socket.connect(Socket.java:529)
at java.net.Socket.connect(Socket.java:478)
at sun.net.NetworkClient.doConnect(NetworkClient.java:163)
at sun.net.www.http.HttpClient.openServer(HttpClient.java:394)
at sun.net.www.http.HttpClient.openServer(HttpClient.java:529)
at sun.net.www.http.HttpClient.(HttpClient.java:233)
at sun.net.www.http.HttpClient.New(HttpClient.java:306)
at sun.net.www.http.HttpClient.New(HttpClient.java:323)
at 
sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:975)
at 
sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:916)
at 
sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:841)
at org.apache.tools.ant.taskdefs.Get.doGet(Get.java:145)
at org.apache.tools.ant.taskdefs.Get.execute(Get.java:78)
at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:288)
at sun.reflect.GeneratedMethodAccessor1.invoke(Unknown Source)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at 
org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106)
at org.apache.tools.ant.Task.perform(Task.java:348)
at org.apache.tools.ant.Target.execute(Target.java:357)
at org.apache.tools.ant.Target.performTasks(Target.java:385)
at org.apache.tools.ant.Project.executeSortedTargets(Project.java:1337)
at org.apache.tools.ant.Project.executeTarget(Project.java:1306)
at 
org.apache.tools.ant.helper.DefaultExecutor.executeTargets(DefaultExecutor.java:41)
at org.apache.tools.ant.Project.executeTargets(Project.java:1189)
at org.apache.tools.ant.Main.runBuild(Main.java:758)
at org.apache.tools.ant.Main.startAnt(Main.java:217)
at org.apache.tools.ant.launch.Launcher.run(Launcher.java:257)
at org.apache.tools.ant.launch.Launcher.main(Launcher.java:104)

Total time: 3 minutes 9 seconds
Publishing Javadoc
Archiving artifacts
Recording test results
Recording fingerprints
Description set: 
Email was triggered for: Failure
Sending email for trigger: Failure



###
## FAILED TESTS (if any) 
##
No tests ran.


Hadoop-Hdfs-22-branch - Build # 52 - Still Failing

2011-05-20 Thread Apache Jenkins Server
See https://builds.apache.org/hudson/job/Hadoop-Hdfs-22-branch/52/

###
## LAST 60 LINES OF THE CONSOLE 
###
[...truncated 3049 lines...]

compile-hdfs-test:
   [delete] Deleting directory 
/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/cache
[mkdir] Created dir: 
/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/cache
 [copy] Copying 1 file to 
/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/cache
 [copy] Copying 1 file to 
/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/cache
 [copy] Copying 1 file to 
/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/cache
 [copy] Copying 1 file to 
/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/cache
 [copy] Copying 1 file to 
/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/cache
 [copy] Copying 1 file to 
/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/cache
 [copy] Copying 1 file to 
/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/cache
 [copy] Copying 1 file to 
/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/cache
 [copy] Copying 1 file to 
/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/cache
 [copy] Copying 1 file to 
/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/cache

run-test-hdfs-excluding-commit-and-smoke:
[mkdir] Created dir: 
/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/data
[mkdir] Created dir: 
/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/logs
 [copy] Copying 1 file to 
/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/extraconf
 [copy] Copying 1 file to 
/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/extraconf
[junit] WARNING: multiple versions of ant detected in path for junit 
[junit]  
jar:file:/homes/hudson/tools/ant/latest/lib/ant.jar!/org/apache/tools/ant/Project.class
[junit]  and 
jar:file:/homes/hudson/.ivy2/cache/ant/ant/jars/ant-1.6.5.jar!/org/apache/tools/ant/Project.class
[junit] Running org.apache.hadoop.fs.TestFiListPath
[junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 2.213 sec
[junit] Running org.apache.hadoop.fs.TestFiRename
[junit] Tests run: 4, Failures: 0, Errors: 0, Time elapsed: 5.582 sec
[junit] Running org.apache.hadoop.hdfs.TestFiHFlush
[junit] Tests run: 9, Failures: 0, Errors: 0, Time elapsed: 15.679 sec
[junit] Running org.apache.hadoop.hdfs.TestFiHftp
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 41.616 sec
[junit] Running org.apache.hadoop.hdfs.TestFiPipelines
[junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 5.787 sec
[junit] Running 
org.apache.hadoop.hdfs.server.datanode.TestFiDataTransferProtocol
[junit] Tests run: 29, Failures: 0, Errors: 0, Time elapsed: 211.716 sec
[junit] Running 
org.apache.hadoop.hdfs.server.datanode.TestFiDataTransferProtocol2
[junit] Tests run: 10, Failures: 0, Errors: 0, Time elapsed: 410.337 sec
[junit] Running org.apache.hadoop.hdfs.server.datanode.TestFiPipelineClose
[junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 35.621 sec

checkfailure:

-run-test-hdfs-fault-inject-withtestcaseonly:

run-test-hdfs-fault-inject:

BUILD FAILED
/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build.xml:746:
 Tests failed!

Total time: 60 minutes 22 seconds
[FINDBUGS] Skipping publisher since build result is FAILURE
Publishing Javadoc
Archiving artifacts
Recording test results
Recording fingerprints
Publishing Clover coverage report...
No Clover report will be published due to a Build Failure
Email was triggered for: Failure
Sending email for trigger: Failure



###
## FAILED TESTS (if any) 
##
3 tests failed.
REGRESSION:  org.apache.hadoop.hdfs.server.balancer.TestBalancer.testBalancer0

Error Message:
127.0.0.1:44072is not an underUtilized node

Stack Trace:
junit.framework.AssertionFailedError: 127.0.0.1:44072is not an underUtilized 
node
at 
org.apache.hadoop.hdfs.server.balancer.Balancer.initNodes(Balancer.java:1011)
at 
org.apache.hadoop.hdfs.server.balancer.Balancer.initNodes(Balancer.java:953)
at 
org.apache.hadoop.hdfs.server.balancer.Balancer.run(Balancer.java:1496)
at 
org.apache.hadoop.hdfs.server.balancer.TestBalancer.runBalancer(TestBalancer.java:247)
at 
org.apache.hadoop.hdfs.ser

Hadoop-Hdfs-trunk - Build # 672 - Still Failing

2011-05-20 Thread Apache Jenkins Server
See https://builds.apache.org/hudson/job/Hadoop-Hdfs-trunk/672/

###
## LAST 60 LINES OF THE CONSOLE 
###
[...truncated 872726 lines...]
[junit] 2011-05-20 12:48:17,904 INFO  datanode.DataNode 
(DataNode.java:shutdown(1620)) - Waiting for threadgroup to exit, active 
threads is 0
[junit] 2011-05-20 12:48:17,904 WARN  datanode.DataNode 
(DataNode.java:offerService(1063)) - BPOfferService for block 
pool=BP-1566714727-127.0.1.1-1305895696214 received 
exception:java.lang.InterruptedException
[junit] 2011-05-20 12:48:17,904 WARN  datanode.DataNode 
(DataNode.java:run(1216)) - DatanodeRegistration(127.0.0.1:59046, 
storageID=DS-1556683808-127.0.1.1-59046-1305895696890, infoPort=44182, 
ipcPort=50322, storageInfo=lv=-35;cid=testClusterID;nsid=795710742;c=0) ending 
block pool service for: BP-1566714727-127.0.1.1-1305895696214
[junit] 2011-05-20 12:48:18,004 INFO  datanode.DataBlockScanner 
(DataBlockScanner.java:removeBlockPool(277)) - Removed 
bpid=BP-1566714727-127.0.1.1-1305895696214 from blockPoolScannerMap
[junit] 2011-05-20 12:48:18,004 INFO  datanode.DataNode 
(FSDataset.java:shutdownBlockPool(2559)) - Removing block pool 
BP-1566714727-127.0.1.1-1305895696214
[junit] 2011-05-20 12:48:18,004 INFO  datanode.FSDatasetAsyncDiskService 
(FSDatasetAsyncDiskService.java:shutdown(133)) - Shutting down all async disk 
service threads...
[junit] 2011-05-20 12:48:18,005 INFO  datanode.FSDatasetAsyncDiskService 
(FSDatasetAsyncDiskService.java:shutdown(142)) - All async disk service threads 
have been shut down.
[junit] 2011-05-20 12:48:18,106 WARN  namenode.DecommissionManager 
(DecommissionManager.java:run(70)) - Monitor interrupted: 
java.lang.InterruptedException: sleep interrupted
[junit] 2011-05-20 12:48:18,106 INFO  namenode.FSEditLog 
(FSEditLog.java:printStatistics(580)) - Number of transactions: 6 Total time 
for transactions(ms): 0Number of transactions batched in Syncs: 0 Number of 
syncs: 3 SyncTimes(ms): 96 14 
[junit] 2011-05-20 12:48:18,106 WARN  namenode.FSNamesystem 
(FSNamesystem.java:run(3016)) - ReplicationMonitor thread received 
InterruptedException.java.lang.InterruptedException: sleep interrupted
[junit] 2011-05-20 12:48:18,108 INFO  ipc.Server (Server.java:stop(1636)) - 
Stopping server on 48611
[junit] 2011-05-20 12:48:18,109 INFO  ipc.Server (Server.java:run(1471)) - 
IPC Server handler 0 on 48611: exiting
[junit] 2011-05-20 12:48:18,109 INFO  ipc.Server (Server.java:run(494)) - 
Stopping IPC Server listener on 48611
[junit] 2011-05-20 12:48:18,109 INFO  ipc.Server (Server.java:run(698)) - 
Stopping IPC Server Responder
[junit] 2011-05-20 12:48:18,109 INFO  impl.MetricsSystemImpl 
(MetricsSystemImpl.java:stop(199)) - Stopping DataNode metrics system...
[junit] 2011-05-20 12:48:18,110 INFO  impl.MetricsSystemImpl 
(MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source JvmMetrics
[junit] 2011-05-20 12:48:18,110 INFO  impl.MetricsSystemImpl 
(MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source 
NameNodeActivity
[junit] 2011-05-20 12:48:18,110 INFO  impl.MetricsSystemImpl 
(MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source 
RpcActivityForPort48611
[junit] 2011-05-20 12:48:18,110 INFO  impl.MetricsSystemImpl 
(MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source 
RpcDetailedActivityForPort48611
[junit] 2011-05-20 12:48:18,110 INFO  impl.MetricsSystemImpl 
(MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source FSNamesystem
[junit] 2011-05-20 12:48:18,111 INFO  impl.MetricsSystemImpl 
(MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source 
RpcActivityForPort50322
[junit] 2011-05-20 12:48:18,111 INFO  impl.MetricsSystemImpl 
(MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source 
RpcDetailedActivityForPort50322
[junit] 2011-05-20 12:48:18,111 INFO  impl.MetricsSystemImpl 
(MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source JvmMetrics-1
[junit] 2011-05-20 12:48:18,111 INFO  impl.MetricsSystemImpl 
(MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source 
DataNodeActivity-h8.grid.sp2.yahoo.net-59046
[junit] 2011-05-20 12:48:18,112 INFO  impl.MetricsSystemImpl 
(MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source 
RpcActivityForPort39462
[junit] 2011-05-20 12:48:18,112 INFO  impl.MetricsSystemImpl 
(MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source 
RpcDetailedActivityForPort39462
[junit] 2011-05-20 12:48:18,112 INFO  impl.MetricsSystemImpl 
(MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source JvmMetrics-2
[junit] 2011-05-20 12:48:18,112 INFO  impl.MetricsSystemImpl 
(MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source 
DataNodeActivity-h8.grid.sp2.yahoo.net-41844
[junit] 201

[jira] [Created] (HDFS-1970) Null pointer exception comes when Namenode recovery happens and there is no response from client to NN more than the hardlimit for NN recovery and the current block is mor

2011-05-20 Thread ramkrishna.s.vasudevan (JIRA)
Null pointer exception comes when Namenode recovery happens and there is no 
response from client to NN more than the hardlimit for NN recovery and the 
current block is more than the prev block size in NN 


 Key: HDFS-1970
 URL: https://issues.apache.org/jira/browse/HDFS-1970
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: name-node
Affects Versions: 0.20-append
Reporter: ramkrishna.s.vasudevan
 Fix For: 0.20-append


Null pointer exception comes when Namenode recovery happens and there is no 
response from client to NN more than the hardlimit for NN recovery and the 
current block is more than the prev block size in NN 
1. Write using a client to 2 datanodes
2. Kill one data node and allow pipeline recovery.
3. write somemore data to the same block
4. Parallely allow the namenode recovery to happen
Null pointer exception will come in addStoreBlock api.

Pls find the logs

Debugging in MachineName.. 
Listening for transport dt_socket at address: 8007
11/05/20 21:38:33 INFO namenode.NameNode: STARTUP_MSG: 
/
STARTUP_MSG: Starting NameNode
STARTUP_MSG:   host = linux76/10.18.52.76
STARTUP_MSG:   args = []
STARTUP_MSG:   version = 0.20.3-SNAPSHOT
STARTUP_MSG:   build =  -r ; compiled by 'G00900856' on Tue Feb  1 11:40:14 IST 
2011
/
11/05/20 21:38:33 INFO metrics.RpcMetrics: Initializing RPC Metrics with 
hostName=NameNode, port=9000
11/05/20 21:38:33 INFO namenode.NameNode: Namenode up at: 
linux76/10.18.52.76:9000
11/05/20 21:38:33 INFO jvm.JvmMetrics: Initializing JVM Metrics with 
processName=NameNode, sessionId=null
11/05/20 21:38:33 INFO metrics.NameNodeMetrics: Initializing NameNodeMeterics 
using context object:org.apache.hadoop.metrics.spi.NullContext
11/05/20 21:38:33 INFO namenode.FSNamesystem: fsOwner=root,root
11/05/20 21:38:33 INFO namenode.FSNamesystem: supergroup=supergroup
11/05/20 21:38:33 INFO namenode.FSNamesystem: isPermissionEnabled=false
11/05/20 21:38:33 INFO metrics.FSNamesystemMetrics: Initializing 
FSNamesystemMetrics using context 
object:org.apache.hadoop.metrics.spi.NullContext
11/05/20 21:38:33 INFO namenode.FSNamesystem: Registered FSNamesystemStatusMBean
11/05/20 21:38:33 INFO common.Storage: Number of files = 1
11/05/20 21:38:33 INFO common.Storage: Number of files under construction = 0
11/05/20 21:38:33 INFO common.Storage: Image file of size 94 loaded in 0 
seconds.
11/05/20 21:38:33 INFO common.Storage: Edits file 
/home/ramkrishna/opensrchadoop/appendbranch/hadoop-0.20.3-SNAPSHOT/bin/../hadoop-root/dfs/name/current/edits
 of size 4 edits # 0 loaded in 0 seconds.
11/05/20 21:38:33 INFO common.Storage: Image file of size 94 saved in 0 seconds.
11/05/20 21:38:34 INFO common.Storage: Image file of size 94 saved in 0 seconds.
11/05/20 21:38:34 INFO namenode.FSNamesystem: Finished loading FSImage in 482 
msecs
11/05/20 21:38:34 INFO namenode.FSNamesystem: Total number of blocks = 0
11/05/20 21:38:34 INFO namenode.FSNamesystem: Number of invalid blocks = 0
11/05/20 21:38:34 INFO namenode.FSNamesystem: Number of under-replicated blocks 
= 0
11/05/20 21:38:34 INFO namenode.FSNamesystem: Number of  over-replicated blocks 
= 0
11/05/20 21:38:34 INFO hdfs.StateChange: STATE* Leaving safe mode after 0 secs.
11/05/20 21:38:34 INFO hdfs.StateChange: STATE* Network topology has 0 racks 
and 0 datanodes
11/05/20 21:38:34 INFO hdfs.StateChange: STATE* UnderReplicatedBlocks has 0 
blocks
11/05/20 21:38:34 INFO mortbay.log: Logging to 
org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via org.mortbay.log.Slf4jLog
11/05/20 21:38:35 INFO http.HttpServer: Port returned by 
webServer.getConnectors()[0].getLocalPort() before open() is -1. Opening the 
listener on 50070
11/05/20 21:38:35 INFO http.HttpServer: listener.getLocalPort() returned 50070 
webServer.getConnectors()[0].getLocalPort() returned 50070
11/05/20 21:38:35 INFO http.HttpServer: Jetty bound to port 50070
11/05/20 21:38:35 INFO mortbay.log: jetty-6.1.14
11/05/20 21:38:37 INFO mortbay.log: Started SelectChannelConnector@linux76:50070
11/05/20 21:38:37 INFO namenode.NameNode: Web-server up at: linux76:50070
11/05/20 21:38:37 INFO ipc.Server: IPC Server Responder: starting
11/05/20 21:38:37 INFO ipc.Server: IPC Server listener on 9000: starting
11/05/20 21:38:37 INFO ipc.Server: IPC Server handler 0 on 9000: starting
11/05/20 21:38:37 INFO ipc.Server: IPC Server handler 1 on 9000: starting
11/05/20 21:38:37 INFO ipc.Server: IPC Server handler 2 on 9000: starting
11/05/20 21:38:37 INFO ipc.Server: IPC Server handler 3 on 9000: starting
11/05/20 21:38:37 INFO ipc.Server: IPC Server handler 4 on 9000: starting
11/05/20 21:3

[jira] [Created] (HDFS-1971) HA: Send block report from datanode to both active and standby namenodes

2011-05-20 Thread Suresh Srinivas (JIRA)
HA: Send block report from datanode to both active and standby namenodes


 Key: HDFS-1971
 URL: https://issues.apache.org/jira/browse/HDFS-1971
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: data-node, name-node
Reporter: Suresh Srinivas
Assignee: Sanjay Radia


To enable hot standby namenode, the standby node must have current information 
for - namenode state (image + edits) and block location information. This jira 
addresses keeping the block location information current in the standby node. 
To do this, the proposed solution is to send block reports from the datanodes 
to both the active and the standby namenode.


--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-1972) HA: Datanode fencing mechanism

2011-05-20 Thread Suresh Srinivas (JIRA)
HA: Datanode fencing mechanism
--

 Key: HDFS-1972
 URL: https://issues.apache.org/jira/browse/HDFS-1972
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: data-node, name-node
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas


In high availability setup, with an active and standby namenode, there is a 
possibility of two namenodes sending commands to the datanode. The datanode 
must honor commands from only the active namenode and reject the commands from 
standby, to prevent corruption. This invariant must be complied with during 
fail over and other states such as split brain. This jira addresses issues 
related to this, design of the solution and implementation.


--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-1974) HA: Introduce active and standby states to the namenode

2011-05-20 Thread Suresh Srinivas (JIRA)
HA: Introduce active and standby states to the namenode
---

 Key: HDFS-1974
 URL: https://issues.apache.org/jira/browse/HDFS-1974
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: name-node
 Environment: Currently namenode supports active, secondary and backup 
roles. To support namenode high availability, active and standby states are 
needed. Note that this is different from the existing notion of namenode role, 
where a namenode cannot transition from one role to the other.

Reporter: Suresh Srinivas
Assignee: Suresh Srinivas




--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-1975) HA: Support for sharing the namenode state from active to standby.

2011-05-20 Thread Suresh Srinivas (JIRA)
HA: Support for sharing the namenode state from active to standby.
--

 Key: HDFS-1975
 URL: https://issues.apache.org/jira/browse/HDFS-1975
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: name-node
Reporter: Suresh Srinivas


To enable hot standby namenode, the standby node must have current information 
for - namenode state (image + edits) and block location information. This jira 
addresses keeping the namenode state current in the standby node. To do this, 
the proposed solution in this jira is to use a shared storage to store the 
namenode state. 

Note one could also build an alternative solution by augmenting the backup 
node. A seperate jira could explore this.


--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-1973) HA: HDFS clients must handle namenode failover and switch over to the new active namenode.

2011-05-20 Thread Suresh Srinivas (JIRA)
HA: HDFS clients must handle namenode failover and switch over to the new 
active namenode.
--

 Key: HDFS-1973
 URL: https://issues.apache.org/jira/browse/HDFS-1973
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Suresh Srinivas


During failover, a client must detect the current active namenode failure and 
switch over to the new active namenode. The switch over might make use of IP 
failover or some thing more elaborate such as zookeeper to discover the new 
active.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-1976) Logging in DataXceiver will sometimes repeat stack traces

2011-05-20 Thread Joey Echeverria (JIRA)
Logging in DataXceiver will sometimes repeat stack traces
-

 Key: HDFS-1976
 URL: https://issues.apache.org/jira/browse/HDFS-1976
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Joey Echeverria
Priority: Minor


The run() method in DataXceiver logs the stack trace of all throwables thrown 
while performing an operation. In some cases, the operations also log stack 
traces despite throwing the exception up the stack. The logging code should try 
to avoid double-logging stack traces where possible.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-1977) Stop using StringUtils.stringifyException()

2011-05-20 Thread Joey Echeverria (JIRA)
Stop using StringUtils.stringifyException()
---

 Key: HDFS-1977
 URL: https://issues.apache.org/jira/browse/HDFS-1977
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Joey Echeverria
Priority: Minor


The old version of the logging APIs didn't support logging stack traces by 
passing exceptions to the logging methods (e.g. Log.error()). A number of log 
statements make use of StringUtils.stringifyException() to get around the old 
behavior. It would be nice if this could get cleaned up to make use of the the 
logger's stack trace printing. This also gives users more control since you can 
configure how the stack traces are written to the logs.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


Hadoop-Hdfs-trunk-Commit - Build # 676 - Still Failing

2011-05-20 Thread Apache Jenkins Server
See https://builds.apache.org/hudson/job/Hadoop-Hdfs-trunk-Commit/676/

###
## LAST 60 LINES OF THE CONSOLE 
###
[...truncated 2642 lines...]
[junit] Running org.apache.hadoop.hdfs.server.datanode.TestDiskError
[junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 8.916 sec
[junit] Running 
org.apache.hadoop.hdfs.server.datanode.TestInterDatanodeProtocol
[junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 5.196 sec
[junit] Running 
org.apache.hadoop.hdfs.server.datanode.TestSimulatedFSDataset
[junit] Tests run: 8, Failures: 0, Errors: 0, Time elapsed: 0.768 sec
[junit] Running org.apache.hadoop.hdfs.server.namenode.TestBackupNode
[junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 18.288 sec
[junit] Running org.apache.hadoop.hdfs.server.namenode.TestCheckpoint
[junit] Tests run: 4, Failures: 0, Errors: 0, Time elapsed: 32.171 sec
[junit] Running 
org.apache.hadoop.hdfs.server.namenode.TestComputeInvalidateWork
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 4.956 sec
[junit] Running 
org.apache.hadoop.hdfs.server.namenode.TestDatanodeDescriptor
[junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 0.172 sec
[junit] Running org.apache.hadoop.hdfs.server.namenode.TestEditLog
[junit] Tests run: 5, Failures: 0, Errors: 0, Time elapsed: 13.244 sec
[junit] Running org.apache.hadoop.hdfs.server.namenode.TestFileLimit
[junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 5.225 sec
[junit] Running org.apache.hadoop.hdfs.server.namenode.TestHeartbeatHandling
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 3.1 sec
[junit] Running org.apache.hadoop.hdfs.server.namenode.TestHost2NodesMap
[junit] Tests run: 4, Failures: 0, Errors: 0, Time elapsed: 0.09 sec
[junit] Running 
org.apache.hadoop.hdfs.server.namenode.TestNamenodeCapacityReport
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 3.175 sec
[junit] Running 
org.apache.hadoop.hdfs.server.namenode.TestOverReplicatedBlocks
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 4.032 sec
[junit] Running 
org.apache.hadoop.hdfs.server.namenode.TestPendingReplication
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 7.3 sec
[junit] Running org.apache.hadoop.hdfs.server.namenode.TestReplicationPolicy
[junit] Tests run: 8, Failures: 0, Errors: 0, Time elapsed: 0.064 sec
[junit] Running org.apache.hadoop.hdfs.server.namenode.TestSafeMode
[junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 8.665 sec
[junit] Running org.apache.hadoop.hdfs.server.namenode.TestStartup
[junit] Tests run: 5, Failures: 0, Errors: 0, Time elapsed: 10.69 sec
[junit] Running org.apache.hadoop.hdfs.server.namenode.TestStorageRestore
[junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 8.412 sec
[junit] Running org.apache.hadoop.net.TestNetworkTopology
[junit] Tests run: 8, Failures: 0, Errors: 0, Time elapsed: 0.111 sec
[junit] Running org.apache.hadoop.security.TestPermission
[junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 5.281 sec

checkfailure:
[touch] Creating 
/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-trunk-Commit/trunk/build/test/testsfailed

BUILD FAILED
/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-trunk-Commit/trunk/build.xml:712:
 The following error occurred while executing this line:
/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-trunk-Commit/trunk/build.xml:669:
 The following error occurred while executing this line:
/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-trunk-Commit/trunk/build.xml:737:
 Tests failed!

Total time: 8 minutes 51 seconds
[FINDBUGS] Skipping publisher since build result is FAILURE
Recording fingerprints
Archiving artifacts
Recording test results
Publishing Javadoc
Publishing Clover coverage report...
No Clover report will be published due to a Build Failure
Email was triggered for: Failure
Sending email for trigger: Failure



###
## FAILED TESTS (if any) 
##
1 tests failed.
FAILED:  org.apache.hadoop.hdfs.TestHDFSTrash.testTrashEmptier

Error Message:
null

Stack Trace:
junit.framework.AssertionFailedError: null
at org.apache.hadoop.fs.TestTrash.testTrashEmptier(TestTrash.java:479)
at junit.extensions.TestDecorator.basicRun(TestDecorator.java:24)
at junit.extensions.TestSetup$1.protect(TestSetup.java:23)
at junit.extensions.TestSetup.run(TestSetup.java:27)