Hadoop-Hdfs-21-Build - Build # 151 - Still Failing
See https://builds.apache.org/hudson/job/Hadoop-Hdfs-21-Build/151/ ### ## LAST 60 LINES OF THE CONSOLE ### [...truncated 895 lines...] clover.setup: clover.info: [echo] [echo] Clover not found. Code coverage reports disabled. [echo] clover: ivy-download: [get] Getting: http://repo2.maven.org/maven2/org/apache/ivy/ivy/2.1.0/ivy-2.1.0.jar [get] To: /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-21-Build/trunk/ivy/ivy-2.1.0.jar [get] Error getting http://repo2.maven.org/maven2/org/apache/ivy/ivy/2.1.0/ivy-2.1.0.jar to /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-21-Build/trunk/ivy/ivy-2.1.0.jar BUILD FAILED java.net.ConnectException: Connection timed out at java.net.PlainSocketImpl.socketConnect(Native Method) at java.net.PlainSocketImpl.doConnect(PlainSocketImpl.java:333) at java.net.PlainSocketImpl.connectToAddress(PlainSocketImpl.java:195) at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:182) at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:366) at java.net.Socket.connect(Socket.java:529) at java.net.Socket.connect(Socket.java:478) at sun.net.NetworkClient.doConnect(NetworkClient.java:163) at sun.net.www.http.HttpClient.openServer(HttpClient.java:394) at sun.net.www.http.HttpClient.openServer(HttpClient.java:529) at sun.net.www.http.HttpClient.(HttpClient.java:233) at sun.net.www.http.HttpClient.New(HttpClient.java:306) at sun.net.www.http.HttpClient.New(HttpClient.java:323) at sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:975) at sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:916) at sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:841) at org.apache.tools.ant.taskdefs.Get.doGet(Get.java:145) at org.apache.tools.ant.taskdefs.Get.execute(Get.java:78) at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:288) at sun.reflect.GeneratedMethodAccessor1.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106) at org.apache.tools.ant.Task.perform(Task.java:348) at org.apache.tools.ant.Target.execute(Target.java:357) at org.apache.tools.ant.Target.performTasks(Target.java:385) at org.apache.tools.ant.Project.executeSortedTargets(Project.java:1337) at org.apache.tools.ant.Project.executeTarget(Project.java:1306) at org.apache.tools.ant.helper.DefaultExecutor.executeTargets(DefaultExecutor.java:41) at org.apache.tools.ant.Project.executeTargets(Project.java:1189) at org.apache.tools.ant.Main.runBuild(Main.java:758) at org.apache.tools.ant.Main.startAnt(Main.java:217) at org.apache.tools.ant.launch.Launcher.run(Launcher.java:257) at org.apache.tools.ant.launch.Launcher.main(Launcher.java:104) Total time: 3 minutes 9 seconds Publishing Javadoc Archiving artifacts Recording test results Recording fingerprints Description set: Email was triggered for: Failure Sending email for trigger: Failure ### ## FAILED TESTS (if any) ## No tests ran.
Hadoop-Hdfs-22-branch - Build # 52 - Still Failing
See https://builds.apache.org/hudson/job/Hadoop-Hdfs-22-branch/52/ ### ## LAST 60 LINES OF THE CONSOLE ### [...truncated 3049 lines...] compile-hdfs-test: [delete] Deleting directory /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/cache [mkdir] Created dir: /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/cache [copy] Copying 1 file to /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/cache [copy] Copying 1 file to /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/cache [copy] Copying 1 file to /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/cache [copy] Copying 1 file to /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/cache [copy] Copying 1 file to /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/cache [copy] Copying 1 file to /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/cache [copy] Copying 1 file to /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/cache [copy] Copying 1 file to /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/cache [copy] Copying 1 file to /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/cache [copy] Copying 1 file to /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/cache run-test-hdfs-excluding-commit-and-smoke: [mkdir] Created dir: /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/data [mkdir] Created dir: /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/logs [copy] Copying 1 file to /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/extraconf [copy] Copying 1 file to /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build-fi/test/extraconf [junit] WARNING: multiple versions of ant detected in path for junit [junit] jar:file:/homes/hudson/tools/ant/latest/lib/ant.jar!/org/apache/tools/ant/Project.class [junit] and jar:file:/homes/hudson/.ivy2/cache/ant/ant/jars/ant-1.6.5.jar!/org/apache/tools/ant/Project.class [junit] Running org.apache.hadoop.fs.TestFiListPath [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 2.213 sec [junit] Running org.apache.hadoop.fs.TestFiRename [junit] Tests run: 4, Failures: 0, Errors: 0, Time elapsed: 5.582 sec [junit] Running org.apache.hadoop.hdfs.TestFiHFlush [junit] Tests run: 9, Failures: 0, Errors: 0, Time elapsed: 15.679 sec [junit] Running org.apache.hadoop.hdfs.TestFiHftp [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 41.616 sec [junit] Running org.apache.hadoop.hdfs.TestFiPipelines [junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 5.787 sec [junit] Running org.apache.hadoop.hdfs.server.datanode.TestFiDataTransferProtocol [junit] Tests run: 29, Failures: 0, Errors: 0, Time elapsed: 211.716 sec [junit] Running org.apache.hadoop.hdfs.server.datanode.TestFiDataTransferProtocol2 [junit] Tests run: 10, Failures: 0, Errors: 0, Time elapsed: 410.337 sec [junit] Running org.apache.hadoop.hdfs.server.datanode.TestFiPipelineClose [junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 35.621 sec checkfailure: -run-test-hdfs-fault-inject-withtestcaseonly: run-test-hdfs-fault-inject: BUILD FAILED /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-22-branch/trunk/build.xml:746: Tests failed! Total time: 60 minutes 22 seconds [FINDBUGS] Skipping publisher since build result is FAILURE Publishing Javadoc Archiving artifacts Recording test results Recording fingerprints Publishing Clover coverage report... No Clover report will be published due to a Build Failure Email was triggered for: Failure Sending email for trigger: Failure ### ## FAILED TESTS (if any) ## 3 tests failed. REGRESSION: org.apache.hadoop.hdfs.server.balancer.TestBalancer.testBalancer0 Error Message: 127.0.0.1:44072is not an underUtilized node Stack Trace: junit.framework.AssertionFailedError: 127.0.0.1:44072is not an underUtilized node at org.apache.hadoop.hdfs.server.balancer.Balancer.initNodes(Balancer.java:1011) at org.apache.hadoop.hdfs.server.balancer.Balancer.initNodes(Balancer.java:953) at org.apache.hadoop.hdfs.server.balancer.Balancer.run(Balancer.java:1496) at org.apache.hadoop.hdfs.server.balancer.TestBalancer.runBalancer(TestBalancer.java:247) at org.apache.hadoop.hdfs.ser
Hadoop-Hdfs-trunk - Build # 672 - Still Failing
See https://builds.apache.org/hudson/job/Hadoop-Hdfs-trunk/672/ ### ## LAST 60 LINES OF THE CONSOLE ### [...truncated 872726 lines...] [junit] 2011-05-20 12:48:17,904 INFO datanode.DataNode (DataNode.java:shutdown(1620)) - Waiting for threadgroup to exit, active threads is 0 [junit] 2011-05-20 12:48:17,904 WARN datanode.DataNode (DataNode.java:offerService(1063)) - BPOfferService for block pool=BP-1566714727-127.0.1.1-1305895696214 received exception:java.lang.InterruptedException [junit] 2011-05-20 12:48:17,904 WARN datanode.DataNode (DataNode.java:run(1216)) - DatanodeRegistration(127.0.0.1:59046, storageID=DS-1556683808-127.0.1.1-59046-1305895696890, infoPort=44182, ipcPort=50322, storageInfo=lv=-35;cid=testClusterID;nsid=795710742;c=0) ending block pool service for: BP-1566714727-127.0.1.1-1305895696214 [junit] 2011-05-20 12:48:18,004 INFO datanode.DataBlockScanner (DataBlockScanner.java:removeBlockPool(277)) - Removed bpid=BP-1566714727-127.0.1.1-1305895696214 from blockPoolScannerMap [junit] 2011-05-20 12:48:18,004 INFO datanode.DataNode (FSDataset.java:shutdownBlockPool(2559)) - Removing block pool BP-1566714727-127.0.1.1-1305895696214 [junit] 2011-05-20 12:48:18,004 INFO datanode.FSDatasetAsyncDiskService (FSDatasetAsyncDiskService.java:shutdown(133)) - Shutting down all async disk service threads... [junit] 2011-05-20 12:48:18,005 INFO datanode.FSDatasetAsyncDiskService (FSDatasetAsyncDiskService.java:shutdown(142)) - All async disk service threads have been shut down. [junit] 2011-05-20 12:48:18,106 WARN namenode.DecommissionManager (DecommissionManager.java:run(70)) - Monitor interrupted: java.lang.InterruptedException: sleep interrupted [junit] 2011-05-20 12:48:18,106 INFO namenode.FSEditLog (FSEditLog.java:printStatistics(580)) - Number of transactions: 6 Total time for transactions(ms): 0Number of transactions batched in Syncs: 0 Number of syncs: 3 SyncTimes(ms): 96 14 [junit] 2011-05-20 12:48:18,106 WARN namenode.FSNamesystem (FSNamesystem.java:run(3016)) - ReplicationMonitor thread received InterruptedException.java.lang.InterruptedException: sleep interrupted [junit] 2011-05-20 12:48:18,108 INFO ipc.Server (Server.java:stop(1636)) - Stopping server on 48611 [junit] 2011-05-20 12:48:18,109 INFO ipc.Server (Server.java:run(1471)) - IPC Server handler 0 on 48611: exiting [junit] 2011-05-20 12:48:18,109 INFO ipc.Server (Server.java:run(494)) - Stopping IPC Server listener on 48611 [junit] 2011-05-20 12:48:18,109 INFO ipc.Server (Server.java:run(698)) - Stopping IPC Server Responder [junit] 2011-05-20 12:48:18,109 INFO impl.MetricsSystemImpl (MetricsSystemImpl.java:stop(199)) - Stopping DataNode metrics system... [junit] 2011-05-20 12:48:18,110 INFO impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source JvmMetrics [junit] 2011-05-20 12:48:18,110 INFO impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source NameNodeActivity [junit] 2011-05-20 12:48:18,110 INFO impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcActivityForPort48611 [junit] 2011-05-20 12:48:18,110 INFO impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcDetailedActivityForPort48611 [junit] 2011-05-20 12:48:18,110 INFO impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source FSNamesystem [junit] 2011-05-20 12:48:18,111 INFO impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcActivityForPort50322 [junit] 2011-05-20 12:48:18,111 INFO impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcDetailedActivityForPort50322 [junit] 2011-05-20 12:48:18,111 INFO impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source JvmMetrics-1 [junit] 2011-05-20 12:48:18,111 INFO impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source DataNodeActivity-h8.grid.sp2.yahoo.net-59046 [junit] 2011-05-20 12:48:18,112 INFO impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcActivityForPort39462 [junit] 2011-05-20 12:48:18,112 INFO impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcDetailedActivityForPort39462 [junit] 2011-05-20 12:48:18,112 INFO impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source JvmMetrics-2 [junit] 2011-05-20 12:48:18,112 INFO impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source DataNodeActivity-h8.grid.sp2.yahoo.net-41844 [junit] 201
[jira] [Created] (HDFS-1970) Null pointer exception comes when Namenode recovery happens and there is no response from client to NN more than the hardlimit for NN recovery and the current block is mor
Null pointer exception comes when Namenode recovery happens and there is no response from client to NN more than the hardlimit for NN recovery and the current block is more than the prev block size in NN Key: HDFS-1970 URL: https://issues.apache.org/jira/browse/HDFS-1970 Project: Hadoop HDFS Issue Type: Bug Components: name-node Affects Versions: 0.20-append Reporter: ramkrishna.s.vasudevan Fix For: 0.20-append Null pointer exception comes when Namenode recovery happens and there is no response from client to NN more than the hardlimit for NN recovery and the current block is more than the prev block size in NN 1. Write using a client to 2 datanodes 2. Kill one data node and allow pipeline recovery. 3. write somemore data to the same block 4. Parallely allow the namenode recovery to happen Null pointer exception will come in addStoreBlock api. Pls find the logs Debugging in MachineName.. Listening for transport dt_socket at address: 8007 11/05/20 21:38:33 INFO namenode.NameNode: STARTUP_MSG: / STARTUP_MSG: Starting NameNode STARTUP_MSG: host = linux76/10.18.52.76 STARTUP_MSG: args = [] STARTUP_MSG: version = 0.20.3-SNAPSHOT STARTUP_MSG: build = -r ; compiled by 'G00900856' on Tue Feb 1 11:40:14 IST 2011 / 11/05/20 21:38:33 INFO metrics.RpcMetrics: Initializing RPC Metrics with hostName=NameNode, port=9000 11/05/20 21:38:33 INFO namenode.NameNode: Namenode up at: linux76/10.18.52.76:9000 11/05/20 21:38:33 INFO jvm.JvmMetrics: Initializing JVM Metrics with processName=NameNode, sessionId=null 11/05/20 21:38:33 INFO metrics.NameNodeMetrics: Initializing NameNodeMeterics using context object:org.apache.hadoop.metrics.spi.NullContext 11/05/20 21:38:33 INFO namenode.FSNamesystem: fsOwner=root,root 11/05/20 21:38:33 INFO namenode.FSNamesystem: supergroup=supergroup 11/05/20 21:38:33 INFO namenode.FSNamesystem: isPermissionEnabled=false 11/05/20 21:38:33 INFO metrics.FSNamesystemMetrics: Initializing FSNamesystemMetrics using context object:org.apache.hadoop.metrics.spi.NullContext 11/05/20 21:38:33 INFO namenode.FSNamesystem: Registered FSNamesystemStatusMBean 11/05/20 21:38:33 INFO common.Storage: Number of files = 1 11/05/20 21:38:33 INFO common.Storage: Number of files under construction = 0 11/05/20 21:38:33 INFO common.Storage: Image file of size 94 loaded in 0 seconds. 11/05/20 21:38:33 INFO common.Storage: Edits file /home/ramkrishna/opensrchadoop/appendbranch/hadoop-0.20.3-SNAPSHOT/bin/../hadoop-root/dfs/name/current/edits of size 4 edits # 0 loaded in 0 seconds. 11/05/20 21:38:33 INFO common.Storage: Image file of size 94 saved in 0 seconds. 11/05/20 21:38:34 INFO common.Storage: Image file of size 94 saved in 0 seconds. 11/05/20 21:38:34 INFO namenode.FSNamesystem: Finished loading FSImage in 482 msecs 11/05/20 21:38:34 INFO namenode.FSNamesystem: Total number of blocks = 0 11/05/20 21:38:34 INFO namenode.FSNamesystem: Number of invalid blocks = 0 11/05/20 21:38:34 INFO namenode.FSNamesystem: Number of under-replicated blocks = 0 11/05/20 21:38:34 INFO namenode.FSNamesystem: Number of over-replicated blocks = 0 11/05/20 21:38:34 INFO hdfs.StateChange: STATE* Leaving safe mode after 0 secs. 11/05/20 21:38:34 INFO hdfs.StateChange: STATE* Network topology has 0 racks and 0 datanodes 11/05/20 21:38:34 INFO hdfs.StateChange: STATE* UnderReplicatedBlocks has 0 blocks 11/05/20 21:38:34 INFO mortbay.log: Logging to org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via org.mortbay.log.Slf4jLog 11/05/20 21:38:35 INFO http.HttpServer: Port returned by webServer.getConnectors()[0].getLocalPort() before open() is -1. Opening the listener on 50070 11/05/20 21:38:35 INFO http.HttpServer: listener.getLocalPort() returned 50070 webServer.getConnectors()[0].getLocalPort() returned 50070 11/05/20 21:38:35 INFO http.HttpServer: Jetty bound to port 50070 11/05/20 21:38:35 INFO mortbay.log: jetty-6.1.14 11/05/20 21:38:37 INFO mortbay.log: Started SelectChannelConnector@linux76:50070 11/05/20 21:38:37 INFO namenode.NameNode: Web-server up at: linux76:50070 11/05/20 21:38:37 INFO ipc.Server: IPC Server Responder: starting 11/05/20 21:38:37 INFO ipc.Server: IPC Server listener on 9000: starting 11/05/20 21:38:37 INFO ipc.Server: IPC Server handler 0 on 9000: starting 11/05/20 21:38:37 INFO ipc.Server: IPC Server handler 1 on 9000: starting 11/05/20 21:38:37 INFO ipc.Server: IPC Server handler 2 on 9000: starting 11/05/20 21:38:37 INFO ipc.Server: IPC Server handler 3 on 9000: starting 11/05/20 21:38:37 INFO ipc.Server: IPC Server handler 4 on 9000: starting 11/05/20 21:3
[jira] [Created] (HDFS-1971) HA: Send block report from datanode to both active and standby namenodes
HA: Send block report from datanode to both active and standby namenodes Key: HDFS-1971 URL: https://issues.apache.org/jira/browse/HDFS-1971 Project: Hadoop HDFS Issue Type: Sub-task Components: data-node, name-node Reporter: Suresh Srinivas Assignee: Sanjay Radia To enable hot standby namenode, the standby node must have current information for - namenode state (image + edits) and block location information. This jira addresses keeping the block location information current in the standby node. To do this, the proposed solution is to send block reports from the datanodes to both the active and the standby namenode. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HDFS-1972) HA: Datanode fencing mechanism
HA: Datanode fencing mechanism -- Key: HDFS-1972 URL: https://issues.apache.org/jira/browse/HDFS-1972 Project: Hadoop HDFS Issue Type: Sub-task Components: data-node, name-node Reporter: Suresh Srinivas Assignee: Suresh Srinivas In high availability setup, with an active and standby namenode, there is a possibility of two namenodes sending commands to the datanode. The datanode must honor commands from only the active namenode and reject the commands from standby, to prevent corruption. This invariant must be complied with during fail over and other states such as split brain. This jira addresses issues related to this, design of the solution and implementation. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HDFS-1974) HA: Introduce active and standby states to the namenode
HA: Introduce active and standby states to the namenode --- Key: HDFS-1974 URL: https://issues.apache.org/jira/browse/HDFS-1974 Project: Hadoop HDFS Issue Type: Sub-task Components: name-node Environment: Currently namenode supports active, secondary and backup roles. To support namenode high availability, active and standby states are needed. Note that this is different from the existing notion of namenode role, where a namenode cannot transition from one role to the other. Reporter: Suresh Srinivas Assignee: Suresh Srinivas -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HDFS-1975) HA: Support for sharing the namenode state from active to standby.
HA: Support for sharing the namenode state from active to standby. -- Key: HDFS-1975 URL: https://issues.apache.org/jira/browse/HDFS-1975 Project: Hadoop HDFS Issue Type: Sub-task Components: name-node Reporter: Suresh Srinivas To enable hot standby namenode, the standby node must have current information for - namenode state (image + edits) and block location information. This jira addresses keeping the namenode state current in the standby node. To do this, the proposed solution in this jira is to use a shared storage to store the namenode state. Note one could also build an alternative solution by augmenting the backup node. A seperate jira could explore this. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HDFS-1973) HA: HDFS clients must handle namenode failover and switch over to the new active namenode.
HA: HDFS clients must handle namenode failover and switch over to the new active namenode. -- Key: HDFS-1973 URL: https://issues.apache.org/jira/browse/HDFS-1973 Project: Hadoop HDFS Issue Type: Sub-task Reporter: Suresh Srinivas During failover, a client must detect the current active namenode failure and switch over to the new active namenode. The switch over might make use of IP failover or some thing more elaborate such as zookeeper to discover the new active. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HDFS-1976) Logging in DataXceiver will sometimes repeat stack traces
Logging in DataXceiver will sometimes repeat stack traces - Key: HDFS-1976 URL: https://issues.apache.org/jira/browse/HDFS-1976 Project: Hadoop HDFS Issue Type: Improvement Reporter: Joey Echeverria Priority: Minor The run() method in DataXceiver logs the stack trace of all throwables thrown while performing an operation. In some cases, the operations also log stack traces despite throwing the exception up the stack. The logging code should try to avoid double-logging stack traces where possible. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HDFS-1977) Stop using StringUtils.stringifyException()
Stop using StringUtils.stringifyException() --- Key: HDFS-1977 URL: https://issues.apache.org/jira/browse/HDFS-1977 Project: Hadoop HDFS Issue Type: Improvement Reporter: Joey Echeverria Priority: Minor The old version of the logging APIs didn't support logging stack traces by passing exceptions to the logging methods (e.g. Log.error()). A number of log statements make use of StringUtils.stringifyException() to get around the old behavior. It would be nice if this could get cleaned up to make use of the the logger's stack trace printing. This also gives users more control since you can configure how the stack traces are written to the logs. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
Hadoop-Hdfs-trunk-Commit - Build # 676 - Still Failing
See https://builds.apache.org/hudson/job/Hadoop-Hdfs-trunk-Commit/676/ ### ## LAST 60 LINES OF THE CONSOLE ### [...truncated 2642 lines...] [junit] Running org.apache.hadoop.hdfs.server.datanode.TestDiskError [junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 8.916 sec [junit] Running org.apache.hadoop.hdfs.server.datanode.TestInterDatanodeProtocol [junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 5.196 sec [junit] Running org.apache.hadoop.hdfs.server.datanode.TestSimulatedFSDataset [junit] Tests run: 8, Failures: 0, Errors: 0, Time elapsed: 0.768 sec [junit] Running org.apache.hadoop.hdfs.server.namenode.TestBackupNode [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 18.288 sec [junit] Running org.apache.hadoop.hdfs.server.namenode.TestCheckpoint [junit] Tests run: 4, Failures: 0, Errors: 0, Time elapsed: 32.171 sec [junit] Running org.apache.hadoop.hdfs.server.namenode.TestComputeInvalidateWork [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 4.956 sec [junit] Running org.apache.hadoop.hdfs.server.namenode.TestDatanodeDescriptor [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 0.172 sec [junit] Running org.apache.hadoop.hdfs.server.namenode.TestEditLog [junit] Tests run: 5, Failures: 0, Errors: 0, Time elapsed: 13.244 sec [junit] Running org.apache.hadoop.hdfs.server.namenode.TestFileLimit [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 5.225 sec [junit] Running org.apache.hadoop.hdfs.server.namenode.TestHeartbeatHandling [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 3.1 sec [junit] Running org.apache.hadoop.hdfs.server.namenode.TestHost2NodesMap [junit] Tests run: 4, Failures: 0, Errors: 0, Time elapsed: 0.09 sec [junit] Running org.apache.hadoop.hdfs.server.namenode.TestNamenodeCapacityReport [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 3.175 sec [junit] Running org.apache.hadoop.hdfs.server.namenode.TestOverReplicatedBlocks [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 4.032 sec [junit] Running org.apache.hadoop.hdfs.server.namenode.TestPendingReplication [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 7.3 sec [junit] Running org.apache.hadoop.hdfs.server.namenode.TestReplicationPolicy [junit] Tests run: 8, Failures: 0, Errors: 0, Time elapsed: 0.064 sec [junit] Running org.apache.hadoop.hdfs.server.namenode.TestSafeMode [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 8.665 sec [junit] Running org.apache.hadoop.hdfs.server.namenode.TestStartup [junit] Tests run: 5, Failures: 0, Errors: 0, Time elapsed: 10.69 sec [junit] Running org.apache.hadoop.hdfs.server.namenode.TestStorageRestore [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 8.412 sec [junit] Running org.apache.hadoop.net.TestNetworkTopology [junit] Tests run: 8, Failures: 0, Errors: 0, Time elapsed: 0.111 sec [junit] Running org.apache.hadoop.security.TestPermission [junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 5.281 sec checkfailure: [touch] Creating /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-trunk-Commit/trunk/build/test/testsfailed BUILD FAILED /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-trunk-Commit/trunk/build.xml:712: The following error occurred while executing this line: /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-trunk-Commit/trunk/build.xml:669: The following error occurred while executing this line: /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-trunk-Commit/trunk/build.xml:737: Tests failed! Total time: 8 minutes 51 seconds [FINDBUGS] Skipping publisher since build result is FAILURE Recording fingerprints Archiving artifacts Recording test results Publishing Javadoc Publishing Clover coverage report... No Clover report will be published due to a Build Failure Email was triggered for: Failure Sending email for trigger: Failure ### ## FAILED TESTS (if any) ## 1 tests failed. FAILED: org.apache.hadoop.hdfs.TestHDFSTrash.testTrashEmptier Error Message: null Stack Trace: junit.framework.AssertionFailedError: null at org.apache.hadoop.fs.TestTrash.testTrashEmptier(TestTrash.java:479) at junit.extensions.TestDecorator.basicRun(TestDecorator.java:24) at junit.extensions.TestSetup$1.protect(TestSetup.java:23) at junit.extensions.TestSetup.run(TestSetup.java:27)