See <https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/491/changes>

Changes:

[eli] HDFS-1487. FSDirectory.removeBlock() should update diskspace count of the 
block owner node. Contributed by Zhong Wang.

[eli] HDFS-1507. TestAbandonBlock should abandon a block. Contributed by Eli 
Collins

[eli] HDFS-259. Remove intentionally corrupt 0.13 directory layout creation. 
Contributed by Todd Lipcon

[omalley] Branching for 0.22

------------------------------------------
[...truncated 756916 lines...]
    [junit]     at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    [junit]     at java.lang.reflect.Method.invoke(Method.java:597)
    [junit]     at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
    [junit]     at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
    [junit]     at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
    [junit]     at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
    [junit]     at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
    [junit]     at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
    [junit]     at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
    [junit]     at 
org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
    [junit]     at 
org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
    [junit]     at 
org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
    [junit]     at 
org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
    [junit]     at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
    [junit]     at 
junit.framework.JUnit4TestAdapter.run(JUnit4TestAdapter.java:39)
    [junit]     at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:420)
    [junit]     at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:911)
    [junit]     at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:768)
    [junit] 2010-11-18 16:32:45,327 INFO  datanode.DataNode 
(DataNode.java:initDataXceiver(467)) - Opened info server at 41536
    [junit] 2010-11-18 16:32:45,327 INFO  datanode.DataNode 
(DataXceiverServer.java:<init>(77)) - Balancing bandwith is 1048576 bytes/s
    [junit] 2010-11-18 16:32:45,329 INFO  common.Storage 
(DataStorage.java:recoverTransitionRead(127)) - Storage directory 
<https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data3>
 is not formatted.
    [junit] 2010-11-18 16:32:45,329 INFO  common.Storage 
(DataStorage.java:recoverTransitionRead(128)) - Formatting ...
    [junit] 2010-11-18 16:32:45,332 INFO  common.Storage 
(DataStorage.java:recoverTransitionRead(127)) - Storage directory 
<https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data4>
 is not formatted.
    [junit] 2010-11-18 16:32:45,332 INFO  common.Storage 
(DataStorage.java:recoverTransitionRead(128)) - Formatting ...
    [junit] 2010-11-18 16:32:45,381 INFO  datanode.DataNode 
(FSDataset.java:registerMBean(1772)) - Registered FSDatasetStatusMBean
    [junit] 2010-11-18 16:32:45,382 INFO  datanode.DirectoryScanner 
(DirectoryScanner.java:<init>(149)) - scan starts at 1290109618382 with 
interval 21600000
    [junit] 2010-11-18 16:32:45,383 INFO  http.HttpServer 
(HttpServer.java:addGlobalFilter(409)) - Added global filtersafety 
(class=org.apache.hadoop.http.HttpServer$QuotingInputFilter)
    [junit] 2010-11-18 16:32:45,384 DEBUG datanode.DataNode 
(DataNode.java:startInfoServer(336)) - Datanode listening on localhost:0
    [junit] 2010-11-18 16:32:45,384 INFO  http.HttpServer 
(HttpServer.java:start(579)) - Port returned by 
webServer.getConnectors()[0].getLocalPort() before open() is -1. Opening the 
listener on 0
    [junit] 2010-11-18 16:32:45,384 INFO  http.HttpServer 
(HttpServer.java:start(584)) - listener.getLocalPort() returned 50189 
webServer.getConnectors()[0].getLocalPort() returned 50189
    [junit] 2010-11-18 16:32:45,385 INFO  http.HttpServer 
(HttpServer.java:start(617)) - Jetty bound to port 50189
    [junit] 2010-11-18 16:32:45,385 INFO  mortbay.log (?:invoke(?)) - 
jetty-6.1.14
    [junit] 2010-11-18 16:32:45,461 INFO  mortbay.log (?:invoke(?)) - Started 
selectchannelconnec...@localhost:50189
    [junit] 2010-11-18 16:32:45,461 INFO  jvm.JvmMetrics 
(JvmMetrics.java:init(71)) - Cannot initialize JVM Metrics with 
processName=DataNode, sessionId=null - already initialized
    [junit] 2010-11-18 16:32:45,462 INFO  ipc.Server (Server.java:run(338)) - 
Starting SocketReader
    [junit] 2010-11-18 16:32:45,462 INFO  metrics.RpcMetrics 
(RpcMetrics.java:<init>(63)) - Initializing RPC Metrics with hostName=DataNode, 
port=43815
    [junit] 2010-11-18 16:32:45,464 INFO  metrics.RpcDetailedMetrics 
(RpcDetailedMetrics.java:<init>(57)) - Initializing RPC Metrics with 
hostName=DataNode, port=43815
    [junit] 2010-11-18 16:32:45,464 INFO  datanode.DataNode 
(DataNode.java:initIpcServer(427)) - dnRegistration = 
DatanodeRegistration(h8.grid.sp2.yahoo.net:41536, storageID=, infoPort=50189, 
ipcPort=43815)
    [junit] 2010-11-18 16:32:45,466 INFO  hdfs.StateChange 
(FSNamesystem.java:registerDatanode(2508)) - BLOCK* 
NameSystem.registerDatanode: node registration from 127.0.0.1:41536 storage 
DS-820573405-127.0.1.1-41536-1290097965465
    [junit] 2010-11-18 16:32:45,466 INFO  net.NetworkTopology 
(NetworkTopology.java:add(331)) - Adding a new node: 
/default-rack/127.0.0.1:41536
    [junit] 2010-11-18 16:32:45,471 INFO  datanode.DataNode 
(DataNode.java:register(697)) - New storage id 
DS-820573405-127.0.1.1-41536-1290097965465 is assigned to data-node 
127.0.0.1:41536
    [junit] 2010-11-18 16:32:45,472 INFO  datanode.DataNode 
(DataNode.java:run(1419)) - DatanodeRegistration(127.0.0.1:41536, 
storageID=DS-820573405-127.0.1.1-41536-1290097965465, infoPort=50189, 
ipcPort=43815)In DataNode.run, data = 
FSDataset{dirpath='<https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data3/current/finalized,/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-trunk/trunk/build-fi/test/data/dfs/data/data4/current/finalized'}>
    [junit] Starting DataNode 2 with dfs.datanode.data.dir: 
<https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data5/,file>:<https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data6/>
    [junit] 2010-11-18 16:32:45,479 INFO  ipc.Server (Server.java:run(608)) - 
IPC Server Responder: starting
    [junit] 2010-11-18 16:32:45,484 INFO  datanode.DataNode 
(DataNode.java:offerService(887)) - using BLOCKREPORT_INTERVAL of 21600000msec 
Initial delay: 0msec
    [junit] 2010-11-18 16:32:45,483 INFO  ipc.Server (Server.java:run(443)) - 
IPC Server listener on 43815: starting
    [junit] 2010-11-18 16:32:45,484 INFO  ipc.Server (Server.java:run(1369)) - 
IPC Server handler 0 on 43815: starting
    [junit] 2010-11-18 16:32:45,512 INFO  datanode.DataNode 
(DataNode.java:blockReport(1126)) - BlockReport of 0 blocks got processed in 16 
msecs
    [junit] 2010-11-18 16:32:45,513 INFO  datanode.DataNode 
(DataNode.java:offerService(929)) - Starting Periodic block scanner.
    [junit] 2010-11-18 16:32:45,536 WARN  datanode.DataNode 
(DataNode.java:registerMXBean(530)) - Failed to register NameNode MXBean
    [junit] javax.management.InstanceAlreadyExistsException: 
HadoopInfo:type=DataNodeInfo
    [junit]     at 
com.sun.jmx.mbeanserver.Repository.addMBean(Repository.java:453)
    [junit]     at 
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.internal_addObject(DefaultMBeanServerInterceptor.java:1484)
    [junit]     at 
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerDynamicMBean(DefaultMBeanServerInterceptor.java:963)
    [junit]     at 
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerObject(DefaultMBeanServerInterceptor.java:917)
    [junit]     at 
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerMBean(DefaultMBeanServerInterceptor.java:312)
    [junit]     at 
com.sun.jmx.mbeanserver.JmxMBeanServer.registerMBean(JmxMBeanServer.java:482)
    [junit]     at 
org.apache.hadoop.hdfs.server.datanode.DataNode.registerMXBean(DataNode.java:528)
    [junit]     at 
org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:498)
    [junit]     at 
org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:281)
    [junit]     at 
org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:266)
    [junit]     at 
org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1556)
    [junit]     at 
org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1499)
    [junit]     at 
org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1466)
    [junit]     at 
org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:614)
    [junit]     at 
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:448)
    [junit]     at 
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:176)
    [junit]     at 
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:71)
    [junit]     at 
org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:168)
    [junit]     at 
org.apache.hadoop.hdfs.server.datanode.TestFiDataTransferProtocol2.writeSeveralPackets(TestFiDataTransferProtocol2.java:91)
    [junit]     at 
org.apache.hadoop.hdfs.server.datanode.TestFiDataTransferProtocol2.runTest17_19(TestFiDataTransferProtocol2.java:138)
    [junit]     at 
org.apache.hadoop.hdfs.server.datanode.TestFiDataTransferProtocol2.pipeline_Fi_19(TestFiDataTransferProtocol2.java:198)
    [junit]     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    [junit]     at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    [junit]     at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    [junit]     at java.lang.reflect.Method.invoke(Method.java:597)
    [junit]     at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
    [junit]     at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
    [junit]     at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
    [junit]     at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
    [junit]     at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
    [junit]     at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
    [junit]     at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
    [junit]     at 
org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
    [junit]     at 
org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
    [junit]     at 
org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
    [junit]     at 
org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
    [junit]     at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
    [junit]     at 
junit.framework.JUnit4TestAdapter.run(JUnit4TestAdapter.java:39)
    [junit]     at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:420)
    [junit]     at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:911)
    [junit]     at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:768)
    [junit] 2010-11-18 16:32:45,538 INFO  datanode.DataNode 
(DataNode.java:initDataXceiver(467)) - Opened info server at 43306
    [junit] 2010-11-18 16:32:45,539 INFO  datanode.DataNode 
(DataXceiverServer.java:<init>(77)) - Balancing bandwith is 1048576 bytes/s
    [junit] 2010-11-18 16:32:45,541 INFO  common.Storage 
(DataStorage.java:recoverTransitionRead(127)) - Storage directory 
<https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data5>
 is not formatted.
    [junit] 2010-11-18 16:32:45,541 INFO  common.Storage 
(DataStorage.java:recoverTransitionRead(128)) - Formatting ...
    [junit] 2010-11-18 16:32:45,543 INFO  common.Storage 
(DataStorage.java:recoverTransitionRead(127)) - Storage directory 
<https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data6>
 is not formatted.
    [junit] 2010-11-18 16:32:45,544 INFO  common.Storage 
(DataStorage.java:recoverTransitionRead(128)) - Formatting ...
    [junit] 2010-11-18 16:32:45,582 INFO  datanode.DataNode 
(FSDataset.java:registerMBean(1772)) - Registered FSDatasetStatusMBean
    [junit] 2010-11-18 16:32:45,582 INFO  datanode.DirectoryScanner 
(DirectoryScanner.java:<init>(149)) - scan starts at 1290115083582 with 
interval 21600000
    [junit] 2010-11-18 16:32:45,584 INFO  http.HttpServer 
(HttpServer.java:addGlobalFilter(409)) - Added global filtersafety 
(class=org.apache.hadoop.http.HttpServer$QuotingInputFilter)
    [junit] 2010-11-18 16:32:45,584 DEBUG datanode.DataNode 
(DataNode.java:startInfoServer(336)) - Datanode listening on localhost:0
    [junit] 2010-11-18 16:32:45,585 INFO  http.HttpServer 
(HttpServer.java:start(579)) - Port returned by 
webServer.getConnectors()[0].getLocalPort() before open() is -1. Opening the 
listener on 0
    [junit] 2010-11-18 16:32:45,585 INFO  http.HttpServer 
(HttpServer.java:start(584)) - listener.getLocalPort() returned 60767 
webServer.getConnectors()[0].getLocalPort() returned 60767
    [junit] 2010-11-18 16:32:45,586 INFO  http.HttpServer 
(HttpServer.java:start(617)) - Jetty bound to port 60767
    [junit] 2010-11-18 16:32:45,586 INFO  mortbay.log (?:invoke(?)) - 
jetty-6.1.14
    [junit] 2010-11-18 16:32:45,675 INFO  mortbay.log (?:invoke(?)) - Started 
selectchannelconnec...@localhost:60767
    [junit] 2010-11-18 16:32:45,676 INFO  jvm.JvmMetrics 
(JvmMetrics.java:init(71)) - Cannot initialize JVM Metrics with 
processName=DataNode, sessionId=null - already initialized
    [junit] 2010-11-18 16:32:45,677 INFO  ipc.Server (Server.java:run(338)) - 
Starting SocketReader
    [junit] 2010-11-18 16:32:45,677 INFO  metrics.RpcMetrics 
(RpcMetrics.java:<init>(63)) - Initializing RPC Metrics with hostName=DataNode, 
port=52910
    [junit] 2010-11-18 16:32:45,678 INFO  metrics.RpcDetailedMetrics 
(RpcDetailedMetrics.java:<init>(57)) - Initializing RPC Metrics with 
hostName=DataNode, port=52910
    [junit] 2010-11-18 16:32:45,678 INFO  datanode.DataNode 
(DataNode.java:initIpcServer(427)) - dnRegistration = 
DatanodeRegistration(h8.grid.sp2.yahoo.net:43306, storageID=, infoPort=60767, 
ipcPort=52910)
    [junit] 2010-11-18 16:32:45,680 INFO  hdfs.StateChange 
(FSNamesystem.java:registerDatanode(2508)) - BLOCK* 
NameSystem.registerDatanode: node registration from 127.0.0.1:43306 storage 
DS-835262486-127.0.1.1-43306-1290097965679
    [junit] 2010-11-18 16:32:45,680 INFO  net.NetworkTopology 
(NetworkTopology.java:add(331)) - Adding a new node: 
/default-rack/127.0.0.1:43306
    [junit] 2010-11-18 16:32:45,684 INFO  datanode.DataNode 
(DataNode.java:register(697)) - New storage id 
DS-835262486-127.0.1.1-43306-1290097965679 is assigned to data-node 
127.0.0.1:43306
    [junit] 2010-11-18 16:32:45,685 INFO  datanode.DataNode 
(DataNode.java:run(1419)) - DatanodeRegistration(127.0.0.1:43306, 
storageID=DS-835262486-127.0.1.1-43306-1290097965679, infoPort=60767, 
ipcPort=52910)In DataNode.run, data = 
FSDataset{dirpath='<https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data5/current/finalized,/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-trunk/trunk/build-fi/test/data/dfs/data/data6/current/finalized'}>
    [junit] 2010-11-18 16:32:45,685 INFO  ipc.Server (Server.java:run(608)) - 
IPC Server Responder: starting
    [junit] 2010-11-18 16:32:45,686 INFO  ipc.Server (Server.java:run(443)) - 
IPC Server listener on 52910: starting
    [junit] 2010-11-18 16:32:45,686 INFO  ipc.Server (Server.java:run(1369)) - 
IPC Server handler 0 on 52910: starting
    [junit] 2010-11-18 16:32:45,687 INFO  datanode.DataNode 
(DataNode.java:offerService(887)) - using BLOCKREPORT_INTERVAL of 21600000msec 
Initial delay: 0msec
    [junit] 2010-11-18 16:32:45,692 INFO  datanode.DataNode 
(DataNode.java:blockReport(1126)) - BlockReport of 0 blocks got processed in 2 
msecs
    [junit] 2010-11-18 16:32:45,692 INFO  datanode.DataNode 
(DataNode.java:offerService(929)) - Starting Periodic block scanner.
    [junit] 2010-11-18 16:32:45,695 DEBUG hdfs.DFSClient 
(DFSClient.java:create(629)) - /pipeline_Fi_19/foo: masked=rwxr-xr-x
    [junit] 2010-11-18 16:32:45,695 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:computePacketChunkSize(1144)) - computePacketChunkSize: 
src=/pipeline_Fi_19/foo, chunkSize=516, chunksPerPacket=2, packetSize=1057
    [junit] 2010-11-18 16:32:45,708 INFO  FSNamesystem.audit 
(FSNamesystem.java:logAuditEvent(148)) - ugi=hudson        ip=/127.0.0.1   
cmd=create      src=/pipeline_Fi_19/foo dst=null        
perm=hudson:supergroup:rw-r--r--
    [junit] 2010-11-18 16:32:45,710 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:writeChunk(1202)) - DFSClient writeChunk allocating new 
packet seqno=0, src=/pipeline_Fi_19/foo, packetSize=1057, chunksPerPacket=2, 
bytesCurBlock=0
    [junit] 2010-11-18 16:32:45,711 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:writeChunk(1221)) - DFSClient writeChunk packet full 
seqno=0, src=/pipeline_Fi_19/foo, bytesCurBlock=1024, blockSize=1048576, 
appendChunk=false
    [junit] 2010-11-18 16:32:45,711 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:queueCurrentPacket(1157)) - Queued packet 0
    [junit] 2010-11-18 16:32:45,712 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:run(444)) - Allocating new block
    [junit] 2010-11-18 16:32:45,712 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:computePacketChunkSize(1144)) - computePacketChunkSize: 
src=/pipeline_Fi_19/foo, chunkSize=516, chunksPerPacket=2, packetSize=1057
    [junit] 2010-11-18 16:32:45,712 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:writeChunk(1202)) - DFSClient writeChunk allocating new 
packet seqno=1, src=/pipeline_Fi_19/foo, packetSize=1057, chunksPerPacket=2, 
bytesCurBlock=1024
    [junit] 2010-11-18 16:32:45,713 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:writeChunk(1221)) - DFSClient writeChunk packet full 
seqno=1, src=/pipeline_Fi_19/foo, bytesCurBlock=2048, blockSize=1048576, 
appendChunk=false
    [junit] 2010-11-18 16:32:45,713 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:queueCurrentPacket(1157)) - Queued packet 1
    [junit] 2010-11-18 16:32:45,713 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:computePacketChunkSize(1144)) - computePacketChunkSize: 
src=/pipeline_Fi_19/foo, chunkSize=516, chunksPerPacket=2, packetSize=1057
    [junit] 2010-11-18 16:32:45,713 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:writeChunk(1202)) - DFSClient writeChunk allocating new 
packet seqno=2, src=/pipeline_Fi_19/foo, packetSize=1057, chunksPerPacket=2, 
bytesCurBlock=2048
    [junit] 2010-11-18 16:32:45,713 INFO  hdfs.StateChange 
(FSNamesystem.java:allocateBlock(1753)) - BLOCK* NameSystem.allocateBlock: 
/pipeline_Fi_19/foo. 
blk_-1087155876419230760_1001{blockUCState=UNDER_CONSTRUCTION, 
primaryNodeIndex=-1, replicas=[ReplicaUnderConstruction[127.0.0.1:41536|RBW], 
ReplicaUnderConstruction[127.0.0.1:43306|RBW], 
ReplicaUnderConstruction[127.0.0.1:36464|RBW]]}
    [junit] 2010-11-18 16:32:45,714 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:writeChunk(1221)) - DFSClient writeChunk packet full 
seqno=2, src=/pipeline_Fi_19/foo, bytesCurBlock=3072, blockSize=1048576, 
appendChunk=false
    [junit] 2010-11-18 16:32:45,714 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:queueCurrentPacket(1157)) - Queued packet 2
    [junit] 2010-11-18 16:32:45,715 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:computePacketChunkSize(1144)) - computePacketChunkSize: 
src=/pipeline_Fi_19/foo, chunkSize=516, chunksPerPacket=2, packetSize=1057
    [junit] 2010-11-18 16:32:45,715 INFO  protocol.ClientProtocolAspects 
(ClientProtocolAspects.aj:ajc$afterReturning$org_apache_hadoop_hdfs_protocol_ClientProtocolAspects$1$7076326d(35))
 - FI: addBlock Pipeline[127.0.0.1:41536, 127.0.0.1:43306, 127.0.0.1:36464]
    [junit] 2010-11-18 16:32:45,715 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:createBlockOutputStream(881)) - pipeline = 127.0.0.1:41536
    [junit] 2010-11-18 16:32:45,715 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:createBlockOutputStream(881)) - pipeline = 127.0.0.1:43306
    [junit] 2010-11-18 16:32:45,715 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:createBlockOutputStream(881)) - pipeline = 127.0.0.1:36464
    [junit] 2010-11-18 16:32:45,716 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:createBlockOutputStream(891)) - Connecting to 
127.0.0.1:41536
    [junit] 2010-11-18 16:32:45,716 DEBUG datanode.DataNode 
(DataXceiver.java:<init>(86)) - Number of active connections is: 1
    [junit] 2010-11-18 16:32:45,716 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:createBlockOutputStream(900)) - Send buf size 131071
    [junit] 2010-11-18 16:32:45,717 INFO  datanode.DataTransferProtocolAspects 
(DataTransferProtocolAspects.aj:ajc$afterReturning$org_apache_hadoop_hdfs_server_datanode_DataTransferProtocolAspects$1$8f59fdd7(51))
 - FI: receiverOp WRITE_BLOCK, datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:45,717 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:writeChunk(1202)) - DFSClient writeChunk allocating new 
packet seqno=3, src=/pipeline_Fi_19/foo, packetSize=1057, chunksPerPacket=2, 
bytesCurBlock=3072
    [junit] 2010-11-18 16:32:45,717 INFO  datanode.DataTransferProtocolAspects 
(DataTransferProtocolAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_DataTransferProtocolAspects$3$3251489(73))
 - FI: receiverOpWriteBlock
    [junit] 2010-11-18 16:32:45,717 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:writeChunk(1221)) - DFSClient writeChunk packet full 
seqno=3, src=/pipeline_Fi_19/foo, bytesCurBlock=4096, blockSize=1048576, 
appendChunk=false
    [junit] 2010-11-18 16:32:45,718 INFO  fi.FiTestUtil 
(DataTransferTestUtil.java:run(344)) - FI: SleepAction:pipeline_Fi_19, index=0, 
duration=[0, 3000), datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:45,718 INFO  fi.FiTestUtil 
(FiTestUtil.java:initialValue(37)) - Thread[DataXceiver for client 
/127.0.0.1:44424 [Waiting for operation],5,dataXceiverServer]: 
seed=-5326246243198308466
    [junit] 2010-11-18 16:32:45,718 INFO  fi.FiTestUtil 
(FiTestUtil.java:sleep(92)) - DataXceiver for client /127.0.0.1:44424 [Waiting 
for operation] sleeps for 1556ms
    [junit] 2010-11-18 16:32:45,718 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:queueCurrentPacket(1157)) - Queued packet 3
    [junit] 2010-11-18 16:32:45,719 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:computePacketChunkSize(1144)) - computePacketChunkSize: 
src=/pipeline_Fi_19/foo, chunkSize=516, chunksPerPacket=2, packetSize=1057
    [junit] 2010-11-18 16:32:45,719 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:writeChunk(1202)) - DFSClient writeChunk allocating new 
packet seqno=4, src=/pipeline_Fi_19/foo, packetSize=1057, chunksPerPacket=2, 
bytesCurBlock=4096
    [junit] 2010-11-18 16:32:45,719 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:writeChunk(1221)) - DFSClient writeChunk packet full 
seqno=4, src=/pipeline_Fi_19/foo, bytesCurBlock=5120, blockSize=1048576, 
appendChunk=false
    [junit] 2010-11-18 16:32:45,720 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:queueCurrentPacket(1157)) - Queued packet 4
    [junit] 2010-11-18 16:32:45,720 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:computePacketChunkSize(1144)) - computePacketChunkSize: 
src=/pipeline_Fi_19/foo, chunkSize=516, chunksPerPacket=2, packetSize=1057
    [junit] 2010-11-18 16:32:45,720 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:writeChunk(1202)) - DFSClient writeChunk allocating new 
packet seqno=5, src=/pipeline_Fi_19/foo, packetSize=1057, chunksPerPacket=2, 
bytesCurBlock=5120
    [junit] 2010-11-18 16:32:45,721 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:writeChunk(1221)) - DFSClient writeChunk packet full 
seqno=5, src=/pipeline_Fi_19/foo, bytesCurBlock=6144, blockSize=1048576, 
appendChunk=false
    [junit] 2010-11-18 16:32:45,721 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:queueCurrentPacket(1157)) - Queued packet 5
    [junit] 2010-11-18 16:32:45,721 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:computePacketChunkSize(1144)) - computePacketChunkSize: 
src=/pipeline_Fi_19/foo, chunkSize=516, chunksPerPacket=2, packetSize=1057
    [junit] 2010-11-18 16:32:45,721 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:writeChunk(1202)) - DFSClient writeChunk allocating new 
packet seqno=6, src=/pipeline_Fi_19/foo, packetSize=1057, chunksPerPacket=2, 
bytesCurBlock=6144
    [junit] 2010-11-18 16:32:45,722 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:queueCurrentPacket(1157)) - Queued packet 6
    [junit] 2010-11-18 16:32:45,722 INFO  hdfs.DFSClientAspects 
(DFSClientAspects.aj:ajc$before$org_apache_hadoop_hdfs_DFSClientAspects$5$5ba7280d(86))
 - FI: before pipelineClose:
    [junit] 2010-11-18 16:32:45,722 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:queueCurrentPacket(1157)) - Queued packet 7
    [junit] 2010-11-18 16:32:45,723 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:waitForAckedSeqno(1408)) - Waiting for ack for: 7
    [junit] 2010-11-18 16:32:47,275 DEBUG datanode.DataNode 
(DataXceiver.java:opWriteBlock(246)) - writeBlock receive buf size 131071 tcp 
no delay true
    [junit] 2010-11-18 16:32:47,275 INFO  datanode.DataNode 
(DataXceiver.java:opWriteBlock(251)) - Receiving block 
blk_-1087155876419230760_1001 src: /127.0.0.1:44424 dest: /127.0.0.1:41536
    [junit] 2010-11-18 16:32:47,276 DEBUG datanode.DataNode 
(ReplicaInPipeline.java:createStreams(176)) - writeTo blockfile is 
<https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data3/current/rbw/blk_-1087155876419230760>
 of size 0
    [junit] 2010-11-18 16:32:47,276 DEBUG datanode.DataNode 
(ReplicaInPipeline.java:createStreams(178)) - writeTo metafile is 
<https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data3/current/rbw/blk_-1087155876419230760_1001.meta>
 of size 0
    [junit] 2010-11-18 16:32:47,277 DEBUG datanode.DataNode 
(DataXceiver.java:<init>(86)) - Number of active connections is: 1
    [junit] 2010-11-18 16:32:47,287 INFO  datanode.DataTransferProtocolAspects 
(DataTransferProtocolAspects.aj:ajc$afterReturning$org_apache_hadoop_hdfs_server_datanode_DataTransferProtocolAspects$1$8f59fdd7(51))
 - FI: receiverOp WRITE_BLOCK, datanode=127.0.0.1:43306
    [junit] 2010-11-18 16:32:47,287 INFO  datanode.DataTransferProtocolAspects 
(DataTransferProtocolAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_DataTransferProtocolAspects$3$3251489(73))
 - FI: receiverOpWriteBlock
    [junit] 2010-11-18 16:32:47,287 INFO  fi.FiTestUtil 
(DataTransferTestUtil.java:run(344)) - FI: SleepAction:pipeline_Fi_19, index=1, 
duration=[0, 3000), datanode=127.0.0.1:43306
    [junit] 2010-11-18 16:32:47,287 INFO  fi.FiTestUtil 
(FiTestUtil.java:initialValue(37)) - Thread[DataXceiver for client 
/127.0.0.1:60263 [Waiting for operation],5,dataXceiverServer]: 
seed=7365113689811878303
    [junit] 2010-11-18 16:32:47,287 INFO  fi.FiTestUtil 
(FiTestUtil.java:sleep(92)) - DataXceiver for client /127.0.0.1:60263 [Waiting 
for operation] sleeps for 506ms
    [junit] 2010-11-18 16:32:47,794 DEBUG datanode.DataNode 
(DataXceiver.java:opWriteBlock(246)) - writeBlock receive buf size 131071 tcp 
no delay true
    [junit] 2010-11-18 16:32:47,794 INFO  datanode.DataNode 
(DataXceiver.java:opWriteBlock(251)) - Receiving block 
blk_-1087155876419230760_1001 src: /127.0.0.1:60263 dest: /127.0.0.1:43306
    [junit] 2010-11-18 16:32:47,795 DEBUG datanode.DataNode 
(ReplicaInPipeline.java:createStreams(176)) - writeTo blockfile is 
<https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data5/current/rbw/blk_-1087155876419230760>
 of size 0
    [junit] 2010-11-18 16:32:47,795 DEBUG datanode.DataNode 
(ReplicaInPipeline.java:createStreams(178)) - writeTo metafile is 
<https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data5/current/rbw/blk_-1087155876419230760_1001.meta>
 of size 0
    [junit] 2010-11-18 16:32:47,796 DEBUG datanode.DataNode 
(DataXceiver.java:<init>(86)) - Number of active connections is: 1
    [junit] 2010-11-18 16:32:47,796 INFO  datanode.DataTransferProtocolAspects 
(DataTransferProtocolAspects.aj:ajc$afterReturning$org_apache_hadoop_hdfs_server_datanode_DataTransferProtocolAspects$1$8f59fdd7(51))
 - FI: receiverOp WRITE_BLOCK, datanode=127.0.0.1:36464
    [junit] 2010-11-18 16:32:47,796 INFO  datanode.DataTransferProtocolAspects 
(DataTransferProtocolAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_DataTransferProtocolAspects$3$3251489(73))
 - FI: receiverOpWriteBlock
    [junit] 2010-11-18 16:32:47,796 INFO  fi.FiTestUtil 
(DataTransferTestUtil.java:run(344)) - FI: SleepAction:pipeline_Fi_19, index=2, 
duration=[0, 3000), datanode=127.0.0.1:36464
    [junit] 2010-11-18 16:32:47,797 INFO  fi.FiTestUtil 
(FiTestUtil.java:initialValue(37)) - Thread[DataXceiver for client 
/127.0.0.1:59520 [Waiting for operation],5,dataXceiverServer]: 
seed=-3002906034623524893
    [junit] 2010-11-18 16:32:47,797 INFO  fi.FiTestUtil 
(FiTestUtil.java:sleep(92)) - DataXceiver for client /127.0.0.1:59520 [Waiting 
for operation] sleeps for 1956ms
    [junit] 2010-11-18 16:32:49,753 DEBUG datanode.DataNode 
(DataXceiver.java:opWriteBlock(246)) - writeBlock receive buf size 131071 tcp 
no delay true
    [junit] 2010-11-18 16:32:49,753 INFO  datanode.DataNode 
(DataXceiver.java:opWriteBlock(251)) - Receiving block 
blk_-1087155876419230760_1001 src: /127.0.0.1:59520 dest: /127.0.0.1:36464
    [junit] 2010-11-18 16:32:49,754 DEBUG datanode.DataNode 
(ReplicaInPipeline.java:createStreams(176)) - writeTo blockfile is 
<https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data1/current/rbw/blk_-1087155876419230760>
 of size 0
    [junit] 2010-11-18 16:32:49,754 DEBUG datanode.DataNode 
(ReplicaInPipeline.java:createStreams(178)) - writeTo metafile is 
<https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data1/current/rbw/blk_-1087155876419230760_1001.meta>
 of size 0
    [junit] 2010-11-18 16:32:49,755 INFO  datanode.DataNode 
(DataXceiver.java:opWriteBlock(371)) - Datanode 0 forwarding connect ack to 
upstream firstbadlink is 
    [junit] 2010-11-18 16:32:49,755 INFO  datanode.DataTransferProtocolAspects 
(DataTransferProtocolAspects.aj:ajc$afterReturning$org_apache_hadoop_hdfs_server_datanode_DataTransferProtocolAspects$2$d4f6605f(61))
 - FI: statusRead SUCCESS, datanode=127.0.0.1:43306
    [junit] 2010-11-18 16:32:49,755 INFO  datanode.BlockReceiverAspects 
(BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
 - FI: callReceivePacket, datanode=127.0.0.1:36464
    [junit] 2010-11-18 16:32:49,755 INFO  fi.FiTestUtil 
(DataTransferTestUtil.java:run(344)) - FI: SleepAction:pipeline_Fi_19, index=1, 
duration=[0, 3000), datanode=127.0.0.1:43306
    [junit] 2010-11-18 16:32:49,756 INFO  fi.FiTestUtil 
(FiTestUtil.java:sleep(92)) - DataXceiver for client /127.0.0.1:60263 
[Receiving block blk_-1087155876419230760_1001 client=DFSClient_-244463401] 
sleeps for 1646ms
    [junit] 2010-11-18 16:32:49,755 DEBUG datanode.DataNode 
(BlockReceiver.java:run(843)) - PacketResponder 0 seqno = -2 for block 
blk_-1087155876419230760_1001 waiting for local datanode to finish write.
    [junit] 2010-11-18 16:32:49,755 INFO  fi.FiTestUtil 
(DataTransferTestUtil.java:run(344)) - FI: SleepAction:pipeline_Fi_19, index=2, 
duration=[0, 3000), datanode=127.0.0.1:36464
    [junit] 2010-11-18 16:32:49,756 INFO  fi.FiTestUtil 
(FiTestUtil.java:sleep(92)) - DataXceiver for client /127.0.0.1:59520 
[Receiving block blk_-1087155876419230760_1001 client=DFSClient_-244463401] 
sleeps for 1012ms
    [junit] 2010-11-18 16:32:51,402 INFO  datanode.DataNode 
(DataXceiver.java:opWriteBlock(338)) - Datanode 1 got response for connect ack  
from downstream datanode with firstbadlink as 
    [junit] 2010-11-18 16:32:51,402 INFO  datanode.DataNode 
(DataXceiver.java:opWriteBlock(371)) - Datanode 1 forwarding connect ack to 
upstream firstbadlink is 
    [junit] 2010-11-18 16:32:51,403 INFO  datanode.DataTransferProtocolAspects 
(DataTransferProtocolAspects.aj:ajc$afterReturning$org_apache_hadoop_hdfs_server_datanode_DataTransferProtocolAspects$2$d4f6605f(61))
 - FI: statusRead SUCCESS, datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:51,403 INFO  datanode.BlockReceiverAspects 
(BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
 - FI: callReceivePacket, datanode=127.0.0.1:43306
    [junit] 2010-11-18 16:32:51,403 INFO  fi.FiTestUtil 
(DataTransferTestUtil.java:run(344)) - FI: SleepAction:pipeline_Fi_19, index=0, 
duration=[0, 3000), datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:51,403 INFO  fi.FiTestUtil 
(DataTransferTestUtil.java:run(344)) - FI: SleepAction:pipeline_Fi_19, index=1, 
duration=[0, 3000), datanode=127.0.0.1:43306
    [junit] 2010-11-18 16:32:51,403 INFO  fi.FiTestUtil 
(FiTestUtil.java:sleep(92)) - DataXceiver for client /127.0.0.1:44424 
[Receiving block blk_-1087155876419230760_1001 client=DFSClient_-244463401] 
sleeps for 1803ms
    [junit] 2010-11-18 16:32:51,403 INFO  fi.FiTestUtil 
(FiTestUtil.java:sleep(92)) - DataXceiver for client /127.0.0.1:60263 
[Receiving block blk_-1087155876419230760_1001 client=DFSClient_-244463401] 
sleeps for 2647ms
    [junit] 2010-11-18 16:32:53,207 INFO  datanode.DataNode 
(DataXceiver.java:opWriteBlock(338)) - Datanode 2 got response for connect ack  
from downstream datanode with firstbadlink as 
    [junit] 2010-11-18 16:32:53,207 INFO  datanode.DataNode 
(DataXceiver.java:opWriteBlock(371)) - Datanode 2 forwarding connect ack to 
upstream firstbadlink is 
    [junit] 2010-11-18 16:32:53,207 INFO  datanode.BlockReceiverAspects 
(BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
 - FI: callReceivePacket, datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:53,207 INFO  hdfs.DFSClientAspects 
(DFSClientAspects.aj:ajc$afterReturning$org_apache_hadoop_hdfs_DFSClientAspects$2$9396d2df(48))
 - FI: after pipelineInitNonAppend: hasError=false errorIndex=-1
    [junit] 2010-11-18 16:32:53,207 INFO  fi.FiTestUtil 
(DataTransferTestUtil.java:run(344)) - FI: SleepAction:pipeline_Fi_19, index=0, 
duration=[0, 3000), datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:53,208 INFO  fi.FiTestUtil 
(FiTestUtil.java:sleep(92)) - DataXceiver for client /127.0.0.1:44424 
[Receiving block blk_-1087155876419230760_1001 client=DFSClient_-244463401] 
sleeps for 2638ms
    [junit] 2010-11-18 16:32:53,208 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:run(496)) - DataStreamer block 
blk_-1087155876419230760_1001 sending packet packet seqno:0 offsetInBlock:0 
lastPacketInBlock:false lastByteOffsetInBlock: 1024
    [junit] 2010-11-18 16:32:53,208 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:run(496)) - DataStreamer block 
blk_-1087155876419230760_1001 sending packet packet seqno:1 offsetInBlock:1024 
lastPacketInBlock:false lastByteOffsetInBlock: 2048
    [junit] 2010-11-18 16:32:53,209 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:run(496)) - DataStreamer block 
blk_-1087155876419230760_1001 sending packet packet seqno:2 offsetInBlock:2048 
lastPacketInBlock:false lastByteOffsetInBlock: 3072
    [junit] 2010-11-18 16:32:53,209 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:run(496)) - DataStreamer block 
blk_-1087155876419230760_1001 sending packet packet seqno:3 offsetInBlock:3072 
lastPacketInBlock:false lastByteOffsetInBlock: 4096
    [junit] 2010-11-18 16:32:53,209 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:run(496)) - DataStreamer block 
blk_-1087155876419230760_1001 sending packet packet seqno:4 offsetInBlock:4096 
lastPacketInBlock:false lastByteOffsetInBlock: 5120
    [junit] 2010-11-18 16:32:53,209 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:run(496)) - DataStreamer block 
blk_-1087155876419230760_1001 sending packet packet seqno:5 offsetInBlock:5120 
lastPacketInBlock:false lastByteOffsetInBlock: 6144
    [junit] 2010-11-18 16:32:53,209 DEBUG hdfs.DFSClient 
(DFSOutputStream.java:run(496)) - DataStreamer block 
blk_-1087155876419230760_1001 sending packet packet seqno:6 offsetInBlock:6144 
lastPacketInBlock:false lastByteOffsetInBlock: 6170
    [junit] 2010-11-18 16:32:55,846 INFO  datanode.BlockReceiverAspects 
(BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
 - FI: callReceivePacket, datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:55,847 INFO  fi.FiTestUtil 
(DataTransferTestUtil.java:run(344)) - FI: SleepAction:pipeline_Fi_19, index=0, 
duration=[0, 3000), datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:55,847 INFO  fi.FiTestUtil 
(FiTestUtil.java:sleep(92)) - DataXceiver for client /127.0.0.1:44424 
[Receiving block blk_-1087155876419230760_1001 client=DFSClient_-244463401] 
sleeps for 863ms
    [junit] 2010-11-18 16:32:56,710 DEBUG datanode.DataNode 
(BlockReceiver.java:receivePacket(456)) - Receiving one packet for block 
blk_-1087155876419230760_1001 of length 1024 seqno 0 offsetInBlock 0 
lastPacketInBlock false
    [junit] 2010-11-18 16:32:56,710 DEBUG datanode.DataNode 
(BlockReceiver.java:enqueue(788)) - PacketResponder 2 adding seqno 0 to ack 
queue.
    [junit] 2010-11-18 16:32:56,710 INFO  datanode.BlockReceiverAspects 
(BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$2$56c32214(71))
 - FI: callWritePacketToDisk
    [junit] 2010-11-18 16:32:56,710 INFO  datanode.BlockReceiverAspects 
(BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
 - FI: callReceivePacket, datanode=127.0.0.1:43306
    [junit] 2010-11-18 16:32:56,710 INFO  datanode.BlockReceiverAspects 
(BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
 - FI: callReceivePacket, datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:56,711 INFO  fi.FiTestUtil 
(DataTransferTestUtil.java:run(344)) - FI: SleepAction:pipeline_Fi_19, index=1, 
duration=[0, 3000), datanode=127.0.0.1:43306
    [junit] 2010-11-18 16:32:56,711 INFO  fi.FiTestUtil 
(DataTransferTestUtil.java:run(344)) - FI: SleepAction:pipeline_Fi_19, index=0, 
duration=[0, 3000), datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:56,711 INFO  fi.FiTestUtil 
(FiTestUtil.java:sleep(92)) - DataXceiver for client /127.0.0.1:60263 
[Receiving block blk_-1087155876419230760_1001 client=DFSClient_-244463401] 
sleeps for 2454ms
    [junit] 2010-11-18 16:32:56,711 INFO  fi.FiTestUtil 
(FiTestUtil.java:sleep(92)) - DataXceiver for client /127.0.0.1:44424 
[Receiving block blk_-1087155876419230760_1001 client=DFSClient_-244463401] 
sleeps for 353ms
    [junit] 2010-11-18 16:32:57,065 INFO  datanode.BlockReceiverAspects 
(BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
 - FI: callReceivePacket, datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:57,065 INFO  fi.FiTestUtil 
(DataTransferTestUtil.java:run(344)) - FI: SleepAction:pipeline_Fi_19, index=0, 
duration=[0, 3000), datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:57,065 INFO  fi.FiTestUtil 
(FiTestUtil.java:sleep(92)) - DataXceiver for client /127.0.0.1:44424 
[Receiving block blk_-1087155876419230760_1001 client=DFSClient_-244463401] 
sleeps for 377ms
    [junit] 2010-11-18 16:32:57,442 DEBUG datanode.DataNode 
(BlockReceiver.java:receivePacket(456)) - Receiving one packet for block 
blk_-1087155876419230760_1001 of length 1024 seqno 1 offsetInBlock 1024 
lastPacketInBlock false
    [junit] 2010-11-18 16:32:57,442 DEBUG datanode.DataNode 
(BlockReceiver.java:enqueue(788)) - PacketResponder 2 adding seqno 1 to ack 
queue.
    [junit] 2010-11-18 16:32:57,443 INFO  datanode.BlockReceiverAspects 
(BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$2$56c32214(71))
 - FI: callWritePacketToDisk
    [junit] 2010-11-18 16:32:57,443 INFO  datanode.BlockReceiverAspects 
(BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
 - FI: callReceivePacket, datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:57,443 INFO  fi.FiTestUtil 
(DataTransferTestUtil.java:run(344)) - FI: SleepAction:pipeline_Fi_19, index=0, 
duration=[0, 3000), datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:57,443 INFO  fi.FiTestUtil 
(FiTestUtil.java:sleep(92)) - DataXceiver for client /127.0.0.1:44424 
[Receiving block blk_-1087155876419230760_1001 client=DFSClient_-244463401] 
sleeps for 1915ms
Build timed out. Aborting
Publishing Javadoc
Archiving artifacts
Recording test results
Recording fingerprints
Publishing Clover coverage report...
No Clover report will be published due to a Build Failure

Reply via email to