See <https://builds.apache.org/job/Pig-trunk/1048/changes>
Changes: [daijy] Make Pig work with hadoop .NEXT [daijy] PIG-2159: New logical plan uses incorrect class for SUM causing for ClassCastException [daijy] PIG-2172: Fix test failure for ant 1.8.x ------------------------------------------ [...truncated 39329 lines...] [junit] at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508) [junit] at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959) [junit] at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955) [junit] at java.security.AccessController.doPrivileged(Native Method) [junit] at javax.security.auth.Subject.doAs(Subject.java:396) [junit] at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953) [junit] [junit] org.apache.hadoop.ipc.RemoteException: java.io.IOException: Could not complete write to file /tmp/TestStore-output-4310659416812167551.txt_cleanupOnFailure_succeeded by DFSClient_1282029458 [junit] at org.apache.hadoop.hdfs.server.namenode.NameNode.complete(NameNode.java:449) [junit] at sun.reflect.GeneratedMethodAccessor18.invoke(Unknown Source) [junit] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) [junit] at java.lang.reflect.Method.invoke(Method.java:597) [junit] at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508) [junit] at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959) [junit] at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955) [junit] at java.security.AccessController.doPrivileged(Native Method) [junit] at javax.security.auth.Subject.doAs(Subject.java:396) [junit] at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953) [junit] [junit] at org.apache.hadoop.ipc.Client.call(Client.java:740) [junit] at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220) [junit] at $Proxy0.complete(Unknown Source) [junit] at sun.reflect.GeneratedMethodAccessor18.invoke(Unknown Source) [junit] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) [junit] at java.lang.reflect.Method.invoke(Method.java:597) [junit] at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82) [junit] at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59) [junit] at $Proxy0.complete(Unknown Source) [junit] at org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.closeInternal(DFSClient.java:3264) [junit] at org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.close(DFSClient.java:3188) [junit] at org.apache.hadoop.hdfs.DFSClient$LeaseChecker.close(DFSClient.java:1043) [junit] at org.apache.hadoop.hdfs.DFSClient.close(DFSClient.java:237) [junit] at org.apache.hadoop.hdfs.DistributedFileSystem.close(DistributedFileSystem.java:269) [junit] at org.apache.pig.test.MiniGenericCluster.shutdownMiniDfsClusters(MiniGenericCluster.java:83) [junit] at org.apache.pig.test.MiniGenericCluster.shutdownMiniDfsAndMrClusters(MiniGenericCluster.java:77) [junit] at org.apache.pig.test.MiniGenericCluster.shutDown(MiniGenericCluster.java:68) [junit] at org.apache.pig.test.TestStore.oneTimeTearDown(TestStore.java:127) [junit] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) [junit] at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) [junit] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) [junit] at java.lang.reflect.Method.invoke(Method.java:597) [junit] at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44) [junit] at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15) [junit] at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41) [junit] at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:37) [junit] at org.junit.runners.ParentRunner.run(ParentRunner.java:220) [junit] at junit.framework.JUnit4TestAdapter.run(JUnit4TestAdapter.java:39) [junit] at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:420) [junit] at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:911) [junit] at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:768) [junit] 11/07/19 10:33:39 WARN hdfs.StateChange: DIR* NameSystem.completeFile: failed to complete /tmp/TestStore-output-6181164642597706789.txt_cleanupOnFailure_succeeded1 because dir.getFileBlocks() is null and pendingFile is null [junit] 11/07/19 10:33:39 INFO ipc.Server: IPC Server handler 2 on 41941, call complete(/tmp/TestStore-output-6181164642597706789.txt_cleanupOnFailure_succeeded1, DFSClient_1282029458) from 127.0.0.1:36483: error: java.io.IOException: Could not complete write to file /tmp/TestStore-output-6181164642597706789.txt_cleanupOnFailure_succeeded1 by DFSClient_1282029458 [junit] java.io.IOException: Could not complete write to file /tmp/TestStore-output-6181164642597706789.txt_cleanupOnFailure_succeeded1 by DFSClient_1282029458 [junit] at org.apache.hadoop.hdfs.server.namenode.NameNode.complete(NameNode.java:449) [junit] at sun.reflect.GeneratedMethodAccessor18.invoke(Unknown Source) [junit] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) [junit] at java.lang.reflect.Method.invoke(Method.java:597) [junit] at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508) [junit] at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959) [junit] at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955) [junit] at java.security.AccessController.doPrivileged(Native Method) [junit] at javax.security.auth.Subject.doAs(Subject.java:396) [junit] at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953) [junit] 11/07/19 10:33:39 ERROR hdfs.DFSClient: Exception closing file /tmp/TestStore-output-6181164642597706789.txt_cleanupOnFailure_succeeded1 : org.apache.hadoop.ipc.RemoteException: java.io.IOException: Could not complete write to file /tmp/TestStore-output-6181164642597706789.txt_cleanupOnFailure_succeeded1 by DFSClient_1282029458 [junit] at org.apache.hadoop.hdfs.server.namenode.NameNode.complete(NameNode.java:449) [junit] at sun.reflect.GeneratedMethodAccessor18.invoke(Unknown Source) [junit] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) [junit] at java.lang.reflect.Method.invoke(Method.java:597) [junit] at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508) [junit] at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959) [junit] at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955) [junit] at java.security.AccessController.doPrivileged(Native Method) [junit] at javax.security.auth.Subject.doAs(Subject.java:396) [junit] at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953) [junit] [junit] org.apache.hadoop.ipc.RemoteException: java.io.IOException: Could not complete write to file /tmp/TestStore-output-6181164642597706789.txt_cleanupOnFailure_succeeded1 by DFSClient_1282029458 [junit] at org.apache.hadoop.hdfs.server.namenode.NameNode.complete(NameNode.java:449) [junit] at sun.reflect.GeneratedMethodAccessor18.invoke(Unknown Source) [junit] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) [junit] at java.lang.reflect.Method.invoke(Method.java:597) [junit] at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508) [junit] at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959) [junit] at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955) [junit] at java.security.AccessController.doPrivileged(Native Method) [junit] at javax.security.auth.Subject.doAs(Subject.java:396) [junit] at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953) [junit] [junit] at org.apache.hadoop.ipc.Client.call(Client.java:740) [junit] at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220) [junit] at $Proxy0.complete(Unknown Source) [junit] at sun.reflect.GeneratedMethodAccessor18.invoke(Unknown Source) [junit] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) [junit] at java.lang.reflect.Method.invoke(Method.java:597) [junit] at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82) [junit] at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59) [junit] at $Proxy0.complete(Unknown Source) [junit] at org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.closeInternal(DFSClient.java:3264) [junit] at org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.close(DFSClient.java:3188) [junit] at org.apache.hadoop.hdfs.DFSClient$LeaseChecker.close(DFSClient.java:1043) [junit] at org.apache.hadoop.hdfs.DFSClient.close(DFSClient.java:237) [junit] at org.apache.hadoop.hdfs.DistributedFileSystem.close(DistributedFileSystem.java:269) [junit] at org.apache.pig.test.MiniGenericCluster.shutdownMiniDfsClusters(MiniGenericCluster.java:83) [junit] at org.apache.pig.test.MiniGenericCluster.shutdownMiniDfsAndMrClusters(MiniGenericCluster.java:77) [junit] at org.apache.pig.test.MiniGenericCluster.shutDown(MiniGenericCluster.java:68) [junit] at org.apache.pig.test.TestStore.oneTimeTearDown(TestStore.java:127) [junit] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) [junit] at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) [junit] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) [junit] at java.lang.reflect.Method.invoke(Method.java:597) [junit] at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44) [junit] at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15) [junit] at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41) [junit] at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:37) [junit] at org.junit.runners.ParentRunner.run(ParentRunner.java:220) [junit] at junit.framework.JUnit4TestAdapter.run(JUnit4TestAdapter.java:39) [junit] at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:420) [junit] at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:911) [junit] at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:768) [junit] Shutting down the Mini HDFS Cluster [junit] Shutting down DataNode 3 [junit] 11/07/19 10:33:39 INFO ipc.Server: Stopping server on 38451 [junit] 11/07/19 10:33:39 INFO ipc.Server: IPC Server handler 0 on 38451: exiting [junit] 11/07/19 10:33:39 INFO datanode.DataNode: Waiting for threadgroup to exit, active threads is 1 [junit] 11/07/19 10:33:39 INFO ipc.Server: Stopping IPC Server Responder [junit] 11/07/19 10:33:39 INFO ipc.Server: Stopping IPC Server listener on 38451 [junit] 11/07/19 10:33:39 INFO ipc.Server: IPC Server handler 2 on 38451: exiting [junit] 11/07/19 10:33:39 INFO ipc.Server: IPC Server handler 1 on 38451: exiting [junit] 11/07/19 10:33:39 WARN datanode.DataNode: DatanodeRegistration(127.0.0.1:48248, storageID=DS-20126214-127.0.1.1-48248-1311071251998, infoPort=40975, ipcPort=38451):DataXceiveServer: java.nio.channels.AsynchronousCloseException [junit] at java.nio.channels.spi.AbstractInterruptibleChannel.end(AbstractInterruptibleChannel.java:185) [junit] at sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:159) [junit] at sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:84) [junit] at org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:130) [junit] at java.lang.Thread.run(Thread.java:662) [junit] [junit] 11/07/19 10:33:39 INFO datanode.DataBlockScanner: Exiting DataBlockScanner thread. [junit] 11/07/19 10:33:40 INFO datanode.DataNode: Waiting for threadgroup to exit, active threads is 0 [junit] 11/07/19 10:33:40 INFO datanode.DataNode: DatanodeRegistration(127.0.0.1:48248, storageID=DS-20126214-127.0.1.1-48248-1311071251998, infoPort=40975, ipcPort=38451):Finishing DataNode in: FSDataset{dirpath='<https://builds.apache.org/job/Pig-trunk/ws/trunk/build/test/data/dfs/data/data7/current,/grid/0/hudson/hudson-slave/workspace/Pig-trunk/trunk/build/test/data/dfs/data/data8/current'}> [junit] 11/07/19 10:33:40 INFO ipc.Server: Stopping server on 38451 [junit] 11/07/19 10:33:40 INFO datanode.DataNode: Waiting for threadgroup to exit, active threads is 0 [junit] Shutting down DataNode 2 [junit] 11/07/19 10:33:40 INFO ipc.Server: Stopping server on 53239 [junit] 11/07/19 10:33:40 INFO ipc.Server: IPC Server handler 0 on 53239: exiting [junit] 11/07/19 10:33:40 INFO ipc.Server: Stopping IPC Server Responder [junit] 11/07/19 10:33:40 INFO datanode.DataNode: Waiting for threadgroup to exit, active threads is 1 [junit] 11/07/19 10:33:40 WARN datanode.DataNode: DatanodeRegistration(127.0.0.1:42326, storageID=DS-1415343713-127.0.1.1-42326-1311071251704, infoPort=54465, ipcPort=53239):DataXceiveServer: java.nio.channels.AsynchronousCloseException [junit] at java.nio.channels.spi.AbstractInterruptibleChannel.end(AbstractInterruptibleChannel.java:185) [junit] at sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:159) [junit] at sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:84) [junit] at org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:130) [junit] at java.lang.Thread.run(Thread.java:662) [junit] [junit] 11/07/19 10:33:40 INFO ipc.Server: IPC Server handler 1 on 53239: exiting [junit] 11/07/19 10:33:40 INFO ipc.Server: IPC Server handler 2 on 53239: exiting [junit] 11/07/19 10:33:40 INFO ipc.Server: Stopping IPC Server listener on 53239 [junit] 11/07/19 10:33:40 INFO datanode.DataNode: BlockReport of 6 blocks got processed in 2 msecs [junit] 11/07/19 10:33:40 INFO datanode.DataNode: Deleting block blk_-2046756786705305301_1124 file build/test/data/dfs/data/data3/current/blk_-2046756786705305301 [junit] 11/07/19 10:33:40 INFO datanode.DataNode: Deleting block blk_464773406130336812_1123 file build/test/data/dfs/data/data4/current/blk_464773406130336812 [junit] 11/07/19 10:33:40 INFO datanode.DataNode: Deleting block blk_5824626566657843290_1122 file build/test/data/dfs/data/data3/current/blk_5824626566657843290 [junit] 11/07/19 10:33:40 INFO datanode.DataBlockScanner: Exiting DataBlockScanner thread. [junit] 11/07/19 10:33:40 INFO datanode.DataNode: DatanodeRegistration(127.0.0.1:42326, storageID=DS-1415343713-127.0.1.1-42326-1311071251704, infoPort=54465, ipcPort=53239):Finishing DataNode in: FSDataset{dirpath='<https://builds.apache.org/job/Pig-trunk/ws/trunk/build/test/data/dfs/data/data5/current,/grid/0/hudson/hudson-slave/workspace/Pig-trunk/trunk/build/test/data/dfs/data/data6/current'}> [junit] 11/07/19 10:33:40 INFO ipc.Server: Stopping server on 53239 [junit] 11/07/19 10:33:40 INFO datanode.DataNode: Waiting for threadgroup to exit, active threads is 0 [junit] Shutting down DataNode 1 [junit] 11/07/19 10:33:41 INFO datanode.DataNode: Waiting for threadgroup to exit, active threads is 0 [junit] 11/07/19 10:33:41 INFO ipc.Server: Stopping server on 54632 [junit] 11/07/19 10:33:41 INFO ipc.Server: IPC Server handler 0 on 54632: exiting [junit] 11/07/19 10:33:41 INFO datanode.DataNode: Waiting for threadgroup to exit, active threads is 1 [junit] 11/07/19 10:33:41 INFO ipc.Server: IPC Server handler 1 on 54632: exiting [junit] 11/07/19 10:33:41 INFO ipc.Server: IPC Server handler 2 on 54632: exiting [junit] 11/07/19 10:33:41 INFO ipc.Server: Stopping IPC Server listener on 54632 [junit] 11/07/19 10:33:41 INFO ipc.Server: Stopping IPC Server Responder [junit] 11/07/19 10:33:41 WARN datanode.DataNode: DatanodeRegistration(127.0.0.1:57264, storageID=DS-1741254656-127.0.1.1-57264-1311071251383, infoPort=50155, ipcPort=54632):DataXceiveServer: java.nio.channels.AsynchronousCloseException [junit] at java.nio.channels.spi.AbstractInterruptibleChannel.end(AbstractInterruptibleChannel.java:185) [junit] at sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:159) [junit] at sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:84) [junit] at org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:130) [junit] at java.lang.Thread.run(Thread.java:662) [junit] [junit] 11/07/19 10:33:41 INFO datanode.DataBlockScanner: Exiting DataBlockScanner thread. [junit] 11/07/19 10:33:41 INFO hdfs.StateChange: BLOCK* ask 127.0.0.1:42326 to delete blk_-4905632516841518502_1126 blk_-3045267291178481173_1127 [junit] 11/07/19 10:33:41 INFO hdfs.StateChange: BLOCK* ask 127.0.0.1:57264 to delete blk_863817712174337058_1121 blk_-4905632516841518502_1126 blk_-3045267291178481173_1127 [junit] Shutting down DataNode 0 [junit] 11/07/19 10:33:42 INFO datanode.DataNode: Waiting for threadgroup to exit, active threads is 0 [junit] 11/07/19 10:33:42 INFO datanode.DataNode: DatanodeRegistration(127.0.0.1:57264, storageID=DS-1741254656-127.0.1.1-57264-1311071251383, infoPort=50155, ipcPort=54632):Finishing DataNode in: FSDataset{dirpath='<https://builds.apache.org/job/Pig-trunk/ws/trunk/build/test/data/dfs/data/data3/current,/grid/0/hudson/hudson-slave/workspace/Pig-trunk/trunk/build/test/data/dfs/data/data4/current'}> [junit] 11/07/19 10:33:42 INFO ipc.Server: Stopping server on 54632 [junit] 11/07/19 10:33:42 INFO datanode.DataNode: Waiting for threadgroup to exit, active threads is 0 [junit] 11/07/19 10:33:42 INFO ipc.Server: Stopping server on 38436 [junit] 11/07/19 10:33:42 INFO ipc.Server: IPC Server handler 0 on 38436: exiting [junit] 11/07/19 10:33:42 INFO datanode.DataNode: Waiting for threadgroup to exit, active threads is 1 [junit] 11/07/19 10:33:42 WARN datanode.DataNode: DatanodeRegistration(127.0.0.1:35943, storageID=DS-2064813100-127.0.1.1-35943-1311071251084, infoPort=47839, ipcPort=38436):DataXceiveServer: java.nio.channels.AsynchronousCloseException [junit] at java.nio.channels.spi.AbstractInterruptibleChannel.end(AbstractInterruptibleChannel.java:185) [junit] at sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:159) [junit] at sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:84) [junit] at org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:130) [junit] at java.lang.Thread.run(Thread.java:662) [junit] [junit] 11/07/19 10:33:42 INFO ipc.Server: Stopping IPC Server Responder [junit] 11/07/19 10:33:42 INFO ipc.Server: IPC Server handler 2 on 38436: exiting [junit] 11/07/19 10:33:42 INFO ipc.Server: Stopping IPC Server listener on 38436 [junit] 11/07/19 10:33:42 INFO ipc.Server: IPC Server handler 1 on 38436: exiting [junit] 11/07/19 10:33:42 INFO datanode.DataBlockScanner: Exiting DataBlockScanner thread. [junit] 11/07/19 10:33:43 INFO datanode.DataNode: Waiting for threadgroup to exit, active threads is 0 [junit] 11/07/19 10:33:43 INFO datanode.DataNode: DatanodeRegistration(127.0.0.1:35943, storageID=DS-2064813100-127.0.1.1-35943-1311071251084, infoPort=47839, ipcPort=38436):Finishing DataNode in: FSDataset{dirpath='<https://builds.apache.org/job/Pig-trunk/ws/trunk/build/test/data/dfs/data/data1/current,/grid/0/hudson/hudson-slave/workspace/Pig-trunk/trunk/build/test/data/dfs/data/data2/current'}> [junit] 11/07/19 10:33:43 INFO ipc.Server: Stopping server on 38436 [junit] 11/07/19 10:33:43 INFO datanode.DataNode: Waiting for threadgroup to exit, active threads is 0 [junit] 11/07/19 10:33:43 WARN namenode.FSNamesystem: ReplicationMonitor thread received InterruptedException.java.lang.InterruptedException: sleep interrupted [junit] 11/07/19 10:33:43 INFO namenode.FSNamesystem: Number of transactions: 694 Total time for transactions(ms): 12Number of transactions batched in Syncs: 114 Number of syncs: 484 SyncTimes(ms): 3257 287 [junit] 11/07/19 10:33:43 INFO namenode.DecommissionManager: Interrupted Monitor [junit] java.lang.InterruptedException: sleep interrupted [junit] at java.lang.Thread.sleep(Native Method) [junit] at org.apache.hadoop.hdfs.server.namenode.DecommissionManager$Monitor.run(DecommissionManager.java:65) [junit] at java.lang.Thread.run(Thread.java:662) [junit] 11/07/19 10:33:43 INFO ipc.Server: Stopping server on 41941 [junit] 11/07/19 10:33:43 INFO ipc.Server: IPC Server handler 0 on 41941: exiting [junit] 11/07/19 10:33:43 INFO ipc.Server: IPC Server handler 2 on 41941: exiting [junit] 11/07/19 10:33:43 INFO ipc.Server: IPC Server handler 8 on 41941: exiting [junit] 11/07/19 10:33:43 INFO ipc.Server: IPC Server handler 4 on 41941: exiting [junit] 11/07/19 10:33:43 INFO ipc.Server: Stopping IPC Server listener on 41941 [junit] 11/07/19 10:33:43 INFO ipc.Server: IPC Server handler 7 on 41941: exiting [junit] 11/07/19 10:33:43 INFO ipc.Server: IPC Server handler 5 on 41941: exiting [junit] 11/07/19 10:33:43 INFO ipc.Server: IPC Server handler 9 on 41941: exiting [junit] 11/07/19 10:33:43 INFO ipc.Server: Stopping IPC Server Responder [junit] Tests run: 17, Failures: 0, Errors: 0, Time elapsed: 368.924 sec [junit] 11/07/19 10:33:43 INFO ipc.Server: IPC Server handler 3 on 41941: exiting [junit] 11/07/19 10:33:43 INFO ipc.Server: IPC Server handler 6 on 41941: exiting [junit] 11/07/19 10:33:43 INFO ipc.Server: IPC Server handler 1 on 41941: exiting [junit] Running org.apache.pig.test.TestStringUDFs [junit] 11/07/19 10:33:43 WARN builtin.SUBSTRING: java.lang.NullPointerException [junit] 11/07/19 10:33:43 WARN builtin.SUBSTRING: java.lang.StringIndexOutOfBoundsException: String index out of range: -2 [junit] 11/07/19 10:33:43 WARN builtin.SUBSTRING: java.lang.StringIndexOutOfBoundsException: String index out of range: -1 [junit] 11/07/19 10:33:43 WARN builtin.SUBSTRING: java.lang.StringIndexOutOfBoundsException: String index out of range: -8 [junit] 11/07/19 10:33:43 WARN builtin.SUBSTRING: java.lang.StringIndexOutOfBoundsException: String index out of range: -2 [junit] 11/07/19 10:33:43 WARN builtin.INDEXOF: Failed to process input; error - null [junit] 11/07/19 10:33:43 WARN builtin.LAST_INDEX_OF: Failed to process input; error - null [junit] Tests run: 11, Failures: 0, Errors: 0, Time elapsed: 0.107 sec [delete] Deleting directory /tmp/pig_junit_tmp1777959676 BUILD FAILED <https://builds.apache.org/job/Pig-trunk/ws/trunk/build.xml>:664: The following error occurred while executing this line: <https://builds.apache.org/job/Pig-trunk/ws/trunk/build.xml>:719: Tests failed! Total time: 21 minutes 13 seconds [FINDBUGS] Skipping publisher since build result is FAILURE Recording test results Publishing Javadoc Archiving artifacts Recording fingerprints Publishing Clover coverage report... No Clover report will be published due to a Build Failure