Lin Zhang created HDFS-14360: -------------------------------- Summary: some excptioins happened while using ISA-L Key: HDFS-14360 URL: https://issues.apache.org/jira/browse/HDFS-14360 Project: Hadoop HDFS Issue Type: Bug Components: ec, erasure-coding Reporter: Lin Zhang
I built my hadoop with isa-l supported. When I try to so some convert job, exception happens. {code:java} //代码占位符 {code} [2019-03-12T11:39:03.183+08:00] [INFO] [1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # [2019-03-12T11:39:03.184+08:00] [INFO] [1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # A fatal error has been detected by the Java Runtime Environment: [2019-03-12T11:39:03.184+08:00] [INFO] [1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # [2019-03-12T11:39:03.184+08:00] [INFO] [1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # SIGSEGV (0xb) at pc=0x00007fc42e182683, pid=17110, tid=0x00007fc40ce9f700 [2019-03-12T11:39:03.184+08:00] [INFO] [1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # [2019-03-12T11:39:03.184+08:00] [INFO] [1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # JRE version: Java(TM) SE Runtime Environment (8.0_121-b13) (build 1.8.0_121-b13) [2019-03-12T11:39:03.184+08:00] [INFO] [1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # Java VM: Java HotSpot(TM) 64-Bit Server VM (25.121-b13 mixed mode linux-amd64 compressed oops) [2019-03-12T11:39:03.184+08:00] [INFO] [1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # Problematic frame: [2019-03-12T11:39:03.184+08:00] [INFO] [1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # V [libjvm.so+0x9bd683] SafepointSynchronize::begin()+0x263 [2019-03-12T11:39:03.185+08:00] [INFO] [1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # [2019-03-12T11:39:03.185+08:00] [INFO] [1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # Failed to write core dump. Core dumps have been disabled. To enable core dumping, try "ulimit -c unlimited" before starting Java again [2019-03-12T11:39:03.185+08:00] [INFO] [1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # [2019-03-12T11:39:03.185+08:00] [INFO] [1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # An error report file with more information is saved as: [2019-03-12T11:39:03.185+08:00] [INFO] [1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # /software/servers/hadoop-2.7.1/hs_err_pid17110.log [2019-03-12T11:39:03.191+08:00] [INFO] [1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # [2019-03-12T11:39:03.191+08:00] [INFO] [1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # If you would like to submit a bug report, please visit: [2019-03-12T11:39:03.191+08:00] [INFO] [1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # http://bugreport.java.com/bugreport/crash.jsp [2019-03-12T11:39:03.191+08:00] [INFO] [1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # [2019-03-12T11:39:07.949+08:00] [ERROR] [pool-10-thread-1] : copy file /test/zhanglin/1g to /test/ttlconverter/factory/test/zhanglin/1gfailed [2019-03-12T11:39:07.949+08:00] [INFO] [DataXceiver for client DFSClient_NONMAPREDUCE_1740978034_1 at /172.22.176.69:40662 [Receiving block BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009]] : Exception for BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:212) at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:213) at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:134) at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:109) at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:529) at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:972) at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:891) at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:171) at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:105) at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:290) at java.lang.Thread.run(Thread.java:745) [2019-03-12T11:39:07.951+08:00] [INFO] [DataXceiver for client DFSClient_NONMAPREDUCE_1740978034_1 at /172.22.176.69:40660 [Sending block BP-442378117-172.16.150.142-1552360340470:blk_1073741825_1001]] : Scheduling a check for /data0/dfs [2019-03-12T11:39:07.954+08:00] [INFO] [PacketResponder: BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009, type=LAST_IN_PIPELINE] : PacketResponder: BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009, type=LAST_IN_PIPELINE: Thread is interrupted. [2019-03-12T11:39:07.954+08:00] [INFO] [PacketResponder: BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009, type=LAST_IN_PIPELINE] : PacketResponder: BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009, type=LAST_IN_PIPELINE terminating [2019-03-12T11:39:07.954+08:00] [INFO] [DataXceiver for client DFSClient_NONMAPREDUCE_1740978034_1 at /172.22.176.69:40662 [Receiving block BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009]] : opWriteBlock BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009 received exception java.io.IOException: Premature EOF from inputStream [2019-03-12T11:39:07.957+08:00] [ERROR] [DataXceiver for client DFSClient_NONMAPREDUCE_1740978034_1 at /172.22.176.69:40662 [Receiving block BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009]] : A01-R02-I176-69-4CY8S12.JD.LOCAL:50010:DataXceiver error processing WRITE_BLOCK operation src: /172.22.176.69:40662 dst: /172.22.176.69:50010 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:212) at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:213) at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:134) at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:109) at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:529) at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:972) at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:891) at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:171) at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:105) at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:290) at java.lang.Thread.run(Thread.java:745) I checked my native env, which showed below {code:java} //代码占位符 {code} Native library checking: hadoop: true /software/servers/hadoop-2.7.1/lib/native/libhadoop.so.1.0.0 zlib: true /lib64/libz.so.1 snappy: true /lib64/libsnappy.so.1 lz4: true revision:99 bzip2: true /lib64/libbz2.so.1 openssl: true /lib64/libcrypto.so ISA-L: true /software/servers/hadoop-2.7.1/lib/native/libisal.so.2 -- This message was sent by Atlassian JIRA (v7.6.3#76005) --------------------------------------------------------------------- To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org