Lin Zhang created HDFS-14360:
--------------------------------

             Summary: some excptioins happened while using ISA-L
                 Key: HDFS-14360
                 URL: https://issues.apache.org/jira/browse/HDFS-14360
             Project: Hadoop HDFS
          Issue Type: Bug
          Components: ec, erasure-coding
            Reporter: Lin Zhang


I built my hadoop with isa-l supported. When I try to so some convert job, 
exception happens.   
{code:java}
//代码占位符
{code}
[2019-03-12T11:39:03.183+08:00] [INFO] 
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # 
[2019-03-12T11:39:03.184+08:00] [INFO] 
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # A fatal error has been 
detected by the Java Runtime Environment: [2019-03-12T11:39:03.184+08:00] 
[INFO] [1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # 
[2019-03-12T11:39:03.184+08:00] [INFO] 
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # SIGSEGV (0xb) at 
pc=0x00007fc42e182683, pid=17110, tid=0x00007fc40ce9f700 
[2019-03-12T11:39:03.184+08:00] [INFO] 
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # 
[2019-03-12T11:39:03.184+08:00] [INFO] 
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # JRE version: Java(TM) 
SE Runtime Environment (8.0_121-b13) (build 1.8.0_121-b13) 
[2019-03-12T11:39:03.184+08:00] [INFO] 
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # Java VM: Java 
HotSpot(TM) 64-Bit Server VM (25.121-b13 mixed mode linux-amd64 compressed 
oops) [2019-03-12T11:39:03.184+08:00] [INFO] 
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # Problematic frame: 
[2019-03-12T11:39:03.184+08:00] [INFO] 
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # V [libjvm.so+0x9bd683] 
SafepointSynchronize::begin()+0x263 [2019-03-12T11:39:03.185+08:00] [INFO] 
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # 
[2019-03-12T11:39:03.185+08:00] [INFO] 
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # Failed to write core 
dump. Core dumps have been disabled. To enable core dumping, try "ulimit -c 
unlimited" before starting Java again [2019-03-12T11:39:03.185+08:00] [INFO] 
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # 
[2019-03-12T11:39:03.185+08:00] [INFO] 
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # An error report file 
with more information is saved as: [2019-03-12T11:39:03.185+08:00] [INFO] 
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # 
/software/servers/hadoop-2.7.1/hs_err_pid17110.log 
[2019-03-12T11:39:03.191+08:00] [INFO] 
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # 
[2019-03-12T11:39:03.191+08:00] [INFO] 
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # If you would like to 
submit a bug report, please visit: [2019-03-12T11:39:03.191+08:00] [INFO] 
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # 
http://bugreport.java.com/bugreport/crash.jsp [2019-03-12T11:39:03.191+08:00] 
[INFO] [1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] : 
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # 
[2019-03-12T11:39:07.949+08:00] [ERROR] [pool-10-thread-1] : copy file 
/test/zhanglin/1g to /test/ttlconverter/factory/test/zhanglin/1gfailed 
[2019-03-12T11:39:07.949+08:00] [INFO] [DataXceiver for client 
DFSClient_NONMAPREDUCE_1740978034_1 at /172.22.176.69:40662 [Receiving block 
BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009]] : 
Exception for 
BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009 
java.io.IOException: Premature EOF from inputStream at 
org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:212) at 
org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:213)
 at 
org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:134)
 at 
org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:109)
 at 
org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:529)
 at 
org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:972)
 at 
org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:891)
 at 
org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:171)
 at 
org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:105)
 at 
org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:290) at 
java.lang.Thread.run(Thread.java:745)

[2019-03-12T11:39:07.951+08:00] [INFO] [DataXceiver for client 
DFSClient_NONMAPREDUCE_1740978034_1 at /172.22.176.69:40660 [Sending block 
BP-442378117-172.16.150.142-1552360340470:blk_1073741825_1001]] : Scheduling a 
check for /data0/dfs
[2019-03-12T11:39:07.954+08:00] [INFO] [PacketResponder: 
BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009, 
type=LAST_IN_PIPELINE] : PacketResponder: 
BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009, 
type=LAST_IN_PIPELINE: Thread is interrupted.
[2019-03-12T11:39:07.954+08:00] [INFO] [PacketResponder: 
BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009, 
type=LAST_IN_PIPELINE] : PacketResponder: 
BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009, 
type=LAST_IN_PIPELINE terminating
[2019-03-12T11:39:07.954+08:00] [INFO] [DataXceiver for client 
DFSClient_NONMAPREDUCE_1740978034_1 at /172.22.176.69:40662 [Receiving block 
BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009]] : 
opWriteBlock 
BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009 
received exception java.io.IOException: Premature EOF from inputStream
[2019-03-12T11:39:07.957+08:00] [ERROR] [DataXceiver for client 
DFSClient_NONMAPREDUCE_1740978034_1 at /172.22.176.69:40662 [Receiving block 
BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009]] : 
A01-R02-I176-69-4CY8S12.JD.LOCAL:50010:DataXceiver error processing WRITE_BLOCK 
operation src: /172.22.176.69:40662 dst: /172.22.176.69:50010
java.io.IOException: Premature EOF from inputStream
 at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:212)
 at 
org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:213)
 at 
org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:134)
 at 
org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:109)
 at 
org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:529)
 at 
org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:972)
 at 
org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:891)
 at 
org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:171)
 at 
org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:105)
 at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:290)
 at java.lang.Thread.run(Thread.java:745)

 

I checked my native env, which showed below
{code:java}
//代码占位符
{code}
Native library checking:

hadoop: true /software/servers/hadoop-2.7.1/lib/native/libhadoop.so.1.0.0

zlib: true /lib64/libz.so.1

snappy: true /lib64/libsnappy.so.1

lz4: true revision:99

bzip2: true /lib64/libbz2.so.1

openssl: true /lib64/libcrypto.so

ISA-L: true /software/servers/hadoop-2.7.1/lib/native/libisal.so.2



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org

Reply via email to