Hi Arthur,

Welcome in our world ;)

For JAVA_LIBRARY_PATH I don't even set it anywhere.

hbase@node3:~/hbase-0.94.3$ echo $JAVA_LIBRARY_PATH

hbase@node3:~/hbase-0.94.3$ grep JAVA_LIBRARY_PATH conf/hbase-env.sh
base@node3:~/hbase-0.94.3$ grep JAVA_LIBRARY_PATH bin/*
bin/hbase:#   HBASE_LIBRARY_PATH  HBase additions to JAVA_LIBRARY_PATH for
adding
bin/hbase:#If avail, add Hadoop to the CLASSPATH and to the
JAVA_LIBRARY_PATH
bin/hbase:  HADOOP_JAVA_LIBRARY_PATH=$(HADOOP_CLASSPATH="$CLASSPATH"
${HADOOP_IN_PATH} \
bin/hbase:  if [ -n "$HADOOP_JAVA_LIBRARY_PATH" ]; then
bin/hbase:    JAVA_LIBRARY_PATH=$(append_path "${JAVA_LIBRARY_PATH}"
"$HADOOP_JAVA_LIBRARY_PATH")
bin/hbase:    JAVA_LIBRARY_PATH=$(append_path "$JAVA_LIBRARY_PATH"
${HBASE_HOME}/build/native/${JAVA_PLATFORM}/lib)
bin/hbase:    JAVA_LIBRARY_PATH=$(append_path "$JAVA_LIBRARY_PATH"
${HBASE_HOME}/lib/native/${JAVA_PLATFORM})
bin/hbase:  JAVA_LIBRARY_PATH=`cygpath -p "$JAVA_LIBRARY_PATH"`
bin/hbase:if [ "x$JAVA_LIBRARY_PATH" != "x" ]; then
bin/hbase:  HBASE_OPTS="$HBASE_OPTS -Djava.library.path=$JAVA_LIBRARY_PATH"
bin/hbase:  export LD_LIBRARY_PATH="$LD_LIBRARY_PATH:$JAVA_LIBRARY_PATH"

It's all the default values everywhere.

But I don't see anything wrong with what you have done.

JM


2014-08-27 7:06 GMT-04:00 [email protected] <[email protected]
>:

> Hi JM,
>
> Thank you so much!
>
> I had not set JAVA_LIBRARY_PATH before.
> Now I added [export
> JAVA_LIBRARY_PATH="$HBASE_HOME/lib/native/Linux-amd64-64”] to hbase-env.sh
> also added [export
> JAVA_LIBRARY_PATH="$HADOOP_HOME/lib/native/Linux-amd64-64”] to hadoop-env.sh
> I hope this is correct way.
>
> Can you please share how you define JAVA_LIBRARY_PATH in your hbase-env.sh
> and hadoop-env.sh as this is new to me (I am also new to HBase :) ) ?
>
> Regards
> Arthur
>
> On 27 Aug, 2014, at 6:41 pm, Jean-Marc Spaggiari <[email protected]>
> wrote:
>
> > Hi Arthur,
> >
> > Glad to hear you got it!
> >
> > Regarding #2, was JAVA_LIBRARY_PATH already set before? If so, that might
> > have been the issue. HBase will append to this path all what it needs (if
> > required) so I don't think there is anything else you need to add.
> >
> > Regarding #1  I don't think it's an error. Might maybe more be a warning.
> > Will look at it to see where it comes form...
> >
> > JM
> >
> >
> > 2014-08-27 4:00 GMT-04:00 [email protected] <
> [email protected]
> >> :
> >
> >> Hi,
> >>
> >> Many thanks for your advices!
> >>
> >> Finally, I managed to make it work.
> >>
> >> I needed to add:
> >> export JAVA_LIBRARY_PATH="$HBASE_HOME/lib/native/Linux-amd64-64”
> >>
> >> then run:
> >> bin/hbase org.apache.hadoop.hbase.util.CompressionTest
> >> file:///tmp/snappy-test snappy
> >> 2014-08-27 15:51:39,459 INFO  [main] Configuration.deprecation:
> >> hadoop.native.lib is deprecated. Instead, use io.native.lib.available
> >> SLF4J: Class path contains multiple SLF4J bindings.
> >> SLF4J: Found binding in
> >>
> [jar:file:/mnt/hadoop/hbase-0.98.4-hadoop2/lib/slf4j-log4j12-1.6.4.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> >> SLF4J: Found binding in
> >>
> [jar:file:/mnt/hadoop/hadoop-2.4.1/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> >> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
> >> explanation.
> >> 2014-08-27 15:51:39,785 INFO  [main] util.ChecksumType: Checksum using
> >> org.apache.hadoop.util.PureJavaCrc32
> >> 2014-08-27 15:51:39,786 INFO  [main] util.ChecksumType: Checksum can use
> >> org.apache.hadoop.util.PureJavaCrc32C
> >> 2014-08-27 15:51:39,926 INFO  [main] compress.CodecPool: Got brand-new
> >> compressor [.snappy]
> >> 2014-08-27 15:51:39,930 INFO  [main] compress.CodecPool: Got brand-new
> >> compressor [.snappy]
> >> 2014-08-27 15:51:39,934 ERROR [main] hbase.KeyValue: Unexpected
> >> getShortMidpointKey result, fakeKey:testkey, firstKeyInBlock:testkey
> >> 2014-08-27 15:51:40,185 INFO  [main] compress.CodecPool: Got brand-new
> >> decompressor [.snappy]
> >> SUCCESS
> >>
> >>
> >> bin/hbase org.apache.hadoop.hbase.util.CompressionTest
> >> file:///tmp/snappy-test gz
> >> 2014-08-27 15:57:18,633 INFO  [main] Configuration.deprecation:
> >> hadoop.native.lib is deprecated. Instead, use io.native.lib.available
> >> SLF4J: Class path contains multiple SLF4J bindings.
> >> SLF4J: Found binding in
> >>
> [jar:file:/mnt/hadoop/hbase-0.98.4-hadoop2/lib/slf4j-log4j12-1.6.4.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> >> SLF4J: Found binding in
> >>
> [jar:file:/mnt/hadoop/hadoop-2.4.1/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> >> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
> >> explanation.
> >> 2014-08-27 15:57:18,969 INFO  [main] util.ChecksumType: Checksum using
> >> org.apache.hadoop.util.PureJavaCrc32
> >> 2014-08-27 15:57:18,970 INFO  [main] util.ChecksumType: Checksum can use
> >> org.apache.hadoop.util.PureJavaCrc32C
> >> 2014-08-27 15:57:19,127 INFO  [main] zlib.ZlibFactory: Successfully
> loaded
> >> & initialized native-zlib library
> >> 2014-08-27 15:57:19,146 INFO  [main] compress.CodecPool: Got brand-new
> >> compressor [.gz]
> >> 2014-08-27 15:57:19,149 INFO  [main] compress.CodecPool: Got brand-new
> >> compressor [.gz]
> >> 2014-08-27 15:57:19,153 ERROR [main] hbase.KeyValue: Unexpected
> >> getShortMidpointKey result, fakeKey:testkey, firstKeyInBlock:testkey
> >> 2014-08-27 15:57:19,401 INFO  [main] compress.CodecPool: Got brand-new
> >> decompressor [.gz]
> >> SUCCESS
> >>
> >>
> >> 2 questions:
> >> 1) Is this OK if “SUCCESS" with "ERROR [main] hbase.KeyValue: Unexpected
> >> getShortMidpointKey result, fakeKey:testkey, firstKeyInBlock:test key”
> >> 2) is this extra setting of “JAVA_LIBRARY_PATH” a good way for setting
> up
> >> snappy with Hadoop 2.4.1 and HBase 0.98.4?
> >>
> >>
> >> Regards
> >> Arthur
> >>
> >>
> >>
> >> On 27 Aug, 2014, at 1:13 pm, [email protected] <
> >> [email protected]> wrote:
> >>
> >>> Hi,
> >>>
> >>> Thanks!  tried but still same error:
> >>>
> >>> rm hadoop-2.4.1-src -Rf
> >>
> >>         // delete all old src files and try again
> >>> tar -vxf hadoop-2.4.1-src.tar.gz
> >>> cd hadoop-2.4.1-src
> >>> mvn -DskipTests clean install -Drequire.snappy=true​-Pnative
> >>                                                      // compile with
> snappy
> >>> [INFO]
> >>> [INFO] Apache Hadoop Main ................................ SUCCESS
> >> [0.887s]
> >>> [INFO] Apache Hadoop Project POM ......................... SUCCESS
> >> [0.306s]
> >>> [INFO] Apache Hadoop Annotations ......................... SUCCESS
> >> [0.859s]
> >>> [INFO] Apache Hadoop Project Dist POM .................... SUCCESS
> >> [0.231s]
> >>> [INFO] Apache Hadoop Assemblies .......................... SUCCESS
> >> [0.071s]
> >>> [INFO] Apache Hadoop Maven Plugins ....................... SUCCESS
> >> [0.960s]
> >>> [INFO] Apache Hadoop MiniKDC ............................. SUCCESS
> >> [0.711s]
> >>> [INFO] Apache Hadoop Auth ................................ SUCCESS
> >> [0.641s]
> >>> [INFO] Apache Hadoop Auth Examples ....................... SUCCESS
> >> [0.528s]
> >>> [INFO] Apache Hadoop Common .............................. SUCCESS
> >> [7.859s]
> >>> [INFO] Apache Hadoop NFS ................................. SUCCESS
> >> [0.282s]
> >>> [INFO] Apache Hadoop Common Project ...................... SUCCESS
> >> [0.013s]
> >>> [INFO] Apache Hadoop HDFS ................................ SUCCESS
> >> [14.210s]
> >>> [INFO] Apache Hadoop HttpFS .............................. SUCCESS
> >> [1.322s]
> >>> [INFO] Apache Hadoop HDFS BookKeeper Journal ............. SUCCESS
> >> [0.418s]
> >>> [INFO] Apache Hadoop HDFS-NFS ............................ SUCCESS
> >> [0.178s]
> >>> [INFO] Apache Hadoop HDFS Project ........................ SUCCESS
> >> [0.016s]
> >>> [INFO] hadoop-yarn ....................................... SUCCESS
> >> [0.014s]
> >>> [INFO] hadoop-yarn-api ................................... SUCCESS
> >> [3.012s]
> >>> [INFO] hadoop-yarn-common ................................ SUCCESS
> >> [1.173s]
> >>> [INFO] hadoop-yarn-server ................................ SUCCESS
> >> [0.029s]
> >>> [INFO] hadoop-yarn-server-common ......................... SUCCESS
> >> [0.379s]
> >>> [INFO] hadoop-yarn-server-nodemanager .................... SUCCESS
> >> [0.612s]
> >>> [INFO] hadoop-yarn-server-web-proxy ...................... SUCCESS
> >> [0.166s]
> >>> [INFO] hadoop-yarn-server-applicationhistoryservice ...... SUCCESS
> >> [0.213s]
> >>> [INFO] hadoop-yarn-server-resourcemanager ................ SUCCESS
> >> [0.970s]
> >>> [INFO] hadoop-yarn-server-tests .......................... SUCCESS
> >> [0.158s]
> >>> [INFO] hadoop-yarn-client ................................ SUCCESS
> >> [0.227s]
> >>> [INFO] hadoop-yarn-applications .......................... SUCCESS
> >> [0.013s]
> >>> [INFO] hadoop-yarn-applications-distributedshell ......... SUCCESS
> >> [0.157s]
> >>> [INFO] hadoop-yarn-applications-unmanaged-am-launcher .... SUCCESS
> >> [0.094s]
> >>> [INFO] hadoop-yarn-site .................................. SUCCESS
> >> [0.024s]
> >>> [INFO] hadoop-yarn-project ............................... SUCCESS
> >> [0.030s]
> >>> [INFO] hadoop-mapreduce-client ........................... SUCCESS
> >> [0.027s]
> >>> [INFO] hadoop-mapreduce-client-core ...................... SUCCESS
> >> [1.206s]
> >>> [INFO] hadoop-mapreduce-client-common .................... SUCCESS
> >> [1.140s]
> >>> [INFO] hadoop-mapreduce-client-shuffle ................... SUCCESS
> >> [0.128s]
> >>> [INFO] hadoop-mapreduce-client-app ....................... SUCCESS
> >> [0.634s]
> >>> [INFO] hadoop-mapreduce-client-hs ........................ SUCCESS
> >> [0.557s]
> >>> [INFO] hadoop-mapreduce-client-jobclient ................. SUCCESS
> >> [0.882s]
> >>> [INFO] hadoop-mapreduce-client-hs-plugins ................ SUCCESS
> >> [0.085s]
> >>> [INFO] Apache Hadoop MapReduce Examples .................. SUCCESS
> >> [0.224s]
> >>> [INFO] hadoop-mapreduce .................................. SUCCESS
> >> [0.030s]
> >>> [INFO] Apache Hadoop MapReduce Streaming ................. SUCCESS
> >> [0.200s]
> >>> [INFO] Apache Hadoop Distributed Copy .................... SUCCESS
> >> [0.656s]
> >>> [INFO] Apache Hadoop Archives ............................ SUCCESS
> >> [0.112s]
> >>> [INFO] Apache Hadoop Rumen ............................... SUCCESS
> >> [0.246s]
> >>> [INFO] Apache Hadoop Gridmix ............................. SUCCESS
> >> [0.283s]
> >>> [INFO] Apache Hadoop Data Join ........................... SUCCESS
> >> [0.111s]
> >>> [INFO] Apache Hadoop Extras .............................. SUCCESS
> >> [0.146s]
> >>> [INFO] Apache Hadoop Pipes ............................... SUCCESS
> >> [0.011s]
> >>> [INFO] Apache Hadoop OpenStack support ................... SUCCESS
> >> [0.283s]
> >>> [INFO] Apache Hadoop Client .............................. SUCCESS
> >> [0.106s]
> >>> [INFO] Apache Hadoop Mini-Cluster ........................ SUCCESS
> >> [0.038s]
> >>> [INFO] Apache Hadoop Scheduler Load Simulator ............ SUCCESS
> >> [0.223s]
> >>> [INFO] Apache Hadoop Tools Dist .......................... SUCCESS
> >> [0.106s]
> >>> [INFO] Apache Hadoop Tools ............................... SUCCESS
> >> [0.010s]
> >>> [INFO] Apache Hadoop Distribution ........................ SUCCESS
> >> [0.034s]
> >>> [INFO]
> >> ------------------------------------------------------------------------
> >>> [INFO] BUILD SUCCESS
> >>> [INFO]
> >> ------------------------------------------------------------------------
> >>> [INFO] Total time: 45.478s
> >>> [INFO] Finished at: Wed Aug 27 12:10:06 HKT 2014
> >>> [INFO] Final Memory: 107M/1898M
> >>> [INFO]
> >> ------------------------------------------------------------------------
> >>> mvn package -Pdist,native -DskipTests -Dtar -Drequire.snappy=true
> >>                                                             // package
> it
> >> with snappy
> >>> [INFO]
> >> ------------------------------------------------------------------------
> >>> [INFO] Reactor Summary:
> >>> [INFO]
> >>> [INFO] Apache Hadoop Main ................................ SUCCESS
> >> [0.727s]
> >>> [INFO] Apache Hadoop Project POM ......................... SUCCESS
> >> [0.555s]
> >>> [INFO] Apache Hadoop Annotations ......................... SUCCESS
> >> [1.011s]
> >>> [INFO] Apache Hadoop Assemblies .......................... SUCCESS
> >> [0.128s]
> >>> [INFO] Apache Hadoop Project Dist POM .................... SUCCESS
> >> [1.342s]
> >>> [INFO] Apache Hadoop Maven Plugins ....................... SUCCESS
> >> [1.251s]
> >>> [INFO] Apache Hadoop MiniKDC ............................. SUCCESS
> >> [1.007s]
> >>> [INFO] Apache Hadoop Auth ................................ SUCCESS
> >> [1.252s]
> >>> [INFO] Apache Hadoop Auth Examples ....................... SUCCESS
> >> [0.929s]
> >>> [INFO] Apache Hadoop Common .............................. SUCCESS
> >> [41.330s]
> >>> [INFO] Apache Hadoop NFS ................................. SUCCESS
> >> [1.986s]
> >>> [INFO] Apache Hadoop Common Project ...................... SUCCESS
> >> [0.015s]
> >>> [INFO] Apache Hadoop HDFS ................................ SUCCESS
> >> [1:08.367s]
> >>> [INFO] Apache Hadoop HttpFS .............................. SUCCESS
> >> [47.198s]
> >>> [INFO] Apache Hadoop HDFS BookKeeper Journal ............. SUCCESS
> >> [2.807s]
> >>> [INFO] Apache Hadoop HDFS-NFS ............................ SUCCESS
> >> [1.350s]
> >>> [INFO] Apache Hadoop HDFS Project ........................ SUCCESS
> >> [0.027s]
> >>> [INFO] hadoop-yarn ....................................... SUCCESS
> >> [0.013s]
> >>> [INFO] hadoop-yarn-api ................................... SUCCESS
> >> [36.848s]
> >>> [INFO] hadoop-yarn-common ................................ SUCCESS
> >> [12.502s]
> >>> [INFO] hadoop-yarn-server ................................ SUCCESS
> >> [0.032s]
> >>> [INFO] hadoop-yarn-server-common ......................... SUCCESS
> >> [3.688s]
> >>> [INFO] hadoop-yarn-server-nodemanager .................... SUCCESS
> >> [8.207s]
> >>> [INFO] hadoop-yarn-server-web-proxy ...................... SUCCESS
> >> [1.048s]
> >>> [INFO] hadoop-yarn-server-applicationhistoryservice ...... SUCCESS
> >> [1.839s]
> >>> [INFO] hadoop-yarn-server-resourcemanager ................ SUCCESS
> >> [4.766s]
> >>> [INFO] hadoop-yarn-server-tests .......................... SUCCESS
> >> [0.247s]
> >>> [INFO] hadoop-yarn-client ................................ SUCCESS
> >> [1.735s]
> >>> [INFO] hadoop-yarn-applications .......................... SUCCESS
> >> [0.013s]
> >>> [INFO] hadoop-yarn-applications-distributedshell ......... SUCCESS
> >> [0.984s]
> >>> [INFO] hadoop-yarn-applications-unmanaged-am-launcher .... SUCCESS
> >> [0.792s]
> >>> [INFO] hadoop-yarn-site .................................. SUCCESS
> >> [0.034s]
> >>> [INFO] hadoop-yarn-project ............................... SUCCESS
> >> [3.327s]
> >>> [INFO] hadoop-mapreduce-client ........................... SUCCESS
> >> [0.090s]
> >>> [INFO] hadoop-mapreduce-client-core ...................... SUCCESS
> >> [7.451s]
> >>> [INFO] hadoop-mapreduce-client-common .................... SUCCESS
> >> [7.081s]
> >>> [INFO] hadoop-mapreduce-client-shuffle ................... SUCCESS
> >> [0.972s]
> >>> [INFO] hadoop-mapreduce-client-app ....................... SUCCESS
> >> [3.085s]
> >>> [INFO] hadoop-mapreduce-client-hs ........................ SUCCESS
> >> [3.119s]
> >>> [INFO] hadoop-mapreduce-client-jobclient ................. SUCCESS
> >> [1.934s]
> >>> [INFO] hadoop-mapreduce-client-hs-plugins ................ SUCCESS
> >> [0.772s]
> >>> [INFO] Apache Hadoop MapReduce Examples .................. SUCCESS
> >> [2.162s]
> >>> [INFO] hadoop-mapreduce .................................. SUCCESS
> >> [2.622s]
> >>> [INFO] Apache Hadoop MapReduce Streaming ................. SUCCESS
> >> [1.744s]
> >>> [INFO] Apache Hadoop Distributed Copy .................... SUCCESS
> >> [4.466s]
> >>> [INFO] Apache Hadoop Archives ............................ SUCCESS
> >> [0.956s]
> >>> [INFO] Apache Hadoop Rumen ............................... SUCCESS
> >> [2.203s]
> >>> [INFO] Apache Hadoop Gridmix ............................. SUCCESS
> >> [1.509s]
> >>> [INFO] Apache Hadoop Data Join ........................... SUCCESS
> >> [0.909s]
> >>> [INFO] Apache Hadoop Extras .............................. SUCCESS
> >> [1.103s]
> >>> [INFO] Apache Hadoop Pipes ............................... SUCCESS
> >> [4.794s]
> >>> [INFO] Apache Hadoop OpenStack support ................... SUCCESS
> >> [2.111s]
> >>> [INFO] Apache Hadoop Client .............................. SUCCESS
> >> [3.919s]
> >>> [INFO] Apache Hadoop Mini-Cluster ........................ SUCCESS
> >> [0.044s]
> >>> [INFO] Apache Hadoop Scheduler Load Simulator ............ SUCCESS
> >> [1.665s]
> >>> [INFO] Apache Hadoop Tools Dist .......................... SUCCESS
> >> [3.936s]
> >>> [INFO] Apache Hadoop Tools ............................... SUCCESS
> >> [0.042s]
> >>> [INFO] Apache Hadoop Distribution ........................ SUCCESS
> >> [15.208s]
> >>> [INFO]
> >> ------------------------------------------------------------------------
> >>> [INFO] BUILD SUCCESS
> >>> [INFO]
> >> ------------------------------------------------------------------------
> >>> [INFO] Total time: 5:22.529s
> >>> [INFO] Finished at: Wed Aug 27 12:17:06 HKT 2014
> >>> [INFO] Final Memory: 86M/755M
> >>> [INFO]
> >> ------------------------------------------------------------------------
> >>>
> >>> ll
> >> hadoop-common-project/hadoop-common/target/native/target/usr/local/lib/
> >>> -rw-rw-r--. 1 hduser hadoop 1062640 Aug 27 12:12 libhadoop.a
> >>> lrwxrwxrwx. 1 hduser hadoop      18 Aug 27 12:12 libhadoop.so ->
> >> libhadoop.so.1.0.0
> >>> -rwxrwxr-x. 1 hduser hadoop  630328 Aug 27 12:12 libhadoop.so.1.0.0
> >>>
> >>> (copy them to $HADOOP_HOME/lib and $HBASE_HOME/lib)
> >>> cp
> >> hadoop-common-project/hadoop-common/target/native/target/usr/local/lib/*
> >> $HADOOP_HOME/lib/native/Linux-amd64-64/
> >>> cp
> >> hadoop-common-project/hadoop-common/target/native/target/usr/local/lib/*
> >> $HBASE_HOME/lib/native/Linux-amd64-64/
> >>>
> >>> ll $HADOOP_HOME/lib/native/Linux-amd64-64/
> >>> total 21236
> >>> -rw-rw-r--. 1 hduser hadoop 1062640 Aug 27 12:19 libhadoop.a
> >>                                                      // new
> >>> lrwxrwxrwx. 1 hduser hadoop      24 Aug 27 06:54 libhadoopsnappy.so ->
> >> libhadoopsnappy.so.0.0.1
> >>> lrwxrwxrwx. 1 hduser hadoop      24 Aug 27 06:54 libhadoopsnappy.so.0
> ->
> >> libhadoopsnappy.so.0.0.1
> >>> -rwxr-xr-x. 1 hduser hadoop   54961 Aug 27 06:54
> libhadoopsnappy.so.0.0.1
> >>> -rwxrwxr-x. 1 hduser hadoop  630328 Aug 27 12:19 libhadoop.so
> >>                                                     // new
> >>> -rwxrwxr-x. 1 hduser hadoop  630328 Aug 27 12:19 libhadoop.so.1.0.0
> >>                                                     // new
> >>> lrwxrwxrwx. 1 hduser hadoop      55 Aug 27 06:54 libjvm.so ->
> >> /usr/lib/jvm/jdk1.6.0_45/jre/lib/amd64/server/libjvm.so
> >>> lrwxrwxrwx. 1 hduser hadoop      25 Aug 27 06:54 libprotobuf-lite.so ->
> >> libprotobuf-lite.so.8.0.0
> >>> lrwxrwxrwx. 1 hduser hadoop      25 Aug 27 06:54 libprotobuf-lite.so.8
> >> -> libprotobuf-lite.so.8.0.0
> >>> -rwxr-xr-x. 1 hduser hadoop  964689 Aug 27 06:54
> >> libprotobuf-lite.so.8.0.0
> >>> lrwxrwxrwx. 1 hduser hadoop      20 Aug 27 06:54 libprotobuf.so ->
> >> libprotobuf.so.8.0.0
> >>> lrwxrwxrwx. 1 hduser hadoop      20 Aug 27 06:54 libprotobuf.so.8 ->
> >> libprotobuf.so.8.0.0
> >>> -rwxr-xr-x. 1 hduser hadoop 8300050 Aug 27 06:54 libprotobuf.so.8.0.0
> >>> lrwxrwxrwx. 1 hduser hadoop      18 Aug 27 06:54 libprotoc.so ->
> >> libprotoc.so.8.0.0
> >>> lrwxrwxrwx. 1 hduser hadoop      18 Aug 27 06:54 libprotoc.so.8 ->
> >> libprotoc.so.8.0.0
> >>> -rwxr-xr-x. 1 hduser hadoop 9935810 Aug 27 06:54 libprotoc.so.8.0.0
> >>> lrwxrwxrwx. 1 hduser hadoop      23 Aug 27 11:31 libsnappy.so ->
> >> /usr/lib64/libsnappy.so
> >>> lrwxrwxrwx. 1 hduser hadoop      23 Aug 27 11:32 libsnappy.so.1 ->
> >> /usr/lib64/libsnappy.so
> >>> -rwxr-xr-x. 1 hduser hadoop  147726 Aug 27 06:54 libsnappy.so.1.2.0
> >>> drwxr-xr-x. 2 hduser hadoop    4096 Aug 27 11:15 pkgconfig
> >>>
> >>>
> >>> ll $HBASE_HOME/lib/native/Linux-amd64-64/
> >>> -rw-rw-r--. 1 hduser hadoop 1062640 Aug 27 12:19 libhadoop.a
> >>                                                      // new
> >>> -rw-rw-r--. 1 hduser hadoop 1487564 Aug 27 11:14 libhadooppipes.a
> >>> lrwxrwxrwx. 1 hduser hadoop      24 Aug 27 07:08 libhadoopsnappy.so ->
> >> libhadoopsnappy.so.0.0.1
> >>> lrwxrwxrwx. 1 hduser hadoop      24 Aug 27 07:08 libhadoopsnappy.so.0
> ->
> >> libhadoopsnappy.so.0.0.1
> >>> -rwxr-xr-x. 1 hduser hadoop   54961 Aug 27 07:08
> libhadoopsnappy.so.0.0.1
> >>> -rwxrwxr-x. 1 hduser hadoop  630328 Aug 27 12:19 libhadoop.so
> >>                                                     // new
> >>> -rwxrwxr-x. 1 hduser hadoop  630328 Aug 27 12:19 libhadoop.so.1.0.0
> >>                                                     // new
> >>> -rw-rw-r--. 1 hduser hadoop  582472 Aug 27 11:14 libhadooputils.a
> >>> -rw-rw-r--. 1 hduser hadoop  298626 Aug 27 11:14 libhdfs.a
> >>> -rwxrwxr-x. 1 hduser hadoop  200370 Aug 27 11:14 libhdfs.so
> >>> -rwxrwxr-x. 1 hduser hadoop  200370 Aug 27 11:14 libhdfs.so.0.0.0
> >>> lrwxrwxrwx. 1 hduser hadoop      55 Aug 27 07:08 libjvm.so ->
> >> /usr/lib/jvm/jdk1.6.0_45/jre/lib/amd64/server/libjvm.so
> >>> lrwxrwxrwx. 1 hduser hadoop      25 Aug 27 07:08 libprotobuf-lite.so ->
> >> libprotobuf-lite.so.8.0.0
> >>> lrwxrwxrwx. 1 hduser hadoop      25 Aug 27 07:08 libprotobuf-lite.so.8
> >> -> libprotobuf-lite.so.8.0.0
> >>> -rwxr-xr-x. 1 hduser hadoop  964689 Aug 27 07:08
> >> libprotobuf-lite.so.8.0.0
> >>> lrwxrwxrwx. 1 hduser hadoop      20 Aug 27 07:08 libprotobuf.so ->
> >> libprotobuf.so.8.0.0
> >>> lrwxrwxrwx. 1 hduser hadoop      20 Aug 27 07:08 libprotobuf.so.8 ->
> >> libprotobuf.so.8.0.0
> >>> -rwxr-xr-x. 1 hduser hadoop 8300050 Aug 27 07:08 libprotobuf.so.8.0.0
> >>> lrwxrwxrwx. 1 hduser hadoop      18 Aug 27 07:08 libprotoc.so ->
> >> libprotoc.so.8.0.0
> >>> lrwxrwxrwx. 1 hduser hadoop      18 Aug 27 07:08 libprotoc.so.8 ->
> >> libprotoc.so.8.0.0
> >>> -rwxr-xr-x. 1 hduser hadoop 9935810 Aug 27 07:08 libprotoc.so.8.0.0
> >>> lrwxrwxrwx. 1 hduser hadoop      23 Aug 27 11:32 libsnappy.so ->
> >> /usr/lib64/libsnappy.so
> >>> lrwxrwxrwx. 1 hduser hadoop      23 Aug 27 11:33 libsnappy.so.1 ->
> >> /usr/lib64/libsnappy.so
> >>> -rwxr-xr-x. 1 hduser hadoop  147726 Aug 27 07:08 libsnappy.so.1.2.0
> >>> drwxr-xr-x. 2 hduser hadoop    4096 Aug 27 07:08 pkgconfig
> >>>
> >>>
> >>>
> >>> sudo yum install snappy snappy-devel
> >>> Loaded plugins: fastestmirror, security
> >>> Loading mirror speeds from cached hostfile
> >>> ...
> >>> Package snappy-1.1.0-1.el6.x86_64 already installed and latest version
> >>> Package snappy-devel-1.1.0-1.el6.x86_64 already installed and latest
> >> version
> >>> Nothing to do
> >>>
> >>>
> >>> ln -sf /usr/lib64/libsnappy.so $HADOOP_HOME/lib/native/Linux-amd64-64/.
> >>> ln -sf /usr/lib64/libsnappy.so $HBASE_HOME/lib/native/Linux-amd64-64/.
> >>>
> >>> ll $HADOOP_HOME/lib/native/Linux-amd64-64/libsnappy.so
> >>> lrwxrwxrwx. 1 hduser hadoop 23 Aug 27 11:31
> >> $HADOOP_HOME/lib/native/Linux-amd64-64/libsnappy.so ->
> >> /usr/lib64/libsnappy.s
> >>> ll $HBASE_HOME/lib/native/Linux-amd64-64/libsnappy.so
> >>> lrwxrwxrwx. 1 hduser hadoop 23 Aug 27 11:32
> >> $HBASE_HOME/lib/native/Linux-amd64-64/libsnappy.so ->
> >> /usr/lib64/libsnappy.so
> >>>
> >>>
> >>>
> >>> ($HADOOP_HOME/etc/hadoop/hadoop-env.sh  added following)
> >>> ### 2014-08-27
> >>> export
> >>
> LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$HADOOP_HOME/lib/native/Linux-amd64-64/:/usr/local/lib/
> >>> ###
> >>>
> >>> ($HBASE_HOME/conf/hbase-env.sh added following)
> >>> ### 2014-08-27
> >>> export
> >>
> LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$HADOOP_HOME/lib/native/Linux-amd64-64/:/usr/local/lib/
> >>> export
> >>
> HBASE_LIBRARY_PATH=$HBASE_LIBRARY_PATH:$HBASE_HOME/lib/native/Linux-amd64-64/:/usr/local/lib/:$HBASE_HOME/lib/hadoop-snappy-0.0.1-SNAPSHOT.jar
> >>> export CLASSPATH=$CLASSPATH:$HBASE_LIBRARY_PATH
> >>> export HBASE_CLASSPATH=$HBASE_CLASSPATH:$HBASE_LIBRARY_PATH
> >>> ###
> >>>
> >>>
> >>> (restarted both HADOOP and HBASE)
> >>> jps
> >>> 26324 HRegionServer
> >>> 26137 HMaster
> >>> 25567 JobHistoryServer
> >>> 25485 NodeManager
> >>> 25913 WebAppProxyServer
> >>> 24831 DataNode
> >>> 24712 NameNode
> >>> 27146 Jps
> >>> 9219 QuorumPeerMain
> >>> 25042 JournalNode
> >>> 25239 DFSZKFailoverController
> >>> 25358 ResourceManager
> >>>
> >>>
> >>> bin/hbase org.apache.hadoop.hbase.util.CompressionTest
> >> file:///tmp/snappy-test snappy
> >>> 2014-08-27 12:24:08,030 INFO  [main] Configuration.deprecation:
> >> hadoop.native.lib is deprecated. Instead, use io.native.lib.available
> >>> SLF4J: Class path contains multiple SLF4J bindings.
> >>> SLF4J: Found binding in
> >>
> [jar:file:/mnt/hadoop/hbase-0.98.4-hadoop2/lib/slf4j-log4j12-1.6.4.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> >>> SLF4J: Found binding in
> >>
> [jar:file:/mnt/hadoop/hadoop-2.4.1/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> >>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
> >> explanation.
> >>> 2014-08-27 12:24:08,387 INFO  [main] util.ChecksumType: Checksum using
> >> org.apache.hadoop.util.PureJavaCrc32
> >>> 2014-08-27 12:24:08,388 INFO  [main] util.ChecksumType: Checksum can
> use
> >> org.apache.hadoop.util.PureJavaCrc32C
> >>> Exception in thread "main" java.lang.RuntimeException: native snappy
> >> library not available: this version of libhadoop was built without
> snappy
> >> support.
> >>>      at
> >>
> org.apache.hadoop.io.compress.SnappyCodec.checkNativeCodeLoaded(SnappyCodec.java:64)
> >>>      at
> >>
> org.apache.hadoop.io.compress.SnappyCodec.getCompressorType(SnappyCodec.java:132)
> >>>      at
> >>
> org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:148)
> >>>      at
> >>
> org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:163)
> >>>      at
> >>
> org.apache.hadoop.hbase.io.compress.Compression$Algorithm.getCompressor(Compression.java:310)
> >>>      at
> >>
> org.apache.hadoop.hbase.io.encoding.HFileBlockDefaultEncodingContext.<init>(HFileBlockDefaultEncodingContext.java:92)
> >>>      at
> >>
> org.apache.hadoop.hbase.io.hfile.HFileBlock$Writer.<init>(HFileBlock.java:690)
> >>>      at
> >>
> org.apache.hadoop.hbase.io.hfile.HFileWriterV2.finishInit(HFileWriterV2.java:117)
> >>>      at
> >>
> org.apache.hadoop.hbase.io.hfile.HFileWriterV2.<init>(HFileWriterV2.java:109)
> >>>      at
> >>
> org.apache.hadoop.hbase.io.hfile.HFileWriterV2$WriterFactoryV2.createWriter(HFileWriterV2.java:97)
> >>>      at
> >>
> org.apache.hadoop.hbase.io.hfile.HFile$WriterFactory.create(HFile.java:393)
> >>>      at
> >>
> org.apache.hadoop.hbase.util.CompressionTest.doSmokeTest(CompressionTest.java:118)
> >>>      at
> >>
> org.apache.hadoop.hbase.util.CompressionTest.main(CompressionTest.java:148)
> >>>
> >>>
> >>> bin/hbase org.apache.hadoop.hbase.util.CompressionTest
> >> file:///tmp/snappy-test gz
> >>> 2014-08-27 12:35:34,485 INFO  [main] Configuration.deprecation:
> >> hadoop.native.lib is deprecated. Instead, use io.native.lib.available
> >>> SLF4J: Class path contains multiple SLF4J bindings.
> >>> SLF4J: Found binding in
> >>
> [jar:file:/mnt/hadoop/hbase-0.98.4-hadoop2/lib/slf4j-log4j12-1.6.4.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> >>> SLF4J: Found binding in
> >>
> [jar:file:/mnt/hadoop/hadoop-2.4.1/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> >>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
> >> explanation.
> >>> 2014-08-27 12:35:35,495 INFO  [main] util.ChecksumType: Checksum using
> >> org.apache.hadoop.util.PureJavaCrc32
> >>> 2014-08-27 12:35:35,495 INFO  [main] util.ChecksumType: Checksum can
> use
> >> org.apache.hadoop.util.PureJavaCrc32C
> >>> 2014-08-27 12:35:35,822 INFO  [main] zlib.ZlibFactory: Successfully
> >> loaded & initialized native-zlib library
> >>> 2014-08-27 12:35:35,851 INFO  [main] compress.CodecPool: Got brand-new
> >> compressor [.gz]
> >>> 2014-08-27 12:35:35,855 INFO  [main] compress.CodecPool: Got brand-new
> >> compressor [.gz]
> >>> 2014-08-27 12:35:35,866 ERROR [main] hbase.KeyValue: Unexpected
> >> getShortMidpointKey result, fakeKey:testkey, firstKeyInBlock:testkey
> >>> 2014-08-27 12:35:36,636 INFO  [main] compress.CodecPool: Got brand-new
> >> decompressor [.gz]
> >>> SUCCESS
> >>>
> >>>
> >>>
> >>>
> >>>
> >>> So still get the same issue,  I feel the issue should come from the
> >> hadoop compilation but no idea where would be wrong. Please help.
> >>>
> >>>
> >>> in my /etc/hadoop/core-site.xml, I have following related to snappy:
> >>>   <property>
> >>>    <name>io.compression.codecs</name>
> >>>    <value>
> >>>      org.apache.hadoop.io.compress.GzipCodec,
> >>>      org.apache.hadoop.io.compress.DefaultCodec,
> >>>      org.apache.hadoop.io.compress.BZip2Codec,
> >>>      org.apache.hadoop.io.compress.SnappyCodec
> >>>    </value>
> >>>   </property>
> >>>
> >>> in my mapred-site.xml, I have following related to snappy:
> >>>   <property>
> >>>    <name>mapred.output.compress</name>
> >>>    <value>false</value>
> >>>    <description>Should the job outputs be compressed?</description>
> >>>   </property>
> >>>   <property>
> >>>    <name>mapred.output.compression.type</name>
> >>>    <value>RECORD</value>
> >>>    <description>If the job outputs are to compressed as SequenceFiles,
> >> how should they be compressed? Should be one of NONE, RECORD or
> >> BLOCK.</description>
> >>>   </property>
> >>>   <property>
> >>>    <name>mapred.output.compression.codec</name>
> >>>    <value>org.apache.hadoop.io.compress.SnappyCodec</value>
> >>>    <description>If the job outputs are compressed, how should they be
> >> compressed?
> >>>    </description>
> >>>   </property>
> >>>   <property>
> >>>    <name>mapred.compress.map.output</name>
> >>>    <value>true</value>
> >>>    <description>Should the outputs of the maps be compressed before
> >> being sent across the network. Uses SequenceFile
> compression.</description>
> >>>   </property>
> >>>   <property>
> >>>    <name>mapred.map.output.compression.codec</name>
> >>>    <value>org.apache.hadoop.io.compress.SnappyCodec</value>
> >>>    <description>If the map outputs are compressed, how should they be
> >> compressed?</description>
> >>>  </property>
> >>>
> >>>  <property>
> >>>   <name>mapreduce.map.output.compress</name>
> >>>   <value>true</value>
> >>>  </property>
> >>>  <property>
> >>>   <name>mapred.map.output.compress.codec</name>
> >>>   <value>org.apache.hadoop.io.compress.SnappyCodec</value>
> >>>  </property>
> >>>
> >>>
> >>> I didn’t add any snappy related property to base-site.xml
> >>>
> >>>
> >>>
> >>> Regards
> >>> Arthur
> >>>
> >>>
> >>>
> >>>
> >>> On 27 Aug, 2014, at 8:07 am, Andrew Purtell <[email protected]>
> wrote:
> >>>
> >>>> On Tue, Aug 26, 2014 at 4:25 PM, [email protected] <
> >>>> [email protected]> wrote:
> >>>>
> >>>>> Exception in thread "main" java.lang.RuntimeException: native snappy
> >>>>> library not available: this version of libhadoop was built without
> >> snappy
> >>>>> support.
> >>>>
> >>>> ​
> >>>> You are almost there. Unfortunately the native Hadoop libraries you
> >> copied
> >>>> into HBase's lib/native/Linux-amd64-64/ directory were
> >>>> ​apparently ​
> >>>> built without snappy support, as the exception indicates. You'll need
> to
> >>>> compile the native Hadoop libraries with snappy support enabled.
> Install
> >>>> snappy-revel as Alex mentioned and then build the Hadoop native
> >> libraries.
> >>>>
> >>>> 1. Get Hadoop sources for the Hadoop version
> >>>> 2. tar xvzf ....
> >>>> 3. cd /path/to/hadoop/src
> >>>> 4. mvn -DskipTests clean install
> >>>> ​ -Drequire.snappy=true​
> >>>> -Pnative
> >>>> 5. cp
> >>>>
> >>
> hadoop-common-project/hadoop-common/target/native/target/usr/local/lib/libhadoop.*
> >>>> /path/to/hbase/lib/native/Linux-amd64-64
> >>>>
> >>>> ​(The -Drequire.snappy=true will fail the build if Snappy link
> libraries
> >>>> are not installed, so you can be sure of this.)​
> >>>>
> >>>>
> >>>> --
> >>>> Best regards,
> >>>>
> >>>>  - Andy
> >>>>
> >>>> Problems worthy of attack prove their worth by hitting back. - Piet
> Hein
> >>>> (via Tom White)
> >>>
> >>
> >>
>
>

Reply via email to