Thanks!

java org.apache.hadoop.util.PlatformName
Linux-amd64-64


On 27 Aug, 2014, at 8:31 am, Jean-Marc Spaggiari <[email protected]> 
wrote:

> This command will give you the exact name:
> 
> java org.apache.hadoop.util.PlatformName | sed -e "s/ /_/g"
> 
> Can you try to run it?
> 
> But it's most probably Linux-amd64-64
> 
> 
> 
> 2014-08-26 20:24 GMT-04:00 [email protected] <
> [email protected]>:
> 
>> Hi,
>> 
>> Thanks!
>> 
>> A question:
>> If I run:
>> $  uname -m
>> x86_64
>> 
>> Should I use " lib/native/Linux-amd64-64” or  "lib/native/x86_64”  in
>> $HADOOP_HOME and $HBASE_HOME?
>> 
>> Arthur
>> 
>> 
>> On 27 Aug, 2014, at 8:10 am, Jean-Marc Spaggiari <[email protected]>
>> wrote:
>> 
>>> Ok.
>>> 
>>> This is the way the lib path is built:
>>> 
>>> JAVA_LIBRARY_PATH=$(append_path "$JAVA_LIBRARY_PATH"
>>> ${HBASE_HOME}/build/native/${JAVA_PLATFORM}/lib)
>>> 
>>> And JAVA_PLATFORM comes from JAVA_PLATFORM=`CLASSPATH=${CLASSPATH}
>> ${JAVA}
>>> org.apache.hadoop.util.PlatformName | sed -e "s/ /_/g"`
>>> 
>>> You can double check it doing:
>>> 
>>> # Adjust to you java_home...
>>> export JAVA_HOME=/usr/local/jdk1.7.0_45/
>>> 
>>> export CLASSPATH=`bin/hbase classpath`
>>> $JAVA_HOME/bin/java org.apache.hadoop.util.PlatformName | sed -e "s/
>> /_/g"
>>> 
>>> Result for me is this: Linux-amd64-64. Might  be different for you.
>>> 
>>> Then you link the libs the way Alex said before:
>>> cd lib/native/Linux-amd64-64
>>> ln -s /home/hbase/snappy-1.0.5/.libs/libsnappy.so .
>>> ln -s /home/hbase/snappy-1.0.5/.libs/libsnappy.so.1 .
>>> 
>>> AND.....
>>> 
>>> The hadoop so too! And I think this is what's missing for you:
>>> ln -s /YOURHADOOPPATH/libhadoop.so .
>>> 
>>> Your folder should look like this:
>>> jmspaggi@node8:~/hbase-0.98.5-hadoop2/lib/native$ tree
>>> .
>>> └── Linux-amd64-64
>>>   ├── libhadoop.so
>>>   ├── libsnappy.so -> /home/hbase/snappy-1.0.5/.libs/libsnappy.so
>>>   └── libsnappy.so.1 -> /home/hbase/snappy-1.0.5/.libs/libsnappy.so.1
>>> 
>>> I copied libhadoop.so instead of doing a link because it was not
>> available
>>> on this computer.
>>> 
>>> Then test it:
>>> jmspaggi@node8:~/hbase-0.98.5-hadoop2$ bin/hbase
>>> org.apache.hadoop.hbase.util.CompressionTest file:///tmp/snappy-test
>> snappy
>>> 2014-08-26 20:06:43,987 INFO  [main] Configuration.deprecation:
>>> hadoop.native.lib is deprecated. Instead, use io.native.lib.available
>>> 2014-08-26 20:06:44,831 INFO  [main] util.ChecksumType: Checksum using
>>> org.apache.hadoop.util.PureJavaCrc32
>>> 2014-08-26 20:06:44,832 INFO  [main] util.ChecksumType: Checksum can use
>>> org.apache.hadoop.util.PureJavaCrc32C
>>> 2014-08-26 20:06:45,125 INFO  [main] compress.CodecPool: Got brand-new
>>> compressor [.snappy]
>>> 2014-08-26 20:06:45,131 INFO  [main] compress.CodecPool: Got brand-new
>>> compressor [.snappy]
>>> 2014-08-26 20:06:45,254 INFO  [main] compress.CodecPool: Got brand-new
>>> decompressor [.snappy]
>>> SUCCESS
>>> 
>>> 
>>> Please let us know if it still doesn't work for you. Without libhadoop.so
>>> it doesn't work for me...
>>> jmspaggi@node8:~/hbase-0.98.5-hadoop2/lib/native$ rm
>>> Linux-amd64-64/libhadoop.so
>>> 
>>> jmspaggi@node8:~/hbase-0.98.5-hadoop2$ bin/hbase
>>> org.apache.hadoop.hbase.util.CompressionTest file:///tmp/snappy-test
>> snappy
>>> 2014-08-26 20:09:28,945 INFO  [main] Configuration.deprecation:
>>> hadoop.native.lib is deprecated. Instead, use io.native.lib.available
>>> 2014-08-26 20:09:29,460 WARN  [main] util.NativeCodeLoader: Unable to
>> load
>>> native-hadoop library for your platform... using builtin-java classes
>> where
>>> applicable
>>> 2014-08-26 20:09:29,775 INFO  [main] util.ChecksumType: Checksum using
>>> org.apache.hadoop.util.PureJavaCrc32
>>> 2014-08-26 20:09:29,776 INFO  [main] util.ChecksumType: Checksum can use
>>> org.apache.hadoop.util.PureJavaCrc32C
>>> Exception in thread "main" java.lang.UnsatisfiedLinkError:
>>> org.apache.hadoop.util.NativeCodeLoader.buildSupportsSnappy()Z
>>>   at org.apache.hadoop.util.NativeCodeLoader.buildSupportsSnappy(Native
>>> Method)
>>> ...
>>> 
>>> 
>>> I did all of that using a brand new extracted
>>> hbase-0.98.5-hadoop2-bin.tar.gz file.
>>> 
>>> JM
>>> 
>>> 
>>> 2014-08-26 19:47 GMT-04:00 [email protected] <
>>> [email protected]>:
>>> 
>>>> $ uname -m
>>>> x86_64
>>>> 
>>>> Arthur
>>>> 
>>>> On 27 Aug, 2014, at 7:45 am, Jean-Marc Spaggiari <
>> [email protected]>
>>>> wrote:
>>>> 
>>>>> Hi Arthur,
>>>>> 
>>>>> What uname -m gives you? you need to check that to create the right
>>>> folder
>>>>> under the lib directory.
>>>>> 
>>>>> JM
>>>>> 
>>>>> 
>>>>> 2014-08-26 19:43 GMT-04:00 Alex Kamil <[email protected]>:
>>>>> 
>>>>>> Something like this worked for me
>>>>>> 1. get hbase binaries
>>>>>> 2. sudo yum install snappy snappy-devel
>>>>>> 3. ln -sf /usr/lib64/libsnappy.so
>>>>>> /var/lib/hadoop/lib/native/Linux-amd64-64/.
>>>>>> 4. ln -sf /usr/lib64/libsnappy.so
>>>>>> /var/lib/hbase/lib/native/Linux-amd64-64/.
>>>>>> 5. add snappy jar under $HADOOP_HOME/lib and $HBASE_HOME/lib
>>>>>> ref: https://issues.apache.org/jira/browse/PHOENIX-877
>>>>>> 
>>>>>> 
>>>>>> On Tue, Aug 26, 2014 at 7:25 PM, [email protected] <
>>>>>> [email protected]> wrote:
>>>>>> 
>>>>>>> Hi,
>>>>>>> 
>>>>>>> I just tried three more steps but was not able to get thru.
>>>>>>> 
>>>>>>> 
>>>>>>> 1) copied  snappy files to $HBASE_HOME/lib
>>>>>>> $ cd $HBASE_HOME
>>>>>>> $ ll lib/*sna*
>>>>>>> -rw-r--r--. 1 hduser hadoop  11526 Aug 27 06:54
>>>>>>> lib/hadoop-snappy-0.0.1-SNAPSHOT.jar
>>>>>>> -rw-rw-r--. 1 hduser hadoop 995968 Aug  3 18:43
>>>>>> lib/snappy-java-1.0.4.1.jar
>>>>>>> 
>>>>>>> ll lib/native/
>>>>>>> drwxrwxr-x. 4 hduser hadoop 4096 Aug 27 06:54 Linux-amd64-64
>>>>>>> 
>>>>>>> ll lib/native/Linux-amd64-64/
>>>>>>> total 18964
>>>>>>> lrwxrwxrwx. 1 hduser Hadoop      24 Aug 27 07:08 libhadoopsnappy.so
>> ->
>>>>>>> libhadoopsnappy.so.0.0.1
>>>>>>> lrwxrwxrwx. 1 hduser Hadoop      24 Aug 27 07:08 libhadoopsnappy.so.0
>>>> ->
>>>>>>> libhadoopsnappy.so.0.0.1
>>>>>>> -rwxr-xr-x. 1 hduser Hadoop   54961 Aug 27 07:08
>>>> libhadoopsnappy.so.0.0.1
>>>>>>> lrwxrwxrwx. 1 hduser Hadoop      55 Aug 27 07:08 libjvm.so ->
>>>>>>> /usr/lib/jvm/jdk1.6.0_45/jre/lib/amd64/server/libjvm.so
>>>>>>> lrwxrwxrwx. 1 hduser Hadoop      25 Aug 27 07:08 libprotobuf-lite.so
>> ->
>>>>>>> libprotobuf-lite.so.8.0.0
>>>>>>> lrwxrwxrwx. 1 hduser Hadoop      25 Aug 27 07:08
>> libprotobuf-lite.so.8
>>>> ->
>>>>>>> libprotobuf-lite.so.8.0.0
>>>>>>> -rwxr-xr-x. 1 hduser Hadoop  964689 Aug 27 07:08
>>>>>> libprotobuf-lite.so.8.0.0
>>>>>>> lrwxrwxrwx. 1 hduser Hadoop      20 Aug 27 07:08 libprotobuf.so ->
>>>>>>> libprotobuf.so.8.0.0
>>>>>>> lrwxrwxrwx. 1 hduser Hadoop      20 Aug 27 07:08 libprotobuf.so.8 ->
>>>>>>> libprotobuf.so.8.0.0
>>>>>>> -rwxr-xr-x. 1 hduser Hadoop 8300050 Aug 27 07:08 libprotobuf.so.8.0.0
>>>>>>> lrwxrwxrwx. 1 hduser Hadoop      18 Aug 27 07:08 libprotoc.so ->
>>>>>>> libprotoc.so.8.0.0
>>>>>>> lrwxrwxrwx. 1 hduser Hadoop      18 Aug 27 07:08 libprotoc.so.8 ->
>>>>>>> libprotoc.so.8.0.0
>>>>>>> -rwxr-xr-x. 1 hduser Hadoop 9935810 Aug 27 07:08 libprotoc.so.8.0.0
>>>>>>> lrwxrwxrwx. 1 hduser Hadoop      18 Aug 27 07:08 libsnappy.so ->
>>>>>>> libsnappy.so.1.2.0
>>>>>>> lrwxrwxrwx. 1 hduser Hadoop      18 Aug 27 07:08 libsnappy.so.1 ->
>>>>>>> libsnappy.so.1.2.0
>>>>>>> -rwxr-xr-x. 1 hduser Hadoop  147726 Aug 27 07:08 libsnappy.so.1.2.0
>>>>>>> drwxr-xr-x. 2 hduser Hadoop    4096 Aug 27 07:08 pkgconfig
>>>>>>> 
>>>>>>> 2)  $HBASE_HOME/conf/hbase-env.sh, added
>>>>>>> 
>>>>>>> ###
>>>>>>> export
>>>>>>> 
>>>>>> 
>>>> 
>> LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$HADOOP_HOME/lib/native/Linux-amd64-64/:/usr/local/lib/
>>>>>>> export
>>>>>>> 
>>>>>> 
>>>> 
>> HBASE_LIBRARY_PATH=$HBASE_LIBRARY_PATH:$HBASE_HOME/lib/native/Linux-amd64-64/:/usr/local/lib/:$HBASE_HOME/lib/hadoop-snappy-0.0.1-SNAPSHOT.jar
>>>>>>> export CLASSPATH=$CLASSPATH:$HBASE_LIBRARY_PATH
>>>>>>> export HBASE_CLASSPATH=$HBASE_CLASSPATH:$HBASE_LIBRARY_PATH
>>>>>>> ###
>>>>>>> 
>>>>>>> 3) restart HBASE and tried again
>>>>>>> $ bin/hbase org.apache.hadoop.hbase.util.CompressionTest
>>>>>>> file:///tmp/snappy-test snappy
>>>>>>> 2014-08-27 07:16:09,490 INFO  [main] Configuration.deprecation:
>>>>>>> hadoop.native.lib is deprecated. Instead, use io.native.lib.available
>>>>>>> SLF4J: Class path contains multiple SLF4J bindings.
>>>>>>> SLF4J: Found binding in
>>>>>>> 
>>>>>> 
>>>> 
>> [jar:file:/edh/hadoop/hbase-0.98.4-hadoop2/lib/slf4j-log4j12-1.6.4.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>>>> SLF4J: Found binding in
>>>>>>> 
>>>>>> 
>>>> 
>> [jar:file:/edh/hadoop/hadoop-2.4.1/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>>>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
>>>>>>> explanation.
>>>>>>> 2014-08-27 07:16:10,323 INFO  [main] util.ChecksumType: Checksum
>> using
>>>>>>> org.apache.hadoop.util.PureJavaCrc32
>>>>>>> 2014-08-27 07:16:10,324 INFO  [main] util.ChecksumType: Checksum can
>>>> use
>>>>>>> org.apache.hadoop.util.PureJavaCrc32C
>>>>>>> Exception in thread "main" java.lang.RuntimeException: native snappy
>>>>>>> library not available: this version of libhadoop was built without
>>>> snappy
>>>>>>> support.
>>>>>>>      at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.io.compress.SnappyCodec.checkNativeCodeLoaded(SnappyCodec.java:64)
>>>>>>>      at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.io.compress.SnappyCodec.getCompressorType(SnappyCodec.java:132)
>>>>>>>      at
>>>>>>> 
>>>> 
>> org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:148)
>>>>>>>      at
>>>>>>> 
>>>> 
>> org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:163)
>>>>>>>      at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.hbase.io.compress.Compression$Algorithm.getCompressor(Compression.java:310)
>>>>>>>      at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.hbase.io.encoding.HFileBlockDefaultEncodingContext.<init>(HFileBlockDefaultEncodingContext.java:92)
>>>>>>>      at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.hbase.io.hfile.HFileBlock$Writer.<init>(HFileBlock.java:690)
>>>>>>>      at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.hbase.io.hfile.HFileWriterV2.finishInit(HFileWriterV2.java:117)
>>>>>>>      at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.hbase.io.hfile.HFileWriterV2.<init>(HFileWriterV2.java:109)
>>>>>>>      at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.hbase.io.hfile.HFileWriterV2$WriterFactoryV2.createWriter(HFileWriterV2.java:97)
>>>>>>>      at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.hbase.io.hfile.HFile$WriterFactory.create(HFile.java:393)
>>>>>>>      at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.hbase.util.CompressionTest.doSmokeTest(CompressionTest.java:118)
>>>>>>>      at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.hbase.util.CompressionTest.main(CompressionTest.java:148)
>>>>>>> 
>>>>>>> 
>>>>>>> Regards
>>>>>>> Arthur
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> On 27 Aug, 2014, at 6:27 am, [email protected] <
>>>>>>> [email protected]> wrote:
>>>>>>> 
>>>>>>>> Hi Sean,
>>>>>>>> 
>>>>>>>> Thanks for your reply.
>>>>>>>> 
>>>>>>>> I tried the following tests
>>>>>>>> 
>>>>>>>> $ bin/hbase org.apache.hadoop.hbase.util.CompressionTest
>>>>>>> file:///tmp/snappy-test gz
>>>>>>>> 2014-08-26 23:06:17,778 INFO  [main] Configuration.deprecation:
>>>>>>> hadoop.native.lib is deprecated. Instead, use io.native.lib.available
>>>>>>>> SLF4J: Class path contains multiple SLF4J bindings.
>>>>>>>> SLF4J: Found binding in
>>>>>>> 
>>>>>> 
>>>> 
>> [jar:file:/edh/hadoop/hbase-0.98.4-hadoop2/lib/slf4j-log4j12-1.6.4.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>>>>> SLF4J: Found binding in
>>>>>>> 
>>>>>> 
>>>> 
>> [jar:file:/edh/hadoop/hadoop-2.4.1/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>>>>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
>>>>>>> explanation.
>>>>>>>> 2014-08-26 23:06:18,103 INFO  [main] util.ChecksumType: Checksum
>> using
>>>>>>> org.apache.hadoop.util.PureJavaCrc32
>>>>>>>> 2014-08-26 23:06:18,104 INFO  [main] util.ChecksumType: Checksum can
>>>>>> use
>>>>>>> org.apache.hadoop.util.PureJavaCrc32C
>>>>>>>> 2014-08-26 23:06:18,260 INFO  [main] zlib.ZlibFactory: Successfully
>>>>>>> loaded & initialized native-zlib library
>>>>>>>> 2014-08-26 23:06:18,276 INFO  [main] compress.CodecPool: Got
>> brand-new
>>>>>>> compressor [.gz]
>>>>>>>> 2014-08-26 23:06:18,280 INFO  [main] compress.CodecPool: Got
>> brand-new
>>>>>>> compressor [.gz]
>>>>>>>> 2014-08-26 23:06:18,921 INFO  [main] compress.CodecPool: Got
>> brand-new
>>>>>>> decompressor [.gz]
>>>>>>>> SUCCESS
>>>>>>>> 
>>>>>>>> 
>>>>>>>> $ bin/hbase org.apache.hadoop.hbase.util.CompressionTest
>>>>>>> file:///tmp/snappy-test snappy
>>>>>>>> 2014-08-26 23:07:08,246 INFO  [main] Configuration.deprecation:
>>>>>>> hadoop.native.lib is deprecated. Instead, use io.native.lib.available
>>>>>>>> SLF4J: Class path contains multiple SLF4J bindings.
>>>>>>>> SLF4J: Found binding in
>>>>>>> 
>>>>>> 
>>>> 
>> [jar:file:/edh/hadoop/hbase-0.98.4-hadoop2/lib/slf4j-log4j12-1.6.4.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>>>>> SLF4J: Found binding in
>>>>>>> 
>>>>>> 
>>>> 
>> [jar:file:/edh/hadoop/hadoop-2.4.1/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>>>>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
>>>>>>> explanation.
>>>>>>>> 2014-08-26 23:07:08,578 INFO  [main] util.ChecksumType: Checksum
>> using
>>>>>>> org.apache.hadoop.util.PureJavaCrc32
>>>>>>>> 2014-08-26 23:07:08,579 INFO  [main] util.ChecksumType: Checksum can
>>>>>> use
>>>>>>> org.apache.hadoop.util.PureJavaCrc32C
>>>>>>>> Exception in thread "main" java.lang.RuntimeException: native snappy
>>>>>>> library not available: this version of libhadoop was built without
>>>> snappy
>>>>>>> support.
>>>>>>>>    at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.io.compress.SnappyCodec.checkNativeCodeLoaded(SnappyCodec.java:64)
>>>>>>>>    at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.io.compress.SnappyCodec.getCompressorType(SnappyCodec.java:132)
>>>>>>>>    at
>>>>>>> 
>>>> 
>> org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:148)
>>>>>>>>    at
>>>>>>> 
>>>> 
>> org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:163)
>>>>>>>>    at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.hbase.io.compress.Compression$Algorithm.getCompressor(Compression.java:310)
>>>>>>>>    at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.hbase.io.encoding.HFileBlockDefaultEncodingContext.<init>(HFileBlockDefaultEncodingContext.java:92)
>>>>>>>>    at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.hbase.io.hfile.HFileBlock$Writer.<init>(HFileBlock.java:690)
>>>>>>>>    at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.hbase.io.hfile.HFileWriterV2.finishInit(HFileWriterV2.java:117)
>>>>>>>>    at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.hbase.io.hfile.HFileWriterV2.<init>(HFileWriterV2.java:109)
>>>>>>>>    at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.hbase.io.hfile.HFileWriterV2$WriterFactoryV2.createWriter(HFileWriterV2.java:97)
>>>>>>>>    at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.hbase.io.hfile.HFile$WriterFactory.create(HFile.java:393)
>>>>>>>>    at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.hbase.util.CompressionTest.doSmokeTest(CompressionTest.java:118)
>>>>>>>>    at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.hbase.util.CompressionTest.main(CompressionTest.java:148)
>>>>>>>> 
>>>>>>>> 
>>>>>>>> $ hbase shell
>>>>>>>> 2014-08-27 06:23:38,707 INFO  [main] Configuration.deprecation:
>>>>>>> hadoop.native.lib is deprecated. Instead, use io.native.lib.available
>>>>>>>> HBase Shell; enter 'help<RETURN>' for list of supported commands.
>>>>>>>> Type "exit<RETURN>" to leave the HBase Shell
>>>>>>>> Version 0.98.4-hadoop2, rUnknown, Sun Aug  3 23:45:36 HKT 2014
>>>>>>>> 
>>>>>>>> hbase(main):001:0>
>>>>>>>> hbase(main):001:0> create 'tsnappy', { NAME => 'f', COMPRESSION =>
>>>>>>> 'snappy'}
>>>>>>>> SLF4J: Class path contains multiple SLF4J bindings.
>>>>>>>> SLF4J: Found binding in
>>>>>>> 
>>>>>> 
>>>> 
>> [jar:file:/edh/hadoop/hbase-0.98.4-hadoop2/lib/slf4j-log4j12-1.6.4.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>>>>> SLF4J: Found binding in
>>>>>>> 
>>>>>> 
>>>> 
>> [jar:file:/edh/hadoop/hadoop-2.4.1/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>>>>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
>>>>>>> explanation.
>>>>>>>> 
>>>>>>>> ERROR: java.io.IOException: Compression algorithm 'snappy'
>> previously
>>>>>>> failed test.
>>>>>>>>    at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.hbase.util.CompressionTest.testCompression(CompressionTest.java:85)
>>>>>>>>    at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.hbase.master.HMaster.checkCompression(HMaster.java:1764)
>>>>>>>>    at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.hbase.master.HMaster.checkCompression(HMaster.java:1757)
>>>>>>>>    at
>>>>>>> org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1739)
>>>>>>>>    at
>>>>>>> org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1774)
>>>>>>>>    at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:40470)
>>>>>>>>    at
>>>>>> org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2027)
>>>>>>>>    at
>> org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:98)
>>>>>>>>    at
>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hadoop.hbase.ipc.FifoRpcScheduler$1.run(FifoRpcScheduler.java:74)
>>>>>>>>    at
>>>>>>> 
>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:439)
>>>>>>>>    at
>>>>>>> java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
>>>>>>>>    at java.util.concurrent.FutureTask.run(FutureTask.java:138)
>>>>>>>>    at
>>>>>>> 
>>>>>> 
>>>> 
>> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:895)
>>>>>>>>    at
>>>>>>> 
>>>>>> 
>>>> 
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918)
>>>>>>>>    at java.lang.Thread.run(Thread.java:662)
>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>>>> Regards
>>>>>>>> Arthur
>>>>>>>> 
>>>>>>>> 
>>>>>>>> On 26 Aug, 2014, at 11:02 pm, Sean Busbey <[email protected]>
>>>> wrote:
>>>>>>>> 
>>>>>>>>> Hi Arthur!
>>>>>>>>> 
>>>>>>>>> Our Snappy build instructions are currently out of date and I'm
>>>>>> working
>>>>>>> on updating them[1]. In short, I don't think there are any special
>>>> build
>>>>>>> steps for using snappy.
>>>>>>>>> 
>>>>>>>>> I'm still working out what needs to be included in our instructions
>>>>>> for
>>>>>>> local and cluster testing.
>>>>>>>>> 
>>>>>>>>> If you use the test for compression options, locally things will
>> fail
>>>>>>> because the native hadoop libs won't be present:
>>>>>>>>> 
>>>>>>>>> bin/hbase org.apache.hadoop.hbase.util.CompressionTest
>>>>>>> file:///tmp/snappy-test snappy
>>>>>>>>> (for comparison, replace "snappy" with "gz" and you will get a
>>>> warning
>>>>>>> about not having native libraries, but the test will succeed.)
>>>>>>>>> 
>>>>>>>>> I believe JM's suggestion is for you to copy the Hadoop native
>>>>>>> libraries into the local HBase lib/native directory, which would
>> allow
>>>>>> the
>>>>>>> local test to pass. If you are running in a deployed Hadoop cluster,
>> I
>>>>>>> would expect the necessary libraries to already be available to
>> HBase.
>>>>>>>>> 
>>>>>>>>> [1]: https://issues.apache.org/jira/browse/HBASE-6189
>>>>>>>>> 
>>>>>>>>> -Sean
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> On Tue, Aug 26, 2014 at 8:30 AM, [email protected] <
>>>>>>> [email protected]> wrote:
>>>>>>>>> Hi JM
>>>>>>>>> 
>>>>>>>>> Below are my commands, tried two cases under same source code
>> folder:
>>>>>>>>> a) compile with snappy parameters(failed),
>>>>>>>>> b) compile without snappy parameters (successful).
>>>>>>>>> 
>>>>>>>>> Regards
>>>>>>>>> Arthur
>>>>>>>>> 
>>>>>>>>> wget
>>>>>>> 
>> http://mirrors.devlib.org/apache/hbase/stable/hbase-0.98.4-src.tar.gz
>>>>>>>>> tar -vxf hbase-0.98.4-src.tar.gz
>>>>>>>>> mv hbase-0.98.4 hbase-0.98.4-src_snappy
>>>>>>>>> cd  hbase-0.98.4-src_snappy
>>>>>>>>> nano dev-support/generate-hadoopX-poms.sh
>>>>>>>>> (change  hbase_home=“/usr/local/hadoop/hbase-0.98.4-src_snappy”)
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> bash -x ./dev-support/generate-hadoopX-poms.sh 0.98.4
>> 0.98.4-hadoop2
>>>>>>>>> a) with snappy parameters
>>>>>>>>> mvn -f pom.xml.hadoop2 install -DskipTests assembly:single
>>>>>>> -Prelease,hadoop-snappy -Dhadoop-snappy.version=0.0.1-SNAPSHOT
>>>>>>>>> [INFO]
>>>>>>> 
>>>> ------------------------------------------------------------------------
>>>>>>>>> [INFO] Building HBase - Server 0.98.4-hadoop2
>>>>>>>>> [INFO]
>>>>>>> 
>>>> ------------------------------------------------------------------------
>>>>>>>>> [WARNING] The POM for
>>>>>>> org.apache.hadoop:hadoop-snappy:jar:0.0.1-SNAPSHOT is missing, no
>>>>>>> dependency information available
>>>>>>>>> [INFO]
>>>>>>> 
>>>> ------------------------------------------------------------------------
>>>>>>>>> [INFO] Reactor Summary:
>>>>>>>>> [INFO]
>>>>>>>>> [INFO] HBase ............................................. SUCCESS
>>>>>>> [8.192s]
>>>>>>>>> [INFO] HBase - Common .................................... SUCCESS
>>>>>>> [5.638s]
>>>>>>>>> [INFO] HBase - Protocol .................................. SUCCESS
>>>>>>> [1.535s]
>>>>>>>>> [INFO] HBase - Client .................................... SUCCESS
>>>>>>> [1.206s]
>>>>>>>>> [INFO] HBase - Hadoop Compatibility ...................... SUCCESS
>>>>>>> [0.193s]
>>>>>>>>> [INFO] HBase - Hadoop Two Compatibility .................. SUCCESS
>>>>>>> [0.798s]
>>>>>>>>> [INFO] HBase - Prefix Tree ............................... SUCCESS
>>>>>>> [0.438s]
>>>>>>>>> [INFO] HBase - Server .................................... FAILURE
>>>>>>> [0.234s]
>>>>>>>>> [INFO] HBase - Testing Util .............................. SKIPPED
>>>>>>>>> [INFO] HBase - Thrift .................................... SKIPPED
>>>>>>>>> [INFO] HBase - Shell ..................................... SKIPPED
>>>>>>>>> [INFO] HBase - Integration Tests ......................... SKIPPED
>>>>>>>>> [INFO] HBase - Examples .................................. SKIPPED
>>>>>>>>> [INFO] HBase - Assembly .................................. SKIPPED
>>>>>>>>> [INFO]
>>>>>>> 
>>>> ------------------------------------------------------------------------
>>>>>>>>> [INFO] BUILD FAILURE
>>>>>>>>> [INFO]
>>>>>>> 
>>>> ------------------------------------------------------------------------
>>>>>>>>> [INFO] Total time: 19.474s
>>>>>>>>> [INFO] Finished at: Tue Aug 26 21:21:13 HKT 2014
>>>>>>>>> [INFO] Final Memory: 51M/1100M
>>>>>>>>> [INFO]
>>>>>>> 
>>>> ------------------------------------------------------------------------
>>>>>>>>> [ERROR] Failed to execute goal on project hbase-server: Could not
>>>>>>> resolve dependencies for project
>>>>>>> org.apache.hbase:hbase-server:jar:0.98.4-hadoop2: Failure to find
>>>>>>> org.apache.hadoop:hadoop-snappy:jar:0.0.1-SNAPSHOT in
>>>>>>> http://maven.oschina.net/content/groups/public/ was cached in the
>>>> local
>>>>>>> repository, resolution will not be reattempted until the update
>>>> interval
>>>>>> of
>>>>>>> nexus-osc has elapsed or updates are forced -> [Help 1]
>>>>>>>>> [ERROR]
>>>>>>>>> [ERROR] To see the full stack trace of the errors, re-run Maven
>> with
>>>>>>> the -e switch.
>>>>>>>>> [ERROR] Re-run Maven using the -X switch to enable full debug
>>>> logging.
>>>>>>>>> [ERROR]
>>>>>>>>> [ERROR] For more information about the errors and possible
>> solutions,
>>>>>>> please read the following articles:
>>>>>>>>> [ERROR] [Help 1]
>>>>>>> 
>>>>>> 
>>>> 
>> http://cwiki.apache.org/confluence/display/MAVEN/DependencyResolutionException
>>>>>>>>> [ERROR]
>>>>>>>>> [ERROR] After correcting the problems, you can resume the build
>> with
>>>>>>> the command
>>>>>>>>> [ERROR]   mvn <goals> -rf :hbase-server
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> b) try again, without snappy parameters
>>>>>>>>> mvn -f pom.xml.hadoop2 install -DskipTests assembly:single
>> -Prelease
>>>>>>>>> [INFO] Building tar:
>>>>>>> 
>>>>>> 
>>>> 
>> /edh/hadoop_all_sources/hbase-0.98.4-src_snappy/hbase-assembly/target/hbase-0.98.4-hadoop2-bin.tar.gz
>>>>>>>>> [INFO]
>>>>>>> 
>>>> ------------------------------------------------------------------------
>>>>>>>>> [INFO] Reactor Summary:
>>>>>>>>> [INFO]
>>>>>>>>> [INFO] HBase ............................................. SUCCESS
>>>>>>> [3.290s]
>>>>>>>>> [INFO] HBase - Common .................................... SUCCESS
>>>>>>> [3.119s]
>>>>>>>>> [INFO] HBase - Protocol .................................. SUCCESS
>>>>>>> [0.972s]
>>>>>>>>> [INFO] HBase - Client .................................... SUCCESS
>>>>>>> [0.920s]
>>>>>>>>> [INFO] HBase - Hadoop Compatibility ...................... SUCCESS
>>>>>>> [0.167s]
>>>>>>>>> [INFO] HBase - Hadoop Two Compatibility .................. SUCCESS
>>>>>>> [0.504s]
>>>>>>>>> [INFO] HBase - Prefix Tree ............................... SUCCESS
>>>>>>> [0.382s]
>>>>>>>>> [INFO] HBase - Server .................................... SUCCESS
>>>>>>> [4.790s]
>>>>>>>>> [INFO] HBase - Testing Util .............................. SUCCESS
>>>>>>> [0.598s]
>>>>>>>>> [INFO] HBase - Thrift .................................... SUCCESS
>>>>>>> [1.536s]
>>>>>>>>> [INFO] HBase - Shell ..................................... SUCCESS
>>>>>>> [0.369s]
>>>>>>>>> [INFO] HBase - Integration Tests ......................... SUCCESS
>>>>>>> [0.443s]
>>>>>>>>> [INFO] HBase - Examples .................................. SUCCESS
>>>>>>> [0.459s]
>>>>>>>>> [INFO] HBase - Assembly .................................. SUCCESS
>>>>>>> [13.240s]
>>>>>>>>> [INFO]
>>>>>>> 
>>>> ------------------------------------------------------------------------
>>>>>>>>> [INFO] BUILD SUCCESS
>>>>>>>>> [INFO]
>>>>>>> 
>>>> ------------------------------------------------------------------------
>>>>>>>>> [INFO] Total time: 31.408s
>>>>>>>>> [INFO] Finished at: Tue Aug 26 21:22:50 HKT 2014
>>>>>>>>> [INFO] Final Memory: 57M/1627M
>>>>>>>>> [INFO]
>>>>>>> 
>>>> ------------------------------------------------------------------------
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> On 26 Aug, 2014, at 8:52 pm, Jean-Marc Spaggiari <
>>>>>>> [email protected]> wrote:
>>>>>>>>> 
>>>>>>>>>> Hi Arthur,
>>>>>>>>>> 
>>>>>>>>>> How have you extracted HBase source and what command do you run to
>>>>>>> build? I
>>>>>>>>>> will do the same here locally so I can provide you the exact step
>> to
>>>>>>>>>> complete.
>>>>>>>>>> 
>>>>>>>>>> JM
>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>>> 2014-08-26 8:42 GMT-04:00 [email protected] <
>>>>>>> [email protected]
>>>>>>>>>>> :
>>>>>>>>>> 
>>>>>>>>>>> Hi JM
>>>>>>>>>>> 
>>>>>>>>>>> Not too sure what you mean, do you mean I should create a new
>>>>>> folder
>>>>>>> in my
>>>>>>>>>>> HBASE_SRC named lib/native/Linux-x86 and copy these files to this
>>>>>>> folder
>>>>>>>>>>> then try to compile it again?
>>>>>>>>>>> 
>>>>>>>>>>> Regards
>>>>>>>>>>> ARthur
>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>>> On 26 Aug, 2014, at 8:17 pm, Jean-Marc Spaggiari <
>>>>>>> [email protected]>
>>>>>>>>>>> wrote:
>>>>>>>>>>> 
>>>>>>>>>>>> Hi Arthur,
>>>>>>>>>>>> 
>>>>>>>>>>>> Almost done! You now need to copy them on the HBase folder.
>>>>>>>>>>>> 
>>>>>>>>>>>> hbase@hbasetest1:~/hbase-0.98.2-hadoop2/lib$ tree | grep -v
>> .jar
>>>>>> |
>>>>>>> grep
>>>>>>>>>>> -v
>>>>>>>>>>>> .rb
>>>>>>>>>>>> .
>>>>>>>>>>>> ├── native
>>>>>>>>>>>> │   └── Linux-x86
>>>>>>>>>>>> │       ├── libsnappy.a
>>>>>>>>>>>> │       ├── libsnappy.la
>>>>>>>>>>>> │       ├── libsnappy.so
>>>>>>>>>>>> │       ├── libsnappy.so.1
>>>>>>>>>>>> │       └── libsnappy.so.1.2.0
>>>>>>>>>>>> 
>>>>>>>>>>>> I don't have any hadoop-snappy lib in my hbase folder and it
>> works
>>>>>>> very
>>>>>>>>>>>> well with Snappy for me...
>>>>>>>>>>>> 
>>>>>>>>>>>> JM
>>>>>>>>>>>> 
>>>>>>>>>>>> 2014-08-26 8:09 GMT-04:00 [email protected] <
>>>>>>>>>>> [email protected]
>>>>>>>>>>>>> :
>>>>>>>>>>>> 
>>>>>>>>>>>>> Hi JM,
>>>>>>>>>>>>> 
>>>>>>>>>>>>> Below are my steps to install snappy lib, do I miss something?
>>>>>>>>>>>>> 
>>>>>>>>>>>>> Regards
>>>>>>>>>>>>> Arthur
>>>>>>>>>>>>> 
>>>>>>>>>>>>> wget https://snappy.googlecode.com/files/snappy-1.1.1.tar.gz
>>>>>>>>>>>>> tar -vxf snappy-1.1.1.tar.gz
>>>>>>>>>>>>> cd snappy-1.1.1
>>>>>>>>>>>>> ./configure
>>>>>>>>>>>>> make
>>>>>>>>>>>>> make install
>>>>>>>>>>>>>    make[1]: Entering directory
>>>>>>>>>>> `/edh/hadoop_all_sources/snappy-1.1.1'
>>>>>>>>>>>>>    test -z "/usr/local/lib" || /bin/mkdir -p "/usr/local/lib"
>>>>>>>>>>>>>     /bin/sh ./libtool   --mode=install /usr/bin/install -c
>>>>>>>>>>>>> libsnappy.la '/usr/local/lib'
>>>>>>>>>>>>>    libtool: install: /usr/bin/install -c
>>>>>>> .libs/libsnappy.so.1.2.0
>>>>>>>>>>>>> /usr/local/lib/libsnappy.so.1.2.0
>>>>>>>>>>>>>    libtool: install: (cd /usr/local/lib && { ln -s -f
>>>>>>>>>>>>> libsnappy.so.1.2.0 libsnappy.so.1 || { rm -f libsnappy.so.1 &&
>> ln
>>>>>>> -s
>>>>>>>>>>>>> libsnappy.so.1.2.0 libsnappy.so.1; }; })
>>>>>>>>>>>>>    libtool: install: (cd /usr/local/lib && { ln -s -f
>>>>>>>>>>>>> libsnappy.so.1.2.0 libsnappy.so || { rm -f libsnappy.so && ln
>> -s
>>>>>>>>>>>>> libsnappy.so.1.2.0 libsnappy.so; }; })
>>>>>>>>>>>>>    libtool: install: /usr/bin/install -c .libs/libsnappy.lai
>>>>>>>>>>>>> /usr/local/lib/libsnappy.la
>>>>>>>>>>>>>    libtool: install: /usr/bin/install -c .libs/libsnappy.a
>>>>>>>>>>>>> /usr/local/lib/libsnappy.a
>>>>>>>>>>>>>    libtool: install: chmod 644 /usr/local/lib/libsnappy.a
>>>>>>>>>>>>>    libtool: install: ranlib /usr/local/lib/libsnappy.a
>>>>>>>>>>>>>    libtool: finish:
>>>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>> 
>>>>>> 
>>>> 
>> PATH="/edh/hadoop/spark/bin:/edh/hadoop/hbase/bin:/edh/hadoop/zookeeper//bin:/edh/hadoop/yarn/hadoop/bin:/edh/hadoop/yarn/hadoop/sbin:/usr/lib64/qt-3.3/bin:/opt/apache-maven-3.1.1/bin:/usr/local/sbin:/usr/local/bin:/sbin:/bin:/usr/sbin:/usr/bin:/edh/hadoop/zookeeper//bin:/edh/hadoop/hive//bin:/usr/lib/jvm/jdk1.6.0_45//bin:/root/bin:/sbin"
>>>>>>>>>>>>> ldconfig -n /usr/local/lib
>>>>>>>>>>>>> 
>>>>>>>>>>>>> 
>>>>>>> 
>> ----------------------------------------------------------------------
>>>>>>>>>>>>>    Libraries have been installed in:
>>>>>>>>>>>>>    /usr/local/lib
>>>>>>>>>>>>>    If you ever happen to want to link against installed
>>>>>>> libraries
>>>>>>>>>>>>>    in a given directory, LIBDIR, you must either use libtool,
>>>>>>> and
>>>>>>>>>>>>>    specify the full pathname of the library, or use the
>>>>>>> `-LLIBDIR'
>>>>>>>>>>>>>    flag during linking and do at least one of the following:
>>>>>>>>>>>>>    - add LIBDIR to the `LD_LIBRARY_PATH' environment variable
>>>>>>>>>>>>>    during execution
>>>>>>>>>>>>>    - add LIBDIR to the `LD_RUN_PATH' environment variable
>>>>>>>>>>>>>    during linking
>>>>>>>>>>>>>    - use the `-Wl,-rpath -Wl,LIBDIR' linker flag
>>>>>>>>>>>>>    - have your system administrator add LIBDIR to
>>>>>>> `/etc/ld.so.conf'
>>>>>>>>>>>>>    See any operating system documentation about shared
>>>>>>> libraries for
>>>>>>>>>>>>>    more information, such as the ld(1) and ld.so(8) manual
>>>>>>> pages.
>>>>>>>>>>>>> 
>>>>>>>>>>>>> 
>>>>>>> 
>> ----------------------------------------------------------------------
>>>>>>>>>>>>>    test -z "/usr/local/share/doc/snappy" || /bin/mkdir -p
>>>>>>>>>>>>> "/usr/local/share/doc/snappy"
>>>>>>>>>>>>>     /usr/bin/install -c -m 644 ChangeLog COPYING INSTALL NEWS
>>>>>>> README
>>>>>>>>>>>>> format_description.txt framing_format.txt
>>>>>>> '/usr/local/share/doc/snappy'
>>>>>>>>>>>>>    test -z "/usr/local/include" || /bin/mkdir -p
>>>>>>>>>>> "/usr/local/include"
>>>>>>>>>>>>>     /usr/bin/install -c -m 644 snappy.h snappy-sinksource.h
>>>>>>>>>>>>> snappy-stubs-public.h snappy-c.h '/usr/local/include'
>>>>>>>>>>>>>    make[1]: Leaving directory
>>>>>>> `/edh/hadoop_all_sources/snappy-1.1.1'
>>>>>>>>>>>>> 
>>>>>>>>>>>>> ll /usr/local/lib
>>>>>>>>>>>>>    -rw-r--r--. 1 root root   233554 Aug 20 00:14 libsnappy.a
>>>>>>>>>>>>>    -rwxr-xr-x. 1 root root      953 Aug 20 00:14 libsnappy.la
>>>>>>>>>>>>>    lrwxrwxrwx. 1 root root       18 Aug 20 00:14 libsnappy.so
>>>>>> ->
>>>>>>>>>>>>> libsnappy.so.1.2.0
>>>>>>>>>>>>>    lrwxrwxrwx. 1 root root       18 Aug 20 00:14
>>>>>> libsnappy.so.1
>>>>>>> ->
>>>>>>>>>>>>> libsnappy.so.1.2.0
>>>>>>>>>>>>>    -rwxr-xr-x. 1 root root   147726 Aug 20 00:14
>>>>>>> libsnappy.so.1.2.0
>>>>>>>>>>>>> 
>>>>>>>>>>>>> 
>>>>>>>>>>>>> 
>>>>>>>>>>>>> On 26 Aug, 2014, at 7:38 pm, Jean-Marc Spaggiari <
>>>>>>>>>>> [email protected]>
>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Hi Arthur,
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Do you have snappy libs installed and configured? HBase
>> doesn't
>>>>>>> come
>>>>>>>>>>> with
>>>>>>>>>>>>>> Snappy. So yo need to have it first.
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Shameless plug:
>>>>>>>>>>>>>> 
>>>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>> 
>>>>>> 
>>>> 
>> http://www.spaggiari.org/index.php/hbase/how-to-install-snappy-with-1#.U_xxSqdZuZY
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> This is for 0.96 but should be very similar for 0.98. I will
>> try
>>>>>>> it
>>>>>>>>>>> soon
>>>>>>>>>>>>>> and post and update, but keep us posted here so we can support
>>>>>>> you...
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> JM
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> 2014-08-26 7:34 GMT-04:00 [email protected] <
>>>>>>>>>>>>> [email protected]
>>>>>>>>>>>>>>> :
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Hi,
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> I need to install snappy to HBase 0.98.4.  (my Hadoop version
>>>>>> is
>>>>>>>>>>> 2.4.1)
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Can you please advise what would be wrong?  Should my pom.xml
>>>>>> be
>>>>>>>>>>>>> incorrect
>>>>>>>>>>>>>>> and missing something?
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Regards
>>>>>>>>>>>>>>> Arthur
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Below are my commands:
>>>>>>>>>>>>>>> bash -x ./dev-support/generate-hadoopX-poms.sh 0.98.4
>>>>>>> 0.98.4-hadoop2
>>>>>>>>>>>>>>> mvn -f pom.xml.hadoop2 install -DskipTests assembly:single
>>>>>>>>>>>>>>> -Prelease,hadoop-snappy
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Iog:
>>>>>>>>>>>>>>> [INFO]
>>>>>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>> 
>>>> ------------------------------------------------------------------------
>>>>>>>>>>>>>>> [INFO] Building HBase - Server 0.98.4-hadoop2
>>>>>>>>>>>>>>> [INFO]
>>>>>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>> 
>>>> ------------------------------------------------------------------------
>>>>>>>>>>>>>>> [WARNING] The POM for
>>>>>>>>>>> org.apache.hadoop:hadoop-snappy:jar:0.0.1-SNAPSHOT
>>>>>>>>>>>>>>> is missing, no dependency information available
>>>>>>>>>>>>>>> [INFO]
>>>>>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>> 
>>>> ------------------------------------------------------------------------
>>>>>>>>>>>>>>> [INFO] Reactor Summary:
>>>>>>>>>>>>>>> [INFO]
>>>>>>>>>>>>>>> [INFO] HBase .............................................
>>>>>>> SUCCESS
>>>>>>>>>>>>> [3.129s]
>>>>>>>>>>>>>>> [INFO] HBase - Common ....................................
>>>>>>> SUCCESS
>>>>>>>>>>>>> [3.105s]
>>>>>>>>>>>>>>> [INFO] HBase - Protocol ..................................
>>>>>>> SUCCESS
>>>>>>>>>>>>> [0.976s]
>>>>>>>>>>>>>>> [INFO] HBase - Client ....................................
>>>>>>> SUCCESS
>>>>>>>>>>>>> [0.925s]
>>>>>>>>>>>>>>> [INFO] HBase - Hadoop Compatibility ......................
>>>>>>> SUCCESS
>>>>>>>>>>>>> [0.183s]
>>>>>>>>>>>>>>> [INFO] HBase - Hadoop Two Compatibility ..................
>>>>>>> SUCCESS
>>>>>>>>>>>>> [0.497s]
>>>>>>>>>>>>>>> [INFO] HBase - Prefix Tree ...............................
>>>>>>> SUCCESS
>>>>>>>>>>>>> [0.407s]
>>>>>>>>>>>>>>> [INFO] HBase - Server ....................................
>>>>>>> FAILURE
>>>>>>>>>>>>> [0.103s]
>>>>>>>>>>>>>>> [INFO] HBase - Testing Util ..............................
>>>>>>> SKIPPED
>>>>>>>>>>>>>>> [INFO] HBase - Thrift ....................................
>>>>>>> SKIPPED
>>>>>>>>>>>>>>> [INFO] HBase - Shell .....................................
>>>>>>> SKIPPED
>>>>>>>>>>>>>>> [INFO] HBase - Integration Tests .........................
>>>>>>> SKIPPED
>>>>>>>>>>>>>>> [INFO] HBase - Examples ..................................
>>>>>>> SKIPPED
>>>>>>>>>>>>>>> [INFO] HBase - Assembly ..................................
>>>>>>> SKIPPED
>>>>>>>>>>>>>>> [INFO]
>>>>>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>> 
>>>> ------------------------------------------------------------------------
>>>>>>>>>>>>>>> [INFO] BUILD FAILURE
>>>>>>>>>>>>>>> [INFO]
>>>>>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>> 
>>>> ------------------------------------------------------------------------
>>>>>>>>>>>>>>> [INFO] Total time: 9.939s
>>>>>>>>>>>>>>> [INFO] Finished at: Tue Aug 26 19:23:14 HKT 2014
>>>>>>>>>>>>>>> [INFO] Final Memory: 61M/2921M
>>>>>>>>>>>>>>> [INFO]
>>>>>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>> 
>>>> ------------------------------------------------------------------------
>>>>>>>>>>>>>>> [ERROR] Failed to execute goal on project hbase-server: Could
>>>>>> not
>>>>>>>>>>>>> resolve
>>>>>>>>>>>>>>> dependencies for project
>>>>>>>>>>>>> org.apache.hbase:hbase-server:jar:0.98.4-hadoop2:
>>>>>>>>>>>>>>> Failure to find
>>>>>>> org.apache.hadoop:hadoop-snappy:jar:0.0.1-SNAPSHOT in
>>>>>>>>>>>>>>> http://maven.oschina.net/content/groups/public/ was cached
>> in
>>>>>>> the
>>>>>>>>>>> local
>>>>>>>>>>>>>>> repository, resolution will not be reattempted until the
>> update
>>>>>>>>>>>>> interval of
>>>>>>>>>>>>>>> nexus-osc has elapsed or updates are forced -> [Help 1]
>>>>>>>>>>>>>>> [ERROR]
>>>>>>>>>>>>>>> [ERROR] To see the full stack trace of the errors, re-run
>> Maven
>>>>>>> with
>>>>>>>>>>> the
>>>>>>>>>>>>>>> -e switch.
>>>>>>>>>>>>>>> [ERROR] Re-run Maven using the -X switch to enable full debug
>>>>>>> logging.
>>>>>>>>>>>>>>> [ERROR]
>>>>>>>>>>>>>>> [ERROR] For more information about the errors and possible
>>>>>>> solutions,
>>>>>>>>>>>>>>> please read the following articles:
>>>>>>>>>>>>>>> [ERROR] [Help 1]
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>> 
>>>>>> 
>>>> 
>> http://cwiki.apache.org/confluence/display/MAVEN/DependencyResolutionException
>>>>>>>>>>>>>>> [ERROR]
>>>>>>>>>>>>>>> [ERROR] After correcting the problems, you can resume the
>> build
>>>>>>> with
>>>>>>>>>>> the
>>>>>>>>>>>>>>> command
>>>>>>>>>>>>>>> [ERROR]   mvn <goals> -rf :hbase-server
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>> 
>>>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> --
>>>>>>>>> Sean
>>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>> 
>>>> 
>>>> 
>> 
>> 

Reply via email to