Something like this worked for me
1. get hbase binaries
2. sudo yum install snappy snappy-devel
3. ln -sf /usr/lib64/libsnappy.so
/var/lib/hadoop/lib/native/Linux-amd64-64/.
4. ln -sf /usr/lib64/libsnappy.so /var/lib/hbase/lib/native/Linux-amd64-64/.
5. add snappy jar under $HADOOP_HOME/lib and $HBASE_HOME/lib
ref: https://issues.apache.org/jira/browse/PHOENIX-877


On Tue, Aug 26, 2014 at 7:25 PM, [email protected] <
[email protected]> wrote:

> Hi,
>
> I just tried three more steps but was not able to get thru.
>
>
> 1) copied  snappy files to $HBASE_HOME/lib
> $ cd $HBASE_HOME
> $ ll lib/*sna*
> -rw-r--r--. 1 hduser hadoop  11526 Aug 27 06:54
> lib/hadoop-snappy-0.0.1-SNAPSHOT.jar
> -rw-rw-r--. 1 hduser hadoop 995968 Aug  3 18:43 lib/snappy-java-1.0.4.1.jar
>
> ll lib/native/
> drwxrwxr-x. 4 hduser hadoop 4096 Aug 27 06:54 Linux-amd64-64
>
> ll lib/native/Linux-amd64-64/
> total 18964
> lrwxrwxrwx. 1 hduser Hadoop      24 Aug 27 07:08 libhadoopsnappy.so ->
> libhadoopsnappy.so.0.0.1
> lrwxrwxrwx. 1 hduser Hadoop      24 Aug 27 07:08 libhadoopsnappy.so.0 ->
> libhadoopsnappy.so.0.0.1
> -rwxr-xr-x. 1 hduser Hadoop   54961 Aug 27 07:08 libhadoopsnappy.so.0.0.1
> lrwxrwxrwx. 1 hduser Hadoop      55 Aug 27 07:08 libjvm.so ->
> /usr/lib/jvm/jdk1.6.0_45/jre/lib/amd64/server/libjvm.so
> lrwxrwxrwx. 1 hduser Hadoop      25 Aug 27 07:08 libprotobuf-lite.so ->
> libprotobuf-lite.so.8.0.0
> lrwxrwxrwx. 1 hduser Hadoop      25 Aug 27 07:08 libprotobuf-lite.so.8 ->
> libprotobuf-lite.so.8.0.0
> -rwxr-xr-x. 1 hduser Hadoop  964689 Aug 27 07:08 libprotobuf-lite.so.8.0.0
> lrwxrwxrwx. 1 hduser Hadoop      20 Aug 27 07:08 libprotobuf.so ->
> libprotobuf.so.8.0.0
> lrwxrwxrwx. 1 hduser Hadoop      20 Aug 27 07:08 libprotobuf.so.8 ->
> libprotobuf.so.8.0.0
> -rwxr-xr-x. 1 hduser Hadoop 8300050 Aug 27 07:08 libprotobuf.so.8.0.0
> lrwxrwxrwx. 1 hduser Hadoop      18 Aug 27 07:08 libprotoc.so ->
> libprotoc.so.8.0.0
> lrwxrwxrwx. 1 hduser Hadoop      18 Aug 27 07:08 libprotoc.so.8 ->
> libprotoc.so.8.0.0
> -rwxr-xr-x. 1 hduser Hadoop 9935810 Aug 27 07:08 libprotoc.so.8.0.0
> lrwxrwxrwx. 1 hduser Hadoop      18 Aug 27 07:08 libsnappy.so ->
> libsnappy.so.1.2.0
> lrwxrwxrwx. 1 hduser Hadoop      18 Aug 27 07:08 libsnappy.so.1 ->
> libsnappy.so.1.2.0
> -rwxr-xr-x. 1 hduser Hadoop  147726 Aug 27 07:08 libsnappy.so.1.2.0
> drwxr-xr-x. 2 hduser Hadoop    4096 Aug 27 07:08 pkgconfig
>
> 2)  $HBASE_HOME/conf/hbase-env.sh, added
>
> ###
> export
> LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$HADOOP_HOME/lib/native/Linux-amd64-64/:/usr/local/lib/
> export
> HBASE_LIBRARY_PATH=$HBASE_LIBRARY_PATH:$HBASE_HOME/lib/native/Linux-amd64-64/:/usr/local/lib/:$HBASE_HOME/lib/hadoop-snappy-0.0.1-SNAPSHOT.jar
> export CLASSPATH=$CLASSPATH:$HBASE_LIBRARY_PATH
> export HBASE_CLASSPATH=$HBASE_CLASSPATH:$HBASE_LIBRARY_PATH
> ###
>
> 3) restart HBASE and tried again
> $ bin/hbase org.apache.hadoop.hbase.util.CompressionTest
> file:///tmp/snappy-test snappy
> 2014-08-27 07:16:09,490 INFO  [main] Configuration.deprecation:
> hadoop.native.lib is deprecated. Instead, use io.native.lib.available
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in
> [jar:file:/edh/hadoop/hbase-0.98.4-hadoop2/lib/slf4j-log4j12-1.6.4.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in
> [jar:file:/edh/hadoop/hadoop-2.4.1/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
> explanation.
> 2014-08-27 07:16:10,323 INFO  [main] util.ChecksumType: Checksum using
> org.apache.hadoop.util.PureJavaCrc32
> 2014-08-27 07:16:10,324 INFO  [main] util.ChecksumType: Checksum can use
> org.apache.hadoop.util.PureJavaCrc32C
> Exception in thread "main" java.lang.RuntimeException: native snappy
> library not available: this version of libhadoop was built without snappy
> support.
>         at
> org.apache.hadoop.io.compress.SnappyCodec.checkNativeCodeLoaded(SnappyCodec.java:64)
>         at
> org.apache.hadoop.io.compress.SnappyCodec.getCompressorType(SnappyCodec.java:132)
>         at
> org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:148)
>         at
> org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:163)
>         at
> org.apache.hadoop.hbase.io.compress.Compression$Algorithm.getCompressor(Compression.java:310)
>         at
> org.apache.hadoop.hbase.io.encoding.HFileBlockDefaultEncodingContext.<init>(HFileBlockDefaultEncodingContext.java:92)
>         at
> org.apache.hadoop.hbase.io.hfile.HFileBlock$Writer.<init>(HFileBlock.java:690)
>         at
> org.apache.hadoop.hbase.io.hfile.HFileWriterV2.finishInit(HFileWriterV2.java:117)
>         at
> org.apache.hadoop.hbase.io.hfile.HFileWriterV2.<init>(HFileWriterV2.java:109)
>         at
> org.apache.hadoop.hbase.io.hfile.HFileWriterV2$WriterFactoryV2.createWriter(HFileWriterV2.java:97)
>         at
> org.apache.hadoop.hbase.io.hfile.HFile$WriterFactory.create(HFile.java:393)
>         at
> org.apache.hadoop.hbase.util.CompressionTest.doSmokeTest(CompressionTest.java:118)
>         at
> org.apache.hadoop.hbase.util.CompressionTest.main(CompressionTest.java:148)
>
>
> Regards
> Arthur
>
>
>
> On 27 Aug, 2014, at 6:27 am, [email protected] <
> [email protected]> wrote:
>
> > Hi Sean,
> >
> > Thanks for your reply.
> >
> > I tried the following tests
> >
> > $ bin/hbase org.apache.hadoop.hbase.util.CompressionTest
> file:///tmp/snappy-test gz
> > 2014-08-26 23:06:17,778 INFO  [main] Configuration.deprecation:
> hadoop.native.lib is deprecated. Instead, use io.native.lib.available
> > SLF4J: Class path contains multiple SLF4J bindings.
> > SLF4J: Found binding in
> [jar:file:/edh/hadoop/hbase-0.98.4-hadoop2/lib/slf4j-log4j12-1.6.4.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> > SLF4J: Found binding in
> [jar:file:/edh/hadoop/hadoop-2.4.1/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
> explanation.
> > 2014-08-26 23:06:18,103 INFO  [main] util.ChecksumType: Checksum using
> org.apache.hadoop.util.PureJavaCrc32
> > 2014-08-26 23:06:18,104 INFO  [main] util.ChecksumType: Checksum can use
> org.apache.hadoop.util.PureJavaCrc32C
> > 2014-08-26 23:06:18,260 INFO  [main] zlib.ZlibFactory: Successfully
> loaded & initialized native-zlib library
> > 2014-08-26 23:06:18,276 INFO  [main] compress.CodecPool: Got brand-new
> compressor [.gz]
> > 2014-08-26 23:06:18,280 INFO  [main] compress.CodecPool: Got brand-new
> compressor [.gz]
> > 2014-08-26 23:06:18,921 INFO  [main] compress.CodecPool: Got brand-new
> decompressor [.gz]
> > SUCCESS
> >
> >
> > $ bin/hbase org.apache.hadoop.hbase.util.CompressionTest
> file:///tmp/snappy-test snappy
> > 2014-08-26 23:07:08,246 INFO  [main] Configuration.deprecation:
> hadoop.native.lib is deprecated. Instead, use io.native.lib.available
> > SLF4J: Class path contains multiple SLF4J bindings.
> > SLF4J: Found binding in
> [jar:file:/edh/hadoop/hbase-0.98.4-hadoop2/lib/slf4j-log4j12-1.6.4.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> > SLF4J: Found binding in
> [jar:file:/edh/hadoop/hadoop-2.4.1/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
> explanation.
> > 2014-08-26 23:07:08,578 INFO  [main] util.ChecksumType: Checksum using
> org.apache.hadoop.util.PureJavaCrc32
> > 2014-08-26 23:07:08,579 INFO  [main] util.ChecksumType: Checksum can use
> org.apache.hadoop.util.PureJavaCrc32C
> > Exception in thread "main" java.lang.RuntimeException: native snappy
> library not available: this version of libhadoop was built without snappy
> support.
> >       at
> org.apache.hadoop.io.compress.SnappyCodec.checkNativeCodeLoaded(SnappyCodec.java:64)
> >       at
> org.apache.hadoop.io.compress.SnappyCodec.getCompressorType(SnappyCodec.java:132)
> >       at
> org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:148)
> >       at
> org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:163)
> >       at
> org.apache.hadoop.hbase.io.compress.Compression$Algorithm.getCompressor(Compression.java:310)
> >       at
> org.apache.hadoop.hbase.io.encoding.HFileBlockDefaultEncodingContext.<init>(HFileBlockDefaultEncodingContext.java:92)
> >       at
> org.apache.hadoop.hbase.io.hfile.HFileBlock$Writer.<init>(HFileBlock.java:690)
> >       at
> org.apache.hadoop.hbase.io.hfile.HFileWriterV2.finishInit(HFileWriterV2.java:117)
> >       at
> org.apache.hadoop.hbase.io.hfile.HFileWriterV2.<init>(HFileWriterV2.java:109)
> >       at
> org.apache.hadoop.hbase.io.hfile.HFileWriterV2$WriterFactoryV2.createWriter(HFileWriterV2.java:97)
> >       at
> org.apache.hadoop.hbase.io.hfile.HFile$WriterFactory.create(HFile.java:393)
> >       at
> org.apache.hadoop.hbase.util.CompressionTest.doSmokeTest(CompressionTest.java:118)
> >       at
> org.apache.hadoop.hbase.util.CompressionTest.main(CompressionTest.java:148)
> >
> >
> > $ hbase shell
> > 2014-08-27 06:23:38,707 INFO  [main] Configuration.deprecation:
> hadoop.native.lib is deprecated. Instead, use io.native.lib.available
> > HBase Shell; enter 'help<RETURN>' for list of supported commands.
> > Type "exit<RETURN>" to leave the HBase Shell
> > Version 0.98.4-hadoop2, rUnknown, Sun Aug  3 23:45:36 HKT 2014
> >
> > hbase(main):001:0>
> > hbase(main):001:0> create 'tsnappy', { NAME => 'f', COMPRESSION =>
> 'snappy'}
> > SLF4J: Class path contains multiple SLF4J bindings.
> > SLF4J: Found binding in
> [jar:file:/edh/hadoop/hbase-0.98.4-hadoop2/lib/slf4j-log4j12-1.6.4.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> > SLF4J: Found binding in
> [jar:file:/edh/hadoop/hadoop-2.4.1/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
> explanation.
> >
> > ERROR: java.io.IOException: Compression algorithm 'snappy' previously
> failed test.
> >       at
> org.apache.hadoop.hbase.util.CompressionTest.testCompression(CompressionTest.java:85)
> >       at
> org.apache.hadoop.hbase.master.HMaster.checkCompression(HMaster.java:1764)
> >       at
> org.apache.hadoop.hbase.master.HMaster.checkCompression(HMaster.java:1757)
> >       at
> org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1739)
> >       at
> org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1774)
> >       at
> org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:40470)
> >       at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2027)
> >       at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:98)
> >       at
> org.apache.hadoop.hbase.ipc.FifoRpcScheduler$1.run(FifoRpcScheduler.java:74)
> >       at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:439)
> >       at
> java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
> >       at java.util.concurrent.FutureTask.run(FutureTask.java:138)
> >       at
> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:895)
> >       at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918)
> >       at java.lang.Thread.run(Thread.java:662)
> >
> >
> >
> >
> > Regards
> > Arthur
> >
> >
> > On 26 Aug, 2014, at 11:02 pm, Sean Busbey <[email protected]> wrote:
> >
> >> Hi Arthur!
> >>
> >> Our Snappy build instructions are currently out of date and I'm working
> on updating them[1]. In short, I don't think there are any special build
> steps for using snappy.
> >>
> >> I'm still working out what needs to be included in our instructions for
> local and cluster testing.
> >>
> >> If you use the test for compression options, locally things will fail
> because the native hadoop libs won't be present:
> >>
> >> bin/hbase org.apache.hadoop.hbase.util.CompressionTest
> file:///tmp/snappy-test snappy
> >> (for comparison, replace "snappy" with "gz" and you will get a warning
> about not having native libraries, but the test will succeed.)
> >>
> >> I believe JM's suggestion is for you to copy the Hadoop native
> libraries into the local HBase lib/native directory, which would allow the
> local test to pass. If you are running in a deployed Hadoop cluster, I
> would expect the necessary libraries to already be available to HBase.
> >>
> >> [1]: https://issues.apache.org/jira/browse/HBASE-6189
> >>
> >> -Sean
> >>
> >>
> >> On Tue, Aug 26, 2014 at 8:30 AM, [email protected] <
> [email protected]> wrote:
> >> Hi JM
> >>
> >> Below are my commands, tried two cases under same source code folder:
> >> a) compile with snappy parameters(failed),
> >> b) compile without snappy parameters (successful).
> >>
> >> Regards
> >> Arthur
> >>
> >> wget
> http://mirrors.devlib.org/apache/hbase/stable/hbase-0.98.4-src.tar.gz
> >> tar -vxf hbase-0.98.4-src.tar.gz
> >> mv hbase-0.98.4 hbase-0.98.4-src_snappy
> >> cd  hbase-0.98.4-src_snappy
> >> nano dev-support/generate-hadoopX-poms.sh
> >>   (change  hbase_home=“/usr/local/hadoop/hbase-0.98.4-src_snappy”)
> >>
> >>
> >> bash -x ./dev-support/generate-hadoopX-poms.sh 0.98.4 0.98.4-hadoop2
> >> a) with snappy parameters
> >> mvn -f pom.xml.hadoop2 install -DskipTests assembly:single
> -Prelease,hadoop-snappy -Dhadoop-snappy.version=0.0.1-SNAPSHOT
> >> [INFO]
> ------------------------------------------------------------------------
> >> [INFO] Building HBase - Server 0.98.4-hadoop2
> >> [INFO]
> ------------------------------------------------------------------------
> >> [WARNING] The POM for
> org.apache.hadoop:hadoop-snappy:jar:0.0.1-SNAPSHOT is missing, no
> dependency information available
> >> [INFO]
> ------------------------------------------------------------------------
> >> [INFO] Reactor Summary:
> >> [INFO]
> >> [INFO] HBase ............................................. SUCCESS
> [8.192s]
> >> [INFO] HBase - Common .................................... SUCCESS
> [5.638s]
> >> [INFO] HBase - Protocol .................................. SUCCESS
> [1.535s]
> >> [INFO] HBase - Client .................................... SUCCESS
> [1.206s]
> >> [INFO] HBase - Hadoop Compatibility ...................... SUCCESS
> [0.193s]
> >> [INFO] HBase - Hadoop Two Compatibility .................. SUCCESS
> [0.798s]
> >> [INFO] HBase - Prefix Tree ............................... SUCCESS
> [0.438s]
> >> [INFO] HBase - Server .................................... FAILURE
> [0.234s]
> >> [INFO] HBase - Testing Util .............................. SKIPPED
> >> [INFO] HBase - Thrift .................................... SKIPPED
> >> [INFO] HBase - Shell ..................................... SKIPPED
> >> [INFO] HBase - Integration Tests ......................... SKIPPED
> >> [INFO] HBase - Examples .................................. SKIPPED
> >> [INFO] HBase - Assembly .................................. SKIPPED
> >> [INFO]
> ------------------------------------------------------------------------
> >> [INFO] BUILD FAILURE
> >> [INFO]
> ------------------------------------------------------------------------
> >> [INFO] Total time: 19.474s
> >> [INFO] Finished at: Tue Aug 26 21:21:13 HKT 2014
> >> [INFO] Final Memory: 51M/1100M
> >> [INFO]
> ------------------------------------------------------------------------
> >> [ERROR] Failed to execute goal on project hbase-server: Could not
> resolve dependencies for project
> org.apache.hbase:hbase-server:jar:0.98.4-hadoop2: Failure to find
> org.apache.hadoop:hadoop-snappy:jar:0.0.1-SNAPSHOT in
> http://maven.oschina.net/content/groups/public/ was cached in the local
> repository, resolution will not be reattempted until the update interval of
> nexus-osc has elapsed or updates are forced -> [Help 1]
> >> [ERROR]
> >> [ERROR] To see the full stack trace of the errors, re-run Maven with
> the -e switch.
> >> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
> >> [ERROR]
> >> [ERROR] For more information about the errors and possible solutions,
> please read the following articles:
> >> [ERROR] [Help 1]
> http://cwiki.apache.org/confluence/display/MAVEN/DependencyResolutionException
> >> [ERROR]
> >> [ERROR] After correcting the problems, you can resume the build with
> the command
> >> [ERROR]   mvn <goals> -rf :hbase-server
> >>
> >>
> >>
> >>
> >> b) try again, without snappy parameters
> >> mvn -f pom.xml.hadoop2 install -DskipTests assembly:single -Prelease
> >> [INFO] Building tar:
> /edh/hadoop_all_sources/hbase-0.98.4-src_snappy/hbase-assembly/target/hbase-0.98.4-hadoop2-bin.tar.gz
> >> [INFO]
> ------------------------------------------------------------------------
> >> [INFO] Reactor Summary:
> >> [INFO]
> >> [INFO] HBase ............................................. SUCCESS
> [3.290s]
> >> [INFO] HBase - Common .................................... SUCCESS
> [3.119s]
> >> [INFO] HBase - Protocol .................................. SUCCESS
> [0.972s]
> >> [INFO] HBase - Client .................................... SUCCESS
> [0.920s]
> >> [INFO] HBase - Hadoop Compatibility ...................... SUCCESS
> [0.167s]
> >> [INFO] HBase - Hadoop Two Compatibility .................. SUCCESS
> [0.504s]
> >> [INFO] HBase - Prefix Tree ............................... SUCCESS
> [0.382s]
> >> [INFO] HBase - Server .................................... SUCCESS
> [4.790s]
> >> [INFO] HBase - Testing Util .............................. SUCCESS
> [0.598s]
> >> [INFO] HBase - Thrift .................................... SUCCESS
> [1.536s]
> >> [INFO] HBase - Shell ..................................... SUCCESS
> [0.369s]
> >> [INFO] HBase - Integration Tests ......................... SUCCESS
> [0.443s]
> >> [INFO] HBase - Examples .................................. SUCCESS
> [0.459s]
> >> [INFO] HBase - Assembly .................................. SUCCESS
> [13.240s]
> >> [INFO]
> ------------------------------------------------------------------------
> >> [INFO] BUILD SUCCESS
> >> [INFO]
> ------------------------------------------------------------------------
> >> [INFO] Total time: 31.408s
> >> [INFO] Finished at: Tue Aug 26 21:22:50 HKT 2014
> >> [INFO] Final Memory: 57M/1627M
> >> [INFO]
> ------------------------------------------------------------------------
> >>
> >>
> >>
> >>
> >>
> >> On 26 Aug, 2014, at 8:52 pm, Jean-Marc Spaggiari <
> [email protected]> wrote:
> >>
> >> > Hi Arthur,
> >> >
> >> > How have you extracted HBase source and what command do you run to
> build? I
> >> > will do the same here locally so I can provide you the exact step to
> >> > complete.
> >> >
> >> > JM
> >> >
> >> >
> >> > 2014-08-26 8:42 GMT-04:00 [email protected] <
> [email protected]
> >> >> :
> >> >
> >> >> Hi JM
> >> >>
> >> >> Not too sure what you mean, do you mean I should create a new folder
> in my
> >> >> HBASE_SRC named lib/native/Linux-x86 and copy these files to this
> folder
> >> >> then try to compile it again?
> >> >>
> >> >> Regards
> >> >> ARthur
> >> >>
> >> >>
> >> >> On 26 Aug, 2014, at 8:17 pm, Jean-Marc Spaggiari <
> [email protected]>
> >> >> wrote:
> >> >>
> >> >>> Hi Arthur,
> >> >>>
> >> >>> Almost done! You now need to copy them on the HBase folder.
> >> >>>
> >> >>> hbase@hbasetest1:~/hbase-0.98.2-hadoop2/lib$ tree | grep -v .jar |
> grep
> >> >> -v
> >> >>> .rb
> >> >>> .
> >> >>> ├── native
> >> >>> │   └── Linux-x86
> >> >>> │       ├── libsnappy.a
> >> >>> │       ├── libsnappy.la
> >> >>> │       ├── libsnappy.so
> >> >>> │       ├── libsnappy.so.1
> >> >>> │       └── libsnappy.so.1.2.0
> >> >>>
> >> >>> I don't have any hadoop-snappy lib in my hbase folder and it works
> very
> >> >>> well with Snappy for me...
> >> >>>
> >> >>> JM
> >> >>>
> >> >>> 2014-08-26 8:09 GMT-04:00 [email protected] <
> >> >> [email protected]
> >> >>>> :
> >> >>>
> >> >>>> Hi JM,
> >> >>>>
> >> >>>> Below are my steps to install snappy lib, do I miss something?
> >> >>>>
> >> >>>> Regards
> >> >>>> Arthur
> >> >>>>
> >> >>>> wget https://snappy.googlecode.com/files/snappy-1.1.1.tar.gz
> >> >>>> tar -vxf snappy-1.1.1.tar.gz
> >> >>>> cd snappy-1.1.1
> >> >>>> ./configure
> >> >>>> make
> >> >>>> make install
> >> >>>>       make[1]: Entering directory
> >> >> `/edh/hadoop_all_sources/snappy-1.1.1'
> >> >>>>       test -z "/usr/local/lib" || /bin/mkdir -p "/usr/local/lib"
> >> >>>>        /bin/sh ./libtool   --mode=install /usr/bin/install -c
> >> >>>> libsnappy.la '/usr/local/lib'
> >> >>>>       libtool: install: /usr/bin/install -c
> .libs/libsnappy.so.1.2.0
> >> >>>> /usr/local/lib/libsnappy.so.1.2.0
> >> >>>>       libtool: install: (cd /usr/local/lib && { ln -s -f
> >> >>>> libsnappy.so.1.2.0 libsnappy.so.1 || { rm -f libsnappy.so.1 && ln
> -s
> >> >>>> libsnappy.so.1.2.0 libsnappy.so.1; }; })
> >> >>>>       libtool: install: (cd /usr/local/lib && { ln -s -f
> >> >>>> libsnappy.so.1.2.0 libsnappy.so || { rm -f libsnappy.so && ln -s
> >> >>>> libsnappy.so.1.2.0 libsnappy.so; }; })
> >> >>>>       libtool: install: /usr/bin/install -c .libs/libsnappy.lai
> >> >>>> /usr/local/lib/libsnappy.la
> >> >>>>       libtool: install: /usr/bin/install -c .libs/libsnappy.a
> >> >>>> /usr/local/lib/libsnappy.a
> >> >>>>       libtool: install: chmod 644 /usr/local/lib/libsnappy.a
> >> >>>>       libtool: install: ranlib /usr/local/lib/libsnappy.a
> >> >>>>       libtool: finish:
> >> >>>>
> >> >>
> PATH="/edh/hadoop/spark/bin:/edh/hadoop/hbase/bin:/edh/hadoop/zookeeper//bin:/edh/hadoop/yarn/hadoop/bin:/edh/hadoop/yarn/hadoop/sbin:/usr/lib64/qt-3.3/bin:/opt/apache-maven-3.1.1/bin:/usr/local/sbin:/usr/local/bin:/sbin:/bin:/usr/sbin:/usr/bin:/edh/hadoop/zookeeper//bin:/edh/hadoop/hive//bin:/usr/lib/jvm/jdk1.6.0_45//bin:/root/bin:/sbin"
> >> >>>> ldconfig -n /usr/local/lib
> >> >>>>
> >> >>>>
> ----------------------------------------------------------------------
> >> >>>>       Libraries have been installed in:
> >> >>>>       /usr/local/lib
> >> >>>>       If you ever happen to want to link against installed
> libraries
> >> >>>>       in a given directory, LIBDIR, you must either use libtool,
> and
> >> >>>>       specify the full pathname of the library, or use the
> `-LLIBDIR'
> >> >>>>       flag during linking and do at least one of the following:
> >> >>>>       - add LIBDIR to the `LD_LIBRARY_PATH' environment variable
> >> >>>>       during execution
> >> >>>>       - add LIBDIR to the `LD_RUN_PATH' environment variable
> >> >>>>       during linking
> >> >>>>       - use the `-Wl,-rpath -Wl,LIBDIR' linker flag
> >> >>>>       - have your system administrator add LIBDIR to
> `/etc/ld.so.conf'
> >> >>>>       See any operating system documentation about shared
> libraries for
> >> >>>>       more information, such as the ld(1) and ld.so(8) manual
> pages.
> >> >>>>
> >> >>>>
> ----------------------------------------------------------------------
> >> >>>>       test -z "/usr/local/share/doc/snappy" || /bin/mkdir -p
> >> >>>> "/usr/local/share/doc/snappy"
> >> >>>>        /usr/bin/install -c -m 644 ChangeLog COPYING INSTALL NEWS
> README
> >> >>>> format_description.txt framing_format.txt
> '/usr/local/share/doc/snappy'
> >> >>>>       test -z "/usr/local/include" || /bin/mkdir -p
> >> >> "/usr/local/include"
> >> >>>>        /usr/bin/install -c -m 644 snappy.h snappy-sinksource.h
> >> >>>> snappy-stubs-public.h snappy-c.h '/usr/local/include'
> >> >>>>       make[1]: Leaving directory
> `/edh/hadoop_all_sources/snappy-1.1.1'
> >> >>>>
> >> >>>> ll /usr/local/lib
> >> >>>>       -rw-r--r--. 1 root root   233554 Aug 20 00:14 libsnappy.a
> >> >>>>       -rwxr-xr-x. 1 root root      953 Aug 20 00:14 libsnappy.la
> >> >>>>       lrwxrwxrwx. 1 root root       18 Aug 20 00:14 libsnappy.so ->
> >> >>>> libsnappy.so.1.2.0
> >> >>>>       lrwxrwxrwx. 1 root root       18 Aug 20 00:14 libsnappy.so.1
> ->
> >> >>>> libsnappy.so.1.2.0
> >> >>>>       -rwxr-xr-x. 1 root root   147726 Aug 20 00:14
> libsnappy.so.1.2.0
> >> >>>>
> >> >>>>
> >> >>>>
> >> >>>> On 26 Aug, 2014, at 7:38 pm, Jean-Marc Spaggiari <
> >> >> [email protected]>
> >> >>>> wrote:
> >> >>>>
> >> >>>>> Hi Arthur,
> >> >>>>>
> >> >>>>> Do you have snappy libs installed and configured? HBase doesn't
> come
> >> >> with
> >> >>>>> Snappy. So yo need to have it first.
> >> >>>>>
> >> >>>>> Shameless plug:
> >> >>>>>
> >> >>>>
> >> >>
> http://www.spaggiari.org/index.php/hbase/how-to-install-snappy-with-1#.U_xxSqdZuZY
> >> >>>>>
> >> >>>>> This is for 0.96 but should be very similar for 0.98. I will try
> it
> >> >> soon
> >> >>>>> and post and update, but keep us posted here so we can support
> you...
> >> >>>>>
> >> >>>>> JM
> >> >>>>>
> >> >>>>>
> >> >>>>> 2014-08-26 7:34 GMT-04:00 [email protected] <
> >> >>>> [email protected]
> >> >>>>>> :
> >> >>>>>
> >> >>>>>> Hi,
> >> >>>>>>
> >> >>>>>> I need to install snappy to HBase 0.98.4.  (my Hadoop version is
> >> >> 2.4.1)
> >> >>>>>>
> >> >>>>>> Can you please advise what would be wrong?  Should my pom.xml be
> >> >>>> incorrect
> >> >>>>>> and missing something?
> >> >>>>>>
> >> >>>>>> Regards
> >> >>>>>> Arthur
> >> >>>>>>
> >> >>>>>>
> >> >>>>>> Below are my commands:
> >> >>>>>> bash -x ./dev-support/generate-hadoopX-poms.sh 0.98.4
> 0.98.4-hadoop2
> >> >>>>>> mvn -f pom.xml.hadoop2 install -DskipTests assembly:single
> >> >>>>>> -Prelease,hadoop-snappy
> >> >>>>>>
> >> >>>>>> Iog:
> >> >>>>>> [INFO]
> >> >>>>>>
> >> >>
> ------------------------------------------------------------------------
> >> >>>>>> [INFO] Building HBase - Server 0.98.4-hadoop2
> >> >>>>>> [INFO]
> >> >>>>>>
> >> >>
> ------------------------------------------------------------------------
> >> >>>>>> [WARNING] The POM for
> >> >> org.apache.hadoop:hadoop-snappy:jar:0.0.1-SNAPSHOT
> >> >>>>>> is missing, no dependency information available
> >> >>>>>> [INFO]
> >> >>>>>>
> >> >>
> ------------------------------------------------------------------------
> >> >>>>>> [INFO] Reactor Summary:
> >> >>>>>> [INFO]
> >> >>>>>> [INFO] HBase .............................................
> SUCCESS
> >> >>>> [3.129s]
> >> >>>>>> [INFO] HBase - Common ....................................
> SUCCESS
> >> >>>> [3.105s]
> >> >>>>>> [INFO] HBase - Protocol ..................................
> SUCCESS
> >> >>>> [0.976s]
> >> >>>>>> [INFO] HBase - Client ....................................
> SUCCESS
> >> >>>> [0.925s]
> >> >>>>>> [INFO] HBase - Hadoop Compatibility ......................
> SUCCESS
> >> >>>> [0.183s]
> >> >>>>>> [INFO] HBase - Hadoop Two Compatibility ..................
> SUCCESS
> >> >>>> [0.497s]
> >> >>>>>> [INFO] HBase - Prefix Tree ...............................
> SUCCESS
> >> >>>> [0.407s]
> >> >>>>>> [INFO] HBase - Server ....................................
> FAILURE
> >> >>>> [0.103s]
> >> >>>>>> [INFO] HBase - Testing Util ..............................
> SKIPPED
> >> >>>>>> [INFO] HBase - Thrift ....................................
> SKIPPED
> >> >>>>>> [INFO] HBase - Shell .....................................
> SKIPPED
> >> >>>>>> [INFO] HBase - Integration Tests .........................
> SKIPPED
> >> >>>>>> [INFO] HBase - Examples ..................................
> SKIPPED
> >> >>>>>> [INFO] HBase - Assembly ..................................
> SKIPPED
> >> >>>>>> [INFO]
> >> >>>>>>
> >> >>
> ------------------------------------------------------------------------
> >> >>>>>> [INFO] BUILD FAILURE
> >> >>>>>> [INFO]
> >> >>>>>>
> >> >>
> ------------------------------------------------------------------------
> >> >>>>>> [INFO] Total time: 9.939s
> >> >>>>>> [INFO] Finished at: Tue Aug 26 19:23:14 HKT 2014
> >> >>>>>> [INFO] Final Memory: 61M/2921M
> >> >>>>>> [INFO]
> >> >>>>>>
> >> >>
> ------------------------------------------------------------------------
> >> >>>>>> [ERROR] Failed to execute goal on project hbase-server: Could not
> >> >>>> resolve
> >> >>>>>> dependencies for project
> >> >>>> org.apache.hbase:hbase-server:jar:0.98.4-hadoop2:
> >> >>>>>> Failure to find
> org.apache.hadoop:hadoop-snappy:jar:0.0.1-SNAPSHOT in
> >> >>>>>> http://maven.oschina.net/content/groups/public/ was cached in
> the
> >> >> local
> >> >>>>>> repository, resolution will not be reattempted until the update
> >> >>>> interval of
> >> >>>>>> nexus-osc has elapsed or updates are forced -> [Help 1]
> >> >>>>>> [ERROR]
> >> >>>>>> [ERROR] To see the full stack trace of the errors, re-run Maven
> with
> >> >> the
> >> >>>>>> -e switch.
> >> >>>>>> [ERROR] Re-run Maven using the -X switch to enable full debug
> logging.
> >> >>>>>> [ERROR]
> >> >>>>>> [ERROR] For more information about the errors and possible
> solutions,
> >> >>>>>> please read the following articles:
> >> >>>>>> [ERROR] [Help 1]
> >> >>>>>>
> >> >>>>
> >> >>
> http://cwiki.apache.org/confluence/display/MAVEN/DependencyResolutionException
> >> >>>>>> [ERROR]
> >> >>>>>> [ERROR] After correcting the problems, you can resume the build
> with
> >> >> the
> >> >>>>>> command
> >> >>>>>> [ERROR]   mvn <goals> -rf :hbase-server
> >> >>>>>>
> >> >>>>>>
> >> >>>>
> >> >>>>
> >> >>
> >> >>
> >>
> >>
> >>
> >>
> >> --
> >> Sean
> >
>
>

Reply via email to