That basically says that snappy isn't working properly on your box. You can
forget about that for now by running:

set spark.io.compression.codec=org.apache.spark.io.LZFCompressionCodec;


On Sat, Dec 5, 2015 at 1:45 PM, Mich Talebzadeh <m...@peridale.co.uk> wrote:

> Great stuff.
>
>
>
> Built the source code for 1.3.1 and generated
> spark-assembly-1.3.1-hadoop2.4.0.jar
>
>
>
> jar tvf spark-assembly-1.3.1-hadoop2.4.0.jar|grep hive | grep -i -v Archive
>
>
>
> so no hive there
>
>
>
> Downloaded prebuilt spark 1.3.1 and started master and slave OK
>
>
>
> Started hive as usual in debug mode  and did a simple select count(1) from
> t;
>
>
>
> Spark app started OK
>
>
>
> hduser@rhes564::/usr/lib/spark-1.3.1-bin-hadoop2.6/logs>
>
>
>
> -rw-r--r-- 1 hduser hadoop  31562 Dec  5 21:18
> spark-hduser-org.apache.spark.deploy.master.Master-1-rhes564.out
>
> -rw-r--r-- 1 hduser hadoop  19684 Dec  5 21:18
> spark-hduser-org.apache.spark.deploy.worker.Worker-1-rhes564.out
>
> -rwxrwx--- 1 hduser hadoop  60491 Dec  5 21:18
> app-20151205211814-0005.inprogress
>
>
>
> Now I get some library error
>
>
>
> 5/12/05 21:18:16 [stderr-redir-1]: INFO client.SparkClientImpl: Caused by:
> java.lang.UnsatisfiedLinkError: /tmp/snappy-1.0.5-libsnappyjava.so:
> /usr/lib64/libstdc++.so.6: version `GLIBCXX_3.4.9' not found (required by
> /tmp/snappy-1.0.5-libsnappyjava.so)
>
>
>
>
>
> strings /usr/lib/libstdc++.so.6 | grep GLIBC
>
> GLIBCXX_3.4
>
> GLIBCXX_3.4.1
>
> GLIBCXX_3.4.2
>
> GLIBCXX_3.4.3
>
> GLIBCXX_3.4.4
>
> GLIBCXX_3.4.5
>
> GLIBCXX_3.4.6
>
> GLIBCXX_3.4.7
>
> GLIBCXX_3.4.8
>
> GLIBC_2.3
>
> GLIBC_2.0
>
> GLIBC_2.3.2
>
> GLIBC_2.4
>
> GLIBC_2.1
>
> GLIBC_2.1.3
>
> GLIBC_2.2
>
> GLIBCXX_FORCE_NEW
>
>
>
> Looking into sorting this out.
>
>
>
> Mich Talebzadeh
>
>
>
> *Sybase ASE 15 Gold Medal Award 2008*
>
> A Winning Strategy: Running the most Critical Financial Data on ASE 15
>
>
> http://login.sybase.com/files/Product_Overviews/ASE-Winning-Strategy-091908.pdf
>
> Author of the books* "A Practitioner’s Guide to Upgrading to Sybase ASE
> 15", ISBN 978-0-9563693-0-7*.
>
> co-author *"Sybase Transact SQL Guidelines Best Practices", ISBN
> 978-0-9759693-0-4*
>
> *Publications due shortly:*
>
> *Complex Event Processing in Heterogeneous Environments*, ISBN:
> 978-0-9563693-3-8
>
> *Oracle and Sybase, Concepts and Contrasts*, ISBN: 978-0-9563693-1-4, volume
> one out shortly
>
>
>
> http://talebzadehmich.wordpress.com
>
>
>
> NOTE: The information in this email is proprietary and confidential. This
> message is for the designated recipient only, if you are not the intended
> recipient, you should destroy it immediately. Any information in this
> message shall not be understood as given or endorsed by Peridale Technology
> Ltd, its subsidiaries or their employees, unless expressly so stated. It is
> the responsibility of the recipient to ensure that this email is virus
> free, therefore neither Peridale Ltd, its subsidiaries nor their employees
> accept any responsibility.
>
>
>
> *From:* Xuefu Zhang [mailto:xzh...@cloudera.com]
> *Sent:* 04 December 2015 17:47
> *To:* user@hive.apache.org
> *Subject:* Re: FW: Getting error when trying to start master node after
> building spark 1.3
>
>
>
> 1.3.1 is what officially supported by Hive 1.2.1. 1.3.0 might be okay too.
>
>
>
> On Fri, Dec 4, 2015 at 9:34 AM, Mich Talebzadeh <m...@peridale.co.uk>
> wrote:
>
> Appreciated the response. Just to clarify the build will be spark 1.3 and
> the pre-build download will be 1.3. this is the version I am attempting to
> make it work.
>
>
>
> Thanks
>
>
>
> Mich
>
>
>
> *From:* Xuefu Zhang [mailto:xzh...@cloudera.com]
> *Sent:* 04 December 2015 17:03
> *To:* user@hive.apache.org
> *Subject:* Re: FW: Getting error when trying to start master node after
> building spark 1.3
>
>
>
> My last attempt:
>
> 1. Make sure the spark-assembly.jar from your own build doesn't contain
> hive classes, using "jar -tf spark-assembly.jar | grep hive" command. Copy
> it to Hive's /lib directory. After this, you can forget everything about
> this build.
>
> 2. Download prebuilt tarball from Spark download site and deploy it.
> Forget about Hive for a moment. Make sure the cluster comes up and
> functions.
>
> 3. Unset environment variable SPARK_HOME before you start Hive. Start
> Hive, and run "set spark.home=/path/to/spark/dir" command. Then run other
> commands as you did previously when trying hive on spark.
>
>
>
>
>
>
>

Reply via email to