@rogthefrog
Were you able to figure out how to fix this issue?
Even I tried all combinations that possible but no luck yet.
Thanks,
Harsha
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/LZO-support-in-Spark-1-0-0-nothing-seems-to-work-tp14494p18349.html
That does appear to be the case. Thanks!
For posterity, I ran my pyspark like this:
$ sudo su yarn
$ pyspark --driver-library-path
/opt/cloudera/parcels/GPLEXTRAS/lib/hadoop/lib/native/
>>> p = sc.textFile("/some/file")
>>> p.count()
everything appears to be working now.
--
View this message
It works for me :
export
JAVA_LIBRARY_PATH=$JAVA_LIBRARY_PATH:/opt/cloudera/parcels/HADOOP_LZO/lib/hadoop/lib/native
export
LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/opt/cloudera/parcels/HADOOP_LZO/lib/hadoop/lib/native
export
SPARK_LIBRARY_PATH=$SPARK_LIBRARY_PATH:/opt/cloudera/parcels/HADOOP_LZO/l
I believe this is a known bug:
https://issues.apache.org/jira/browse/SPARK-1719
On Wed, Sep 17, 2014 at 5:40 PM, rogthefrog wrote:
> I have a HDFS cluster managed with CDH Manager. Version is CDH 5.1 with
> matching GPLEXTRAS parcel. LZO works with Hive and Pig, but I can't make it
> work with Sp
I have a HDFS cluster managed with CDH Manager. Version is CDH 5.1 with
matching GPLEXTRAS parcel. LZO works with Hive and Pig, but I can't make it
work with Spark 1.0.0. I've tried:
* Setting this:
HADOOP_OPTS="-Djava.net.preferIPv4Stack=true $HADOOP_CLIENT_OPTS
-Djava.library.path=/opt/cloudera