Hi,

You’d need to specify the complete path along with HDFS URI. for eg. 
hdfs://namenode:<port>/path/to/file. 

- Nikaash 
> On 14-Apr-2016, at 3:39 PM, Rohit Jain <[email protected]> wrote:
> 
> Hey Nikaash,
> 
> I ran this command but its working local files not for hadoop input file.
> How to make it work on hadoop uri?
> 
> /Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/bin/mahout
> spark-itemsimilarity -i spdemo.txt -o /spout --filter1 like -fc 1 -ic 2
> -D:spark.io.compression.codec=lzf
> 
> On Thu, Apr 14, 2016 at 3:19 PM, Nikaash Puri <[email protected]> wrote:
> 
>> Cool. Even if you have problems setting up the library, just change the
>> compression codec from snappy to lzf (-D:spark.io.compression.codec=lzf)
>> like in the command. It might hurt performance, but it’ll get it to work
>> 
>> - Nikaash
>>> On 14-Apr-2016, at 3:17 PM, Rohit Jain <[email protected]> wrote:
>>> 
>>> yes. i am downloading the that library and putting it java/lib/ path.
>>> Anythingelse need to be done?
>>> 
>>> On Thu, Apr 14, 2016 at 3:08 PM, Nikaash Puri <[email protected]>
>> wrote:
>>> 
>>>> Hi Rohit,
>>>> 
>>>> It’s a snappy library issue. Are you running on Mac?
>>>> Anyhow, run it like this:
>>>> ./bin/mahout spark-itemsimilarity --input <in> --output <out>
>>>> -D:spark.io.compression.codec=lzf
>>>> 
>>>> Nikaash
>>>> 
>>>>> On 14-Apr-2016, at 12:25 PM, Rohit Jain <[email protected]>
>> wrote:
>>>>> 
>>>>> Hello everyone,
>>>>> I am new to mahout. I am trying to run this command
>>>>> 
>>>>> /Documents/apache-mahout-distribution-0.11.2/bin/mahout
>>>>> spark-itemsimilarity -i spdemo1.txt -o /spout --filter1 like -fc 1 -ic
>> 2
>>>>> 
>>>>> before running the command i uploaded file to hdfs
>>>>> 
>>>>> bin/hadoop dfs -put spdemo.txt /spdemo1.txt
>>>>> 
>>>>> but i am geting this long list of error.Do I need to install spark too?
>>>>> 
>>>>> 
>>>>> SLF4J: Class path contains multiple SLF4J bindings.
>>>>> 
>>>>> SLF4J: Found binding in
>>>>> 
>>>> 
>> [jar:file:/Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/mahout-examples-0.11.2-job.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>> 
>>>>> SLF4J: Found binding in
>>>>> 
>>>> 
>> [jar:file:/Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/mahout-mr-0.11.2-job.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>> 
>>>>> SLF4J: Found binding in
>>>>> 
>>>> 
>> [jar:file:/Users/rohitjain/Documents/spark-1.6.1-bin-hadoop2.6/lib/spark-assembly-1.6.1-hadoop2.6.0.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>> 
>>>>> SLF4J: Found binding in
>>>>> 
>>>> 
>> [jar:file:/Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/lib/slf4j-log4j12-1.7.12.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>> 
>>>>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
>>>>> explanation.
>>>>> 
>>>>> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
>>>>> 
>>>>> 16/04/14 10:53:09 WARN SparkConf: The configuration key
>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark 1.4
>> and
>>>>> may be removed in the future. Please use spark.kryoserializer.buffer
>>>>> instead. The default value for spark.kryoserializer.buffer.mb was
>>>>> previously specified as '0.064'. Fractional values are no longer
>>>> accepted.
>>>>> To specify the equivalent now, one may use '64k'.
>>>>> 
>>>>> 16/04/14 10:53:09 WARN SparkConf: The configuration key
>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark 1.4
>> and
>>>>> and may be removed in the future. Please use the new key
>>>>> 'spark.kryoserializer.buffer' instead.
>>>>> 
>>>>> 16/04/14 10:53:10 INFO SparkContext: Running Spark version 1.6.1
>>>>> 
>>>>> 16/04/14 10:53:10 WARN NativeCodeLoader: Unable to load native-hadoop
>>>>> library for your platform... using builtin-java classes where
>> applicable
>>>>> 
>>>>> 16/04/14 10:53:10 WARN SparkConf: The configuration key
>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark 1.4
>> and
>>>>> may be removed in the future. Please use spark.kryoserializer.buffer
>>>>> instead. The default value for spark.kryoserializer.buffer.mb was
>>>>> previously specified as '0.064'. Fractional values are no longer
>>>> accepted.
>>>>> To specify the equivalent now, one may use '64k'.
>>>>> 
>>>>> 16/04/14 10:53:10 WARN SparkConf: The configuration key
>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark 1.4
>> and
>>>>> and may be removed in the future. Please use the new key
>>>>> 'spark.kryoserializer.buffer' instead.
>>>>> 
>>>>> 16/04/14 10:53:10 INFO SecurityManager: Changing view acls to:
>> rohitjain
>>>>> 
>>>>> 16/04/14 10:53:10 INFO SecurityManager: Changing modify acls to:
>>>> rohitjain
>>>>> 
>>>>> 16/04/14 10:53:10 INFO SecurityManager: SecurityManager: authentication
>>>>> disabled; ui acls disabled; users with view permissions:
>> Set(rohitjain);
>>>>> users with modify permissions: Set(rohitjain)
>>>>> 
>>>>> 16/04/14 10:53:11 WARN SparkConf: The configuration key
>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark 1.4
>> and
>>>>> may be removed in the future. Please use spark.kryoserializer.buffer
>>>>> instead. The default value for spark.kryoserializer.buffer.mb was
>>>>> previously specified as '0.064'. Fractional values are no longer
>>>> accepted.
>>>>> To specify the equivalent now, one may use '64k'.
>>>>> 
>>>>> 16/04/14 10:53:11 WARN SparkConf: The configuration key
>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark 1.4
>> and
>>>>> and may be removed in the future. Please use the new key
>>>>> 'spark.kryoserializer.buffer' instead.
>>>>> 
>>>>> 16/04/14 10:53:11 WARN SparkConf: The configuration key
>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark 1.4
>> and
>>>>> may be removed in the future. Please use spark.kryoserializer.buffer
>>>>> instead. The default value for spark.kryoserializer.buffer.mb was
>>>>> previously specified as '0.064'. Fractional values are no longer
>>>> accepted.
>>>>> To specify the equivalent now, one may use '64k'.
>>>>> 
>>>>> 16/04/14 10:53:11 WARN SparkConf: The configuration key
>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark 1.4
>> and
>>>>> and may be removed in the future. Please use the new key
>>>>> 'spark.kryoserializer.buffer' instead.
>>>>> 
>>>>> 16/04/14 10:53:11 INFO Utils: Successfully started service
>> 'sparkDriver'
>>>> on
>>>>> port 64062.
>>>>> 
>>>>> 16/04/14 10:53:11 INFO Slf4jLogger: Slf4jLogger started
>>>>> 
>>>>> 16/04/14 10:53:11 INFO Remoting: Starting remoting
>>>>> 
>>>>> 16/04/14 10:53:11 INFO Remoting: Remoting started; listening on
>> addresses
>>>>> :[akka.tcp://[email protected]:64063]
>>>>> 
>>>>> 16/04/14 10:53:11 INFO Utils: Successfully started service
>>>>> 'sparkDriverActorSystem' on port 64063.
>>>>> 
>>>>> 16/04/14 10:53:11 INFO SparkEnv: Registering MapOutputTracker
>>>>> 
>>>>> 16/04/14 10:53:11 WARN SparkConf: The configuration key
>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark 1.4
>> and
>>>>> may be removed in the future. Please use spark.kryoserializer.buffer
>>>>> instead. The default value for spark.kryoserializer.buffer.mb was
>>>>> previously specified as '0.064'. Fractional values are no longer
>>>> accepted.
>>>>> To specify the equivalent now, one may use '64k'.
>>>>> 
>>>>> 16/04/14 10:53:11 WARN SparkConf: The configuration key
>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark 1.4
>> and
>>>>> and may be removed in the future. Please use the new key
>>>>> 'spark.kryoserializer.buffer' instead.
>>>>> 
>>>>> 16/04/14 10:53:11 WARN SparkConf: The configuration key
>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark 1.4
>> and
>>>>> may be removed in the future. Please use spark.kryoserializer.buffer
>>>>> instead. The default value for spark.kryoserializer.buffer.mb was
>>>>> previously specified as '0.064'. Fractional values are no longer
>>>> accepted.
>>>>> To specify the equivalent now, one may use '64k'.
>>>>> 
>>>>> 16/04/14 10:53:11 WARN SparkConf: The configuration key
>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark 1.4
>> and
>>>>> and may be removed in the future. Please use the new key
>>>>> 'spark.kryoserializer.buffer' instead.
>>>>> 
>>>>> 16/04/14 10:53:11 INFO SparkEnv: Registering BlockManagerMaster
>>>>> 
>>>>> 16/04/14 10:53:11 INFO DiskBlockManager: Created local directory at
>>>>> 
>>>> 
>> /private/var/folders/17/vhrqnx4x1n10yqbl1vnch9000000gn/T/blockmgr-6cdfd6f1-56c8-445c-95e0-70305b0f5f5a
>>>>> 
>>>>> 16/04/14 10:53:11 INFO MemoryStore: MemoryStore started with capacity
>>>> 2.4 GB
>>>>> 
>>>>> 16/04/14 10:53:12 INFO SparkEnv: Registering OutputCommitCoordinator
>>>>> 
>>>>> 16/04/14 10:53:12 INFO Server: jetty-8.y.z-SNAPSHOT
>>>>> 
>>>>> 16/04/14 10:53:12 INFO AbstractConnector: Started
>>>>> [email protected]:4040
>>>>> 
>>>>> 16/04/14 10:53:12 INFO Utils: Successfully started service 'SparkUI' on
>>>>> port 4040.
>>>>> 
>>>>> 16/04/14 10:53:12 INFO SparkUI: Started SparkUI at
>>>> http://192.168.1.122:4040
>>>>> 
>>>>> 16/04/14 10:53:12 INFO HttpFileServer: HTTP File server directory is
>>>>> 
>>>> 
>> /private/var/folders/17/vhrqnx4x1n10yqbl1vnch9000000gn/T/spark-ad6a714b-6dcb-410c-aef3-2b098695743b/httpd-11f32419-5f8a-4b61-9654-9ea2d32cab5d
>>>>> 
>>>>> 16/04/14 10:53:12 INFO HttpServer: Starting HTTP Server
>>>>> 
>>>>> 16/04/14 10:53:12 INFO Server: jetty-8.y.z-SNAPSHOT
>>>>> 
>>>>> 16/04/14 10:53:12 INFO AbstractConnector: Started
>>>>> [email protected]:64064
>>>>> 
>>>>> 16/04/14 10:53:12 INFO Utils: Successfully started service 'HTTP file
>>>>> server' on port 64064.
>>>>> 
>>>>> 16/04/14 10:53:12 INFO SparkContext: Added JAR
>>>>> 
>>>> 
>> /Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/mahout-hdfs-0.11.2.jar
>>>>> at http://192.168.1.122:64064/jars/mahout-hdfs-0.11.2.jar with
>> timestamp
>>>>> 1460611392248
>>>>> 
>>>>> 16/04/14 10:53:12 INFO SparkContext: Added JAR
>>>>> 
>>>> 
>> /Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/mahout-math-0.11.2.jar
>>>>> at http://192.168.1.122:64064/jars/mahout-math-0.11.2.jar with
>> timestamp
>>>>> 1460611392253
>>>>> 
>>>>> 16/04/14 10:53:12 INFO SparkContext: Added JAR
>>>>> 
>>>> 
>> /Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/mahout-math-scala_2.10-0.11.2.jar
>>>>> at http://192.168.1.122:64064/jars/mahout-math-scala_2.10-0.11.2.jar
>>>> with
>>>>> timestamp 1460611392255
>>>>> 
>>>>> 16/04/14 10:53:12 INFO SparkContext: Added JAR
>>>>> 
>>>> 
>> /Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/mahout-spark_2.10-0.11.2-dependency-reduced.jar
>>>>> at
>>>>> 
>>>> 
>> http://192.168.1.122:64064/jars/mahout-spark_2.10-0.11.2-dependency-reduced.jar
>>>>> with timestamp 1460611392303
>>>>> 
>>>>> 16/04/14 10:53:12 INFO SparkContext: Added JAR
>>>>> 
>>>> 
>> /Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/mahout-spark_2.10-0.11.2.jar
>>>>> at http://192.168.1.122:64064/jars/mahout-spark_2.10-0.11.2.jar with
>>>>> timestamp 1460611392304
>>>>> 
>>>>> 16/04/14 10:53:12 WARN SparkConf: The configuration key
>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark 1.4
>> and
>>>>> may be removed in the future. Please use spark.kryoserializer.buffer
>>>>> instead. The default value for spark.kryoserializer.buffer.mb was
>>>>> previously specified as '0.064'. Fractional values are no longer
>>>> accepted.
>>>>> To specify the equivalent now, one may use '64k'.
>>>>> 
>>>>> 16/04/14 10:53:12 WARN SparkConf: The configuration key
>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark 1.4
>> and
>>>>> and may be removed in the future. Please use the new key
>>>>> 'spark.kryoserializer.buffer' instead.
>>>>> 
>>>>> 16/04/14 10:53:12 WARN SparkConf: The configuration key
>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark 1.4
>> and
>>>>> may be removed in the future. Please use spark.kryoserializer.buffer
>>>>> instead. The default value for spark.kryoserializer.buffer.mb was
>>>>> previously specified as '0.064'. Fractional values are no longer
>>>> accepted.
>>>>> To specify the equivalent now, one may use '64k'.
>>>>> 
>>>>> 16/04/14 10:53:12 WARN SparkConf: The configuration key
>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark 1.4
>> and
>>>>> and may be removed in the future. Please use the new key
>>>>> 'spark.kryoserializer.buffer' instead.
>>>>> 
>>>>> 16/04/14 10:53:12 INFO Executor: Starting executor ID driver on host
>>>>> localhost
>>>>> 
>>>>> 16/04/14 10:53:12 INFO Utils: Successfully started service
>>>>> 'org.apache.spark.network.netty.NettyBlockTransferService' on port
>> 64065.
>>>>> 
>>>>> 16/04/14 10:53:12 INFO NettyBlockTransferService: Server created on
>> 64065
>>>>> 
>>>>> 16/04/14 10:53:12 INFO BlockManagerMaster: Trying to register
>>>> BlockManager
>>>>> 
>>>>> 16/04/14 10:53:12 INFO BlockManagerMasterEndpoint: Registering block
>>>>> manager localhost:64065 with 2.4 GB RAM, BlockManagerId(driver,
>>>> localhost,
>>>>> 64065)
>>>>> 
>>>>> 16/04/14 10:53:12 INFO BlockManagerMaster: Registered BlockManager
>>>>> 
>>>>> java.lang.reflect.InvocationTargetException
>>>>> 
>>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>>> 
>>>>> at java.lang.reflect.Method.invoke(Method.java:497)
>>>>> 
>>>>> at
>>>> org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:317)
>>>>> 
>>>>> at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219)
>>>>> 
>>>>> at org.xerial.snappy.Snappy.<clinit>(Snappy.java:44)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.io.SnappyCompressionCodec$.liftedTree1$1(CompressionCodec.scala:169)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.io.SnappyCompressionCodec$.org$apache$spark$io$SnappyCompressionCodec$$version$lzycompute(CompressionCodec.scala:168)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.io.SnappyCompressionCodec$.org$apache$spark$io$SnappyCompressionCodec$$version(CompressionCodec.scala:168)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.io.SnappyCompressionCodec.<init>(CompressionCodec.scala:152)
>>>>> 
>>>>> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
>> Method)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>>>>> 
>>>>> at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.io.CompressionCodec$.createCodec(CompressionCodec.scala:72)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.io.CompressionCodec$.createCodec(CompressionCodec.scala:65)
>>>>> 
>>>>> at org.apache.spark.broadcast.TorrentBroadcast.org
>>>>> 
>>>> 
>> $apache$spark$broadcast$TorrentBroadcast$$setConf(TorrentBroadcast.scala:73)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.broadcast.TorrentBroadcast.<init>(TorrentBroadcast.scala:80)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:34)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:63)
>>>>> 
>>>>> at org.apache.spark.SparkContext.broadcast(SparkContext.scala:1326)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:1014)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:1011)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
>>>>> 
>>>>> at org.apache.spark.SparkContext.withScope(SparkContext.scala:714)
>>>>> 
>>>>> at org.apache.spark.SparkContext.hadoopFile(SparkContext.scala:1011)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:832)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:830)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
>>>>> 
>>>>> at org.apache.spark.SparkContext.withScope(SparkContext.scala:714)
>>>>> 
>>>>> at org.apache.spark.SparkContext.textFile(SparkContext.scala:830)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.drivers.TDIndexedDatasetReader$class.elementReader(TextDelimitedReaderWriter.scala:61)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.drivers.TextDelimitedIndexedDatasetReader.elementReader(TextDelimitedReaderWriter.scala:304)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.drivers.TextDelimitedIndexedDatasetReader.elementReader(TextDelimitedReaderWriter.scala:304)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.math.indexeddataset.Reader$class.readElementsFrom(ReaderWriter.scala:75)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.drivers.TextDelimitedIndexedDatasetReader.readElementsFrom(TextDelimitedReaderWriter.scala:304)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.sparkbindings.SparkEngine$.indexedDatasetDFSReadElements(SparkEngine.scala:382)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.sparkbindings.SparkEngine$.indexedDatasetDFSReadElements(SparkEngine.scala:39)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.math.indexeddataset.package$.indexedDatasetDFSReadElements(package.scala:336)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.drivers.ItemSimilarityDriver$.readIndexedDatasets(ItemSimilarityDriver.scala:152)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.drivers.ItemSimilarityDriver$.process(ItemSimilarityDriver.scala:201)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.drivers.ItemSimilarityDriver$$anonfun$main$1.apply(ItemSimilarityDriver.scala:112)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.drivers.ItemSimilarityDriver$$anonfun$main$1.apply(ItemSimilarityDriver.scala:110)
>>>>> 
>>>>> at scala.Option.map(Option.scala:145)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.drivers.ItemSimilarityDriver$.main(ItemSimilarityDriver.scala:110)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.drivers.ItemSimilarityDriver.main(ItemSimilarityDriver.scala)
>>>>> 
>>>>> Caused by: java.lang.UnsatisfiedLinkError: no snappyjava in
>>>>> java.library.path
>>>>> 
>>>>> at java.lang.ClassLoader.loadLibrary(ClassLoader.java:1864)
>>>>> 
>>>>> at java.lang.Runtime.loadLibrary0(Runtime.java:870)
>>>>> 
>>>>> at java.lang.System.loadLibrary(System.java:1122)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.xerial.snappy.SnappyNativeLoader.loadLibrary(SnappyNativeLoader.java:52)
>>>>> 
>>>>> ... 49 more
>>>>> 
>>>>> Exception in thread "main" java.lang.reflect.InvocationTargetException
>>>>> 
>>>>> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
>> Method)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>>>>> 
>>>>> at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.io.CompressionCodec$.createCodec(CompressionCodec.scala:72)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.io.CompressionCodec$.createCodec(CompressionCodec.scala:65)
>>>>> 
>>>>> at org.apache.spark.broadcast.TorrentBroadcast.org
>>>>> 
>>>> 
>> $apache$spark$broadcast$TorrentBroadcast$$setConf(TorrentBroadcast.scala:73)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.broadcast.TorrentBroadcast.<init>(TorrentBroadcast.scala:80)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:34)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:63)
>>>>> 
>>>>> at org.apache.spark.SparkContext.broadcast(SparkContext.scala:1326)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:1014)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:1011)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
>>>>> 
>>>>> at org.apache.spark.SparkContext.withScope(SparkContext.scala:714)
>>>>> 
>>>>> at org.apache.spark.SparkContext.hadoopFile(SparkContext.scala:1011)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:832)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:830)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
>>>>> 
>>>>> at org.apache.spark.SparkContext.withScope(SparkContext.scala:714)
>>>>> 
>>>>> at org.apache.spark.SparkContext.textFile(SparkContext.scala:830)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.drivers.TDIndexedDatasetReader$class.elementReader(TextDelimitedReaderWriter.scala:61)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.drivers.TextDelimitedIndexedDatasetReader.elementReader(TextDelimitedReaderWriter.scala:304)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.drivers.TextDelimitedIndexedDatasetReader.elementReader(TextDelimitedReaderWriter.scala:304)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.math.indexeddataset.Reader$class.readElementsFrom(ReaderWriter.scala:75)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.drivers.TextDelimitedIndexedDatasetReader.readElementsFrom(TextDelimitedReaderWriter.scala:304)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.sparkbindings.SparkEngine$.indexedDatasetDFSReadElements(SparkEngine.scala:382)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.sparkbindings.SparkEngine$.indexedDatasetDFSReadElements(SparkEngine.scala:39)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.math.indexeddataset.package$.indexedDatasetDFSReadElements(package.scala:336)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.drivers.ItemSimilarityDriver$.readIndexedDatasets(ItemSimilarityDriver.scala:152)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.drivers.ItemSimilarityDriver$.process(ItemSimilarityDriver.scala:201)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.drivers.ItemSimilarityDriver$$anonfun$main$1.apply(ItemSimilarityDriver.scala:112)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.drivers.ItemSimilarityDriver$$anonfun$main$1.apply(ItemSimilarityDriver.scala:110)
>>>>> 
>>>>> at scala.Option.map(Option.scala:145)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.drivers.ItemSimilarityDriver$.main(ItemSimilarityDriver.scala:110)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.mahout.drivers.ItemSimilarityDriver.main(ItemSimilarityDriver.scala)
>>>>> 
>>>>> Caused by: java.lang.IllegalArgumentException:
>>>>> org.xerial.snappy.SnappyError: [FAILED_TO_LOAD_NATIVE_LIBRARY] null
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.io.SnappyCompressionCodec$.liftedTree1$1(CompressionCodec.scala:171)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.io.SnappyCompressionCodec$.org$apache$spark$io$SnappyCompressionCodec$$version$lzycompute(CompressionCodec.scala:168)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.io.SnappyCompressionCodec$.org$apache$spark$io$SnappyCompressionCodec$$version(CompressionCodec.scala:168)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.io.SnappyCompressionCodec.<init>(CompressionCodec.scala:152)
>>>>> 
>>>>> ... 38 more
>>>>> 
>>>>> Caused by: org.xerial.snappy.SnappyError:
>> [FAILED_TO_LOAD_NATIVE_LIBRARY]
>>>>> null
>>>>> 
>>>>> at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:229)
>>>>> 
>>>>> at org.xerial.snappy.Snappy.<clinit>(Snappy.java:44)
>>>>> 
>>>>> at
>>>>> 
>>>> 
>> org.apache.spark.io.SnappyCompressionCodec$.liftedTree1$1(CompressionCodec.scala:169)
>>>>> 
>>>>> ... 41 more
>>>>> 
>>>>> 16/04/14 10:53:13 INFO SparkContext: Invoking stop() from shutdown hook
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
>>>>> o.s.j.s.ServletContextHandler{/metrics/json,null}
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
>>>>> o.s.j.s.ServletContextHandler{/stages/stage/kill,null}
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
>>>>> o.s.j.s.ServletContextHandler{/api,null}
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
>>>>> o.s.j.s.ServletContextHandler{/,null}
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
>>>>> o.s.j.s.ServletContextHandler{/static,null}
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
>>>>> o.s.j.s.ServletContextHandler{/executors/threadDump/json,null}
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
>>>>> o.s.j.s.ServletContextHandler{/executors/threadDump,null}
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
>>>>> o.s.j.s.ServletContextHandler{/executors/json,null}
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
>>>>> o.s.j.s.ServletContextHandler{/executors,null}
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
>>>>> o.s.j.s.ServletContextHandler{/environment/json,null}
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
>>>>> o.s.j.s.ServletContextHandler{/environment,null}
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
>>>>> o.s.j.s.ServletContextHandler{/storage/rdd/json,null}
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
>>>>> o.s.j.s.ServletContextHandler{/storage/rdd,null}
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
>>>>> o.s.j.s.ServletContextHandler{/storage/json,null}
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
>>>>> o.s.j.s.ServletContextHandler{/storage,null}
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
>>>>> o.s.j.s.ServletContextHandler{/stages/pool/json,null}
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
>>>>> o.s.j.s.ServletContextHandler{/stages/pool,null}
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
>>>>> o.s.j.s.ServletContextHandler{/stages/stage/json,null}
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
>>>>> o.s.j.s.ServletContextHandler{/stages/stage,null}
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
>>>>> o.s.j.s.ServletContextHandler{/stages/json,null}
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
>>>>> o.s.j.s.ServletContextHandler{/stages,null}
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
>>>>> o.s.j.s.ServletContextHandler{/jobs/job/json,null}
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
>>>>> o.s.j.s.ServletContextHandler{/jobs/job,null}
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
>>>>> o.s.j.s.ServletContextHandler{/jobs/json,null}
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
>>>>> o.s.j.s.ServletContextHandler{/jobs,null}
>>>>> 
>>>>> 16/04/14 10:53:13 INFO SparkUI: Stopped Spark web UI at
>>>>> http://192.168.1.122:4040
>>>>> 
>>>>> 16/04/14 10:53:13 INFO MapOutputTrackerMasterEndpoint:
>>>>> MapOutputTrackerMasterEndpoint stopped!
>>>>> 
>>>>> 16/04/14 10:53:13 INFO MemoryStore: MemoryStore cleared
>>>>> 
>>>>> 16/04/14 10:53:13 INFO BlockManager: BlockManager stopped
>>>>> 
>>>>> 16/04/14 10:53:13 INFO BlockManagerMaster: BlockManagerMaster stopped
>>>>> 
>>>>> 16/04/14 10:53:13 INFO
>>>>> OutputCommitCoordinator$OutputCommitCoordinatorEndpoint:
>>>>> OutputCommitCoordinator stopped!
>>>>> 
>>>>> 16/04/14 10:53:13 INFO SparkContext: Successfully stopped SparkContext
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ShutdownHookManager: Shutdown hook called
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ShutdownHookManager: Deleting directory
>>>>> 
>>>> 
>> /private/var/folders/17/vhrqnx4x1n10yqbl1vnch9000000gn/T/spark-ad6a714b-6dcb-410c-aef3-2b098695743b/httpd-11f32419-5f8a-4b61-9654-9ea2d32cab5d
>>>>> 
>>>>> 16/04/14 10:53:13 INFO RemoteActorRefProvider$RemotingTerminator:
>>>> Shutting
>>>>> down remote daemon.
>>>>> 
>>>>> 16/04/14 10:53:13 INFO ShutdownHookManager: Deleting directory
>>>>> 
>>>> 
>> /private/var/folders/17/vhrqnx4x1n10yqbl1vnch9000000gn/T/spark-ad6a714b-6dcb-410c-aef3-2b098695743b
>>>>> 
>>>>> --
>>>>> Thanks & Regards,
>>>>> 
>>>>> *Rohit Jain*
>>>>> Web developer | Consultant
>>>>> Mob +91 8097283931
>>>> 
>>>> 
>>> 
>>> 
>>> --
>>> Thanks & Regards,
>>> 
>>> *Rohit Jain*
>>> Web developer | Consultant
>>> Mob +91 8097283931
>> 
>> 
> 
> 
> -- 
> Thanks & Regards,
> 
> *Rohit Jain*
> Web developer | Consultant
> Mob +91 8097283931

Reply via email to