spark-itemsimilarity: Exception in thread "main" com.typesafe.config.ConfigException$Missing: No configuration setting found for key 'akka.event-handlers'

2015-01-27 Thread Kevin Zhang
Hi,

I'm new to Spark, Mahout. Just tried to run the spark-itemsimilarity but no 
luck. 

Here is what I did.

1. clone git Mahout git project
2. run "mvn install"
3. set $MAHOUT_HOME to the project path
4. added $MAHOUT_HOME/bin to the PATH variable
5. download Spark, extract it to a dir
6. set $SPARK_HOME to the dir /spark-1.2.0-bin-hadoop2.4
7. run command  "mahout spark-itemsimilarity -i ./mahout-input/order_item.tsv 
-o ./output -f1 purchase -f2 view -os -ic 2 -fc 1 -td ,"

Below is the output. It says "Exception in thread "main" 
com.typesafe.config.ConfigException$Missing: No configuration setting found for 
key 'akka.event-handlers'"

Any help or suggestion is highly appreciated.

Thanks,
Kevin

--
MAHOUT_LOCAL is set, so we don't add HADOOP_CONF_DIR to classpath.
log4j:WARN No appenders could be found for logger 
(org.apache.mahout.sparkbindings).
log4j:WARN Please initialize the log4j system properly.
log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more 
info.
Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in 
[jar:file:/Users/Zhang_Kevin/Documents/mine/big/projects/mahout/mrlegacy/target/mahout-mrlegacy-1.0-SNAPSHOT-job.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in 
[jar:file:/Users/Zhang_Kevin/Documents/mine/big/projects/mahout/spark/target/mahout-spark_2.10-1.0-SNAPSHOT-job.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in 
[jar:file:/Users/Zhang_Kevin/Applications/spark-1.2.0-bin-hadoop2.4/lib/spark-assembly-1.2.0-hadoop2.4.0.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
15/01/27 12:01:58 INFO SecurityManager: Changing view acls to: Zhang_Kevin,
15/01/27 12:01:58 INFO SecurityManager: Changing modify acls to: Zhang_Kevin,
15/01/27 12:01:58 INFO SecurityManager: SecurityManager: authentication 
disabled; ui acls disabled; users with view permissions: Set(Zhang_Kevin, ); 
users with modify permissions: Set(Zhang_Kevin, )
Exception in thread "main" com.typesafe.config.ConfigException$Missing: No 
configuration setting found for key 'akka.event-handlers'
at com.typesafe.config.impl.SimpleConfig.findKey(SimpleConfig.java:115)
at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:136)
at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:142)
at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:150)
at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:155)
at com.typesafe.config.impl.SimpleConfig.getList(SimpleConfig.java:203)
at 
com.typesafe.config.impl.SimpleConfig.getHomogeneousUnwrappedList(SimpleConfig.java:260)
at com.typesafe.config.impl.SimpleConfig.getStringList(SimpleConfig.java:318)
at akka.actor.ActorSystem$Settings.(ActorSystem.scala:150)
at akka.actor.ActorSystemImpl.(ActorSystem.scala:470)
at akka.actor.ActorSystem$.apply(ActorSystem.scala:111)
at akka.actor.ActorSystem$.apply(ActorSystem.scala:104)
at 
org.apache.spark.util.AkkaUtils$.org$apache$spark$util$AkkaUtils$$doCreateActorSystem(AkkaUtils.scala:121)
at org.apache.spark.util.AkkaUtils$$anonfun$1.apply(AkkaUtils.scala:54)
at org.apache.spark.util.AkkaUtils$$anonfun$1.apply(AkkaUtils.scala:53)
at 
org.apache.spark.util.Utils$$anonfun$startServiceOnPort$1.apply$mcVI$sp(Utils.scala:1446)
at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:141)
at org.apache.spark.util.Utils$.startServiceOnPort(Utils.scala:1442)
at org.apache.spark.util.AkkaUtils$.createActorSystem(AkkaUtils.scala:56)
at org.apache.spark.SparkEnv$.create(SparkEnv.scala:153)
at org.apache.spark.SparkContext.(SparkContext.scala:203)
at org.apache.mahout.sparkbindings.package$.mahoutSparkContext(package.scala:95)
at org.apache.mahout.drivers.MahoutSparkDriver.start(MahoutSparkDriver.scala:81)
at 
org.apache.mahout.drivers.ItemSimilarityDriver$.start(ItemSimilarityDriver.scala:128)
at 
org.apache.mahout.drivers.ItemSimilarityDriver$.process(ItemSimilarityDriver.scala:211)
at 
org.apache.mahout.drivers.ItemSimilarityDriver$$anonfun$main$1.apply(ItemSimilarityDriver.scala:116)
at 
org.apache.mahout.drivers.ItemSimilarityDriver$$anonfun$main$1.apply(ItemSimilarityDriver.scala:114)
at scala.Option.map(Option.scala:145)
at 
org.apache.mahout.drivers.ItemSimilarityDriver$.main(ItemSimilarityDriver.scala:114)
at 
org.apache.mahout.drivers.ItemSimilarityDriver.main(ItemSimilarityDriver.scala)


Re: spark-itemsimilarity: Exception in thread "main" com.typesafe.config.ConfigException$Missing: No configuration setting found for key 'akka.event-handlers'

2015-01-27 Thread Dmitriy Lyubimov
spark 1.2 not supported (yet). current head runs on 1.1.0 (but i guess you
can take a pull request #71 and compile it for 1.1.1 too, and perhaps even
1.2)

On Tue, Jan 27, 2015 at 12:04 PM, Kevin Zhang <
zhangyongji...@yahoo.com.invalid> wrote:

> Hi,
>
> I'm new to Spark, Mahout. Just tried to run the spark-itemsimilarity but
> no luck.
>
> Here is what I did.
>
> 1. clone git Mahout git project
> 2. run "mvn install"
> 3. set $MAHOUT_HOME to the project path
> 4. added $MAHOUT_HOME/bin to the PATH variable
> 5. download Spark, extract it to a dir
> 6. set $SPARK_HOME to the dir /spark-1.2.0-bin-hadoop2.4
> 7. run command  "mahout spark-itemsimilarity -i
> ./mahout-input/order_item.tsv -o ./output -f1 purchase -f2 view -os -ic 2
> -fc 1 -td ,"
>
> Below is the output. It says "Exception in thread "main"
> com.typesafe.config.ConfigException$Missing: No configuration setting found
> for key 'akka.event-handlers'"
>
> Any help or suggestion is highly appreciated.
>
> Thanks,
> Kevin
>
>
> --
> MAHOUT_LOCAL is set, so we don't add HADOOP_CONF_DIR to classpath.
> log4j:WARN No appenders could be found for logger
> (org.apache.mahout.sparkbindings).
> log4j:WARN Please initialize the log4j system properly.
> log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for
> more info.
> Using Spark's default log4j profile:
> org/apache/spark/log4j-defaults.properties
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in
> [jar:file:/Users/Zhang_Kevin/Documents/mine/big/projects/mahout/mrlegacy/target/mahout-mrlegacy-1.0-SNAPSHOT-job.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in
> [jar:file:/Users/Zhang_Kevin/Documents/mine/big/projects/mahout/spark/target/mahout-spark_2.10-1.0-SNAPSHOT-job.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in
> [jar:file:/Users/Zhang_Kevin/Applications/spark-1.2.0-bin-hadoop2.4/lib/spark-assembly-1.2.0-hadoop2.4.0.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
> explanation.
> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
> 15/01/27 12:01:58 INFO SecurityManager: Changing view acls to: Zhang_Kevin,
> 15/01/27 12:01:58 INFO SecurityManager: Changing modify acls to:
> Zhang_Kevin,
> 15/01/27 12:01:58 INFO SecurityManager: SecurityManager: authentication
> disabled; ui acls disabled; users with view permissions: Set(Zhang_Kevin,
> ); users with modify permissions: Set(Zhang_Kevin, )
> Exception in thread "main" com.typesafe.config.ConfigException$Missing: No
> configuration setting found for key 'akka.event-handlers'
> at com.typesafe.config.impl.SimpleConfig.findKey(SimpleConfig.java:115)
> at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:136)
> at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:142)
> at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:150)
> at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:155)
> at com.typesafe.config.impl.SimpleConfig.getList(SimpleConfig.java:203)
> at
> com.typesafe.config.impl.SimpleConfig.getHomogeneousUnwrappedList(SimpleConfig.java:260)
> at
> com.typesafe.config.impl.SimpleConfig.getStringList(SimpleConfig.java:318)
> at akka.actor.ActorSystem$Settings.(ActorSystem.scala:150)
> at akka.actor.ActorSystemImpl.(ActorSystem.scala:470)
> at akka.actor.ActorSystem$.apply(ActorSystem.scala:111)
> at akka.actor.ActorSystem$.apply(ActorSystem.scala:104)
> at
> org.apache.spark.util.AkkaUtils$.org$apache$spark$util$AkkaUtils$$doCreateActorSystem(AkkaUtils.scala:121)
> at org.apache.spark.util.AkkaUtils$$anonfun$1.apply(AkkaUtils.scala:54)
> at org.apache.spark.util.AkkaUtils$$anonfun$1.apply(AkkaUtils.scala:53)
> at
> org.apache.spark.util.Utils$$anonfun$startServiceOnPort$1.apply$mcVI$sp(Utils.scala:1446)
> at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:141)
> at org.apache.spark.util.Utils$.startServiceOnPort(Utils.scala:1442)
> at org.apache.spark.util.AkkaUtils$.createActorSystem(AkkaUtils.scala:56)
> at org.apache.spark.SparkEnv$.create(SparkEnv.scala:153)
> at org.apache.spark.SparkContext.(SparkContext.scala:203)
> at
> org.apache.mahout.sparkbindings.package$.mahoutSparkContext(package.scala:95)
> at
> org.apache.mahout.drivers.MahoutSparkDriver.start(MahoutSparkDriver.scala:81)
> at
> org.apache.mahout.drivers.ItemSimilarityDriver$.start(ItemSimilarityDriver.scala:128)
> at
> org.apache.mahout.drivers.ItemSimilarityDriver$.process(ItemSimilarityDriver.scala:211)
> at
> org.apache.mahout.drivers.ItemSimilarityDriver$$anonfun$main$1.apply(ItemSimilarityDriver.scala:116)
> at
> org.apache.mahout.drivers.ItemSimilarityDriver$$anonfun$main$1.apply(ItemSimilarityDriver.scala:114)
> at scala.Option.map(Option.scala:145)
> at
> org.apache.mahout.drivers.ItemSimilarityDriver$.main(ItemSimilarityDri

java.lang.UnsatisfiedLinkError: no snappyjava in java.library.path

2015-01-27 Thread Kevin Zhang
Thanks to Dmitriy for answering my previous question regarding the Spark 
version. I just downgraded the version to spark-1.1.0-bin-hadoop2.4 and run my 
commane "mahout spark-itemsimilarity -i ./mahout-input/order_item.tsv -o 
./output -f1 purchase -f2 view -os -ic 2 -fc 1 -td ," again. This time I got 
error "java.lang.UnsatisfiedLinkError: no snappyjava in java.library.path" as 
attached. I'm using Mac.

Thanks for the help
-Kevin


at 
org.apache.spark.shuffle.hash.HashShuffleWriter.write(HashShuffleWriter.scala:65)

at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
at org.apache.spark.scheduler.Task.run(Task.scala:54)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:177)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.UnsatisfiedLinkError: no snappyjava in java.library.path
at java.lang.ClassLoader.loadLibrary(ClassLoader.java:1886)
at java.lang.Runtime.loadLibrary0(Runtime.java:849)
at java.lang.System.loadLibrary(System.java:1088)
at org.xerial.snappy.SnappyNativeLoader.loadLibrary(SnappyNativeLoader.java:52)
... 26 more
15/01/27 14:54:24 ERROR Executor: Exception in task 0.0 in stage 1.0 (TID 0)
org.xerial.snappy.SnappyError: [FAILED_TO_LOAD_NATIVE_LIBRARY] null
at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:229)
at org.xerial.snappy.Snappy.(Snappy.java:44)
at org.xerial.snappy.SnappyOutputStream.(SnappyOutputStream.java:79)
at 
org.apache.spark.io.SnappyCompressionCodec.compressedOutputStream(CompressionCodec.scala:125)
at 
org.apache.spark.storage.BlockManager.wrapForCompression(BlockManager.scala:1029)
at 
org.apache.spark.storage.BlockManager$$anonfun$8.apply(BlockManager.scala:608)
at 
org.apache.spark.storage.BlockManager$$anonfun$8.apply(BlockManager.scala:608)

Re: java.lang.UnsatisfiedLinkError: no snappyjava in java.library.path

2015-01-27 Thread Dmitriy Lyubimov
This looks like hadoop or spark -specific thing (snappy codec is used by
spark by default). There should be a way to disable this to a more
palatable library but you will need to investigate it a little bit since i
don't think anybody here knows mac specifics.

Better yet is to figure how to install native snappy codec on your Mac.
There should be a way. Ask on spark list.

To switch codec to something else you may try to get something like this
into system properties of the driver process:
-Dspark.io.compression.codec=lzf

Now, normally it would be as easy as adding it to MAHOUT_OPTS environment;
however, i think our current head is broken w.r.t. MAHOUT_OPTS for spark
processes (i have a fix for it elsewhere but not in public branch). So if
you decide to switch codec, you may need hack bin/mahout script a little,
not sure.





On Tue, Jan 27, 2015 at 3:07 PM, Kevin Zhang <
zhangyongji...@yahoo.com.invalid> wrote:

> Thanks to Dmitriy for answering my previous question regarding the Spark
> version. I just downgraded the version to spark-1.1.0-bin-hadoop2.4 and run
> my commane "mahout spark-itemsimilarity -i ./mahout-input/order_item.tsv -o
> ./output -f1 purchase -f2 view -os -ic 2 -fc 1 -td ," again. This time I
> got error "java.lang.UnsatisfiedLinkError: no snappyjava in
> java.library.path" as attached. I'm using Mac.
>
> Thanks for the help
> -Kevin
>
>
> at
> org.apache.spark.shuffle.hash.HashShuffleWriter.write(HashShuffleWriter.scala:65)
>
> at
> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68)
> at
> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
> at org.apache.spark.scheduler.Task.run(Task.scala:54)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:177)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> Caused by: java.lang.UnsatisfiedLinkError: no snappyjava in
> java.library.path
> at java.lang.ClassLoader.loadLibrary(ClassLoader.java:1886)
> at java.lang.Runtime.loadLibrary0(Runtime.java:849)
> at java.lang.System.loadLibrary(System.java:1088)
> at
> org.xerial.snappy.SnappyNativeLoader.loadLibrary(SnappyNativeLoader.java:52)
> ... 26 more
> 15/01/27 14:54:24 ERROR Executor: Exception in task 0.0 in stage 1.0 (TID
> 0)
> org.xerial.snappy.SnappyError: [FAILED_TO_LOAD_NATIVE_LIBRARY] null
> at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:229)
> at org.xerial.snappy.Snappy.(Snappy.java:44)
> at org.xerial.snappy.SnappyOutputStream.(SnappyOutputStream.java:79)
> at
> org.apache.spark.io.SnappyCompressionCodec.compressedOutputStream(CompressionCodec.scala:125)
> at
> org.apache.spark.storage.BlockManager.wrapForCompression(BlockManager.scala:1029)
> at
> org.apache.spark.storage.BlockManager$$anonfun$8.apply(BlockManager.scala:608)
> at
> org.apache.spark.storage.BlockManager$$anonfun$8.apply(BlockManager.scala:608)