Re: spark 1.1.0 (w/ hadoop 2.4) vs aws java sdk 1.7.2

2014-09-19 Thread Aniket
Looks like the same issue as http://mail-archives.apache.org/mod_mbox/spark-dev/201409.mbox/%3ccajob8btdxks-7-spjj5jmnw0xsnrjwdpcqqtjht1hun6j4z...@mail.gmail.com%3E On Sep 20, 2014 11:09 AM, "tian zhang [via Apache Spark Developers List]" < ml-node+s1001551n8481...@n3.nabble.com> wrote: > > > Hi,

spark 1.1.0 (w/ hadoop 2.4) vs aws java sdk 1.7.2

2014-09-19 Thread tian zhang
Hi, Spark experts, I have the following issue when using aws java sdk in my spark application. Here I narrowed down the following steps to reproduce the problem 1) I have Spark 1.1.0 with hadoop 2.4 installed on 3 nodes cluster 2) from the master node, I did the following steps. spark-shell --

guava version conflicts

2014-09-19 Thread Cody Koeninger
After the recent spark project changes to guava shading, I'm seeing issues with the datastax spark cassandra connector (which depends on guava 15.0) and the datastax cql driver (which depends on guava 16.0.1) Building an assembly for a job (with spark marked as provided) that includes either guava

Re: I want to contribute MLlib two quality measures(ARHR and HR) for top N recommendation system. Is this meaningful?

2014-09-19 Thread Debasish Das
Thanks Christoph. Are these numbers for mllib als implicit and explicit feedback on movielens/netflix datasets documented on JIRA ? On Sep 19, 2014 1:16 PM, "Christoph Sawade" < christoph.saw...@googlemail.com> wrote: > Hey Deb, > > NDCG is the "Normalized Discounted Cumulative Gain" [1]. Anothe

Re: I want to contribute MLlib two quality measures(ARHR and HR) for top N recommendation system. Is this meaningful?

2014-09-19 Thread Christoph Sawade
Hey Deb, NDCG is the "Normalized Discounted Cumulative Gain" [1]. Another popular measure is "Expected Reciprocal Rank" (ERR) [2]; it is based on a probabilistic user model, where the user scans the presented list of search results or recommendations and chooses the first that is sufficiently rele

Re: Gaussian Mixture Model clustering

2014-09-19 Thread Evan R. Sparks
Hey Meethu - what are you setting "K" to in the benchmarks you show? This can greatly affect the runtime. On Thu, Sep 18, 2014 at 10:38 PM, Meethu Mathew wrote: > Hi all, > Please find attached the image of benchmark results. The table in the > previous mail got messed up. Thanks. > > > > On Fr

Re: I want to contribute MLlib two quality measures(ARHR and HR) for top N recommendation system. Is this meaningful?

2014-09-19 Thread Debasish Das
Hi Xiangrui, Could you please point to some reference for calculating prec@k and ndcg@k ? prec is precision I suppose but ndcg I have no idea about... Thanks. Deb On Mon, Aug 25, 2014 at 12:28 PM, Xiangrui Meng wrote: > The evaluation metrics are definitely useful. How do they differ from >