Re: spark.shuffle.consolidateFiles seems not working

2014-08-01 Thread Jianshi Huang
file number is the same as basic hash-based shuffle. >>> >>> >>> >>> Thanks >>> >>> Jerry >>> >>> >>> >>> *From:* Jianshi Huang [mailto:jianshi.hu...@gmail.com] >>> *Sent:* Thursday, July 31, 2014 10:3

Re: spark.shuffle.consolidateFiles seems not working

2014-07-31 Thread Aaron Davidson
mailto:jianshi.hu...@gmail.com] >> *Sent:* Thursday, July 31, 2014 10:34 AM >> *To:* user@spark.apache.org >> *Cc:* xia...@sjtu.edu.cn >> *Subject:* Re: spark.shuffle.consolidateFiles seems not working >> >> >> >> Ok... but my question is why spark.shuffle.cons

Re: spark.shuffle.consolidateFiles seems not working

2014-07-31 Thread Jianshi Huang
> *To:* user@spark.apache.org > *Cc:* xia...@sjtu.edu.cn > *Subject:* Re: spark.shuffle.consolidateFiles seems not working > > > > Ok... but my question is why spark.shuffle.consolidateFiles is working > (or is it)? Is this a bug? > > > > On Wed, Jul 30, 2014 at 4:2

RE: spark.shuffle.consolidateFiles seems not working

2014-07-30 Thread Shao, Saisai
: user@spark.apache.org Cc: xia...@sjtu.edu.cn Subject: Re: spark.shuffle.consolidateFiles seems not working Ok... but my question is why spark.shuffle.consolidateFiles is working (or is it)? Is this a bug? On Wed, Jul 30, 2014 at 4:29 PM, Larry Xiao mailto:xia...@sjtu.edu.cn>> wrote: Hi J

Re: spark.shuffle.consolidateFiles seems not working

2014-07-30 Thread Jianshi Huang
Ok... but my question is why spark.shuffle.consolidateFiles is working (or is it)? Is this a bug? On Wed, Jul 30, 2014 at 4:29 PM, Larry Xiao wrote: > Hi Jianshi, > > I've met similar situation before. > And my solution was 'ulimit', you can use > > -a to see your current settings > -n to set o

Re: spark.shuffle.consolidateFiles seems not working

2014-07-30 Thread Larry Xiao
Hi Jianshi, I've met similar situation before. And my solution was 'ulimit', you can use -a to see your current settings -n to set open files limit (and other limits also) And I set -n to 10240. I see spark.shuffle.consolidateFiles helps by reusing open files. (so I don't know to what extend d

spark.shuffle.consolidateFiles seems not working

2014-07-30 Thread Jianshi Huang
I'm using Spark 1.0.1 on Yarn-Client mode. SortByKey always reports a FileNotFoundExceptions with messages says "too many open files". I already set spark.shuffle.consolidateFiles to true: conf.set("spark.shuffle.consolidateFiles", "true") But it seems not working. What are the other possible