Hi Jianshi,

I've met similar situation before.
And my solution was 'ulimit', you can use

-a to see your current settings
-n to set open files limit
(and other limits also)

And I set -n to 10240.

I see spark.shuffle.consolidateFiles helps by reusing open files.
(so I don't know to what extend does it help)

Hope it helps.

Larry

On 7/30/14, 4:01 PM, Jianshi Huang wrote:
I'm using Spark 1.0.1 on Yarn-Client mode.

SortByKey always reports a FileNotFoundExceptions with messages says "too many open files".

I already set spark.shuffle.consolidateFiles to true:

  conf.set("spark.shuffle.consolidateFiles", "true")

But it seems not working. What are the other possible reasons? How to fix it?

Jianshi

--
Jianshi Huang

LinkedIn: jianshi
Twitter: @jshuang
Github & Blog: http://huangjs.github.com/

Reply via email to