Re: BUILD FAILURE for Scala 2.11?

2016-01-06 Thread Jacek Laskowski
Hi, Done. See https://github.com/apache/spark/pull/10636 Pozdrawiam, Jacek Jacek Laskowski | https://medium.com/@jaceklaskowski/ Mastering Apache Spark ==> https://jaceklaskowski.gitbooks.io/mastering-apache-spark/ Follow me at https://twitter.com/jaceklaskowski On Thu, Jan 7, 2016 at 8:10 AM,

Re: java.io.FileNotFoundException(Too many open files) in Spark streaming

2016-01-06 Thread Priya Ch
The line of code which I highlighted in the screenshot is within the spark source code. Spark implements sort-based shuffle implementation and the spilled files are merged using the merge sort. Here is the link https://issues.apache.org/jira/secure/attachment/12655884/Sort-basedshuffledesign.pdf w

Re: GraphX does not unpersist RDDs

2016-01-06 Thread Alexander Pivovarov
The same issue exists in spark-1.6.0 I've opened Jira ticket for that https://issues.apache.org/jira/browse/SPARK-12655 On Mon, Jan 4, 2016 at 11:30 PM, Alexander Pivovarov wrote: > // open spark-shell 1.5.2 > // run > > import org.apache.spark.graphx._ > > val vert = sc.parallelize(List((1L, 1

Spark 2.x Java API issues: Optional, and Iterator/Iterable

2016-01-06 Thread Sean Owen
Here are two interesting issues (with PRs) concerning Java APIs for Spark 2.x. Details and discussion inside, and comments requested. https://issues.apache.org/jira/browse/SPARK-3369 This concerns fixing Iterator/Iterable problems in some Java API methods, to make them consistent with Scala. Thi

Re: java.io.FileNotFoundException(Too many open files) in Spark streaming

2016-01-06 Thread Hamel Kothari
If you're filling up the number of open files, odds are there's one code path that's opening most of these files. If that's the case, these files will likely be named similarly and easy to pick out if you just sort the output of "lsof" once you find the group that is clearly the largest, you can th

Re: java.io.FileNotFoundException(Too many open files) in Spark streaming

2016-01-06 Thread Priya Ch
Running 'lsof' will let us know the open files but how do we come to know the root cause behind opening too many files. Thanks, Padma CH On Wed, Jan 6, 2016 at 8:39 AM, Hamel Kothari wrote: > The "Too Many Files" part of the exception is just indicative of the fact > that when that call was mad