Could you provide a short script to reproduce this issue? On Tue, Jan 20, 2015 at 9:00 PM, TJ Klein <tjkl...@gmail.com> wrote: > Hi, > > I just recently tried to migrate from Spark 1.1 to Spark 1.2 - using > PySpark. Initially, I was super glad, noticing that Spark 1.2 is way faster > than Spark 1.1. However, the initial joy faded quickly when I noticed that > all my stuff didn't successfully terminate operations anymore. Using Spark > 1.1 it still works perfectly fine, though. > Specifically, the execution just freezes without any error output at one > point, when calling a joint map() and collect() statement (after having it > called many times successfully before in a loop). > > Any clue? Or do I have to wait for the next version? > > Best, > Tassilo > > > > -- > View this message in context: > http://apache-spark-user-list.1001560.n3.nabble.com/Spark-1-1-slow-working-Spark-1-2-fast-freezing-tp21278.html > Sent from the Apache Spark User List mailing list archive at Nabble.com. > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > For additional commands, e-mail: user-h...@spark.apache.org >
--------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org