Problem solved: for i in range(1,6): L=L.cartesian(D) L.unpersist() L=L.reduceByKey(min).coalesce(6).map(lambda (l,n):l).cache() L.collect() Number of partitions should be constant
-- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/cartesian-in-the-loop-runtime-grows-tp25303p26069.html Sent from the Apache Spark User List mailing list archive at Nabble.com. --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org