I'm getting a lot of task lost with this build in a large mesos cluster. Happens with both hash and sort shuffles.
14/11/20 18:08:38 WARN TaskSetManager: Lost task 9.1 in stage 1.0 (TID 897, i-d4d6553a.inst.aws.airbnb.com): FetchFailed(null, shuffleId=1, mapId=-1, reduceId=9, message= org.apache.spark.shuffle.MetadataFetchFailedException: Missing an output location for shuffle 1 at org.apache.spark.MapOutputTracker$$anonfun$org$apache$spark$MapOutputTracker$$convertMapStatuses$1.apply(MapOutputTracker.scala:386) at org.apache.spark.MapOutputTracker$$anonfun$org$apache$spark$MapOutputTracker$$convertMapStatuses$1.apply(MapOutputTracker.scala:383) at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244) at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244) at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33) at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:108) at scala.collection.TraversableLike$class.map(TraversableLike.scala:244) at scala.collection.mutable.ArrayOps$ofRef.map(ArrayOps.scala:108) at org.apache.spark.MapOutputTracker$.org$apache$spark$MapOutputTracker$$convertMapStatuses(MapOutputTracker.scala:382) at org.apache.spark.MapOutputTracker.getServerStatuses(MapOutputTracker.scala:178) at org.apache.spark.shuffle.hash.BlockStoreShuffleFetcher$.fetch(BlockStoreShuffleFetcher.scala:42) at org.apache.spark.shuffle.hash.HashShuffleReader.read(HashShuffleReader.scala:40) at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:92) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:263) at org.apache.spark.rdd.RDD.iterator(RDD.scala:230) On Thu, Nov 20, 2014 at 7:42 AM, Nan Zhu <zhunanmcg...@gmail.com> wrote: > BTW, this PR https://github.com/apache/spark/pull/2524 is related to a > blocker level bug, > > and this is actually close to be merged (have been reviewed for several > rounds) > > I would appreciated if anyone can continue the process, > > @mateiz > > -- > Nan Zhu > http://codingcat.me > > > On Thursday, November 20, 2014 at 10:17 AM, Corey Nolet wrote: > > > I was actually about to post this myself- I have a complex join that > could > > benefit from something like a GroupComparator vs having to do multiple > > grouyBy operations. This is probably the wrong thread for a full > discussion > > on this but I didn't see a JIRA ticket for this or anything similar- any > > reasons why this would not make sense given Spark's design? > > > > On Thu, Nov 20, 2014 at 9:39 AM, Madhu <ma...@madhu.com (mailto: > ma...@madhu.com)> wrote: > > > > > Thanks Patrick. > > > > > > I've been testing some 1.2 features, looks good so far. > > > I have some example code that I think will be helpful for certain > MR-style > > > use cases (secondary sort). > > > Can I still add that to the 1.2 documentation, or is that frozen at > this > > > point? > > > > > > > > > > > > ----- > > > -- > > > Madhu > > > https://www.linkedin.com/in/msiddalingaiah > > > -- > > > View this message in context: > > > > http://apache-spark-developers-list.1001551.n3.nabble.com/ANNOUNCE-Spark-1-2-0-Release-Preview-Posted-tp9400p9449.html > > > Sent from the Apache Spark Developers List mailing list archive at > > > Nabble.com (http://Nabble.com). > > > > > > --------------------------------------------------------------------- > > > To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org (mailto: > dev-unsubscr...@spark.apache.org) > > > For additional commands, e-mail: dev-h...@spark.apache.org (mailto: > dev-h...@spark.apache.org) > > > > > > > > > > > > > > -- Yee Yang Li Hector <http://google.com/+HectorYee> *google.com/+HectorYee <http://google.com/+HectorYee>*