Hi,
I was running a job (on Spark 1.5 + Yarn + java 8). In a stage that
lookup
(org.apache.spark.rdd.PairRDDFunctions.lookup(PairRDDFunctions.scala:873))
there was an executor that took the executor computing time > 6 times of
median. This executor had almost the same shuffle read size and low gc
time as others.
What can impact the executor computing time? Any suggestions what
parameters I should monitor/configure?
BR,
Patcharee
---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org