and even the same process where the data might be cached.
these are the different locality levels:
PROCESS_LOCAL
NODE_LOCAL
RACK_LOCAL
ANY
relevant code:
https://github.com/apache/spark/blob/7712e724ad69dd0b83754e938e9799d13a4d43b9/core/src/test/scala/org/apache/spark/scheduler/TaskSetManagerSu
Hi Anish,
Spark, like MapReduce, makes an effort to schedule tasks on the same nodes
and racks that the input blocks reside on.
-Sandy
On Tue, Jul 8, 2014 at 12:27 PM, anishs...@yahoo.co.in <
anishs...@yahoo.co.in> wrote:
> Hi All
>
> My apologies for very basic question, do we have full suppo