Hi,

We have a streaming application that read microbatch csv files and involves the foreachBatch call.  Each microbatch can be processed independently.  I noticed that only one worker node is being utilized.  Is there anyway or any explicit method to distribute the batch work load to multiple workers?  I would think Spark would execute foreachBatch method on different workers since each batch can be treated as atomic?

Thanks!

ND


---------------------------------------------------------------------
To unsubscribe e-mail: user-unsubscr...@spark.apache.org

Reply via email to