My settings are: Running Spark 2.1 on 3 node YARN cluster with 160 GB.
Dynamic allocation turned on. spark.executor.memory=6G,
spark.executor.cores=6
First, I am reading hive tables: orders(329MB) and lineitems(1.43GB) and
doing left outer join.
Next, I apply 7 different filter conditions based on
Spark 2.1
My settings are: Running Spark 2.1 on 3 node YARN cluster with 160 GB. Dynamic
allocation turned on. spark.executor.memory=6G, spark.executor.cores=6
First, I am reading hive tables: orders(329MB) and lineitems(1.43GB) and doing
left outer join.Next, I apply 7 different filter condition