Re: Memory issue in pyspark for 1.6 mb file

2017-06-17 Thread Pralabh Kumar
Hi Naga Is it failing because of driver memory full or executor memory full ? can you please try setting this property spark.cleaner.ttl ? . So that older RDDs /metadata should also get clear automatically. Can you please provide the complete error stacktrace and code snippet ?. Regards Prala

Memory issue in pyspark for 1.6 mb file

2017-06-17 Thread Naga Guduru
Hi, I am trying to load 1.6 mb excel file which has 16 tabs. We converted excel to csv and loaded 16 csv files to 8 tables. Job was running successful in 1st run in pyspark. When trying to run the same job 2 time, container getting killed due to memory issues. I am using unpersist and clearcache