Hi , I am running pyspark in dockerized jupyter environment , I am constantly getting this error :
``` Py4JJavaError: An error occurred while calling z:org.apache.spark.api.python.PythonRDD.runJob. : org.apache.spark.SparkException: Job aborted due to stage failure: Task 33 in stage 25.0 failed 1 times, most recent failure: Lost task 33.0 in stage 25.0 (TID 35067, localhost, executor driver) : org.apache.hadoop.hdfs.BlockMissingException : Could not obtain block: BP-1742911633-10.225.201.50-1479296658503:blk_1233169822_159765693 ``` Please can anyone help me with how to handle such exception in pyspark. -- Best Regards *Divay Jindal*