You can try increasing the parallelism, can you be more specific about the task that you are doing? May be pasting the piece of code would help. On 18 Jan 2015 13:22, "Deep Pradhan" <pradhandeep1...@gmail.com> wrote:
> The error in the log file says: > > *java.lang.OutOfMemoryError: GC overhead limit exceeded* > > with certain task ID and the error repeats for further task IDs. > > What could be the problem? > > On Sun, Jan 18, 2015 at 2:45 PM, Deep Pradhan <pradhandeep1...@gmail.com> > wrote: > >> Updating the Spark version means setting up the entire cluster once more? >> Or can we update it in some other way? >> >> On Sat, Jan 17, 2015 at 3:22 PM, Akhil Das <ak...@sigmoidanalytics.com> >> wrote: >> >>> Can you paste the code? Also you can try updating your spark version. >>> >>> Thanks >>> Best Regards >>> >>> On Sat, Jan 17, 2015 at 2:40 PM, Deep Pradhan <pradhandeep1...@gmail.com >>> > wrote: >>> >>>> Hi, >>>> I am using Spark-1.0.0 in a single node cluster. When I run a job with >>>> small data set it runs perfectly but when I use a data set of 350 KB, no >>>> output is being produced and when I try to run it the second time it is >>>> giving me an exception telling that SparkContext was shut down. >>>> Can anyone help me on this? >>>> >>>> Thank You >>>> >>> >>> >> >