Yes, I have increased the driver memory in spark-default.conf to 2g. Still the error persists.
On Tue, Jan 20, 2015 at 10:18 AM, Ted Yu <yuzhih...@gmail.com> wrote: > Have you seen these threads ? > > http://search-hadoop.com/m/JW1q5tMFlb > http://search-hadoop.com/m/JW1q5dabji1 > > Cheers > > On Mon, Jan 19, 2015 at 8:33 PM, Deep Pradhan <pradhandeep1...@gmail.com> > wrote: > >> Hi Ted, >> When I am running the same job with small data, I am able to run. But >> when I run it with relatively bigger set of data, it is giving me >> OutOfMemoryError: GC overhead limit exceeded. >> The first time I run the job, no output. When I run for second time, I am >> getting this error. I am aware that, the memory is getting full, but is >> there any way to avoid this? >> I have a single node Spark cluster. >> >> Thank You >> >> On Tue, Jan 20, 2015 at 9:52 AM, Deep Pradhan <pradhandeep1...@gmail.com> >> wrote: >> >>> I had the Spark Shell running through out. Is it because of that? >>> >>> On Tue, Jan 20, 2015 at 9:47 AM, Ted Yu <yuzhih...@gmail.com> wrote: >>> >>>> Was there another instance of Spark running on the same machine ? >>>> >>>> Can you pastebin the full stack trace ? >>>> >>>> Cheers >>>> >>>> On Mon, Jan 19, 2015 at 8:11 PM, Deep Pradhan < >>>> pradhandeep1...@gmail.com> wrote: >>>> >>>>> Hi, >>>>> I am running a Spark job. I get the output correctly but when I see >>>>> the logs file I see the following: >>>>> AbstractLifeCycle: FAILED.....: java.net.BindException: Address >>>>> already in use... >>>>> >>>>> What could be the reason for this? >>>>> >>>>> Thank You >>>>> >>>> >>>> >>> >> >