Yes, I have increased the driver memory in spark-default.conf to 2g. Still
the error persists.

On Tue, Jan 20, 2015 at 10:18 AM, Ted Yu <yuzhih...@gmail.com> wrote:

> Have you seen these threads ?
>
> http://search-hadoop.com/m/JW1q5tMFlb
> http://search-hadoop.com/m/JW1q5dabji1
>
> Cheers
>
> On Mon, Jan 19, 2015 at 8:33 PM, Deep Pradhan <pradhandeep1...@gmail.com>
> wrote:
>
>> Hi Ted,
>> When I am running the same job with small data, I am able to run. But
>> when I run it with relatively bigger set of data, it is giving me
>> OutOfMemoryError: GC overhead limit exceeded.
>> The first time I run the job, no output. When I run for second time, I am
>> getting this error. I am aware that, the memory is getting full, but is
>> there any way to avoid this?
>> I have a single node Spark cluster.
>>
>> Thank You
>>
>> On Tue, Jan 20, 2015 at 9:52 AM, Deep Pradhan <pradhandeep1...@gmail.com>
>> wrote:
>>
>>> I had the Spark Shell running through out. Is it because of that?
>>>
>>> On Tue, Jan 20, 2015 at 9:47 AM, Ted Yu <yuzhih...@gmail.com> wrote:
>>>
>>>> Was there another instance of Spark running on the same machine ?
>>>>
>>>> Can you pastebin the full stack trace ?
>>>>
>>>> Cheers
>>>>
>>>> On Mon, Jan 19, 2015 at 8:11 PM, Deep Pradhan <
>>>> pradhandeep1...@gmail.com> wrote:
>>>>
>>>>> Hi,
>>>>> I am running a Spark job. I get the output correctly but when I see
>>>>> the logs file I see the following:
>>>>> AbstractLifeCycle: FAILED.....: java.net.BindException: Address
>>>>> already in use...
>>>>>
>>>>> What could be the reason for this?
>>>>>
>>>>> Thank You
>>>>>
>>>>
>>>>
>>>
>>
>

Reply via email to