>From log file I noticed that the ExecutorLostFailure happens after the
memory used by Executor becomes more than the Executor memory value.
However, even if I increase the value of Executor Memory the Executor fails
- only that it takes longer time.

I'm wondering that for joining 2 Hive tables, one with 100 MB data (around
1 M rows) and another with 20 KB data (around 100 rows) why an executor is
consuming so much of memory. Even if I increase the memory to 20 GB. The
same failure happens.

Regards,
Sourav

On Tue, Jun 9, 2015 at 12:58 PM, Sourav Mazumder <
sourav.mazumde...@gmail.com> wrote:

> Hi,
>
> I'm just doing a select statement which is supposed to return 10 MB data
> maximum. The driver memory is 2G and executor memory is 20 G.
>
> The query I'm trying to run is something like
>
> SELECT PROJECT_LIVE_DT, FLOORPLAN_NM, FLOORPLAN_DB_KEY
> FROM POG_PRE_EXT P, PROJECT_CALENDAR_EXT C
> WHERE PROJECT_TYPE = 'CR'
>
> Not sure what exactly you mean by physical plan.
>
> Here is he stack trace from the machine where the thrift process is
> running.
>
> Regards,
> Sourav
>
> On Mon, Jun 8, 2015 at 11:18 PM, Cheng, Hao <hao.ch...@intel.com> wrote:
>
>>  Is it the large result set return from the Thrift Server? And can you
>> paste the SQL and physical plan?
>>
>>
>>
>> *From:* Ted Yu [mailto:yuzhih...@gmail.com]
>> *Sent:* Tuesday, June 9, 2015 12:01 PM
>> *To:* Sourav Mazumder
>> *Cc:* user
>> *Subject:* Re: Spark SQL with Thrift Server is very very slow and
>> finally failing
>>
>>
>>
>> Which Spark release are you using ?
>>
>>
>>
>> Can you pastebin the stack trace w.r.t. ExecutorLostFailure ?
>>
>>
>>
>> Thanks
>>
>>
>>
>> On Mon, Jun 8, 2015 at 8:52 PM, Sourav Mazumder <
>> sourav.mazumde...@gmail.com> wrote:
>>
>>      Hi,
>>
>> I am trying to run a SQL form a JDBC driver using Spark's Thrift Server.
>>
>> I'm doing a join between a Hive Table of size around 100 GB and another
>> Hive Table with 10 KB, with a filter on a particular column
>>
>> The query takes more than 45 minutes and then I get ExecutorLostFailure.
>> That is because of memory as once I increase the memory the failure happens
>> but after a long time.
>>
>> I'm having executor memory 20 GB, Spark DRiver Memory 2 GB, Executor
>> Instances 2 and Executor Core 2.
>>
>> Running the job using Yarn with master as 'yarn-client'.
>>
>> Any idea if I'm missing any other configuration ?
>>
>> Regards,
>>
>> Sourav
>>
>>
>>
>
>

Reply via email to