What value should I set there?
Currently I have set it as

spark.driver.memory  -Xms4096m -Xmx4096m -XX:MaxPermSize=2048m

But still same error

On Mar 26, 2017 1:19 PM, "Eric Charles" <e...@apache.org> wrote:

> You also have to check the memory you give to the spark driver
> (spark.driver.memory property)
>
> On 26/03/17 07:40, RUSHIKESH RAUT wrote:
>
>> Yes I know it inevitable if the data is large. I want to know how do I
>> increase the interpreter memory to handle large data?
>>
>> Thanks,
>> Rushikesh Raut
>>
>> On Mar 26, 2017 8:56 AM, "Jianfeng (Jeff) Zhang" <jzh...@hortonworks.com
>> <mailto:jzh...@hortonworks.com>> wrote:
>>
>>
>>     How large is your data ? This problem is inevitable if your data is
>>     too large, you can try to use spark data frame if that works for you.
>>
>>
>>
>>
>>
>>     Best Regard,
>>     Jeff Zhang
>>
>>
>>     From: RUSHIKESH RAUT <rushikeshraut...@gmail.com
>>     <mailto:rushikeshraut...@gmail.com>>
>>     Reply-To: "users@zeppelin.apache.org
>>     <mailto:users@zeppelin.apache.org>" <users@zeppelin.apache.org
>>     <mailto:users@zeppelin.apache.org>>
>>     Date: Saturday, March 25, 2017 at 5:06 PM
>>     To: "users@zeppelin.apache.org <mailto:users@zeppelin.apache.org>"
>>     <users@zeppelin.apache.org <mailto:users@zeppelin.apache.org>>
>>     Subject: Zeppelin out of memory issue - (GC overhead limit exceeded)
>>
>>     Hi everyone,
>>
>>     I am trying to load some data from hive table into my notebook and
>>     then convert this dataframe into r dataframe using spark.r
>>     interpreter. This works perfectly for small amount of data.
>>     But if the data is increased then it gives me error
>>
>>     java.lang.OutOfMemoryError: GC overhead limit exceeded
>>
>>     I have tried increasing the ZEPPELIN_MEM and ZEPPELIN_INTP_MEM in
>>     the zeppelin-env.cmd file but i am still facing this issue. I have
>>     used the following configuration
>>
>>     set ZEPPELIN_MEM="-Xms4096m -Xmx4096m -XX:MaxPermSize=2048m"
>>     set ZEPPELIN_INTP_MEM="-Xmx4096m -Xms4096m -XX:MaxPermSize=2048m"
>>
>>     I am sure that this much size should be sufficient for my data but
>>     still i am getting this same error. Any guidance will be much
>>     appreciated.
>>
>>     Thanks,
>>     Rushikesh Raut
>>
>>

Reply via email to