You don't have to set spark.driver.memory with -X... but simply with
memory size.
Look at http://spark.apache.org/docs/latest/configuration.html
spark.driver.memory 1g Amount of memory to use for the driver process,
i.e. where SparkContext is initialized. (e.g. 1g, 2g).
Note: In client mode, this config must not be set through the SparkConf
directly in your application, because the driver JVM has already started
at that point. Instead, please set this through the --driver-memory
command line option or in your default properties file.
On 26/03/17 09:57, RUSHIKESH RAUT wrote:
What value should I set there?
Currently I have set it as
spark.driver.memory -Xms4096m -Xmx4096m -XX:MaxPermSize=2048m
But still same error
On Mar 26, 2017 1:19 PM, "Eric Charles" <e...@apache.org
<mailto:e...@apache.org>> wrote:
You also have to check the memory you give to the spark driver
(spark.driver.memory property)
On 26/03/17 07:40, RUSHIKESH RAUT wrote:
Yes I know it inevitable if the data is large. I want to know
how do I
increase the interpreter memory to handle large data?
Thanks,
Rushikesh Raut
On Mar 26, 2017 8:56 AM, "Jianfeng (Jeff) Zhang"
<jzh...@hortonworks.com <mailto:jzh...@hortonworks.com>
<mailto:jzh...@hortonworks.com <mailto:jzh...@hortonworks.com>>>
wrote:
How large is your data ? This problem is inevitable if your
data is
too large, you can try to use spark data frame if that works
for you.
Best Regard,
Jeff Zhang
From: RUSHIKESH RAUT <rushikeshraut...@gmail.com
<mailto:rushikeshraut...@gmail.com>
<mailto:rushikeshraut...@gmail.com
<mailto:rushikeshraut...@gmail.com>>>
Reply-To: "users@zeppelin.apache.org
<mailto:users@zeppelin.apache.org>
<mailto:users@zeppelin.apache.org
<mailto:users@zeppelin.apache.org>>" <users@zeppelin.apache.org
<mailto:users@zeppelin.apache.org>
<mailto:users@zeppelin.apache.org
<mailto:users@zeppelin.apache.org>>>
Date: Saturday, March 25, 2017 at 5:06 PM
To: "users@zeppelin.apache.org
<mailto:users@zeppelin.apache.org>
<mailto:users@zeppelin.apache.org
<mailto:users@zeppelin.apache.org>>"
<users@zeppelin.apache.org
<mailto:users@zeppelin.apache.org>
<mailto:users@zeppelin.apache.org
<mailto:users@zeppelin.apache.org>>>
Subject: Zeppelin out of memory issue - (GC overhead limit
exceeded)
Hi everyone,
I am trying to load some data from hive table into my
notebook and
then convert this dataframe into r dataframe using spark.r
interpreter. This works perfectly for small amount of data.
But if the data is increased then it gives me error
java.lang.OutOfMemoryError: GC overhead limit exceeded
I have tried increasing the ZEPPELIN_MEM and
ZEPPELIN_INTP_MEM in
the zeppelin-env.cmd file but i am still facing this issue.
I have
used the following configuration
set ZEPPELIN_MEM="-Xms4096m -Xmx4096m -XX:MaxPermSize=2048m"
set ZEPPELIN_INTP_MEM="-Xmx4096m -Xms4096m
-XX:MaxPermSize=2048m"
I am sure that this much size should be sufficient for my
data but
still i am getting this same error. Any guidance will be much
appreciated.
Thanks,
Rushikesh Raut