Hi, I have set driver memory 10 GB and job ran with intermediate failure which is recovered back by spark.
But I still what to know if no of parts increases git driver ram need to be increased and what is ration of no of parts/RAM. @RK : I am using cache on RDD. Is this reason of high RAM utilization. Thanks, Saurav Sinha On Tue, Jul 19, 2016 at 10:14 PM, RK Aduri <rkad...@collectivei.com> wrote: > Just want to see if this helps. > > Are you doing heavy collects and persist that? If that is so, you might > want to parallelize that collection by converting to an RDD. > > Thanks, > RK > > On Tue, Jul 19, 2016 at 12:09 AM, Saurav Sinha <sauravsinh...@gmail.com> > wrote: > >> Hi Mich, >> >> 1. In what mode are you running the spark standalone, yarn-client, >> yarn cluster etc >> >> Ans: spark standalone >> >> 1. You have 4 nodes with each executor having 10G. How many actual >> executors do you see in UI (Port 4040 by default) >> >> Ans: There are 4 executor on which am using 8 cores >> (--total-executor-core 32) >> >> 1. What is master memory? Are you referring to diver memory? May be I >> am misunderstanding this >> >> Ans: Driver memory is set as --drive-memory 5g >> >> 1. The only real correlation I see with the driver memory is when you >> are running in local mode where worker lives within JVM process that you >> start with spark-shell etc. In that case driver memory matters. However, >> it >> appears that you are running in another mode with 4 nodes? >> >> Ans: I am running my job as spark-submit and on my worker(executor) node >> there is no OOM issue ,it only happening on driver app. >> >> Thanks, >> Saurav Sinha >> >> On Tue, Jul 19, 2016 at 2:42 AM, Mich Talebzadeh < >> mich.talebza...@gmail.com> wrote: >> >>> can you please clarify: >>> >>> >>> 1. In what mode are you running the spark standalone, yarn-client, >>> yarn cluster etc >>> 2. You have 4 nodes with each executor having 10G. How many actual >>> executors do you see in UI (Port 4040 by default) >>> 3. What is master memory? Are you referring to diver memory? May be >>> I am misunderstanding this >>> 4. The only real correlation I see with the driver memory is when >>> you are running in local mode where worker lives within JVM process that >>> you start with spark-shell etc. In that case driver memory matters. >>> However, it appears that you are running in another mode with 4 nodes? >>> >>> Can you get a snapshot of your environment tab in UI and send the output >>> please? >>> >>> HTH >>> >>> >>> Dr Mich Talebzadeh >>> >>> >>> >>> LinkedIn * >>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw >>> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* >>> >>> >>> >>> http://talebzadehmich.wordpress.com >>> >>> >>> *Disclaimer:* Use it at your own risk. Any and all responsibility for >>> any loss, damage or destruction of data or any other property which may >>> arise from relying on this email's technical content is explicitly >>> disclaimed. The author will in no case be liable for any monetary damages >>> arising from such loss, damage or destruction. >>> >>> >>> >>> On 18 July 2016 at 11:50, Saurav Sinha <sauravsinh...@gmail.com> wrote: >>> >>>> I have set --drive-memory 5g. I need to understand that as no of >>>> partition increase drive-memory need to be increased. What will be >>>> best ration of No of partition/drive-memory. >>>> >>>> On Mon, Jul 18, 2016 at 4:07 PM, Zhiliang Zhu <zchl.j...@yahoo.com> >>>> wrote: >>>> >>>>> try to set --drive-memory xg , x would be as large as can be set . >>>>> >>>>> >>>>> On Monday, July 18, 2016 6:31 PM, Saurav Sinha < >>>>> sauravsinh...@gmail.com> wrote: >>>>> >>>>> >>>>> Hi, >>>>> >>>>> I am running spark job. >>>>> >>>>> Master memory - 5G >>>>> executor memort 10G(running on 4 node) >>>>> >>>>> My job is getting killed as no of partition increase to 20K. >>>>> >>>>> 16/07/18 14:53:13 INFO DAGScheduler: Got job 17 (foreachPartition at >>>>> WriteToKafka.java:45) with 13524 output partitions (allowLocal=false) >>>>> 16/07/18 14:53:13 INFO DAGScheduler: Final stage: ResultStage >>>>> 640(foreachPartition at WriteToKafka.java:45) >>>>> 16/07/18 14:53:13 INFO DAGScheduler: Parents of final stage: >>>>> List(ShuffleMapStage 518, ShuffleMapStage 639) >>>>> 16/07/18 14:53:23 INFO DAGScheduler: Missing parents: List() >>>>> 16/07/18 14:53:23 INFO DAGScheduler: Submitting ResultStage 640 >>>>> (MapPartitionsRDD[271] at map at BuildSolrDocs.java:209), which has no >>>>> missing >>>>> parents >>>>> 16/07/18 14:53:23 INFO MemoryStore: ensureFreeSpace(8248) called with >>>>> curMem=41923262, maxMem=2778778828 >>>>> 16/07/18 14:53:23 INFO MemoryStore: Block broadcast_90 stored as >>>>> values in memory (estimated size 8.1 KB, free 2.5 GB) >>>>> Exception in thread "dag-scheduler-event-loop" >>>>> java.lang.OutOfMemoryError: Java heap space >>>>> at >>>>> org.apache.spark.util.io.ByteArrayChunkOutputStream.allocateNewChunkIfNeeded(ByteArrayChunkOutputStream.scala:66) >>>>> at >>>>> org.apache.spark.util.io.ByteArrayChunkOutputStream.write(ByteArrayChunkOutputStream.scala:55) >>>>> at >>>>> org.xerial.snappy.SnappyOutputStream.dumpOutput(SnappyOutputStream.java:294) >>>>> at >>>>> org.xerial.snappy.SnappyOutputStream.flush(SnappyOutputStream.java:273) >>>>> at >>>>> org.apache.spark.io.SnappyOutputStreamWrapper.flush(CompressionCodec.scala:197) >>>>> at >>>>> java.io.ObjectOutputStream$BlockDataOutputStream.flush(ObjectOutputStream.java:1822) >>>>> >>>>> >>>>> Help needed. >>>>> >>>>> -- >>>>> Thanks and Regards, >>>>> >>>>> Saurav Sinha >>>>> >>>>> Contact: 9742879062 >>>>> >>>>> >>>>> >>>> >>>> >>>> -- >>>> Thanks and Regards, >>>> >>>> Saurav Sinha >>>> >>>> Contact: 9742879062 >>>> >>> >>> >> >> >> -- >> Thanks and Regards, >> >> Saurav Sinha >> >> Contact: 9742879062 >> > > > Collective[i] dramatically improves sales and marketing performance using > technology, applications and a revolutionary network designed to provide > next generation analytics and decision-support directly to business users. > Our goal is to maximize human potential and minimize mistakes. In most > cases, the results are astounding. We cannot, however, stop emails from > sometimes being sent to the wrong person. If you are not the intended > recipient, please notify us by replying to this email's sender and deleting > it (and any attachments) permanently from your system. If you are, please > respect the confidentiality of this communication's contents. -- Thanks and Regards, Saurav Sinha Contact: 9742879062