Re: Spark 1.6.0 Streaming + Persistance Bug?

2016-02-02 Thread mkhaitman
Actually disregard! Forgot that spark.dynamicAllocation.cachedExecutorIdleTimeout was defaulted to Infinity, so lowering that should solve the problem :) Mark. -- View this message in context: http://apache-spark-developers-list.1001551.n3.nabble.com/Spark-1-6-0-Streaming-Persistance-Bug-tp161

Spark 1.6.0 Streaming + Persistance Bug?

2016-02-02 Thread mkhaitman
Calling unpersist on an RDD in a spark streaming application does not actually unpersist the blocks from memory and/or disk. After the RDD has been processed in a .foreach(rdd) call, I attempt to unpersist the rdd since it is no longer useful to store in memory/disk. This mainly causes a problem wi

Re: [VOTE] Release Apache Spark 1.6.0 (RC1)

2015-12-03 Thread mkhaitman
I reported this in the 1.6 preview thread, but wouldn't mind if someone can confirm that ctrl-c is not keyboard interrupting / clearing the current line of input anymore in the pyspark shell. I saw the change that would kill the currently running job when using ctrl+c, but now the only way to clear

Re: [ANNOUNCE] Spark 1.6.0 Release Preview

2015-11-23 Thread mkhaitman
Nice! Built and testing on CentOS 7 on a Hadoop 2.7.1 cluster. One thing I've noticed is that KeyboardInterrupts are now ignored? Is that intended? I starting typing a line out and then changed my mind and wanted to issue the good old ctrl+c to interrupt, but that didn't work. Otherwise haven't s

Re: Spark.Executor.Cores question

2015-10-28 Thread mkhaitman
Unfortunately setting the executor memory to prevent multiple executors from the same framework would inherently mean that we'd need to set just over half the available worker memory for each node. So if each node had 32GB of worker memory, then the application would need to set 17GB to absolutely

Re: Spark.Executor.Cores question

2015-10-27 Thread mkhaitman
Hi Richard, Thanks for the response. I should have added that the specific case where this becomes a problem is when one of the executors for that application is lost/killed prematurely, and the application attempts to spawn up a new executor without consideration as to whether an executor alrea

Spark.Executor.Cores question

2015-10-23 Thread mkhaitman
Regarding the 'spark.executor.cores' config option in a Standalone spark environment, I'm curious about whether there's a way to enforce the following logic: *- Max cores per executor = 4* ** Max executors PER application PER worker = 1* In order to force better balance across all workers, I want

Re: If you use Spark 1.5 and disabled Tungsten mode ...

2015-10-15 Thread mkhaitman
My apologies for mixing up what was being referred to in that case! :) Mark. -- View this message in context: http://apache-spark-developers-list.1001551.n3.nabble.com/If-you-use-Spark-1-5-and-disabled-Tungsten-mode-tp14604p14629.html Sent from the Apache Spark Developers List mailing list a

Re: If you use Spark 1.5 and disabled Tungsten mode ...

2015-10-15 Thread mkhaitman
Are you referring to spark.shuffle.manager=tungsten-sort? If so, we saw the default value as still being as the regular sort, and since it was only first introduced in 1.5, were actually waiting a bit to see if anyone ENABLED it as opposed to DISABLING it since - it's disabled by default! :) I rec

Re: [VOTE] Release Apache Spark 1.5.0 (RC3)

2015-09-03 Thread mkhaitman
Built and tested on CentOS 7, Hadoop 2.7.1 (Built for 2.6 profile), Standalone without any problems. Re-tested dynamic allocation specifically. "Lost executor" messages are still an annoyance since they're expected to occur with dynamic allocation, and shouldn't WARN/ERROR as they do now, however

Re: [VOTE] Release Apache Spark 1.5.0 (RC1)

2015-08-21 Thread mkhaitman
Just a heads up that this RC1 release is still appearing as "1.5.0-SNAPSHOT" (Not just me right..?) -- View this message in context: http://apache-spark-developers-list.1001551.n3.nabble.com/VOTE-Release-Apache-Spark-1-5-0-RC1-tp13780p13792.html Sent from the Apache Spark Developers List mailin

Re: Fwd: [ANNOUNCE] Spark 1.5.0-preview package

2015-08-20 Thread mkhaitman
Turns out it was a mix of user-error as well as a bug in the sbt/sbt build that has since been fixed in the current 1.5 branch (I built from this commit: b4f4e91c395cb69ced61d9ff1492d1b814f96828) I've been testing out the dynamic allocation specifically and it's looking pretty solid! Haven't come

Re: Fwd: [ANNOUNCE] Spark 1.5.0-preview package

2015-08-14 Thread mkhaitman
Has anyone had success using this preview? We were able to build the preview, and able to start the spark-master, however, unable to connect any spark workers to it. Kept receiving "AkkaRpcEnv address in use" while attempting to connect the spark-worker to the master. Also confirmed that the work

Re: "Spree": Live-updating web UI for Spark

2015-07-29 Thread mkhaitman
We tested this out on our dev cluster (Hadoop 2.7.1 + Spark 1.4.0), and it looks great! I might also be interested in contributing to it when I get a chance! Keep up the awesome work! :) Mark. -- View this message in context: http://apache-spark-developers-list.1001551.n3.nabble.com/Spree-Live

RE: StreamingContext textFileStream question

2015-02-23 Thread mkhaitman
Hi Jerry, Thanks for the quick response! Looks like I'll need to come up with an alternative solution in the meantime, since I'd like to avoid the other input streams + WAL approach. :) Thanks again, Mark. -- View this message in context: http://apache-spark-developers-list.1001551.n3.nabble

StreamingContext textFileStream question

2015-02-23 Thread mkhaitman
Hello, I was interested in creating a StreamingContext textFileStream based job, which runs for long durations, and can also recover from prolonged driver failure... It seems like StreamingContext checkpointing is mainly used for the case when the driver dies during the processing of an RDD, and t

pyspark.daemon issues?

2015-02-09 Thread mkhaitman
I've noticed a couple oddities with the pyspark.daemons which are causing us a bit of memory problems within some of our heavy spark jobs, especially when they run at the same time... It seems that there is typically a 1-to-1 ratio of pyspark.daemons to cores per executor during aggregations. By d