Long(20+ seconds) startup delay for jobs when running Spark on YARN

2016-04-21 Thread Akmal Abbasov
Hi, I'm running Spark(1.6.1) on YARN(2.5.1), cluster mode. It's taking 20+ seconds for application to move from ACCEPTED to RUNNING state, here's logs 16/04/21 09:06:56 INFO impl.YarnClientImpl: Submitted application application_1461229289298_0001 16/04/21 09:06:57 INFO yarn.Client: Application

Re: Running Spark on Yarn-Client/Cluster mode

2016-04-12 Thread Jon Kjær Amundsen
in yarn-client mode > > Any suggestion ? > > > > -- > View this message in context: > http://apache-spark-user-list.1001560.n3.nabble.com/Running-Spark-on-Yarn-Client-Cluster-mode-tp26691p26752.html > Sent from the Apache Spa

Re: Running Spark on Yarn-Client/Cluster mode

2016-04-11 Thread ashesh_28
I have updated all my nodes in the Cluster to have 4GB RAM memory , but still face the same error when trying to launch Spark-Shell in yarn-client mode Any suggestion ? -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Running-Spark-on-Yarn-Client-Cluster

Re: Running Spark on Yarn-Client/Cluster mode

2016-04-11 Thread ashesh_28
-Spark-on-Yarn-Client-Cluster-mode-tp26691p26739.html Sent from the Apache Spark User List mailing list archive at Nabble.com. - To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h

Re: Running Spark on Yarn-Client/Cluster mode

2016-04-08 Thread ashesh_28
0.n3.nabble.com/Running-Spark-on-Yarn-Client-Cluster-mode-tp26691p26717.html Sent from the Apache Spark User List mailing list archive at Nabble.com. - To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional command

Re: Running Spark on Yarn-Client/Cluster mode

2016-04-08 Thread ashesh_28
Few more added information with Nodes Memory and Core ptfhadoop01v - 4GB ntpcam01v - 1GB ntpcam03v - 2GB Each of the VM has only 1 core CPU -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Running-Spark-on-Yarn-Client-Cluster-mode-tp26691p26714.html Sent

Re: Running Spark on Yarn-Client/Cluster mode

2016-04-08 Thread ashesh_28
his message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Running-Spark-on-Yarn-Client-Cluster-mode-tp26691p26713.html Sent from the Apache Spark User List mailing list archive at Nabble.com. - To unsubscribe,

Re: Running Spark on Yarn-Client/Cluster mode

2016-04-07 Thread ashesh_28
560.n3.nabble.com/Running-Spark-on-Yarn-Client-Cluster-mode-tp26691p26710.html Sent from the Apache Spark User List mailing list archive at Nabble.com. - To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands,

Re: Running Spark on Yarn-Client/Cluster mode

2016-04-07 Thread ashesh_28
quick thoughts on this issue. Regards Ashesh -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Running-Spark-on-Yarn-Client-Cluster-mode-tp26691p26709.html Sent from the Apache Spark User List mailing list

Re: Running Spark on Yarn-Client/Cluster mode

2016-04-07 Thread JasmineGeorge
sembly.jar -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Running-Spark-on-Yarn-Client-Cluster-mode-tp26691p26704.html Sent from the Apache Spark User List mailing list archive at Nabble.com. --

Running Spark on Yarn-Client/Cluster mode

2016-04-06 Thread ashesh_28
n each node in cluster ? How do i start the spark-shell in yarn-client mode. Thanks in advance. -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Running-Spark-on-Yarn-Client-Cluster-mode-tp26691.html Sent from the Apache Spark User List mailin

Re: Running Spark on Yarn

2016-03-30 Thread Vineet Mishra
RM NM logs traced below, RM --> 2016-03-30 14:59:15,498 INFO org.apache.hadoop.yarn.server.resourcemanager.amlauncher.AMLauncher: Setting up container Container: [ContainerId: container_1459326455972_0004_01_01, NodeId: myhost:60653, NodeHttpAddress: myhost:8042, Resource: , Priority: 0, Toke

Re: Running Spark on Yarn

2016-03-29 Thread Alexander Pivovarov
ok, start EMR-4.3.0 or 4.2.0 cluster and look at how to configure spark on yarn properly

Re: Running Spark on Yarn

2016-03-29 Thread Vineet Mishra
:~/Downloads/package/spark-1.6.1-bin-hadoop2.6$ bin/spark-shell --master yarn-client 16/03/30 03:24:43 DEBUG ipc.Client: IPC Client (111576772) connection to myhost/192.168.1.108:8032 from myhost sending #138 16/03/30 03:24:43 DEBUG ipc.Client: IPC Client (111576772) connection to myhost/192.168.1

Re: Running Spark on Yarn

2016-03-29 Thread Vineet Mishra
Looks like still the same while the other MR application is working fine, On Wed, Mar 30, 2016 at 3:15 AM, Alexander Pivovarov wrote: > for small cluster set the following settings > > yarn-site.xml > > > yarn.scheduler.minimum-allocation-mb > 32 > > > > capacity-scheduler.xml > > >

Re: Running Spark on Yarn

2016-03-29 Thread Alexander Pivovarov
for small cluster set the following settings yarn-site.xml yarn.scheduler.minimum-allocation-mb 32 capacity-scheduler.xml yarn.scheduler.capacity.maximum-am-resource-percent 0.5 Maximum percent of resources in the cluster which can be used to run application m

Re: Running Spark on Yarn

2016-03-29 Thread Vineet Mishra
Yarn seems to be running fine, I have successful MR jobs completed on the same, *Cluster Metrics* *Apps Submitted Apps Pending Apps Running Apps Completed Containers Running Memory Used Memory Total Memory Reserved VCores Used VCores Total VCores Reserved Active Nodes Decommissioned Nodes Lost Nod

Re: Running Spark on Yarn

2016-03-29 Thread Alexander Pivovarov
check resource manager and node manager logs. Maybe you find smth explaining why 1 app is pending do you have any app run successfully? *Apps Completed is 0 on the UI* On Tue, Mar 29, 2016 at 2:13 PM, Vineet Mishra wrote: > Hi Alex/Surendra, > > Hadoop is up and running fine and I am able to r

Re: Running Spark on Yarn

2016-03-29 Thread Vineet Mishra
Hi Alex/Surendra, Hadoop is up and running fine and I am able to run example on the same. *Cluster Metrics* *Apps Submitted Apps Pending Apps Running Apps Completed Containers Running Memory Used Memory Total Memory Reserved VCores Used VCores Total VCores Reserved Active Nodes Decommissioned Nod

Re: Running Spark on Yarn

2016-03-29 Thread Alexander Pivovarov
check 8088 ui - how many cores and memory available - how many slaves are active run teragen or pi from hadoop examples to make sure that yarn works On Tue, Mar 29, 2016 at 1:25 PM, Surendra , Manchikanti < surendra.manchika...@gmail.com> wrote: > Hi Vineeth, > > Can you please check resource(RA

Re: Running Spark on Yarn

2016-03-29 Thread Surendra , Manchikanti
Hi Vineeth, Can you please check resource(RAM,Cores) availability in your local cluster, And change accordingly. Regards, Surendra M -- Surendra Manchikanti On Tue, Mar 29, 2016 at 1:15 PM, Vineet Mishra wrote: > Hi All, > > While starting Spark on Yarn on local cluster(Single Node Hadoop 2.6

Running Spark on Yarn

2016-03-29 Thread Vineet Mishra
Hi All, While starting Spark on Yarn on local cluster(Single Node Hadoop 2.6 yarn) I am facing some issues. As I try to start the Spark Shell it keeps on iterating in a endless loop while initiating, *6/03/30 01:32:38 DEBUG ipc.Client: IPC Client (1782965120) connection to myhost/192.168.1.108:8

Re: How to display the web ui when running Spark on YARN?

2016-03-09 Thread Shady Xu
Thanks for the reply. I am now trying to configure yarn.web-proxy.address according to https://issues.apache.org/jira/browse/SPARK-5837, but cannot start the standalone web proxy server. I am using CDH 5.0.1 and below is the error log: sbin/yarn-daemon.sh: line 44: /opt/cloudera/parcels/CDH/lib/

Re: How to display the web ui when running Spark on YARN?

2016-03-04 Thread Steve Loughran
On 3 Mar 2016, at 09:17, Shady Xu mailto:shad...@gmail.com>> wrote: Hi all, I am running Spark in yarn-client mode, but every time I access the web ui, the browser redirect me to one of the worker nodes and shows nothing. The url looks like http://hadoop-node31.company.com:8088/proxy/applica

How to display the web ui when running Spark on YARN?

2016-03-03 Thread Shady Xu
Hi all, I am running Spark in yarn-client mode, but every time I access the web ui, the browser redirect me to one of the worker nodes and shows nothing. The url looks like http://hadoop-node31.company.com:8088/proxy/application_1453797301246_120264 . I googled a lot and found some possible bugs

Re: Is it possible to see Spark jobs on MapReduce job history ? (running Spark on YARN cluster)

2015-06-12 Thread Steve Loughran
For that you need SPARK-1537 and the patch to go with it It is still the spark web UI, it just hands off storage and retrieval of the history to the underlying Yarn timeline server, rather than through the filesystem. You'll get to see things as they go along too. If you do want to try it, ple

Is it possible to see Spark jobs on MapReduce job history ? (running Spark on YARN cluster)

2015-06-11 Thread Elkhan Dadashov
Hi all, I wonder if anyone has used use MapReduce Job History to show Spark jobs. I can see my Spark jobs (Spark running on Yarn cluster) on Resource manager (RM). I start Spark History server, and then through Spark's web-based user interface I can monitor the cluster (and track cluster and job

running spark on yarn

2015-05-21 Thread Nathan Kronenfeld
Hello, folks. We just recently switched to using Yarn on our cluster (when upgrading to cloudera 5.4.1) I'm trying to run a spark job from within a broader application (a web service running on Jetty), so I can't just start it using spark-submit. Does anyone know of an instructions page on how t

Re: Who manage the log4j appender while running spark on yarn?

2014-12-22 Thread Marcelo Vanzin
but I think spark code has changed a lot since then. > > Any one could offer some guide? Thanks. > > > > > > -- > View this message in context: > http://apache-spark-user-list.1001560.n3.nabble.com/Who-manage-the-log4j-appender-while-running-spark-on-y

Re: Who manage the log4j appender while running spark on yarn?

2014-12-22 Thread WangTaoTheTonic
ext: http://apache-spark-user-list.1001560.n3.nabble.com/Who-manage-the-log4j-appender-while-running-spark-on-yarn-tp20778p20818.html Sent from the Apache Spark User List mailing list archive at Nabble.com. - To unsu

Who manage the log4j appender while running spark on yarn?

2014-12-19 Thread WangTaoTheTonic
01560.n3.nabble.com/Who-manage-the-log4j-appender-while-running-spark-on-yarn-tp20778.html Sent from the Apache Spark User List mailing list archive at Nabble.com. - To unsubscribe, e-mail: user-unsubscr...@spark.apache.or

Re: "vcores used" in cluster metrics(yarn resource manager ui) when running spark on yarn

2014-12-08 Thread Sandy Ryza
Hi yuemeng, Are you possibly running the Capacity Scheduler with the default resource calculator? -Sandy On Sat, Dec 6, 2014 at 7:29 PM, yuemeng1 wrote: > Hi, all > When i running an app with this cmd: ./bin/spark-sql --master > yarn-client --num-executors 2 --executor-cores 3, i noticed

"vcores used" in cluster metrics(yarn resource manager ui) when running spark on yarn

2014-12-06 Thread yuemeng1
Hi, all When i running an app with this cmd: ./bin/spark-sql --master yarn-client --num-executors 2 --executor-cores 3, i noticed that yarn resource manager ui shows the `vcores used` in cluster metrics is 3. It seems `vcores used` show wrong num (should be 7?)? Or i miss something? Tha

Re: hadoop_conf_dir when running spark on yarn

2014-11-03 Thread Amey Chaugule
I'm using org.apache.spark.deploy.yarn.Client object to run my spark job. I guess this is what spark-submit wraps really. - Amey On Mon, Nov 3, 2014 at 5:25 PM, Tobias Pfeiffer wrote: > Hi, > > On Mon, Nov 3, 2014 at 1:29 PM, Amey Chaugule wrote: > >> I thought that only applied when you're tr

Re: hadoop_conf_dir when running spark on yarn

2014-11-03 Thread Tobias Pfeiffer
Hi, On Mon, Nov 3, 2014 at 1:29 PM, Amey Chaugule wrote: > I thought that only applied when you're trying to run a job using > spark-submit or in the shell... > And how are you starting your Yarn job, if not via spark-submit? Tobias

Re: hadoop_conf_dir when running spark on yarn

2014-11-02 Thread Amey Chaugule
> -- >> View this message in context: >> http://apache-spark-user-list.1001560.n3.nabble.com/hadoop-conf-dir-when-running-spark-on-yarn-tp17872.html >> Sent from the Apache Spark User List mailing list archive at Nabble.com. >> >> --

Re: hadoop_conf_dir when running spark on yarn

2014-11-02 Thread Akhil Das
iguration that I pull from sc.hadoopConfiguration() is incorrect. > > > > -- > View this message in context: > http://apache-spark-user-list.1001560.n3.nabble.com/hadoop-conf-dir-when-running-spark-on-yarn-tp17872.html > Sent from the Apache Spark User List

hadoop_conf_dir when running spark on yarn

2014-10-31 Thread ameyc
60.n3.nabble.com/hadoop-conf-dir-when-running-spark-on-yarn-tp17872.html Sent from the Apache Spark User List mailing list archive at Nabble.com. - To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e

Re: Running Spark On Yarn without Spark-Submit

2014-08-29 Thread Chester @work
Archit We are using yarn-cluster mode , and calling spark via Client class directly from servlet server. It works fine. To establish a communication channel to give further requests, It should be possible with yarn client, but not with yarn server. Yarn client mode, spark driver i

Re: Running Spark On Yarn without Spark-Submit

2014-08-29 Thread Archit Thakur
including user@spark.apache.org. On Fri, Aug 29, 2014 at 2:03 PM, Archit Thakur wrote: > Hi, > > My requirement is to run Spark on Yarn without using the script > spark-submit. > > I have a servlet and a tomcat server. As and when request comes, it > creates a new SC and keeps it alive for the

Running Spark On Yarn without Spark-Submit

2014-08-29 Thread Archit Thakur
Hi, My requirement is to run Spark on Yarn without using the script spark-submit. I have a servlet and a tomcat server. As and when request comes, it creates a new SC and keeps it alive for the further requests, I ma setting my master in sparkConf as sparkConf.setMaster("yarn-cluster") but the

Running Spark on Yarn vs Mesos

2014-07-10 Thread k.tham
.nabble.com/Running-Spark-on-Yarn-vs-Mesos-tp9320.html Sent from the Apache Spark User List mailing list archive at Nabble.com.

Re: Question about running spark on yarn

2014-04-22 Thread sandy . ryza
I currently don't have plans to work on that. -Sandy > On Apr 22, 2014, at 8:06 PM, Gordon Wang wrote: > > Thanks I see. Do you guys have plan to port this to sbt? > > >> On Wed, Apr 23, 2014 at 10:24 AM, Sandy Ryza wrote: >> Right, it only works for Maven >> >> >>> On Tue, Apr 22, 2014 at

Re: Question about running spark on yarn

2014-04-22 Thread Gordon Wang
Thanks I see. Do you guys have plan to port this to sbt? On Wed, Apr 23, 2014 at 10:24 AM, Sandy Ryza wrote: > Right, it only works for Maven > > > On Tue, Apr 22, 2014 at 6:23 PM, Gordon Wang wrote: > >> Hi Sandy, >> >> Thanks for your reply ! >> >> Does this work for sbt ? >> >> I checked the

Re: Question about running spark on yarn

2014-04-22 Thread Sandy Ryza
Right, it only works for Maven On Tue, Apr 22, 2014 at 6:23 PM, Gordon Wang wrote: > Hi Sandy, > > Thanks for your reply ! > > Does this work for sbt ? > > I checked the commit, looks like only maven build has such option. > > > > On Wed, Apr 23, 2014 at 12:38 AM, Sandy Ryza wrote: > >> Hi Gord

Re: Question about running spark on yarn

2014-04-22 Thread Gordon Wang
Hi Sandy, Thanks for your reply ! Does this work for sbt ? I checked the commit, looks like only maven build has such option. On Wed, Apr 23, 2014 at 12:38 AM, Sandy Ryza wrote: > Hi Gordon, > > We recently handled this in SPARK-1064. As of 1.0.0, you'll be able to > pass -Phadoop-provided

Re: Question about running spark on yarn

2014-04-22 Thread Sandy Ryza
Hi Gordon, We recently handled this in SPARK-1064. As of 1.0.0, you'll be able to pass -Phadoop-provided to Maven and avoid including Hadoop and its dependencies in the assembly jar. -Sandy On Tue, Apr 22, 2014 at 2:43 AM, Gordon Wang wrote: > In this page http://spark.apache.org/docs/0.9.0/

Question about running spark on yarn

2014-04-22 Thread Gordon Wang
In this page http://spark.apache.org/docs/0.9.0/running-on-yarn.html We have to use spark assembly to submit spark apps to yarn cluster. And I checked the assembly jars of spark. It contains some yarn classes which are added during compile time. The yarn classes are not what I want. My question i