Meant this one: https://docs.databricks.com/api/latest/jobs.html
On Thu, Mar 28, 2019 at 5:06 PM Pat Ferrel <p...@occamsmachete.com> wrote: > Thanks, are you referring to > https://github.com/spark-jobserver/spark-jobserver or the undocumented > REST job server included in Spark? > > > From: Jason Nerothin <jasonnerot...@gmail.com> <jasonnerot...@gmail.com> > Reply: Jason Nerothin <jasonnerot...@gmail.com> <jasonnerot...@gmail.com> > Date: March 28, 2019 at 2:53:05 PM > To: Pat Ferrel <p...@occamsmachete.com> <p...@occamsmachete.com> > Cc: Felix Cheung <felixcheun...@hotmail.com> <felixcheun...@hotmail.com>, > Marcelo > Vanzin <van...@cloudera.com> <van...@cloudera.com>, user > <user@spark.apache.org> <user@spark.apache.org> > Subject: Re: spark.submit.deployMode: cluster > > Check out the Spark Jobs API... it sits behind a REST service... > > > On Thu, Mar 28, 2019 at 12:29 Pat Ferrel <p...@occamsmachete.com> wrote: > >> ;-) >> >> Great idea. Can you suggest a project? >> >> Apache PredictionIO uses spark-submit (very ugly) and Apache Mahout only >> launches trivially in test apps since most uses are as a lib. >> >> >> From: Felix Cheung <felixcheun...@hotmail.com> >> <felixcheun...@hotmail.com> >> Reply: Felix Cheung <felixcheun...@hotmail.com> >> <felixcheun...@hotmail.com> >> Date: March 28, 2019 at 9:42:31 AM >> To: Pat Ferrel <p...@occamsmachete.com> <p...@occamsmachete.com>, Marcelo >> Vanzin <van...@cloudera.com> <van...@cloudera.com> >> Cc: user <user@spark.apache.org> <user@spark.apache.org> >> Subject: Re: spark.submit.deployMode: cluster >> >> If anyone wants to improve docs please create a PR. >> >> lol >> >> >> But seriously you might want to explore other projects that manage job >> submission on top of spark instead of rolling your own with spark-submit. >> >> >> ------------------------------ >> *From:* Pat Ferrel <p...@occamsmachete.com> >> *Sent:* Tuesday, March 26, 2019 2:38 PM >> *To:* Marcelo Vanzin >> *Cc:* user >> *Subject:* Re: spark.submit.deployMode: cluster >> >> Ahh, thank you indeed! >> >> It would have saved us a lot of time if this had been documented. I know, >> OSS so contributions are welcome… I can also imagine your next comment; “If >> anyone wants to improve docs see the Apache contribution rules and create a >> PR.” or something like that. >> >> BTW the code where the context is known and can be used is what I’d call >> a Driver and since all code is copied to nodes and is know in jars, it was >> not obvious to us that this rule existed but it does make sense. >> >> We will need to refactor our code to use spark-submit it appears. >> >> Thanks again. >> >> >> From: Marcelo Vanzin <van...@cloudera.com> <van...@cloudera.com> >> Reply: Marcelo Vanzin <van...@cloudera.com> <van...@cloudera.com> >> Date: March 26, 2019 at 1:59:36 PM >> To: Pat Ferrel <p...@occamsmachete.com> <p...@occamsmachete.com> >> Cc: user <user@spark.apache.org> <user@spark.apache.org> >> Subject: Re: spark.submit.deployMode: cluster >> >> If you're not using spark-submit, then that option does nothing. >> >> If by "context creation API" you mean "new SparkContext()" or an >> equivalent, then you're explicitly creating the driver inside your >> application. >> >> On Tue, Mar 26, 2019 at 1:56 PM Pat Ferrel <p...@occamsmachete.com> wrote: >> > >> > I have a server that starts a Spark job using the context creation API. >> It DOES NOY use spark-submit. >> > >> > I set spark.submit.deployMode = “cluster” >> > >> > In the GUI I see 2 workers with 2 executors. The link for running >> application “name” goes back to my server, the machine that launched the >> job. >> > >> > This is spark.submit.deployMode = “client” according to the docs. I set >> the Driver to run on the cluster but it runs on the client, ignoring the >> spark.submit.deployMode. >> > >> > Is this as expected? It is documented nowhere I can find. >> > >> >> >> -- >> Marcelo >> >> -- > Thanks, > Jason > > -- Thanks, Jason