Ah, right. So only the launch script has changed. Everything else is still
essentially binary compatible?

Well, that makes it too easy! Thanks!


On Wed, Jun 18, 2014 at 2:35 PM, Patrick Wendell <pwend...@gmail.com> wrote:

> Actually you'll just want to clone the 1.0 branch then use the
> spark-ec2 script in there to launch your cluster. The --spark-git-repo
> flag is if you want to launch with a different version of Spark on the
> cluster. In your case you just need a different version of the launch
> script itself, which will be present in the 1.0 branch of Spark.
>
> - Patrick
>
> On Tue, Jun 17, 2014 at 9:29 PM, Jeremy Lee
> <unorthodox.engine...@gmail.com> wrote:
> > I am about to spin up some new clusters, so I may give that a go... any
> > special instructions for making them work? I assume I use the "
> > --spark-git-repo=" option on the spark-ec2 command. Is it as easy as
> > concatenating your string as the value?
> >
> > On cluster management GUIs... I've been looking around at Amabari,
> Datastax,
> > Cloudera, OpsCenter etc. Not totally convinced by any of them yet. Anyone
> > using a good one I should know about? I'm really beginning to lean in the
> > direction of Cassandra as the distributed data store...
> >
> >
> > On Wed, Jun 18, 2014 at 1:46 PM, Patrick Wendell <pwend...@gmail.com>
> wrote:
> >>
> >> By the way, in case it's not clear, I mean our maintenance branches:
> >>
> >> https://github.com/apache/spark/tree/branch-1.0
> >>
> >> On Tue, Jun 17, 2014 at 8:35 PM, Patrick Wendell <pwend...@gmail.com>
> >> wrote:
> >> > Hey Jeremy,
> >> >
> >> > This is patched in the 1.0 and 0.9 branches of Spark. We're likely to
> >> > make a 1.0.1 release soon (this patch being one of the main reasons),
> >> > but if you are itching for this sooner, you can just checkout the head
> >> > of branch-1.0 and you will be able to use r3.XXX instances.
> >> >
> >> > - Patrick
> >> >
> >> > On Tue, Jun 17, 2014 at 4:17 PM, Jeremy Lee
> >> > <unorthodox.engine...@gmail.com> wrote:
> >> >> Some people (me included) might have wondered why all our m1.large
> spot
> >> >> instances (in us-west-1) shut down a few hours ago...
> >> >>
> >> >> Simple reason: The EC2 spot price for Spark's default "m1.large"
> >> >> instances
> >> >> just jumped from 0.016 per hour, to about 0.750. Yes, Fifty times.
> >> >> Probably
> >> >> something to do with world cup.
> >> >>
> >> >> So far this is just us-west-1, but prices have a tendency to equalize
> >> >> across
> >> >> centers as the days pass. Time to make backups and plans.
> >> >>
> >> >> "m3" spot prices are still down at $0.02 (and being new, will be
> >> >> bypassed by
> >> >> older systems), so it would be REAAALLYY nice if there had been some
> >> >> progress on that issue. Let me know if I can help with testing and
> >> >> whatnot.
> >> >>
> >> >>
> >> >> --
> >> >> Jeremy Lee  BCompSci(Hons)
> >> >>   The Unorthodox Engineers
> >
> >
> >
> >
> > --
> > Jeremy Lee  BCompSci(Hons)
> >   The Unorthodox Engineers
>



-- 
Jeremy Lee  BCompSci(Hons)
  The Unorthodox Engineers

Reply via email to