Hey, do you have a blog post or url I can share ?

This is a quite cool experiment !

E/


2014-03-20 15:01 GMT+01:00 Chanwit Kaewkasi <chan...@gmail.com>:

> Hi Chester,
>
> It is on our todo-list but it doesn't work at the moment. The
> Parallela cores can not be utilized by the JVM. So, Spark will just
> use its ARM cores. We'll be looking at Parallela again when the JVM
> supports it.
>
> Best regards,
>
> -chanwit
>
> --
> Chanwit Kaewkasi
> linkedin.com/in/chanwit
>
>
> On Thu, Mar 20, 2014 at 8:52 PM, Chester <chesterxgc...@yahoo.com> wrote:
> > I am curious  to see if you have tried on Parallela supercomputer (16 or
> 64 cores) cluster, run spark on that should be fun.
> >
> > Chester
> >
> > Sent from my iPad
> >
> > On Mar 19, 2014, at 9:18 AM, Chanwit Kaewkasi <chan...@gmail.com> wrote:
> >
> >> Hi Koert,
> >>
> >> There's some NAND flash built-in each node. We mount the NAND flash as
> >> a local directory for Spark to spill data out.
> >> A DZone article, also written by me, will tell more about the cluster.
> >> We really appreciate the design of Spark's RDD done by the Spark team.
> >> It turned out to be perfect for ARM clusters.
> >>
> >> http://www.dzone.com/articles/big-data-processing-arm-0
> >>
> >> Another great thing is that our cluster can operate at the room
> >> temperature (25C / 77F) too.
> >>
> >> The board is Cubieboard here it is:
> >> https://en.wikipedia.org/wiki/Cubieboard#Specification
> >>
> >> Best regards,
> >>
> >> -chanwit
> >>
> >> --
> >> Chanwit Kaewkasi
> >> linkedin.com/in/chanwit
> >>
> >>
> >> On Wed, Mar 19, 2014 at 9:43 PM, Koert Kuipers <ko...@tresata.com>
> wrote:
> >>> i dont know anything about arm clusters.... but it looks great. what
> are the
> >>> specs? the nodes have no local disk at all?
> >>>
> >>>
> >>> On Tue, Mar 18, 2014 at 10:36 PM, Chanwit Kaewkasi <chan...@gmail.com>
> >>> wrote:
> >>>>
> >>>> Hi all,
> >>>>
> >>>> We are a small team doing a research on low-power (and low-cost) ARM
> >>>> clusters. We built a 20-node ARM cluster that be able to start Hadoop.
> >>>> But as all of you've known, Hadoop is performing on-disk operations,
> >>>> so it's not suitable for a constraint machine powered by ARM.
> >>>>
> >>>> We then switched to Spark and had to say wow!!
> >>>>
> >>>> Spark / HDFS enables us to crush Wikipedia articles (of year 2012) of
> >>>> size 34GB in 1h50m. We have identified the bottleneck and it's our
> >>>> 100M network.
> >>>>
> >>>> Here's the cluster:
> >>>>
> https://dl.dropboxusercontent.com/u/381580/aiyara_cluster/Mk-I_SSD.png
> >>>>
> >>>> And this is what we got from Spark's shell:
> >>>>
> https://dl.dropboxusercontent.com/u/381580/aiyara_cluster/result_00.png
> >>>>
> >>>> I think it's the first ARM cluster that can process a non-trivial size
> >>>> of Big Data.
> >>>> (Please correct me if I'm wrong)
> >>>> I really want to thank the Spark team that makes this possible !!
> >>>>
> >>>> Best regards,
> >>>>
> >>>> -chanwit
> >>>>
> >>>> --
> >>>> Chanwit Kaewkasi
> >>>> linkedin.com/in/chanwit
> >>>
> >>>
>

Reply via email to