I am a beginner to Scala/Spark. Could you please elaborate on how to make
RDD of results of func() and collect?


On Tue, Feb 24, 2015 at 2:27 PM, Sean Owen <so...@cloudera.com> wrote:

> They aren't the same 'lst'. One is on your driver. It gets copied to
> executors when the tasks are executed. Those copies are updated. But
> the updates will never reflect in the local copy back in the driver.
>
> You may just wish to make an RDD of the results of func() and
> collect() them back to the driver.
>
> On Tue, Feb 24, 2015 at 7:20 PM, kvvt <kvi...@vt.edu> wrote:
> > I am working on the below piece of code.
> >
> > var lst = scala.collection.mutable.MutableList[VertexId]()
> > graph.edges.groupBy[VertexId](f).foreach {
> >   edgesBySrc => {
> >       lst ++= func(edgesBySrc)
> >   }
> > }
> >
> > println(lst.length)
> >
> > Here, the final println() always says that the length of the list is 0.
> The
> > list is non-empty (correctly prints the length of the returned list
> inside
> > func()).
> >
> > I am not sure if I am doing the append correctly. Can someone point out
> what
> > I am doing wrong?
> >
> >
> >
> >
> >
> > --
> > View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/Not-able-to-update-collections-tp21790.html
> > Sent from the Apache Spark User List mailing list archive at Nabble.com.
> >
> > ---------------------------------------------------------------------
> > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> > For additional commands, e-mail: user-h...@spark.apache.org
> >
>

Reply via email to