Hi Jay,

I just came across SPARK-720 Statically guarantee serialization will succeed
<https://issues.apache.org/jira/browse/SPARK-720> which sounds like exactly
what you're referring to.  Like Reynold I think it's not possible at this
time but it would be good to get your feedback on that ticket.

Andrew


On Sun, Nov 16, 2014 at 4:37 PM, Reynold Xin <r...@databricks.com> wrote:

> That's a great idea and it is also a pain point for some users. However, it
> is not possible to solve this problem at compile time, because the content
> of serialization can only be determined at runtime.
>
> There are some efforts in Scala to help users avoid mistakes like this. One
> example project that is more researchy is Spore:
> http://docs.scala-lang.org/sips/pending/spores.html
>
>
>
> On Sun, Nov 16, 2014 at 4:12 PM, jay vyas <jayunit100.apa...@gmail.com>
> wrote:
>
> > This is more a curiosity than an immediate problem.
> >
> > Here is my question: I ran into this easily solved issue
> >
> >
> http://stackoverflow.com/questions/22592811/task-not-serializable-java-io-notserializableexception-when-calling-function-ou
> > recently.  The solution was to replace my "class" with a scala singleton,
> > which i guess is readily serializable.
> >
> > So its clear that spark needs to serialize objects which carry the driver
> > methods for an app, in order to run... but I'm wondering,,, maybe there
> is
> > a way to change or update the spark API to catch unserializable spark
> apps
> > at compile time?
> >
> >
> > --
> > jay vyas
> >
>

Reply via email to