The closer I look @ the stack trace in the Scala shell, it appears to be
the call to toString() that is causing the construction of the Job object
to fail. Is there a ways to suppress this output since it appears to be
hindering my ability to new up this object?

On Wed, Nov 5, 2014 at 5:49 PM, Corey Nolet <cjno...@gmail.com> wrote:

> I'm trying to use a custom input format with SparkContext.newAPIHadoopRDD.
> Creating the new RDD works fine but setting up the configuration file via
> the static methods on input formats that require a Hadoop Job object is
> proving to be difficult.
>
> Trying to new up my own Job object with the
> SparkContext.hadoopConfiguration is throwing the exception on line 283 of
> this grepcode:
>
>
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hadoop/hadoop-mapreduce-client-core/2.5.0/org/apache/hadoop/mapreduce/Job.java#Job
>
> Looking in the SparkContext code, I'm seeing that it's newing up Job
> objects just fine using nothing but the configuraiton. Using
> SparkContext.textFile() appears to be working for me. Any ideas? Has anyone
> else run into this as well? Is it possible to have a method like
> SparkContext.getJob() or something similar?
>
> Thanks.
>
>

Reply via email to