I did not change driver program. I just shutdown the context and again
started.

BTW, I see this ticket already open in unassigned state - SPARK-6892
<https://issues.apache.org/jira/browse/SPARK-6892> that talks about this
issue.
Is this a known issue?
Also, any workarounds?







On Wed, Jun 10, 2015 at 9:18 AM, Akhil Das <ak...@sigmoidanalytics.com>
wrote:

> Delete the checkpoint directory, you might have modified your driver
> program.
>
> Thanks
> Best Regards
>
> On Wed, Jun 10, 2015 at 9:44 PM, Ashish Nigam <ashnigamt...@gmail.com>
> wrote:
>
>> Hi,
>> If checkpoint data is already present in HDFS, driver fails to load as it
>> is performing lookup on previous application directory. As that folder
>> already exists, it fails to start context.
>> Failed job's application id was application_1432284018452_0635 and job
>> was performing lookup on application_1432284018452_0633 folder.
>>
>> Here's snippet of exception stack trace-
>>
>> 15/06/10 05:28:36 INFO yarn.ApplicationMaster: Final app status: FAILED,
>> exitCode: 15, (reason: User class threw exception: Log directory
>> hdfs://x.x.x.x:8020/user/myuser/spark/applicationHistory/application_1432284018452_0633
>> already exists!)
>>
>> Exception in thread "Driver" java.io.IOException: Log directory
>> hdfs://x.x.x.x
>> :8020/user/myuswer/spark/applicationHistory/application_1432284018452_0633
>> <http://172.16.201.171:8020/user/shn/spark/applicationHistory/application_1432284018452_0633>
>> already exists!
>>
>>         at
>> org.apache.spark.util.FileLogger.createLogDir(FileLogger.scala:129)
>>
>>         at org.apache.spark.util.FileLogger.start(FileLogger.scala:115)
>>
>>         at
>> org.apache.spark.scheduler.EventLoggingListener.start(EventLoggingListener.scala:74)
>>
>>         at org.apache.spark.SparkContext.<init>(SparkContext.scala:368)
>>
>>         at
>> org.apache.spark.streaming.StreamingContext.<init>(StreamingContext.scala:118)
>>
>>         at
>> org.apache.spark.streaming.StreamingContext$$anonfun$getOrCreate$1.apply(StreamingContext.scala:561)
>>
>>         at
>> org.apache.spark.streaming.StreamingContext$$anonfun$getOrCreate$1.apply(StreamingContext.scala:561)
>>
>>         at scala.Option.map(Option.scala:145)
>>
>>         at
>> org.apache.spark.streaming.StreamingContext$.getOrCreate(StreamingContext.scala:561)
>>
>>         at
>>
>>
>>
>> Any idea on how to fix this issue?
>>
>>
>> Thanks
>>
>> Ashish
>>
>
>

Reply via email to