there might be data loss.
>>>>>>>>>>>>
>>>>>>>>>>>> 2.Catch that exception and somehow force things to "reset" for
>>>>>>>>>>>> that
>>>>>>>>>>>>
there is one)?
>>>>>>>>>>>
>>>>>>>>>>> 3.Track the offsets separately, restart the job by providing the
>>>>>>>>>>> offsets.
>>>>>>>>>>>
>>>>>>>>&
t; 3.Track the offsets separately, restart the job by providing the
>>>>>>>>>> offsets.
>>>>>>>>>>
>>>>>>>>>> 4.Or a straightforward approach would be to monitor the log for
>>>>>>>>>> this error,
>>>&g
l the job, remove the
>>>>>>>>> checkpoint
>>>>>>>>> directory, and restart.
>>>>>>>>>
>>>>>>>>> ERROR DirectKafkaInputDStream:
>>>>>>>>> ArrayBuffer(kafka.common.UnknownException,
>>
>>>
>>>>>>>>
>>>>>>>> java.lang.ClassNotFoundException:
>>>>>>>> kafka.common.NotLeaderForPartitionException
>>>>
gt;>>>>>
>>>>>>> at
>>>>>>> org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1699)
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> java.util.concurrent.RejectedExecutionExc
ointWriter$CheckpointWriteHandler@a48c5a8
>>>>>> rejected from java.util.concurrent.ThreadPoolExecutor@543258e0
>>>>>> [Terminated,
>>>>>> pool size = 0, active threads = 0, qu
t;>>>>
>>>>>
>>>>> org.apache.spark.SparkException: Job aborted due to stage failure:
>>>>> Task 10
>>>>> in stage 52.0 failed 4 times, most recent failure: Lost task 10.3 in
>>>>> stage
>>>>
times, most recent failure: Lost task 10.3 in
>>>> stage
>>>> 52.0 (TID 255, 172.16.97.97): UnknownReason
>>>>
>>>> Exception in thread "streaming-job-executor-0" java.lang.Error:
>>>> java.lang.InterruptedException
>>>>
&g
hread "streaming-job-executor-0" java.lang.Error:
>>> java.lang.InterruptedException
>>>
>>> Caused by: java.lang.InterruptedException
>>>
>>> java.lang.ClassNotFoundException: kafka.common.OffsetOutOfRangeException
>>>
>>> at org.apache.spark.util.Ut
;>
>>> Caused by: java.lang.InterruptedException
>>>
>>> java.lang.ClassNotFoundException: kafka.common.OffsetOutOfRangeException
>>>
>>> at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1699)
>>>
>>>
>>>
>>>
e 33.0 failed 4 times, most recent failure: Lost task 7.3 in stage
>> 33.0
>> (TID 283, 172.16.97.103): UnknownReason
>>
>> java.lang.ClassNotFoundException: kafka.common.OffsetOutOfRangeException
>>
>> at org.apache.spark.util.Utils$.logUncaughtExceptions(Ut
> java.lang.ClassNotFoundException: kafka.common.OffsetOutOfRangeException
>
> at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1699)
>
> java.lang.ClassNotFoundException: kafka.common.OffsetOutOfRangeException
>
> at org.apache.spark.util.Utils$.logUncaugh
etOutOfRangeException
at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1699)
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Recovery-for-Spark-Streaming-Kafka-Direct-in-case-of-issues-with-Kafka-tp25524.html
Sent from the Apache Spark User Lis
14 matches
Mail list logo