Hi Cody,

KafkaUtils.createRDD totally make sense now I can run my spark job once in
15 minutes extract data out of kafka and stop ..., I rely on kafka offset
for Incremental data am I right ? so no duplicate data will be returned.


Thanks
Sri





On Fri, Dec 18, 2015 at 2:41 PM, Cody Koeninger <c...@koeninger.org> wrote:

> If you're really doing a daily batch job, have you considered just using
> KafkaUtils.createRDD rather than a streaming job?
>
> On Fri, Dec 18, 2015 at 5:04 AM, kali.tumm...@gmail.com <
> kali.tumm...@gmail.com> wrote:
>
>> Hi All,
>>
>> Imagine I have a Production spark streaming kafka (direct connection)
>> subscriber and publisher jobs running which publish and subscriber
>> (receive)
>> data from a kafka topic and I save one day's worth of data using
>> dstream.slice to Cassandra daily table (so I create daily table before
>> running spark streaming job).
>>
>> My question if all the above code runs in some scheduler like autosys how
>> should I say to spark publisher to stop publishing as it is End of day and
>> to spark subscriber to stop receiving to stop receiving without killing
>> the
>> jobs ? if I kill my autosys scheduler turns red saying the job had failed
>> etc ...
>> Is there a way to stop both subscriber and publisher with out killing or
>> terminating the code.
>>
>> Thanks
>>
>>
>>
>>
>> --
>> View this message in context:
>> http://apache-spark-user-list.1001560.n3.nabble.com/how-to-turn-off-spark-streaming-gracefully-tp25734.html
>> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>>
>> ---------------------------------------------------------------------
>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>> For additional commands, e-mail: user-h...@spark.apache.org
>>
>>
>


-- 
Thanks & Regards
Sri Tummala

Reply via email to