Hi Cody,

First of all thanks for the note about spark.streaming.concurrentJobs. I
guess this is why it's not mentioned in the actual spark streaming doc.
Since those 3 topics contain completely different data on which I need to
apply different kind of transformations, I am not sure joining them would
be really efficient, unless you know something that I don't.

As I really don't need any interaction between those streams, I think I
might end up running 3 different streaming apps instead of one.

Thanks again!

On Thu, Dec 17, 2015 at 11:43 AM, Cody Koeninger <c...@koeninger.org> wrote:

> Using spark.streaming.concurrentJobs for this probably isn't a good idea,
> as it allows the next batch to start processing before current one is
> finished, which may have unintended consequences.
>
> Why can't you use a single stream with all the topics you care about, or
> multiple streams if you're e.g. joining them?
>
>
>
> On Wed, Dec 16, 2015 at 3:00 PM, jpocalan <jpoca...@gmail.com> wrote:
>
>> Nevermind, I found the answer to my questions.
>> The following spark configuration property will allow you to process
>> multiple KafkaDirectStream in parallel:
>> --conf spark.streaming.concurrentJobs=<something greater than 1>
>>
>>
>>
>>
>>
>> --
>> View this message in context:
>> http://apache-spark-user-list.1001560.n3.nabble.com/Kafka-streaming-from-multiple-topics-tp8678p25723.html
>> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>>
>> ---------------------------------------------------------------------
>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>> For additional commands, e-mail: user-h...@spark.apache.org
>>
>>
>


-- 
jean-pierre ocalan
jpoca...@gmail.com

Reply via email to