This sounds really really weird. Can you give me a piece of code that I can
run to reproduce this issue myself?

TD


On Tue, Jul 15, 2014 at 12:02 AM, Walrus theCat <walrusthe...@gmail.com>
wrote:

> This is (obviously) spark streaming, by the way.
>
>
> On Mon, Jul 14, 2014 at 8:27 PM, Walrus theCat <walrusthe...@gmail.com>
> wrote:
>
>> Hi,
>>
>> I've got a socketTextStream through which I'm reading input.  I have
>> three Dstreams, all of which are the same window operation over that
>> socketTextStream.  I have a four core machine.  As we've been covering
>> lately, I have to give a "cores" parameter to my StreamingSparkContext:
>>
>> ssc = new StreamingContext("local[4]" /**TODO change once a cluster is up
>> **/,
>>       "AppName", Seconds(1))
>>
>> Now, I have three dstreams, and all I ask them to do is print or count.
>> I should preface this with the statement that they all work on their own.
>>
>> dstream1 // 1 second window
>> dstream2 // 2 second window
>> dstream3 // 5 minute window
>>
>>
>> If I construct the ssc with "local[8]", and put these statements in this
>> order, I get prints on the first one, and zero counts on the second one:
>>
>> ssc(local[8])  // hyperthread dat sheezy
>> dstream1.print // works
>> dstream2.count.print // always prints 0
>>
>>
>>
>> If I do this, this happens:
>> ssc(local[4])
>> dstream1.print // doesn't work, just gives me the Time: .... ms message
>> dstream2.count.print // doesn't work, prints 0
>>
>> ssc(local[6])
>> dstream1.print // doesn't work, just gives me the Time: .... ms message
>> dstream2.count.print // works, prints 1
>>
>> Sometimes these results switch up, seemingly at random. How can I get
>> things to the point where I can develop and test my application locally?
>>
>> Thanks
>>
>>
>>
>>
>>
>>
>>
>

Reply via email to