This sounds really really weird. Can you give me a piece of code that I can run to reproduce this issue myself?
TD On Tue, Jul 15, 2014 at 12:02 AM, Walrus theCat <walrusthe...@gmail.com> wrote: > This is (obviously) spark streaming, by the way. > > > On Mon, Jul 14, 2014 at 8:27 PM, Walrus theCat <walrusthe...@gmail.com> > wrote: > >> Hi, >> >> I've got a socketTextStream through which I'm reading input. I have >> three Dstreams, all of which are the same window operation over that >> socketTextStream. I have a four core machine. As we've been covering >> lately, I have to give a "cores" parameter to my StreamingSparkContext: >> >> ssc = new StreamingContext("local[4]" /**TODO change once a cluster is up >> **/, >> "AppName", Seconds(1)) >> >> Now, I have three dstreams, and all I ask them to do is print or count. >> I should preface this with the statement that they all work on their own. >> >> dstream1 // 1 second window >> dstream2 // 2 second window >> dstream3 // 5 minute window >> >> >> If I construct the ssc with "local[8]", and put these statements in this >> order, I get prints on the first one, and zero counts on the second one: >> >> ssc(local[8]) // hyperthread dat sheezy >> dstream1.print // works >> dstream2.count.print // always prints 0 >> >> >> >> If I do this, this happens: >> ssc(local[4]) >> dstream1.print // doesn't work, just gives me the Time: .... ms message >> dstream2.count.print // doesn't work, prints 0 >> >> ssc(local[6]) >> dstream1.print // doesn't work, just gives me the Time: .... ms message >> dstream2.count.print // works, prints 1 >> >> Sometimes these results switch up, seemingly at random. How can I get >> things to the point where I can develop and test my application locally? >> >> Thanks >> >> >> >> >> >> >> >