You can use 
foreachRDD<http://spark.apache.org/docs/latest/streaming-programming-guide.html#output-operations-on-dstreams>
 to get access to the batch API in streaming jobs.


From: Amir Rahnama [mailto:amirrahn...@gmail.com]
Sent: Thursday, November 12, 2015 12:11 AM
To: ayan guha <guha.a...@gmail.com>
Cc: user <user@spark.apache.org>
Subject: Re: How can you sort wordcounts by counts in 
stateful_network_wordcount.py example

Hi Ayan,

Thanks for help,

Your example is not the streaming example. There we don't have sortByKey.



On Wed, Nov 11, 2015 at 11:35 PM, ayan guha 
<guha.a...@gmail.com<mailto:guha.a...@gmail.com>> wrote:
how about this?

sorted = running_counts.map(lambda t: t[1],t[0]).sortByKey()

Basically swap key and value of the RDD and then sort?

On Thu, Nov 12, 2015 at 8:53 AM, Amir Rahnama 
<amirrahn...@gmail.com<mailto:amirrahn...@gmail.com>> wrote:
Hey,

Anybody knows how can one sort the result in the stateful example?

Python would be prefered.

https://github.com/apache/spark/blob/859dff56eb0f8c63c86e7e900a12340c199e6247/examples/src/main/python/streaming/stateful_network_wordcount.py
--
Thanks and Regards,

Amir Hossein Rahnama

Tel: +46 (0) 761 681 102
Website: www.ambodi.com<http://www.ambodi.com>
Twitter: @_ambodi<https://twitter.com/_ambodi>



--
Best Regards,
Ayan Guha



--
Thanks and Regards,

Amir Hossein Rahnama

Tel: +46 (0) 761 681 102
Website: www.ambodi.com<http://www.ambodi.com>
Twitter: @_ambodi<https://twitter.com/_ambodi>

Reply via email to