I agree with Jorn et al on this. An alternative approach would be best. as
a  24 hour operation sounds  like a classic batch job more suitable for
later reporting. This happens all the time in RDBMS.

As I understand and within Spark the sliding interval can only be used
after that window length (in this case 24 hours) has elapsed. You might as
well use normal storage for it. It may be slower but would be far more
manageable.

Otherwise use other suggestions I made.


HTH

Dr Mich Talebzadeh



LinkedIn * 
https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
<https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*



http://talebzadehmich.wordpress.com



On 9 May 2016 at 15:15, firemonk9 <dhiraj.peech...@gmail.com> wrote:

> I have not come across official docs in this regard how ever if you use 24
> hour window size, you will need to have memory big enough to fit the stream
> data for 24 hours. Usually memory is the limiting factor for the window
> size.
>
> Dhiraj Peechara
>
>
>
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/How-big-the-spark-stream-window-could-be-tp26899p26903.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>
>

Reply via email to