I don't see this happening without a store. You can try parquet on top of hdfs. This will at least avoid third party systems burden. On 09 Apr 2016 9:04 AM, "Daniela S" <daniela_4...@gmx.at> wrote:
> Hi, > > I would like to cache values and to use only the latest "valid" values to > build a sum. > In more detail, I receive values from devices periodically. I would like > to add up all the valid values each minute. But not every device sends a > new value every minute. And as long as there is no new value the old one > should be used for the sum. As soon as I receive a new value from a device > I would like to overwrite the old value and to use the new one for the sum. > Would that be possible with Spark Streaming only? Or would I need a kind of > distributed cache, like Redis? I also need to group the sums per region. > Should that be done before I store the values in the cache or afterwards? > > Thank you in advance. > > Regards, > Daniela > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > For additional commands, e-mail: user-h...@spark.apache.org > >