Hi Sergi,

Thanks for your thoughts on this. 

I think everything is already available in Ignite to react upon cache 
operations and what needs to be done is to check if that operation should also 
change the resultset of a query. I think this is also how "Continuous Queries" 
work in Ignite.
What I see as a hurdle though, is the ACID compliance (ie. inventory is updated 
within a transaction and the materialized view 'cache' is updated at the same 
time or maybe eventual consistent with a step in between by eg writing to a WAL 
or a temp cache)

Let me elaborate a little bit on my example, also giving an example with 
streaming of IoT data and refer to how Kafka uses this idea (they call it 
Interactive Queries):


--- EXAMPLE 1: simple shopping basket---
(I know I could do some manual denormalisation in the example below by putting 
inventory in shopping basket cache, but we have a much more challenging use 
case that would be too hard to describe in a mail, so if you could follow the 
reasoning and imagine that manual denormalisation is not possible, thinking 
CQRS would also help)

Coming back to the shopping basket example, where you would like to show the 
end user the amount of items still available, the resultset of the materialized 
view is not that big. The join would be between 'shoppingbasket' and 
'inventory' where shoppingbasket is 'active', ie. materialized view cache (on 
ignite node A) would have the fields

shoppingbasketID | productItem | amount | inventoryAvailability

The shoppingbasket cache (on ignite node X) contains all active and inactive 
(finalized, discarded,... whatever reason) shoppingbaskets and the inventory 
cache (on ignite node Y) all product items with the amount still available. 

The materialized view would then only contain active shoppingbaskets with for 
each productitem in the shoppingbasket the available items.

Whenever a shoppingbasket is updated, the 'query' (ie. materialized view) is 
triggered (--> an event on the shoppingbasket cache), the where clause is 
checked (ie. shoppingbasket must be active) and the materialized view is 
updated (ie. item is added to cache or updated and inventory cache is checked 
for abailability of that item)

Same when inventory is updated, the 'query' (ie.materialized view) is triggered 
(--> an event on inventory cache), each entry of the productitem in the 
materialized view is updated (ie. availbility for that productitem is updated 
in each active shoppingbasket in the view)

This is a kind of 'automagical' denormalization based on queries known at 
development time and hence deploy time in the grid (ie. a query for 
denormalized views needs to be registered/deployed). Seeing what is happening 
in the market with other tools (Kafka, Samza), this would be a very powerful 
feature for Ignite and would be a huge speed optimisation for querying.


--- EXAMPLE 2: streaming from IoT devices ---
This would be even a better example, where a materialized view is actually a 
join of two or more streams, each filling up a cache in Ignite. 

Today in ignite it is already possible to have continuous queries to get new 
incoming data transparantly.

The materialized view would only contain relevant *state* for certain things 
coming from that stream. This is especially for IoT interesting because you 
could keep for example the average or median of the temperature of thousands of 
devices in the materialized view cache while data on the stream can be 
discarded very quickly.


--- SAMZA/KAFKA ---
This is how Apache Samza works 
(https://www.confluent.io/blog/turning-the-database-inside-out-with-apache-samza/
 ) and those ideas are also the basis for how the Interactive Queries from 
Kafka works 
(https://www.confluent.io/blog/unifying-stream-processing-and-interactive-queries-in-apache-kafka/
 ), but the big advandage with Ignite is that everything is already in place 
and you don't need an extra embedded database like RocksDB. All the cons they 
describe could be solved in Ignite out of the box.



Regards,

Sven



 
SVEN BEAUPREZ 
 
L e a d   A r c h i t e c t

 

De Kleetlaan 5, B-1831 Diegem

www.theglue.com <http://www.theglue.com/>

On 18/09/2018, 15:10, "Sergi Vladykin" <sergi.vlady...@gmail.com> wrote:

    Sven,
    
    Support of materialized views sounds like a huge project. I would not
    expect it to appear in Ignite soon.
    
    As far as I see you have problems with data collocation. If you can not
    store the original data in replicated caches, then these views will be huge
    as well and thus must be partitioned with some collocation. So it does not
    change the picture too much actually.
    
    I would suggest to go for a manual denormalization: put the same value with
    different affinity keys.
    
    Sergi
    
    пн, 17 сент. 2018 г. в 14:22, Sven Beauprez <sven.beaup...@theglue.com>:
    
    > Hi Dmitry,
    >
    > Yes we can use those solutions in some cases, but not always.
    >
    > Replication is indeed the simplest, but sadly enough replicated caches are
    > too much overhead. We have often a minimum of 12 nodes and all data must
    > stay in sync 12x then. We do use it for small caches that don't need a lot
    > of updates.
    >
    > We use colocation all over the place. Colocation based on affinity keys is
    > not possible though for distinct data sets with only some very specific
    > relationships with _some other_ dataset, well known before hand.
    >  (eg. for example -not our exact use case which is more complex- items in
    > a shopping basket with items from product inventory, both are in different
    > caches managed on other nodes and it is not possible to denormalize such
    > that the shopping basket knows the amount of availble items)
    >
    >
    > Regards,
    >
    > Sven
    >
    >
    >
    >
    > SVEN BEAUPREZ
    >
    > L e a d   A r c h i t e c t
    >
    >
    >
    > De Kleetlaan 5, B-1831 Diegem
    >
    > www.theglue.com <http://www.theglue.com/>
    >
    > On 17/09/2018, 10:37, "Dmitriy Setrakyan" <dsetrak...@apache.org> wrote:
    >
    >     Hi Sven,
    >
    >     I will let others comment on the materialized view suggestion, but I
    > have
    >     another question.
    >
    >     *As we all know, joins are a nightmare in a distributed system*
    >
    >
    >     Have you considered collocation or replication? If a table is
    > replicated,
    >     then it will be present on all the nodes and all joins will be fast.
    > If two
    >     partitioned tables are colocated based on some affinity key, then
    > joins on
    >     that affinity key will be fast as well.
    >
    >     Both, colocation and replication are supported by Ignite. Will any of
    > these
    >     approaches work for you?
    >
    >     D.
    >
    >     On Mon, Sep 17, 2018 at 11:04 AM Sven Beauprez <
    > sven.beaup...@theglue.com>
    >     wrote:
    >
    >     > All,
    >     >
    >     >
    >     >
    >     > We are in a situation where we have to query data over several
    > caches. As
    >     > we all know, joins are a nightmare in a distributed system and I
    > know there
    >     > are other means like denormalisation, but it is not sufficient
    > anymore in
    >     > some cases we have and we need the joins.
    >     >
    >     >
    >     >
    >     > We mainly work in an OLTP context, where queries are known in
    > advance (ie
    >     > dev time) and inpsired by following blog of several years ago, I was
    >     > wondering if the concept of “materialized views” could make it into
    > Apache
    >     > Ignite.
    >     >
    >     > (
    >     >
    > 
https://www.confluent.io/blog/turning-the-database-inside-out-with-apache-samza/
    >     > )
    >     >
    >     >
    >     >
    >     > It would work as follows:
    >     >
    >     >    - A query must register itself in Ignite at startup time (eg. via
    >     >    configuration) or during run time (eg. API call)
    >     >    - The registered query is parsed and a new “view” cache is 
created
    >     >    which will ‘cache’ the resultset of the query (could take a
    > while, but
    >     >    intermediate status can be “warming up” and “hot” when ready)
    >     >    - All caches involved in the joins are now monitored for CUD
    >     >    operations and relevant data is stored in the new “view” cache so
    > the view
    >     >    gets updated in real time
    >     >    - All operations must be ACID compliant
    >     >    - The view is queried via a very trivial select statement
    >     >
    >     >
    >     >
    >     > Would that be feasible as a new feature?
    >     >
    >     >
    >     >
    >     >
    >     >
    >     > Regards,
    >     >
    >     >
    >     >
    >     > Sven
    >     >
    >     >
    >     >
    >     >
    >     >
    >     >
    >     >
    >     > [image: cid:image001.png@01D3007B.4D007790]
    >     >
    >     >
    >     >
    >     > SVEN BEAUPREZ
    >     >
    >     >
    >     >
    >     > L e a d   A r c h i t e c t
    >     >
    >     >
    >     >
    >     > De Kleetlaan 5, B-1831 Diegem
    >     >
    >     > www.theglue.com
    >     >
    >
    >
    >
    

Reply via email to