Hi Guozhang,
thank you very much for the reply. It explained a lot more of your
reasoning to me
once again!
I have to disagree with you on the first point. As you mentioned the
Join Case.
A Join is usually a "logically" materialized table and its
KTableValueGetterSupplier
is to be used when one wants todo a lookup. But this is not at all what
is currently
Happening. The join merge processor currently maintains its own new
statestore
when join is invoked with Storename or supplier.
This describes the Issue I want to address perfectly. A Joined-Table
doesn't become
querieable because it is a JOINEDtable but because it is a joinedTABLE.
the emphasis here
is that we put the store logic with the join and not the table. It is
part of the join() method invocation and not the KTable Interface. This
abstraction is wrong.
This will always show its ugly face. Lets check your example:
stream.groupByKey(..).aggregate(.., Materializedas("store1")) //
this resulted KTable is materialized in order to complete the aggregation
operation
.filter(Materialized.as("store2"))
// this restuled KTable is not materialized but its
GetterSupplier is implemented to get values from "store1"
Currently this is only half true. For IQ a store is used that is maintained
by the KTableFilterProcessor, for downstream gets like joins the
ValueGetterSupplier is used
and indeed uses store1.
With the String overload (that you picked here on purpose I guess) it works
easier
as you can logically map those. But with the StateStoreSupplier it wouldn't.
you could not optimize this away as the user is expecting puts and gets to be
called
on what he supplied.
table1.filter(() -> true, InMemoryStore).filter(()->true,SQlLiteStore)
There is no way to optimize these away.
The same argument with the join holds for filter. Its not querrieable because
it got filtered
it is querrieable because its a KTable. That's where the emphasis needs to go.
The second point was new to me. So I had to think about this in more detail.
For me the breaking of the flow comes in very natural.
One Stream app I put most of my heart in has the these key metrics:
It has:
8 input topics.
3 1:n Joins
6 Group bys
2 built in Joins
2 built in left joins
some filters and mappers.
this is spanning 390 lines, counting java imports and some more stuff.
The whole topology forms a tree in wich the input topics usually get joined and
then collected to maps
and then joined again and collected to maps again. until they get send to 1
final output topic for consumption in our application servers.
I would argue it is impossible to express this topology as a chain of calls.
What happened is that
usually each join + groupBy tuple became its method taking in the builder and
return the Table
expressing the result of the sub topology. All Ktables that meet each other
with the same key in the
process get joined (most of that happening on the top level). This leads to
breaking in the fluent interface
quite naturally. especially if you have 2 KTables expressing sub-topologies
joined together. One subtopology had to go into the method call which is
unreasonable IMHO.
Even inside these methods we broke the chains. The variable names we used give
intermediate KTables really helped in making the semantics clear. They are much
like CTE's in hive or the required name in Mysql Subquerries. They help to mark
milestones inside the topology.
I would argue that for big topologies. (I haven't seen others but I think its
big) these milestones would
be the most important ones for IQ aswell. So i would argue breaking the chains
is not really a problem in
reality and it can help in many cases. As I laid out, we broke our chained
calls intuitively and it helped
other developers debugging the logic a lot. Even without detailed streams
understanding.
If one really do not want to stop the flow. I could argue that one could either
do something like this
KTable joinresult;
KTable<Integer,Integer> t1 = b.table("laa");
KTable<Integer,Integer> t2 = b.table("luu");
(joinresult = t1.join(t2, (value1, value2) -> value1 + value2))
.filter((key, value) -> false);
or write a little snitch like that
KTable<Integer,Integer> rememberTableandContinue(KTable<Integer,Integer> t){
joinresult = t;
return t;
}
for usuage as such
rememberTableandContinue(t1.join(t2, (value1, value2) -> value1 + value2))
.filter((key, value) -> false);
These suggestions might not looks so pretty. But in the context of breaking
bigger topology at milestones.
I think everything becomes acceptable really. Probably user would store that
intermediate KTable anyways just for clarity.
To summarize to give a KTable a name. I would always opt to the host language
variable names.
Tables used for IQ are probably tables that are of some sort more important to
the topology than
others and saving them separatly will increase the readability of topologies by
a lot IMO.
For these quick example Topologies that we have floating around in all places:
I am pretty sure one can go unbroken on them and usually the last table will be
the one that
is needed for IQ then.
Thanks again. The second point really got me thinking, as your perspective on
the importance
of "not break the fluent interface" was not clear to me. I hope I managed to
line out why I
think it shouldn't have such a big weight in the discussion.
PS.: check out Hive CTE, everyone loves them and our Analytic team is crazy for
them
because you can name them and that brings clarity. and you get rid of the
nesting and can
split everything into logical chunks of SQL. KTable variables are the CTE of
kafka streams.
One can probably sell this to people :)
Best Jan
Enjoyed your feedback! hope mine makes sense
On 03.08.2017 00:10, Guozhang Wang wrote:
Hello Jan,
Thanks for your proposal. As Bill mentioned the main difference is that we
extract the user-customizable materialization logic out of the topology
building DSL workflow. And the main motivations are in two folds:
1) efficiency wise, it allows some KTables to not be materialized if
unnecessary, saving one state store instance and changelog topic.
2) programming wise, it looks nicer to separate the topology construction
code from the KTable materialization for IQ uses code.
Here are my thoughts regarding these two points:
Regarding 1), I think with whichever the public APIs (either Damian's
proposal or yours), we can always apply the internal optimization to not
physically materialize the KTable. You can take a look at the internal
interface of "KTableValueGetterSupplier", which is used exactly for this
purposes such that a get call on a "logically" materialized KTable can be
traced back to its parent KTables that are physically materialized in a
state store. So following proposed APIs, for example:
stream.groupByKey(..).aggregate(.., Materializedas("store1")) //
this resulted KTable is materialized in order to complete the aggregation
operation
.filter(Materialized.as("store2"))
// this restuled KTable is not materialized but its
GetterSupplier is implemented to get values from "store1"
Or
table1 = stream.groupByKey(..).aggregate(..);
table2 = table1.filter();
tabel1.queryHandle("store1"); // this resulted KTable is materialized
in order to complete the aggregation operation
tabel1.queryHandle("store2") // this restuled KTable is not
materialized but its GetterSupplier is implemented to get values from
"store1"
When user query a value for "store2" which is not actually materialized
into a state store, the GetterSupplier will be triggered to in turn query
the store for "store1", and then apply the filter operator on-the-fly to
return the value. So the bottom line is, we can achieve the same efficiency
optimization with either of the public APIs.
Regarding 2), I actually have proposed a similar API to yours earlier in
this discussion thread:
---------------------------------------
// specifying the topology, should be concise and conveniently
concatenated, no specs of materialization at all
KStream stream1 = builder.stream();
KTable table1 = stream1.groupby(...).aggregate(initializer, aggregator,
sessionMerger, sessionWindows); // do not allow to pass-in a state store
supplier here any more
// additional code to the topology above, could be more prescriptive
than descriptive
// only advanced users would want to code in both parts above; while other
users would only code the topology as above.
table1.materialize("queryableStoreName"); // or..
table1.materialize("queryableStoreName").enableCaching().enableLogging();
// or..
table1.materialize(stateStoreSupplier); // we check type (key-value types,
windowed or not etc) at starting time and add the metrics / logging /
caching / windowing wrapper on top of the store, or..
table1.materialize(stateStoreSupplier).enableCaching().enableLogging(); //
etc..
---------------------------------------
But one caveat of that, as illustrated above, is that you need to have
separate object of the KTable in order to call either "queryHandle" or
"materialize" (whatever the function name is) for the specifications of
materialization options. This can break the concatenation of the topology
construction part of the code, that you cannot simply add one operator
directly after another. So I think this is a trade-off we have to make and
the current approach looks better in this regard.
Guozhang
On Wed, Aug 2, 2017 at 2:07 PM, Jan Filipiak<jan.filip...@trivago.com>
wrote:
Hi Bill,
totally! So in the original discussion it was mentioned that the overloads
are nasty when implementing new features. So we wanted to get rid of them.
But what I felt was that the
copy & pasted code in the KTableProcessors for maintaining IQ stores was
as big as a hurdle as the overloads.
With this proposal I try to shift things into the direction of getting IQ
for free if
KTableValueGetterSupplier is properly implemented (like getting join for
free then). Instead of having the code for maintaining IQ stores all the
places. I realized I can do that while getting rid of the overloads, that
makes me feel my proposal is superior.
Further I try to optimize by using as few stores as possible to give the
user what he needs. That should save all sorts of resources while allowing
faster rebalances.
The target ultimately is to only have KTableSource and the Aggregators
maintain a Store and provide a ValueGetterSupplier.
Does this makes sense to you?
Best Jan
On 02.08.2017 18:09, Bill Bejeck wrote:
Hi Jan,
Thanks for the effort in putting your thoughts down on paper.
Comparing what I see from your proposal and what is presented in KIP-182,
one of the main differences is the exclusion of an`Materialized` instance
in the `KTable` methods.
Can you go into more detail why this is so and the specific problems is
avoids and or solves with this approach?
Thanks!
Bill
On Wed, Aug 2, 2017 at 4:19 AM, Damian Guy <damian....@gmail.com <mailto:
damian....@gmail.com>> wrote:
Hi Jan,
Thanks for taking the time to put this together, appreciated. For the
benefit of others would you mind explaining a bit about your
motivation?
Cheers,
Damian
On Wed, 2 Aug 2017 at 01:40 Jan Filipiak <jan.filip...@trivago.com
<mailto:jan.filip...@trivago.com>> wrote:
> Hi all,
>
> after some further discussions, the best thing to show my Idea
of how it
> should evolve would be a bigger mock/interface description.
> The goal is to reduce the store maintaining processors to only the
> Aggregators + and KTableSource. While having KTableSource optionally
> materialized.
>
> Introducing KTable:copy() will allow users to maintain state
twice if
> they really want to. KStream::join*() wasn't touched. I never
personally
> used that so I didn't feel
> comfortable enough touching it. Currently still making up my
mind. None
> of the suggestions made it querieable so far. Gouzhangs
'Buffered' idea
> seems ideal here.
>
> please have a look. Looking forward for your opinions.
>
> Best Jan
>
>
>
> On 21.06.2017 17 <tel:21.06.2017%2017>:24, Eno Thereska wrote:
> > (cc’ing user-list too)
> >
> > Given that we already have StateStoreSuppliers that are
configurable
> using the fluent-like API, probably it’s worth discussing the other
> examples with joins and serdes first since those have many
overloads and
> are in need of some TLC.
> >
> > So following your example, I guess you’d have something like:
> > .join()
> > .withKeySerdes(…)
> > .withValueSerdes(…)
> > .withJoinType(“outer”)
> >
> > etc?
> >
> > I like the approach since it still remains declarative and
it’d reduce
> the number of overloads by quite a bit.
> >
> > Eno
> >
> >> On Jun 21, 2017, at 3:37 PM, Damian Guy <damian....@gmail.com
<mailto:damian....@gmail.com>> wrote:
> >>
> >> Hi,
> >>
> >> I'd like to get a discussion going around some of the API
choices we've
> >> made in the DLS. In particular those that relate to stateful
operations
> >> (though this could expand).
> >> As it stands we lean heavily on overloaded methods in the
API, i.e,
> there
> >> are 9 overloads for KGroupedStream.count(..)! It is becoming
noisy and i
> >> feel it is only going to get worse as we add more optional
params. In
> >> particular we've had some requests to be able to turn caching
off, or
> >> change log configs, on a per operator basis (note this can
be done now
> if
> >> you pass in a StateStoreSupplier, but this can be a bit
cumbersome).
> >>
> >> So this is a bit of an open question. How can we change the DSL
> overloads
> >> so that it flows, is simple to use and understand, and is easily
> extended
> >> in the future?
> >>
> >> One option would be to use a fluent API approach for
providing the
> optional
> >> params, so something like this:
> >>
> >> groupedStream.count()
> >> .withStoreName("name")
> >> .withCachingEnabled(false)
> >> .withLoggingEnabled(config)
> >> .table()
> >>
> >>
> >>
> >> Another option would be to provide a Builder to the count
method, so it
> >> would look something like this:
> >> groupedStream.count(new
> >> CountBuilder("storeName").withCachingEnabled(false).build())
> >>
> >> Another option is to say: Hey we don't need this, what are
you on about!
> >>
> >> The above has focussed on state store related overloads, but
the same
> ideas
> >> could be applied to joins etc, where we presently have many join
> methods
> >> and many overloads.
> >>
> >> Anyway, i look forward to hearing your opinions.
> >>
> >> Thanks,
> >> Damian
>
>