1. How can I create a kafka table that can use headers and map them to columns? 
Currently, I am using KafkaDeserilizationSchema to create a DataStream, and 
then I convert that DataStream into a Table. I would like to use a more direct 
approach.

2. What is the recommended way to enrich a kafka table or data-stream with 
data-from postgres?
        a) kafka table and JDBC temporal dimension table with temporal join and 
lookup cache setup
        b) data-stream with async io which connects via JDBC.  (note that 
asycio does not support Keyed State cache)
        c) data-stream rich function or process function that uses Keyed State.

3. When using a kafka told and JDBC temporal dimension table how do I prevent N 
+ 1 queries per join row?

        When I issued a query such as this:

        SELECT k.name, t1.id, t2.metadata, SUM(k.cost)
        FROM kafka_table AS k
        JOIN jdbc_table_one AS t1 ON k.t1_id = t1.ID
        LEFT JOIN jdbc_table_two FOR SYSTEM_TIME AS OF k.proc_time AS t2 ON 
t1.t2_id = t2.id AND t2.name = k.name
        GROUP BY TUMBLE (k.proc_time, INTERVAL '3' MINUTE), k.name, t1.id, 
t2.metadata

        My PostgreSQL sql logs show that jdbc_table_two has a query per each 
distinct t2.name.

        In a real production system, that would be 200,000 queries!

4. When using a JDBC temporal dimension table does Flink retrieve the from the 
database asynchronously , or is it possible for Flink to multiple join rows at 
time with a IN (subquery) syntax?



Reply via email to