Hi Nihat, at the current state, Flink's SQL and Table APIs require a static schema. You could use an JSON object as value and implement scalar functions to extract fields, but that would not be very usable.
Best, Fabian 2017-01-19 2:59 GMT+01:00 Nihat Hosgur <nhos...@gmail.com>: > Hi there, > We are evaluating flink SQL to understand if it would be a better fit > instead of Spark. So far we loved how natural it is to consume streams on > Flink. > > We do read bunch of Kafka topics and like to join those streams and > eventually run some SQL queries. We've used Kafka tables yet if I'm not > mistaken we must provide the schema up front. However on many use cases of > ours we don't have schema information available up front to us. With spark > we didn't have to have the schema. I wonder if it is doable with Flink if > so any reference or even some sample code would be appreciated. > > Thanks, > Nihat > -- > Thanks, > Nihat >