The first level must be static, but a field can hold a complex object with nested data. So you could have a schema with a single JSON object as field.
Best, Fabian 2017-01-19 19:31 GMT+01:00 Nihat Hosgur <nhos...@gmail.com>: > Hi Fabian, > I just want to make sure there is no misunderstanding. So what I've > understood from your response is that regardless table source is KafkaTable > or not we need to provide static schema. > Best, > Nihat > > > On Thu, Jan 19, 2017 at 2:50 AM Fabian Hueske <fhue...@gmail.com> wrote: > > Hi Nihat, > > at the current state, Flink's SQL and Table APIs require a static schema. > You could use an JSON object as value and implement scalar functions to > extract fields, but that would not be very usable. > > Best, Fabian > > 2017-01-19 2:59 GMT+01:00 Nihat Hosgur <nhos...@gmail.com>: > > Hi there, > We are evaluating flink SQL to understand if it would be a better fit > instead of Spark. So far we loved how natural it is to consume streams on > Flink. > > We do read bunch of Kafka topics and like to join those streams and > eventually run some SQL queries. We've used Kafka tables yet if I'm not > mistaken we must provide the schema up front. However on many use cases of > ours we don't have schema information available up front to us. With spark > we didn't have to have the schema. I wonder if it is doable with Flink if > so any reference or even some sample code would be appreciated. > > Thanks, > Nihat > -- > Thanks, > Nihat > > > -- > Thanks, > Nihat >