Thank you all for you answers. It's ok with BatchTableSource<Row>
All the best François 2018-08-26 17:40 GMT+02:00 Rong Rong <walter...@gmail.com>: > Yes you should be able to use Row instead of Tuple in your > BatchTableSink<T>. > There's sections in Flink documentation regarding mapping of data types to > table schemas [1]. and table can be converted into various typed DataStream > [2] as well. Hope these are helpful. > > Thanks, > Rong > > [1] https://ci.apache.org/projects/flink/flink-docs- > release-1.6/dev/table/common.html#mapping-of-data-types-to-table-schema > [2] https://ci.apache.org/projects/flink/flink-docs- > release-1.6/dev/table/common.html#convert-a-table-into-a- > datastream-or-dataset > > > > On Fri, Aug 24, 2018 at 8:04 AM françois lacombe < > francois.laco...@dcbrain.com> wrote: > >> Hi Timo, >> >> Thanks for your answer >> I was looking for a Tuple as to feed a BatchTableSink<T> subclass, but it >> may be achived with a Row instead? >> >> All the best >> >> François >> >> 2018-08-24 10:21 GMT+02:00 Timo Walther <twal...@apache.org>: >> >>> Hi, >>> >>> tuples are just a sub category of rows. Because the tuple arity is >>> limited to 25 fields. I think the easiest solution would be to write your >>> own converter that maps rows to tuples if you know that you will not need >>> more than 25 fields. Otherwise it might be easier to just use a >>> TextInputFormat and do the parsing yourself with a library. >>> >>> Regards, >>> Timo >>> >>> >>> Am 23.08.18 um 18:54 schrieb françois lacombe: >>> >>> Hi all, >>>> >>>> I'm looking for best practices regarding Tuple<T> instances creation. >>>> >>>> I have a TypeInformation object produced by AvroSchemaConverter. >>>> convertToTypeInfo("{...}"); >>>> Is this possible to define a corresponding Tuple<T> instance with it? >>>> (get the T from the TypeInformation) >>>> >>>> Example : >>>> { >>>> "type": "record", >>>> "fields": [ >>>> { "name": "field1", "type": "int" }, >>>> { "name": "field2", "type": "string"} >>>> ]} >>>> = Tuple2<Int,String> >>>> >>>> The same question rises with DataSet or other any record handling class >>>> with parametrized types. >>>> >>>> My goal is to parse several CsvFiles with different structures >>>> described in an Avro schema. >>>> It would be great to not hard-code structures in my Java code and only >>>> get types information at runtime from Avro schemas >>>> >>>> Is this possible? >>>> >>>> Thanks in advance >>>> >>>> François Lacombe >>>> >>> >>> >>> >>