Hi Jose,
you have to add additional Maven modules depending on the
connector/format you want to use. See this page [1] for more information.
Feel free to ask further questions if the description is not enough for you.
Regards,
Timo
[1]
https://ci.apache.org/projects/flink/flink-docs-release-1.6/dev/table/connect.html#further-tablesources-and-tablesinks
Am 13.09.18 um 11:50 schrieb jose farfan:
Hi
I am checking the documentation
https://ci.apache.org/projects/flink/flink-docs-release-1.6/dev/table/common.html#register-a-tablesink
Register a TableSink
A registered |TableSink| can be used to emit the result of a Table API
or SQL query
<https://ci.apache.org/projects/flink/flink-docs-master/dev/table/common.html#emit-a-table> to
an external storage system, such as a database, key-value store,
message queue, or file system (in different encodings, e.g., CSV,
Apache [Parquet, Avro, ORC], …).
Flink aims to provide TableSinks for common data formats and storage
systems. Please see the documentation about Table Sources and Sinks
<https://ci.apache.org/projects/flink/flink-docs-master/dev/table/sourceSinks.html> page
for details about available sinks and instructions for how to
implement a custom |TableSink|.
You can read that we can define different encodigos CSV, ORC, etc.
But in the source code, I can only find CsvTableFlink
How I can get a OrcTableFlink? Do I need to extend the TableSinkBase,
of there is another place to find that implementation
BR
Jose