Bassically I want to "monitor" a bucket on S3 and every file that gets
created in that bucket read it and stream it.

If I understand correctly, I can just use env.readCsvFile() and config to
continuously read a folder path?


On Tue., Jul. 28, 2020, 1:38 a.m. Jingsong Li, <jingsongl...@gmail.com>
wrote:

> Hi John,
>
> Do you mean you want to read S3 CSV files using partition/bucket pruning?
>
> If just using the DataSet API, you can use CsvInputFormat to read csv
> files.
>
> If you want to use Table/Sql API, In 1.10, Csv format in table not support
> partitioned table. So the only way is specific the partition/bucket path,
> and read single directory.
>
> In 1.11, the Table/Sql filesystem connector with csv format supports
> partitioned table, complete support partition semantics.
>
> [1]
> https://ci.apache.org/projects/flink/flink-docs-release-1.11/dev/table/connectors/filesystem.html
>
> Best,
> Jingsong
>
> On Mon, Jul 27, 2020 at 10:54 PM John Smith <java.dev....@gmail.com>
> wrote:
>
>> Hi, using Flink 1.10
>>
>> 1- How do we go about reading CSV files that are copied to s3 buckets?
>> 2- Is there a source that can tail S3 and start reading a CSV when it is
>> copied to S3?
>> 3- Is that part of the table APIs?
>>
>
>
> --
> Best, Jingsong Lee
>

Reply via email to