Hi Timothy,

The issue would require a refactor FileSystems abstraction to allow
multiple FileSystems objects of the same FileSystem type being configured
differently. While this would improve the code quality and enable such use
cases, I currently have no capacity to work on it or guide it.

If you are interested in working on it, I can try to find someone to
shepherd.

On Mon, Dec 13, 2021 at 7:13 PM Timothy James <t...@decodable.co> wrote:

> Thank you Timo.  Hi Arvid!
>
> I note that that ticket proposes two alternatives for solution.  The first
>
> > Either we allow a properties map similar to Kafka or Kinesis properties
> to our connectors.
>
> seems to solve our problem.  The second, much more detailed, appears
> unrelated to our needs:
>
> > Or something like:
> > Management of two properties related S3 Object management:
> > ...
>
> The ticket is unassigned and has been open for more than a year.  It looks
> like you increased the ticket priority, thank you.
>
> Tim
>
> On Mon, Dec 13, 2021 at 6:52 AM Timo Walther <twal...@apache.org> wrote:
>
>> Hi Timothy,
>>
>> unfortunetaly, this is not supported yet. However, the effort will be
>> tracked under the following ticket:
>>
>> https://issues.apache.org/jira/browse/FLINK-19589
>>
>> I will loop-in Arvid (in CC) which might help you in contributing the
>> missing functioniality.
>>
>> Regards,
>> Timo
>>
>>
>> On 10.12.21 23:48, Timothy James wrote:
>> > Hi,
>> >
>> > The Hadoop s3a library itself supports some properties we need, but the
>> > "FileSystem SQL Connector" (via FileSystemTableFactory) does not pass
>> > connector options for these to the "Hadoop/Presto S3 File Systems
>> > plugins" (via S3FileSystemFactory).
>> >
>> > Instead, only Job-global Flink config values are passed to Hadoop s3a.
>> > That won't work for us: we need to vary these values per Flink SQL
>> > table, and not override our config for other use of S3 (such as Flink
>> > checkpointing).
>> >
>> > Contrast this with the Kafka connector, which supports an analogous
>> > "properties.*" prefixed pass-through mechanism, and the Kinesis
>> > connector, which supports all the specific properties we would need out
>> > of the box.
>> >
>> > Our current intent is to alter FileSystemTableFactory to follow the
>> > "properties.*" approach used by the Kafka connector.
>> >
>> > *** ➡️ Our questions for you: ⬅️
>> > - Know of anything like this? Anybody solved this?
>> > - Know of anything that's going to break this approach?
>> > - What are we missing?
>> >
>> > For context, our particular use case requires options like:
>> > - fs.s3a.assumed.role.arn
>> > - fs.s3a.aws.credentials.provider, (or some other mechanism to pass
>> > externalId)
>> >
>> > We imagine there would be other use cases for this, and if we build it
>> > ourselves there's the possibility of contributing it to the Flink repo
>> > for everybody.
>> >
>> > Relevant documentation:
>> > -
>> >
>> https://nightlies.apache.org/flink/flink-docs-release-1.13/docs/connectors/table/filesystem/
>> > <
>> https://nightlies.apache.org/flink/flink-docs-release-1.13/docs/connectors/table/filesystem/
>> >
>> > -
>> >
>> https://nightlies.apache.org/flink/flink-docs-master/docs/deployment/filesystems/s3/#hadooppresto-s3-file-systems-plugins
>> > <
>> https://nightlies.apache.org/flink/flink-docs-master/docs/deployment/filesystems/s3/#hadooppresto-s3-file-systems-plugins
>> >
>> > -
>> >
>> https://hadoop.apache.org/docs/stable/hadoop-aws/tools/hadoop-aws/index.html
>> > <
>> https://hadoop.apache.org/docs/stable/hadoop-aws/tools/hadoop-aws/index.html
>> >
>> > -
>> >
>> https://nightlies.apache.org/flink/flink-docs-release-1.13/docs/connectors/table/kafka/#properties
>> > <
>> https://nightlies.apache.org/flink/flink-docs-release-1.13/docs/connectors/table/kafka/#properties
>> >
>> > -
>> >
>> https://nightlies.apache.org/flink/flink-docs-release-1.13/docs/connectors/table/kinesis/#aws-credentials-role-externalid
>> > <
>> https://nightlies.apache.org/flink/flink-docs-release-1.13/docs/connectors/table/kinesis/#aws-credentials-role-externalid
>> >
>> >
>> > Thank you!
>> >
>> > Tim James
>> > Decodable.co
>> >
>>
>>

Reply via email to