Hi,

it seems that this is what you have to do for now. However, I see that it
would be nice if Flink would allow reading from multiple configuration
files, so that you can have a "common configuration" and a "per cluster"
configuration.

I filed a JIRA ticket for a feature request:
https://issues.apache.org/jira/browse/FLINK-19828


On Tue, Oct 27, 2020 at 10:54 AM Shachar Carmeli <carmeli....@gmail.com>
wrote:

> Hi,
> Thank you for your reply,
> WE are deploying on kubernetes and the xml is part of the  common config
> map to all flink jobs we have(or at least was for previous versions)
>
> This means that we need to duplicate the configuration in the
> flink-conf.yaml for each job
> instead of having a common configmap
>
> Thanks,
> Shachar
>
> On 2020/10/27 08:48:17, Robert Metzger <rmetz...@apache.org> wrote:
> > Hi Shachar,
> >
> > Why do you want to use the core-site.xml to configure the file system?
> >
> > Since we are adding the file systems as plugins, their initialization is
> > customized. It might be the case that we are intentionally ignoring xml
> > configurations from the classpath.
> > You can configure the filesystem in the flink-conf.yaml file.
> >
> >
> > On Sun, Oct 25, 2020 at 7:56 AM Shachar Carmeli <carmeli....@gmail.com>
> > wrote:
> >
> > > Hi,
> > > I'm trying to define filesystem to flink 1.11 using core-site.xml
> > > I tried adding in the flink-conf.yaml env.hadoop.conf.dir and I see it
> is
> > > added to the classpath
> > > also adding environment variable HADOOP_CONF_DIR didn't help
> > >
> > > The flink 1.11.2 is running on docker using kubernetes
> > >
> > > I added hadoop using plugin as mentioned in
> > >
> https://ci.apache.org/projects/flink/flink-docs-stable/ops/filesystems/s3.html#hadooppresto-s3-file-systems-plugins
> > >
> > > when configure the parameters manually I can connect to the local s3a
> > > server
> > > So it looks like the flink is not reading the core-site.xml file
> > >
> > > please advise
> > >
> > > Thanks,
> > > Shachar
> > >
> >
>

Reply via email to