Sweet, this was the issue. I got this to work by copying the s3 jar over to plugins for the client container.
Thanks for all of the help! The Table API is sweet! On Mon, Sep 14, 2020 at 11:14 PM Dan Hill <quietgol...@gmail.com> wrote: > Yes, the client runs in K8. It uses a different K8 config than the Helm > chart and does not load the plugins. Does the client use the same plugin > structure as the Flink job/task manager? I can try using it tomorrow. > > Cool, that link would work too. > > Thanks, Arvid! > > > On Mon, Sep 14, 2020 at 10:59 PM Arvid Heise <ar...@ververica.com> wrote: > >> Hi Dan, >> >> Are you running the client also in K8s? If so you need an initialization >> step, where you add the library to the plugins directory. Putting it into >> lib or into the user jar doesn't work anymore as we removed the shading in >> s3 in Flink 1.10. >> >> The official Flink docker image has an easy way to add these plugins [1]. >> >> [1] >> https://ci.apache.org/projects/flink/flink-docs-stable/ops/deployment/docker.html#using-plugins >> >> On Tue, Sep 15, 2020 at 6:40 AM Dan Hill <quietgol...@gmail.com> wrote: >> >>> Thanks for the update! >>> >>> I'm trying a bunch of combinations on the client side to get the S3 >>> Filesystem to be picked up correctly. Most of my attempts involved >>> building into the job jar (which I'm guessing won't work). I then started >>> getting issues with ClassCastExceptions. >>> >>> I might try a little more tomorrow (e.g. modifying the custom image). >>> If I can't get it, I'll roll back to a previous Flink version that works. >>> >>> Caused by: java.lang.ClassCastException: >>> org.codehaus.janino.CompilerFactory cannot be cast to >>> org.codehaus.commons.compiler.ICompilerFactory >>> >>> at >>> org.codehaus.commons.compiler.CompilerFactoryFactory.getCompilerFactory(CompilerFactoryFactory.java:129) >>> >>> at >>> org.codehaus.commons.compiler.CompilerFactoryFactory.getDefaultCompilerFactory(CompilerFactoryFactory.java:79) >>> >>> at >>> org.apache.calcite.rel.metadata.JaninoRelMetadataProvider.compile(JaninoRelMetadataProvider.java:431) >>> >>> ... 51 more >>> >>> >>> On Mon, Sep 14, 2020 at 7:03 PM Jingsong Li <jingsongl...@gmail.com> >>> wrote: >>> >>>> Hi Dan, >>>> >>>> I think Arvid and Dawid are right, as a workaround, you can try making >>>> S3Filesystem works in the client. But for a long term solution, we can fix >>>> it. >>>> >>>> I created https://issues.apache.org/jira/browse/FLINK-19228 for >>>> tracking this. >>>> >>>> Best, >>>> Jingsong >>>> >>>> On Mon, Sep 14, 2020 at 3:57 PM Dawid Wysakowicz < >>>> dwysakow...@apache.org> wrote: >>>> >>>>> Hi Dan, >>>>> >>>>> As far as I checked in the code, the FileSystemSink will try to create >>>>> staging directories from the client. I think it might be problematic, as >>>>> your case shows. We might need to revisit that part. I am cc'ing Jingsong >>>>> who worked on the FileSystemSink. >>>>> >>>>> As a workaround you might try putting the s3 plugin on the CLI >>>>> classpath (not sure if plugins work for the CLI through the /plugins >>>>> directory). >>>>> >>>>> Best, >>>>> >>>>> Dawid >>>>> On 10/09/2020 22:13, Dan Hill wrote: >>>>> >>>>> This is running on my local minikube and is trying to hit minio. >>>>> >>>>> On Thu, Sep 10, 2020 at 1:10 PM Dan Hill <quietgol...@gmail.com> >>>>> wrote: >>>>> >>>>>> I'm using this Helm chart >>>>>> <https://github.com/riskfocus/helm-charts-public/tree/master/flink>. >>>>>> I start the job by building an image with the job jar and using kubectl >>>>>> apply to do a flink run with the jar. >>>>>> >>>>>> The log4j.properties on jobmanager and taskmanager have debug level >>>>>> set and are pretty embedded into the Helm chart. My log4j-cli.properties >>>>>> is hacked on the CLI side. >>>>>> >>>>>> I thought I just needed the s3 plugins in the jobmanager and >>>>>> taskmanager. Do I need to have a similar plugin structure from the image >>>>>> where I run 'flink run'? >>>>>> >>>>>> >>>>>> On Thu, Sep 10, 2020 at 1:03 PM Dan Hill <quietgol...@gmail.com> >>>>>> wrote: >>>>>> >>>>>>> Copying more of the log >>>>>>> >>>>>>> 2020-09-10 19:50:17,712 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend [] - >>>>>>> -------------------------------------------------------------------------------- >>>>>>> >>>>>>> 2020-09-10 19:50:17,718 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend [] - >>>>>>> Starting >>>>>>> Command Line Client (Version: 1.11.1, Scala: 2.12, Rev:7eb514a, >>>>>>> Date:2020-07-15T07:02:09+02:00) >>>>>>> >>>>>>> 2020-09-10 19:50:17,719 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend [] - OS >>>>>>> current user: root >>>>>>> >>>>>>> 2020-09-10 19:50:17,719 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend [] - >>>>>>> Current >>>>>>> Hadoop/Kerberos user: <no hadoop dependency found> >>>>>>> >>>>>>> 2020-09-10 19:50:17,719 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend [] - JVM: >>>>>>> OpenJDK 64-Bit Server VM - Oracle Corporation - 1.8/25.265-b01 >>>>>>> >>>>>>> 2020-09-10 19:50:17,719 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend [] - >>>>>>> Maximum >>>>>>> heap size: 2167 MiBytes >>>>>>> >>>>>>> tail: log/flink--client-flink-jobmanager-0.log: file truncated >>>>>>> >>>>>>> 2020-09-10 19:50:17,720 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend [] - >>>>>>> JAVA_HOME: >>>>>>> /usr/local/openjdk-8 >>>>>>> >>>>>>> 2020-09-10 19:50:17,720 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend [] - No >>>>>>> Hadoop Dependency available >>>>>>> >>>>>>> 2020-09-10 19:50:17,720 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend [] - JVM >>>>>>> Options: >>>>>>> >>>>>>> 2020-09-10 19:50:17,720 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend [] - >>>>>>> -Djava.security.properties=/opt/flink/conf/security.properties >>>>>>> >>>>>>> 2020-09-10 19:50:17,720 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend [] - >>>>>>> -Dlog.file=/opt/flink/log/flink--client-flink-jobmanager-0.log >>>>>>> >>>>>>> 2020-09-10 19:50:17,720 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend [] - >>>>>>> -Dlog4j.configuration=file:/opt/flink/conf/log4j-cli.properties >>>>>>> >>>>>>> 2020-09-10 19:50:17,721 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend [] - >>>>>>> -Dlog4j.configurationFile= >>>>>>> file:/opt/flink/conf/log4j-cli.properties >>>>>>> >>>>>>> 2020-09-10 19:50:17,721 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend [] - >>>>>>> -Dlogback.configurationFile=file:/opt/flink/conf/logback.xml >>>>>>> >>>>>>> 2020-09-10 19:50:17,721 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend [] - >>>>>>> Program >>>>>>> Arguments: >>>>>>> >>>>>>> 2020-09-10 19:50:17,721 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend [] - >>>>>>> list >>>>>>> >>>>>>> 2020-09-10 19:50:17,721 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend [] - >>>>>>> --jobmanager >>>>>>> >>>>>>> 2020-09-10 19:50:17,721 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend [] - >>>>>>> localhost:8081 >>>>>>> >>>>>>> 2020-09-10 19:50:17,721 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend [] - >>>>>>> Classpath: >>>>>>> /opt/flink/lib/flink-csv-1.11.1.jar:/opt/flink/lib/flink-json-1.11.1.jar:/opt/flink/lib/flink-shaded-zookeeper-3.4.14.jar:/opt/flink/lib/flink-table-blink_2.12-1.11.1.jar:/opt/flink/lib/flink-table_2.12-1.11.1.jar:/opt/flink/lib/jna-5.4.0.jar:/opt/flink/lib/jna-platform-5.4.0.jar:/opt/flink/lib/log4j-1.2-api-2.12.1.jar:/opt/flink/lib/log4j-api-2.12.1.jar:/opt/flink/lib/log4j-core-2.12.1.jar:/opt/flink/lib/log4j-slf4j-impl-2.12.1.jar:/opt/flink/lib/oshi-core-3.4.0.jar:/opt/flink/lib/flink-dist_2.12-1.11.1.jar::: >>>>>>> >>>>>>> 2020-09-10 19:50:17,722 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend [] - >>>>>>> -------------------------------------------------------------------------------- >>>>>>> >>>>>>> 2020-09-10 19:50:17,731 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: >>>>>>> taskmanager.numberOfTaskSlots, 2 >>>>>>> >>>>>>> 2020-09-10 19:50:17,732 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: blob.server.port, >>>>>>> 6124 >>>>>>> >>>>>>> 2020-09-10 19:50:17,732 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: >>>>>>> taskmanager.rpc.port, 6122 >>>>>>> >>>>>>> 2020-09-10 19:50:17,732 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: >>>>>>> jobmanager.heap.size, 1g >>>>>>> >>>>>>> 2020-09-10 19:50:17,732 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: >>>>>>> taskmanager.memory.process.size, 1g >>>>>>> >>>>>>> 2020-09-10 19:50:17,733 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: state.backend, >>>>>>> rocksdb >>>>>>> >>>>>>> 2020-09-10 19:50:17,733 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: >>>>>>> state.checkpoints.dir, file:///flink_state/checkpoints >>>>>>> >>>>>>> 2020-09-10 19:50:17,733 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: >>>>>>> state.savepoints.dir, file:///flink_state/savepoints >>>>>>> >>>>>>> 2020-09-10 19:50:17,733 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: state.backend.async, >>>>>>> true >>>>>>> >>>>>>> 2020-09-10 19:50:17,733 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: >>>>>>> state.backend.fs.memory-threshold, 1024 >>>>>>> >>>>>>> 2020-09-10 19:50:17,734 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: >>>>>>> state.backend.fs.write-buffer-size, 4096 >>>>>>> >>>>>>> 2020-09-10 19:50:17,734 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: >>>>>>> state.backend.incremental, true >>>>>>> >>>>>>> 2020-09-10 19:50:17,734 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: >>>>>>> state.backend.local-recovery, true >>>>>>> >>>>>>> 2020-09-10 19:50:17,734 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: >>>>>>> state.checkpoints.num-retained, 1 >>>>>>> >>>>>>> 2020-09-10 19:50:17,734 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: >>>>>>> taskmanager.state.local.root-dirs, >>>>>>> file:///flink_state/local-recovery >>>>>>> >>>>>>> 2020-09-10 19:50:17,735 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: >>>>>>> state.backend.rocksdb.checkpoint.transfer.thread.num, 1 >>>>>>> >>>>>>> 2020-09-10 19:50:17,735 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: >>>>>>> state.backend.rocksdb.localdir, /flink_state/rocksdb >>>>>>> >>>>>>> 2020-09-10 19:50:17,735 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: >>>>>>> state.backend.rocksdb.options-factory, >>>>>>> org.apache.flink.contrib.streaming.state.DefaultConfigurableOptionsFactory >>>>>>> >>>>>>> 2020-09-10 19:50:17,735 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: >>>>>>> state.backend.rocksdb.predefined-options, DEFAULT >>>>>>> >>>>>>> 2020-09-10 19:50:17,735 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: >>>>>>> state.backend.rocksdb.timer-service.factory, HEAP >>>>>>> >>>>>>> 2020-09-10 19:50:17,735 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: >>>>>>> state.backend.rocksdb.ttl.compaction.filter.enabled, false >>>>>>> >>>>>>> 2020-09-10 19:50:17,736 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: >>>>>>> jobmanager.rpc.address, flink-jobmanager >>>>>>> >>>>>>> 2020-09-10 19:50:17,736 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: jobmanager.rpc.port, >>>>>>> 6123 >>>>>>> >>>>>>> 2020-09-10 19:50:17,736 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: >>>>>>> taskmanager.memory.jvm-metaspace.size, 256mb >>>>>>> >>>>>>> 2020-09-10 19:50:17,736 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: s3a.endpoint, >>>>>>> http://minio:9000 >>>>>>> >>>>>>> 2020-09-10 19:50:17,736 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: >>>>>>> s3a.path.style.access, true >>>>>>> >>>>>>> 2020-09-10 19:50:17,736 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: s3a.access-key, >>>>>>> YOURACCESSKEY >>>>>>> >>>>>>> 2020-09-10 19:50:17,737 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: s3a.secret-key, >>>>>>> ****** >>>>>>> >>>>>>> 2020-09-10 19:50:17,737 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration >>>>>>> [] - Loading configuration property: >>>>>>> s3a.aws.credentials.provider, >>>>>>> org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider >>>>>>> >>>>>>> 2020-09-10 19:50:17,802 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend [] - >>>>>>> Loading FallbackYarnSessionCli >>>>>>> >>>>>>> 2020-09-10 19:50:17,929 INFO org.apache.flink.core.fs.FileSystem >>>>>>> [] - Hadoop is not in the >>>>>>> classpath/dependencies. The extended set of supported File Systems via >>>>>>> Hadoop is not available. >>>>>>> >>>>>>> 2020-09-10 19:50:18,102 INFO >>>>>>> org.apache.flink.runtime.security.modules.HadoopModuleFactory >>>>>>> [] - Cannot create Hadoop Security Module because Hadoop cannot be >>>>>>> found in >>>>>>> the Classpath. >>>>>>> >>>>>>> 2020-09-10 19:50:18,126 INFO >>>>>>> org.apache.flink.runtime.security.modules.JaasModule >>>>>>> [] - Jaas file will be created as >>>>>>> /tmp/jaas-1506212733867615019.conf. >>>>>>> >>>>>>> 2020-09-10 19:50:18,161 INFO >>>>>>> org.apache.flink.runtime.security.contexts.HadoopSecurityContextFactory >>>>>>> [] - Cannot install HadoopSecurityContext because Hadoop cannot be >>>>>>> found in >>>>>>> the Classpath. >>>>>>> >>>>>>> 2020-09-10 19:50:18,163 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend [] - >>>>>>> Running 'list' command. >>>>>>> >>>>>>> 2020-09-10 19:50:18,226 INFO >>>>>>> org.apache.flink.client.deployment.DefaultClusterClientServiceLoader >>>>>>> [] - Could not load factory due to missing dependencies. >>>>>>> >>>>>>> 2020-09-10 19:50:19,107 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend [] - >>>>>>> Waiting for response... >>>>>>> >>>>>>> 2020-09-10 19:50:19,414 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend [] - >>>>>>> Successfully retrieved list of jobs >>>>>>> >>>>>>> >>>>>>> On Thu, Sep 10, 2020 at 1:02 PM Arvid Heise <ar...@ververica.com> >>>>>>> wrote: >>>>>>> >>>>>>>> Hi Dan, >>>>>>>> >>>>>>>> somehow enabling debug statements did not work. >>>>>>>> >>>>>>>> However, the logs helps to narrow down the issue. The exception >>>>>>>> occurs neither on jobmanager nor on taskmanager. It occurs wherever you >>>>>>>> execute the command line interface. >>>>>>>> >>>>>>>> How do you execute the job? Do you start it from your machine? Can >>>>>>>> you try out to also add the respective s3 plugin there? >>>>>>>> >>>>>>>> Best, >>>>>>>> >>>>>>>> Arvid >>>>>>>> >>>>>>>> On Thu, Sep 10, 2020 at 7:50 PM Dan Hill <quietgol...@gmail.com> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> I changed the levels to DEBUG. I don't see useful data in the >>>>>>>>> logs. >>>>>>>>> >>>>>>>>> >>>>>>>>> https://drive.google.com/file/d/1ua1zsr3BInY_8xdsWwA__F0uloAqy-vG/view?usp=sharing >>>>>>>>> >>>>>>>>> On Thu, Sep 10, 2020 at 8:45 AM Arvid Heise <ar...@ververica.com> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> Could you try 1) or 2) and enable debug logging* and share the >>>>>>>>>> log with us? >>>>>>>>>> >>>>>>>>>> *Usually by adjusting FLINK_HOME/conf/log4j.properties. >>>>>>>>>> >>>>>>>>>> On Thu, Sep 10, 2020 at 5:38 PM Dan Hill <quietgol...@gmail.com> >>>>>>>>>> wrote: >>>>>>>>>> >>>>>>>>>>> Ah, sorry, it's a copy/paste issue with this email. I've tried >>>>>>>>>>> both: >>>>>>>>>>> 1) using s3a uri with flink-s3-fs-hadoop jar >>>>>>>>>>> in /opt/flink/plugins/s3-fs-hadoop. >>>>>>>>>>> 2) using s3p uri with flink-s3-fs-presto jar >>>>>>>>>>> in /opt/flink/plugins/s3-fs-presto. >>>>>>>>>>> 3) loading both 1 and 2 >>>>>>>>>>> 4) trying s3 uri. >>>>>>>>>>> >>>>>>>>>>> When doing 1) >>>>>>>>>>> >>>>>>>>>>> Caused by: >>>>>>>>>>> org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: >>>>>>>>>>> Could not >>>>>>>>>>> find a file system implementation for scheme 's3a'. The scheme is >>>>>>>>>>> directly >>>>>>>>>>> supported by Flink through the following plugin: >>>>>>>>>>> flink-s3-fs-hadoop. Please >>>>>>>>>>> ensure that each plugin resides within its own subfolder within the >>>>>>>>>>> plugins >>>>>>>>>>> directory. See >>>>>>>>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/plugins.html >>>>>>>>>>> for more information. If you want to use a Hadoop file system for >>>>>>>>>>> that >>>>>>>>>>> scheme, please add the scheme to the configuration >>>>>>>>>>> fs.allowed-fallback-filesystems. For a full list of supported file >>>>>>>>>>> systems, >>>>>>>>>>> please see >>>>>>>>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/filesystems/ >>>>>>>>>>> . >>>>>>>>>>> >>>>>>>>>>> When doing 2) >>>>>>>>>>> >>>>>>>>>>> Caused by: >>>>>>>>>>> org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: >>>>>>>>>>> Could not >>>>>>>>>>> find a file system implementation for scheme 's3p'. The scheme is >>>>>>>>>>> directly >>>>>>>>>>> supported by Flink through the following plugin: >>>>>>>>>>> flink-s3-fs-presto. Please >>>>>>>>>>> ensure that each plugin resides within its own subfolder within the >>>>>>>>>>> plugins >>>>>>>>>>> directory. See >>>>>>>>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/plugins.html >>>>>>>>>>> for more information. If you want to use a Hadoop file system for >>>>>>>>>>> that >>>>>>>>>>> scheme, please add the scheme to the configuration >>>>>>>>>>> fs.allowed-fallback-filesystems. For a full list of supported file >>>>>>>>>>> systems, >>>>>>>>>>> please see >>>>>>>>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/filesystems/ >>>>>>>>>>> . >>>>>>>>>>> >>>>>>>>>>> etc >>>>>>>>>>> >>>>>>>>>>> On Thu, Sep 10, 2020 at 8:15 AM Arvid Heise <ar...@ververica.com> >>>>>>>>>>> wrote: >>>>>>>>>>> >>>>>>>>>>>> Hi Dan, >>>>>>>>>>>> >>>>>>>>>>>> s3p is only provided by flink-s3-fs-presto plugin. The plugin >>>>>>>>>>>> you used provides s3a. >>>>>>>>>>>> (and both provide s3, but it's good to use the more specific >>>>>>>>>>>> prefix). >>>>>>>>>>>> >>>>>>>>>>>> Best, >>>>>>>>>>>> >>>>>>>>>>>> Arvid >>>>>>>>>>>> >>>>>>>>>>>> On Thu, Sep 10, 2020 at 9:24 AM Dan Hill <quietgol...@gmail.com> >>>>>>>>>>>> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> *Background* >>>>>>>>>>>>> I'm converting some prototype Flink v1.11.1 code that uses >>>>>>>>>>>>> DataSet/DataTable APIs to use the Table API. >>>>>>>>>>>>> >>>>>>>>>>>>> *Problem* >>>>>>>>>>>>> When switching to using the Table API, my s3 plugins stopped >>>>>>>>>>>>> working. I don't know why. I've added the required maven table >>>>>>>>>>>>> dependencies to the job. >>>>>>>>>>>>> >>>>>>>>>>>>> I've tried us moving both the presto and/or the hadoop s3 jars >>>>>>>>>>>>> to plugin subfolders. No luck. >>>>>>>>>>>>> >>>>>>>>>>>>> Any ideas what is wrong? I'm guessing I'm missing something >>>>>>>>>>>>> simple. >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> *Error* >>>>>>>>>>>>> >>>>>>>>>>>>> Caused by: >>>>>>>>>>>>> org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: >>>>>>>>>>>>> Could not >>>>>>>>>>>>> find a file system implementation for scheme 's3p'. The scheme is >>>>>>>>>>>>> directly >>>>>>>>>>>>> supported by Flink through the following plugin: >>>>>>>>>>>>> flink-s3-fs-presto. Please >>>>>>>>>>>>> ensure that each plugin resides within its own subfolder within >>>>>>>>>>>>> the plugins >>>>>>>>>>>>> directory. See >>>>>>>>>>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/plugins.html >>>>>>>>>>>>> for more information. If you want to use a Hadoop file system for >>>>>>>>>>>>> that >>>>>>>>>>>>> scheme, please add the scheme to the configuration >>>>>>>>>>>>> fs.allowed-fallback-filesystems. For a full list of supported >>>>>>>>>>>>> file systems, >>>>>>>>>>>>> please see >>>>>>>>>>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/filesystems/ >>>>>>>>>>>>> . >>>>>>>>>>>>> >>>>>>>>>>>>> at >>>>>>>>>>>>> org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:473) >>>>>>>>>>>>> >>>>>>>>>>>>> at org.apache.flink.core.fs.FileSystem.get(FileSystem.java:389) >>>>>>>>>>>>> >>>>>>>>>>>>> at org.apache.flink.core.fs.Path.getFileSystem(Path.java:292) >>>>>>>>>>>>> >>>>>>>>>>>>> at >>>>>>>>>>>>> org.apache.flink.table.filesystem.FileSystemTableSink.toStagingPath(FileSystemTableSink.java:232) >>>>>>>>>>>>> >>>>>>>>>>>>> ... 35 more >>>>>>>>>>>>> >>>>>>>>>>>>> *ls of plugins directory (same for taskmanager)* >>>>>>>>>>>>> >>>>>>>>>>>>> kubectl exec pod/flink-jobmanager-0 -- ls -l >>>>>>>>>>>>> /opt/flink/plugins/s3-fs-hadoop >>>>>>>>>>>>> >>>>>>>>>>>>> total 19520 >>>>>>>>>>>>> >>>>>>>>>>>>> -rw-r--r-- 1 root root 19985452 Sep 10 06:27 >>>>>>>>>>>>> flink-s3-fs-hadoop-1.11.1.jar >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> >>>>>>>>>>>> Arvid Heise | Senior Java Developer >>>>>>>>>>>> >>>>>>>>>>>> <https://www.ververica.com/> >>>>>>>>>>>> >>>>>>>>>>>> Follow us @VervericaData >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> >>>>>>>>>>>> Join Flink Forward <https://flink-forward.org/> - The Apache >>>>>>>>>>>> Flink Conference >>>>>>>>>>>> >>>>>>>>>>>> Stream Processing | Event Driven | Real Time >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> >>>>>>>>>>>> Ververica GmbH | Invalidenstrasse 115, 10115 Berlin, Germany >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> Ververica GmbH Registered at Amtsgericht Charlottenburg: HRB >>>>>>>>>>>> 158244 B Managing Directors: Timothy Alexander Steinert, Yip >>>>>>>>>>>> Park Tung Jason, Ji (Toni) Cheng >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> >>>>>>>>>> Arvid Heise | Senior Java Developer >>>>>>>>>> >>>>>>>>>> <https://www.ververica.com/> >>>>>>>>>> >>>>>>>>>> Follow us @VervericaData >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> >>>>>>>>>> Join Flink Forward <https://flink-forward.org/> - The Apache >>>>>>>>>> Flink Conference >>>>>>>>>> >>>>>>>>>> Stream Processing | Event Driven | Real Time >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> >>>>>>>>>> Ververica GmbH | Invalidenstrasse 115, 10115 Berlin, Germany >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> Ververica GmbH Registered at Amtsgericht Charlottenburg: HRB >>>>>>>>>> 158244 B Managing Directors: Timothy Alexander Steinert, Yip >>>>>>>>>> Park Tung Jason, Ji (Toni) Cheng >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> >>>>>>>> Arvid Heise | Senior Java Developer >>>>>>>> >>>>>>>> <https://www.ververica.com/> >>>>>>>> >>>>>>>> Follow us @VervericaData >>>>>>>> >>>>>>>> -- >>>>>>>> >>>>>>>> Join Flink Forward <https://flink-forward.org/> - The Apache Flink >>>>>>>> Conference >>>>>>>> >>>>>>>> Stream Processing | Event Driven | Real Time >>>>>>>> >>>>>>>> -- >>>>>>>> >>>>>>>> Ververica GmbH | Invalidenstrasse 115, 10115 Berlin, Germany >>>>>>>> >>>>>>>> -- >>>>>>>> Ververica GmbH Registered at Amtsgericht Charlottenburg: HRB >>>>>>>> 158244 B Managing Directors: Timothy Alexander Steinert, Yip Park >>>>>>>> Tung Jason, Ji (Toni) Cheng >>>>>>>> >>>>>>> >>>> >>>> -- >>>> Best, Jingsong Lee >>>> >>> >> >> -- >> >> Arvid Heise | Senior Java Developer >> >> <https://www.ververica.com/> >> >> Follow us @VervericaData >> >> -- >> >> Join Flink Forward <https://flink-forward.org/> - The Apache Flink >> Conference >> >> Stream Processing | Event Driven | Real Time >> >> -- >> >> Ververica GmbH | Invalidenstrasse 115, 10115 Berlin, Germany >> >> -- >> Ververica GmbH >> Registered at Amtsgericht Charlottenburg: HRB 158244 B >> Managing Directors: Timothy Alexander Steinert, Yip Park Tung Jason, Ji >> (Toni) Cheng >> >