Yes, the client runs in K8.  It uses a different K8 config than the Helm
chart and does not load the plugins.  Does the client use the same plugin
structure as the Flink job/task manager?  I can try using it tomorrow.

Cool, that link would work too.

Thanks, Arvid!


On Mon, Sep 14, 2020 at 10:59 PM Arvid Heise <ar...@ververica.com> wrote:

> Hi Dan,
>
> Are you running the client also in K8s? If so you need an initialization
> step, where you add the library to the plugins directory. Putting it into
> lib or into the user jar doesn't work anymore as we removed the shading in
> s3 in Flink 1.10.
>
> The official Flink docker image has an easy way to add these plugins [1].
>
> [1]
> https://ci.apache.org/projects/flink/flink-docs-stable/ops/deployment/docker.html#using-plugins
>
> On Tue, Sep 15, 2020 at 6:40 AM Dan Hill <quietgol...@gmail.com> wrote:
>
>> Thanks for the update!
>>
>> I'm trying a bunch of combinations on the client side to get the S3
>> Filesystem to be picked up correctly.  Most of my attempts involved
>> building into the job jar (which I'm guessing won't work).  I then started
>> getting issues with ClassCastExceptions.
>>
>> I might try a little more tomorrow (e.g. modifying the custom image).  If
>> I can't get it, I'll roll back to a previous Flink version that works.
>>
>> Caused by: java.lang.ClassCastException:
>> org.codehaus.janino.CompilerFactory cannot be cast to
>> org.codehaus.commons.compiler.ICompilerFactory
>>
>> at
>> org.codehaus.commons.compiler.CompilerFactoryFactory.getCompilerFactory(CompilerFactoryFactory.java:129)
>>
>> at
>> org.codehaus.commons.compiler.CompilerFactoryFactory.getDefaultCompilerFactory(CompilerFactoryFactory.java:79)
>>
>> at
>> org.apache.calcite.rel.metadata.JaninoRelMetadataProvider.compile(JaninoRelMetadataProvider.java:431)
>>
>> ... 51 more
>>
>>
>> On Mon, Sep 14, 2020 at 7:03 PM Jingsong Li <jingsongl...@gmail.com>
>> wrote:
>>
>>> Hi Dan,
>>>
>>> I think Arvid and Dawid are right, as a workaround, you can try making
>>> S3Filesystem works in the client. But for a long term solution, we can fix
>>> it.
>>>
>>> I created https://issues.apache.org/jira/browse/FLINK-19228 for
>>> tracking this.
>>>
>>> Best,
>>> Jingsong
>>>
>>> On Mon, Sep 14, 2020 at 3:57 PM Dawid Wysakowicz <dwysakow...@apache.org>
>>> wrote:
>>>
>>>> Hi Dan,
>>>>
>>>> As far as I checked in the code, the FileSystemSink will try to create
>>>> staging directories from the client. I think it might be problematic, as
>>>> your case shows. We might need to revisit that part. I am cc'ing Jingsong
>>>> who worked on the FileSystemSink.
>>>>
>>>> As a workaround you might try putting the s3 plugin on the CLI
>>>> classpath (not sure if plugins work for the CLI through the /plugins
>>>> directory).
>>>>
>>>> Best,
>>>>
>>>> Dawid
>>>> On 10/09/2020 22:13, Dan Hill wrote:
>>>>
>>>> This is running on my local minikube and is trying to hit minio.
>>>>
>>>> On Thu, Sep 10, 2020 at 1:10 PM Dan Hill <quietgol...@gmail.com> wrote:
>>>>
>>>>> I'm using this Helm chart
>>>>> <https://github.com/riskfocus/helm-charts-public/tree/master/flink>.
>>>>> I start the job by building an image with the job jar and using kubectl
>>>>> apply to do a flink run with the jar.
>>>>>
>>>>> The log4j.properties on jobmanager and taskmanager have debug level
>>>>> set and are pretty embedded into the Helm chart.  My log4j-cli.properties
>>>>> is hacked on the CLI side.
>>>>>
>>>>> I thought I just needed the s3 plugins in the jobmanager and
>>>>> taskmanager.  Do I need to have a similar plugin structure from the image
>>>>> where I run 'flink run'?
>>>>>
>>>>>
>>>>> On Thu, Sep 10, 2020 at 1:03 PM Dan Hill <quietgol...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> Copying more of the log
>>>>>>
>>>>>> 2020-09-10 19:50:17,712 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                     [] -
>>>>>> --------------------------------------------------------------------------------
>>>>>>
>>>>>> 2020-09-10 19:50:17,718 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                     [] -  Starting Command Line Client (Version:
>>>>>> 1.11.1, Scala: 2.12, Rev:7eb514a, Date:2020-07-15T07:02:09+02:00)
>>>>>>
>>>>>> 2020-09-10 19:50:17,719 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                     [] -  OS current user: root
>>>>>>
>>>>>> 2020-09-10 19:50:17,719 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                     [] -  Current Hadoop/Kerberos user: <no hadoop
>>>>>> dependency found>
>>>>>>
>>>>>> 2020-09-10 19:50:17,719 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                     [] -  JVM: OpenJDK 64-Bit Server VM - Oracle
>>>>>> Corporation - 1.8/25.265-b01
>>>>>>
>>>>>> 2020-09-10 19:50:17,719 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                     [] -  Maximum heap size: 2167 MiBytes
>>>>>>
>>>>>> tail: log/flink--client-flink-jobmanager-0.log: file truncated
>>>>>>
>>>>>> 2020-09-10 19:50:17,720 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                     [] -  JAVA_HOME: /usr/local/openjdk-8
>>>>>>
>>>>>> 2020-09-10 19:50:17,720 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                     [] -  No Hadoop Dependency available
>>>>>>
>>>>>> 2020-09-10 19:50:17,720 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                     [] -  JVM Options:
>>>>>>
>>>>>> 2020-09-10 19:50:17,720 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                     [] -
>>>>>> -Djava.security.properties=/opt/flink/conf/security.properties
>>>>>>
>>>>>> 2020-09-10 19:50:17,720 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                     [] -
>>>>>> -Dlog.file=/opt/flink/log/flink--client-flink-jobmanager-0.log
>>>>>>
>>>>>> 2020-09-10 19:50:17,720 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                     [] -     -Dlog4j.configuration=
>>>>>> file:/opt/flink/conf/log4j-cli.properties
>>>>>>
>>>>>> 2020-09-10 19:50:17,721 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                     [] -     -Dlog4j.configurationFile=
>>>>>> file:/opt/flink/conf/log4j-cli.properties
>>>>>>
>>>>>> 2020-09-10 19:50:17,721 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                     [] -     -Dlogback.configurationFile=
>>>>>> file:/opt/flink/conf/logback.xml
>>>>>>
>>>>>> 2020-09-10 19:50:17,721 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                     [] -  Program Arguments:
>>>>>>
>>>>>> 2020-09-10 19:50:17,721 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                     [] -     list
>>>>>>
>>>>>> 2020-09-10 19:50:17,721 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                     [] -     --jobmanager
>>>>>>
>>>>>> 2020-09-10 19:50:17,721 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                     [] -     localhost:8081
>>>>>>
>>>>>> 2020-09-10 19:50:17,721 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                     [] -  Classpath:
>>>>>> /opt/flink/lib/flink-csv-1.11.1.jar:/opt/flink/lib/flink-json-1.11.1.jar:/opt/flink/lib/flink-shaded-zookeeper-3.4.14.jar:/opt/flink/lib/flink-table-blink_2.12-1.11.1.jar:/opt/flink/lib/flink-table_2.12-1.11.1.jar:/opt/flink/lib/jna-5.4.0.jar:/opt/flink/lib/jna-platform-5.4.0.jar:/opt/flink/lib/log4j-1.2-api-2.12.1.jar:/opt/flink/lib/log4j-api-2.12.1.jar:/opt/flink/lib/log4j-core-2.12.1.jar:/opt/flink/lib/log4j-slf4j-impl-2.12.1.jar:/opt/flink/lib/oshi-core-3.4.0.jar:/opt/flink/lib/flink-dist_2.12-1.11.1.jar:::
>>>>>>
>>>>>> 2020-09-10 19:50:17,722 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                     [] -
>>>>>> --------------------------------------------------------------------------------
>>>>>>
>>>>>> 2020-09-10 19:50:17,731 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property:
>>>>>> taskmanager.numberOfTaskSlots, 2
>>>>>>
>>>>>> 2020-09-10 19:50:17,732 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property: blob.server.port, 6124
>>>>>>
>>>>>> 2020-09-10 19:50:17,732 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property: taskmanager.rpc.port,
>>>>>> 6122
>>>>>>
>>>>>> 2020-09-10 19:50:17,732 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property: jobmanager.heap.size,
>>>>>> 1g
>>>>>>
>>>>>> 2020-09-10 19:50:17,732 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property:
>>>>>> taskmanager.memory.process.size, 1g
>>>>>>
>>>>>> 2020-09-10 19:50:17,733 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property: state.backend, rocksdb
>>>>>>
>>>>>> 2020-09-10 19:50:17,733 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property:
>>>>>> state.checkpoints.dir, file:///flink_state/checkpoints
>>>>>>
>>>>>> 2020-09-10 19:50:17,733 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property: state.savepoints.dir,
>>>>>> file:///flink_state/savepoints
>>>>>>
>>>>>> 2020-09-10 19:50:17,733 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property: state.backend.async,
>>>>>> true
>>>>>>
>>>>>> 2020-09-10 19:50:17,733 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property:
>>>>>> state.backend.fs.memory-threshold, 1024
>>>>>>
>>>>>> 2020-09-10 19:50:17,734 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property:
>>>>>> state.backend.fs.write-buffer-size, 4096
>>>>>>
>>>>>> 2020-09-10 19:50:17,734 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property:
>>>>>> state.backend.incremental, true
>>>>>>
>>>>>> 2020-09-10 19:50:17,734 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property:
>>>>>> state.backend.local-recovery, true
>>>>>>
>>>>>> 2020-09-10 19:50:17,734 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property:
>>>>>> state.checkpoints.num-retained, 1
>>>>>>
>>>>>> 2020-09-10 19:50:17,734 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property:
>>>>>> taskmanager.state.local.root-dirs, file:///flink_state/local-recovery
>>>>>>
>>>>>> 2020-09-10 19:50:17,735 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property:
>>>>>> state.backend.rocksdb.checkpoint.transfer.thread.num, 1
>>>>>>
>>>>>> 2020-09-10 19:50:17,735 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property:
>>>>>> state.backend.rocksdb.localdir, /flink_state/rocksdb
>>>>>>
>>>>>> 2020-09-10 19:50:17,735 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property:
>>>>>> state.backend.rocksdb.options-factory,
>>>>>> org.apache.flink.contrib.streaming.state.DefaultConfigurableOptionsFactory
>>>>>>
>>>>>> 2020-09-10 19:50:17,735 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property:
>>>>>> state.backend.rocksdb.predefined-options, DEFAULT
>>>>>>
>>>>>> 2020-09-10 19:50:17,735 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property:
>>>>>> state.backend.rocksdb.timer-service.factory, HEAP
>>>>>>
>>>>>> 2020-09-10 19:50:17,735 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property:
>>>>>> state.backend.rocksdb.ttl.compaction.filter.enabled, false
>>>>>>
>>>>>> 2020-09-10 19:50:17,736 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property:
>>>>>> jobmanager.rpc.address, flink-jobmanager
>>>>>>
>>>>>> 2020-09-10 19:50:17,736 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property: jobmanager.rpc.port,
>>>>>> 6123
>>>>>>
>>>>>> 2020-09-10 19:50:17,736 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property:
>>>>>> taskmanager.memory.jvm-metaspace.size, 256mb
>>>>>>
>>>>>> 2020-09-10 19:50:17,736 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property: s3a.endpoint,
>>>>>> http://minio:9000
>>>>>>
>>>>>> 2020-09-10 19:50:17,736 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property:
>>>>>> s3a.path.style.access, true
>>>>>>
>>>>>> 2020-09-10 19:50:17,736 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property: s3a.access-key,
>>>>>> YOURACCESSKEY
>>>>>>
>>>>>> 2020-09-10 19:50:17,737 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property: s3a.secret-key, ******
>>>>>>
>>>>>> 2020-09-10 19:50:17,737 INFO  
>>>>>> org.apache.flink.configuration.GlobalConfiguration
>>>>>>           [] - Loading configuration property:
>>>>>> s3a.aws.credentials.provider,
>>>>>> org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider
>>>>>>
>>>>>> 2020-09-10 19:50:17,802 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                     [] - Loading FallbackYarnSessionCli
>>>>>>
>>>>>> 2020-09-10 19:50:17,929 INFO  org.apache.flink.core.fs.FileSystem
>>>>>>                       [] - Hadoop is not in the
>>>>>> classpath/dependencies. The extended set of supported File Systems via
>>>>>> Hadoop is not available.
>>>>>>
>>>>>> 2020-09-10 19:50:18,102 INFO  
>>>>>> org.apache.flink.runtime.security.modules.HadoopModuleFactory
>>>>>> [] - Cannot create Hadoop Security Module because Hadoop cannot be found 
>>>>>> in
>>>>>> the Classpath.
>>>>>>
>>>>>> 2020-09-10 19:50:18,126 INFO  
>>>>>> org.apache.flink.runtime.security.modules.JaasModule
>>>>>>         [] - Jaas file will be created as
>>>>>> /tmp/jaas-1506212733867615019.conf.
>>>>>>
>>>>>> 2020-09-10 19:50:18,161 INFO  
>>>>>> org.apache.flink.runtime.security.contexts.HadoopSecurityContextFactory
>>>>>> [] - Cannot install HadoopSecurityContext because Hadoop cannot be found 
>>>>>> in
>>>>>> the Classpath.
>>>>>>
>>>>>> 2020-09-10 19:50:18,163 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                     [] - Running 'list' command.
>>>>>>
>>>>>> 2020-09-10 19:50:18,226 INFO  
>>>>>> org.apache.flink.client.deployment.DefaultClusterClientServiceLoader
>>>>>> [] - Could not load factory due to missing dependencies.
>>>>>>
>>>>>> 2020-09-10 19:50:19,107 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                     [] - Waiting for response...
>>>>>>
>>>>>> 2020-09-10 19:50:19,414 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                     [] - Successfully retrieved list of jobs
>>>>>>
>>>>>>
>>>>>> On Thu, Sep 10, 2020 at 1:02 PM Arvid Heise <ar...@ververica.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Hi Dan,
>>>>>>>
>>>>>>> somehow enabling debug statements did not work.
>>>>>>>
>>>>>>> However, the logs helps to narrow down the issue. The exception
>>>>>>> occurs neither on jobmanager nor on taskmanager. It occurs wherever you
>>>>>>> execute the command line interface.
>>>>>>>
>>>>>>> How do you execute the job? Do you start it from your machine? Can
>>>>>>> you try out to also add the respective s3 plugin there?
>>>>>>>
>>>>>>> Best,
>>>>>>>
>>>>>>> Arvid
>>>>>>>
>>>>>>> On Thu, Sep 10, 2020 at 7:50 PM Dan Hill <quietgol...@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> I changed the levels to DEBUG.  I don't see useful data in the
>>>>>>>> logs.
>>>>>>>>
>>>>>>>>
>>>>>>>> https://drive.google.com/file/d/1ua1zsr3BInY_8xdsWwA__F0uloAqy-vG/view?usp=sharing
>>>>>>>>
>>>>>>>> On Thu, Sep 10, 2020 at 8:45 AM Arvid Heise <ar...@ververica.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Could you try 1) or 2) and enable debug logging* and share the log
>>>>>>>>> with us?
>>>>>>>>>
>>>>>>>>> *Usually by adjusting FLINK_HOME/conf/log4j.properties.
>>>>>>>>>
>>>>>>>>> On Thu, Sep 10, 2020 at 5:38 PM Dan Hill <quietgol...@gmail.com>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>> Ah, sorry, it's a copy/paste issue with this email.  I've tried
>>>>>>>>>> both:
>>>>>>>>>> 1) using s3a uri with flink-s3-fs-hadoop jar
>>>>>>>>>> in /opt/flink/plugins/s3-fs-hadoop.
>>>>>>>>>> 2) using s3p uri with flink-s3-fs-presto jar
>>>>>>>>>> in /opt/flink/plugins/s3-fs-presto.
>>>>>>>>>> 3) loading both 1 and 2
>>>>>>>>>> 4) trying s3 uri.
>>>>>>>>>>
>>>>>>>>>> When doing 1)
>>>>>>>>>>
>>>>>>>>>> Caused by:
>>>>>>>>>> org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Could 
>>>>>>>>>> not
>>>>>>>>>> find a file system implementation for scheme 's3a'. The scheme is 
>>>>>>>>>> directly
>>>>>>>>>> supported by Flink through the following plugin: flink-s3-fs-hadoop. 
>>>>>>>>>> Please
>>>>>>>>>> ensure that each plugin resides within its own subfolder within the 
>>>>>>>>>> plugins
>>>>>>>>>> directory. See
>>>>>>>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/plugins.html
>>>>>>>>>> for more information. If you want to use a Hadoop file system for 
>>>>>>>>>> that
>>>>>>>>>> scheme, please add the scheme to the configuration
>>>>>>>>>> fs.allowed-fallback-filesystems. For a full list of supported file 
>>>>>>>>>> systems,
>>>>>>>>>> please see
>>>>>>>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/filesystems/
>>>>>>>>>> .
>>>>>>>>>>
>>>>>>>>>> When doing 2)
>>>>>>>>>>
>>>>>>>>>> Caused by:
>>>>>>>>>> org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Could 
>>>>>>>>>> not
>>>>>>>>>> find a file system implementation for scheme 's3p'. The scheme is 
>>>>>>>>>> directly
>>>>>>>>>> supported by Flink through the following plugin: flink-s3-fs-presto. 
>>>>>>>>>> Please
>>>>>>>>>> ensure that each plugin resides within its own subfolder within the 
>>>>>>>>>> plugins
>>>>>>>>>> directory. See
>>>>>>>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/plugins.html
>>>>>>>>>> for more information. If you want to use a Hadoop file system for 
>>>>>>>>>> that
>>>>>>>>>> scheme, please add the scheme to the configuration
>>>>>>>>>> fs.allowed-fallback-filesystems. For a full list of supported file 
>>>>>>>>>> systems,
>>>>>>>>>> please see
>>>>>>>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/filesystems/
>>>>>>>>>> .
>>>>>>>>>>
>>>>>>>>>> etc
>>>>>>>>>>
>>>>>>>>>> On Thu, Sep 10, 2020 at 8:15 AM Arvid Heise <ar...@ververica.com>
>>>>>>>>>> wrote:
>>>>>>>>>>
>>>>>>>>>>> Hi Dan,
>>>>>>>>>>>
>>>>>>>>>>> s3p is only provided by flink-s3-fs-presto plugin. The plugin
>>>>>>>>>>> you used provides s3a.
>>>>>>>>>>> (and both provide s3, but it's good to use the more specific
>>>>>>>>>>> prefix).
>>>>>>>>>>>
>>>>>>>>>>> Best,
>>>>>>>>>>>
>>>>>>>>>>> Arvid
>>>>>>>>>>>
>>>>>>>>>>> On Thu, Sep 10, 2020 at 9:24 AM Dan Hill <quietgol...@gmail.com>
>>>>>>>>>>> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> *Background*
>>>>>>>>>>>> I'm converting some prototype Flink v1.11.1 code that uses
>>>>>>>>>>>> DataSet/DataTable APIs to use the Table API.
>>>>>>>>>>>>
>>>>>>>>>>>> *Problem*
>>>>>>>>>>>> When switching to using the Table API, my s3 plugins stopped
>>>>>>>>>>>> working.  I don't know why.  I've added the required maven table
>>>>>>>>>>>> dependencies to the job.
>>>>>>>>>>>>
>>>>>>>>>>>> I've tried us moving both the presto and/or the hadoop s3 jars
>>>>>>>>>>>> to plugin subfolders.  No luck.
>>>>>>>>>>>>
>>>>>>>>>>>> Any ideas what is wrong?  I'm guessing I'm missing something
>>>>>>>>>>>> simple.
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> *Error*
>>>>>>>>>>>>
>>>>>>>>>>>> Caused by:
>>>>>>>>>>>> org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: 
>>>>>>>>>>>> Could not
>>>>>>>>>>>> find a file system implementation for scheme 's3p'. The scheme is 
>>>>>>>>>>>> directly
>>>>>>>>>>>> supported by Flink through the following plugin: 
>>>>>>>>>>>> flink-s3-fs-presto. Please
>>>>>>>>>>>> ensure that each plugin resides within its own subfolder within 
>>>>>>>>>>>> the plugins
>>>>>>>>>>>> directory. See
>>>>>>>>>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/plugins.html
>>>>>>>>>>>> for more information. If you want to use a Hadoop file system for 
>>>>>>>>>>>> that
>>>>>>>>>>>> scheme, please add the scheme to the configuration
>>>>>>>>>>>> fs.allowed-fallback-filesystems. For a full list of supported file 
>>>>>>>>>>>> systems,
>>>>>>>>>>>> please see
>>>>>>>>>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/filesystems/
>>>>>>>>>>>> .
>>>>>>>>>>>>
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:473)
>>>>>>>>>>>>
>>>>>>>>>>>> at org.apache.flink.core.fs.FileSystem.get(FileSystem.java:389)
>>>>>>>>>>>>
>>>>>>>>>>>> at org.apache.flink.core.fs.Path.getFileSystem(Path.java:292)
>>>>>>>>>>>>
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.flink.table.filesystem.FileSystemTableSink.toStagingPath(FileSystemTableSink.java:232)
>>>>>>>>>>>>
>>>>>>>>>>>> ... 35 more
>>>>>>>>>>>>
>>>>>>>>>>>> *ls of plugins directory (same for taskmanager)*
>>>>>>>>>>>>
>>>>>>>>>>>> kubectl exec pod/flink-jobmanager-0  -- ls -l
>>>>>>>>>>>> /opt/flink/plugins/s3-fs-hadoop
>>>>>>>>>>>>
>>>>>>>>>>>> total 19520
>>>>>>>>>>>>
>>>>>>>>>>>> -rw-r--r-- 1 root root 19985452 Sep 10 06:27
>>>>>>>>>>>> flink-s3-fs-hadoop-1.11.1.jar
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>>
>>>>>>>>>>> Arvid Heise | Senior Java Developer
>>>>>>>>>>>
>>>>>>>>>>> <https://www.ververica.com/>
>>>>>>>>>>>
>>>>>>>>>>> Follow us @VervericaData
>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>>
>>>>>>>>>>> Join Flink Forward <https://flink-forward.org/> - The Apache
>>>>>>>>>>> Flink Conference
>>>>>>>>>>>
>>>>>>>>>>> Stream Processing | Event Driven | Real Time
>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>>
>>>>>>>>>>> Ververica GmbH | Invalidenstrasse 115, 10115 Berlin, Germany
>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>> Ververica GmbH Registered at Amtsgericht Charlottenburg: HRB
>>>>>>>>>>> 158244 B Managing Directors: Timothy Alexander Steinert, Yip
>>>>>>>>>>> Park Tung Jason, Ji (Toni) Cheng
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>>
>>>>>>>>> Arvid Heise | Senior Java Developer
>>>>>>>>>
>>>>>>>>> <https://www.ververica.com/>
>>>>>>>>>
>>>>>>>>> Follow us @VervericaData
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>>
>>>>>>>>> Join Flink Forward <https://flink-forward.org/> - The Apache Flink
>>>>>>>>> Conference
>>>>>>>>>
>>>>>>>>> Stream Processing | Event Driven | Real Time
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>>
>>>>>>>>> Ververica GmbH | Invalidenstrasse 115, 10115 Berlin, Germany
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> Ververica GmbH Registered at Amtsgericht Charlottenburg: HRB
>>>>>>>>> 158244 B Managing Directors: Timothy Alexander Steinert, Yip Park
>>>>>>>>> Tung Jason, Ji (Toni) Cheng
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>>
>>>>>>> Arvid Heise | Senior Java Developer
>>>>>>>
>>>>>>> <https://www.ververica.com/>
>>>>>>>
>>>>>>> Follow us @VervericaData
>>>>>>>
>>>>>>> --
>>>>>>>
>>>>>>> Join Flink Forward <https://flink-forward.org/> - The Apache Flink
>>>>>>> Conference
>>>>>>>
>>>>>>> Stream Processing | Event Driven | Real Time
>>>>>>>
>>>>>>> --
>>>>>>>
>>>>>>> Ververica GmbH | Invalidenstrasse 115, 10115 Berlin, Germany
>>>>>>>
>>>>>>> --
>>>>>>> Ververica GmbH Registered at Amtsgericht Charlottenburg: HRB 158244
>>>>>>> B Managing Directors: Timothy Alexander Steinert, Yip Park Tung
>>>>>>> Jason, Ji (Toni) Cheng
>>>>>>>
>>>>>>
>>>
>>> --
>>> Best, Jingsong Lee
>>>
>>
>
> --
>
> Arvid Heise | Senior Java Developer
>
> <https://www.ververica.com/>
>
> Follow us @VervericaData
>
> --
>
> Join Flink Forward <https://flink-forward.org/> - The Apache Flink
> Conference
>
> Stream Processing | Event Driven | Real Time
>
> --
>
> Ververica GmbH | Invalidenstrasse 115, 10115 Berlin, Germany
>
> --
> Ververica GmbH
> Registered at Amtsgericht Charlottenburg: HRB 158244 B
> Managing Directors: Timothy Alexander Steinert, Yip Park Tung Jason, Ji
> (Toni) Cheng
>

Reply via email to