I changed the levels to DEBUG. I don't see useful data in the logs. https://drive.google.com/file/d/1ua1zsr3BInY_8xdsWwA__F0uloAqy-vG/view?usp=sharing
On Thu, Sep 10, 2020 at 8:45 AM Arvid Heise <ar...@ververica.com> wrote: > Could you try 1) or 2) and enable debug logging* and share the log with us? > > *Usually by adjusting FLINK_HOME/conf/log4j.properties. > > On Thu, Sep 10, 2020 at 5:38 PM Dan Hill <quietgol...@gmail.com> wrote: > >> Ah, sorry, it's a copy/paste issue with this email. I've tried both: >> 1) using s3a uri with flink-s3-fs-hadoop jar >> in /opt/flink/plugins/s3-fs-hadoop. >> 2) using s3p uri with flink-s3-fs-presto jar >> in /opt/flink/plugins/s3-fs-presto. >> 3) loading both 1 and 2 >> 4) trying s3 uri. >> >> When doing 1) >> >> Caused by: org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: >> Could not find a file system implementation for scheme 's3a'. The scheme is >> directly supported by Flink through the following plugin: >> flink-s3-fs-hadoop. Please ensure that each plugin resides within its own >> subfolder within the plugins directory. See >> https://ci.apache.org/projects/flink/flink-docs-stable/ops/plugins.html >> for more information. If you want to use a Hadoop file system for that >> scheme, please add the scheme to the configuration >> fs.allowed-fallback-filesystems. For a full list of supported file systems, >> please see >> https://ci.apache.org/projects/flink/flink-docs-stable/ops/filesystems/. >> >> When doing 2) >> >> Caused by: org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: >> Could not find a file system implementation for scheme 's3p'. The scheme is >> directly supported by Flink through the following plugin: >> flink-s3-fs-presto. Please ensure that each plugin resides within its own >> subfolder within the plugins directory. See >> https://ci.apache.org/projects/flink/flink-docs-stable/ops/plugins.html >> for more information. If you want to use a Hadoop file system for that >> scheme, please add the scheme to the configuration >> fs.allowed-fallback-filesystems. For a full list of supported file systems, >> please see >> https://ci.apache.org/projects/flink/flink-docs-stable/ops/filesystems/. >> >> etc >> >> On Thu, Sep 10, 2020 at 8:15 AM Arvid Heise <ar...@ververica.com> wrote: >> >>> Hi Dan, >>> >>> s3p is only provided by flink-s3-fs-presto plugin. The plugin you used >>> provides s3a. >>> (and both provide s3, but it's good to use the more specific prefix). >>> >>> Best, >>> >>> Arvid >>> >>> On Thu, Sep 10, 2020 at 9:24 AM Dan Hill <quietgol...@gmail.com> wrote: >>> >>>> *Background* >>>> I'm converting some prototype Flink v1.11.1 code that uses >>>> DataSet/DataTable APIs to use the Table API. >>>> >>>> *Problem* >>>> When switching to using the Table API, my s3 plugins stopped working. >>>> I don't know why. I've added the required maven table dependencies to the >>>> job. >>>> >>>> I've tried us moving both the presto and/or the hadoop s3 jars to >>>> plugin subfolders. No luck. >>>> >>>> Any ideas what is wrong? I'm guessing I'm missing something simple. >>>> >>>> >>>> *Error* >>>> >>>> Caused by: >>>> org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Could not >>>> find a file system implementation for scheme 's3p'. The scheme is directly >>>> supported by Flink through the following plugin: flink-s3-fs-presto. Please >>>> ensure that each plugin resides within its own subfolder within the plugins >>>> directory. See >>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/plugins.html >>>> for more information. If you want to use a Hadoop file system for that >>>> scheme, please add the scheme to the configuration >>>> fs.allowed-fallback-filesystems. For a full list of supported file systems, >>>> please see >>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/filesystems/ >>>> . >>>> >>>> at >>>> org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:473) >>>> >>>> at org.apache.flink.core.fs.FileSystem.get(FileSystem.java:389) >>>> >>>> at org.apache.flink.core.fs.Path.getFileSystem(Path.java:292) >>>> >>>> at >>>> org.apache.flink.table.filesystem.FileSystemTableSink.toStagingPath(FileSystemTableSink.java:232) >>>> >>>> ... 35 more >>>> >>>> *ls of plugins directory (same for taskmanager)* >>>> >>>> kubectl exec pod/flink-jobmanager-0 -- ls -l >>>> /opt/flink/plugins/s3-fs-hadoop >>>> >>>> total 19520 >>>> >>>> -rw-r--r-- 1 root root 19985452 Sep 10 06:27 >>>> flink-s3-fs-hadoop-1.11.1.jar >>>> >>>> >>>> >>> >>> -- >>> >>> Arvid Heise | Senior Java Developer >>> >>> <https://www.ververica.com/> >>> >>> Follow us @VervericaData >>> >>> -- >>> >>> Join Flink Forward <https://flink-forward.org/> - The Apache Flink >>> Conference >>> >>> Stream Processing | Event Driven | Real Time >>> >>> -- >>> >>> Ververica GmbH | Invalidenstrasse 115, 10115 Berlin, Germany >>> >>> -- >>> Ververica GmbH >>> Registered at Amtsgericht Charlottenburg: HRB 158244 B >>> Managing Directors: Timothy Alexander Steinert, Yip Park Tung Jason, Ji >>> (Toni) Cheng >>> >> > > -- > > Arvid Heise | Senior Java Developer > > <https://www.ververica.com/> > > Follow us @VervericaData > > -- > > Join Flink Forward <https://flink-forward.org/> - The Apache Flink > Conference > > Stream Processing | Event Driven | Real Time > > -- > > Ververica GmbH | Invalidenstrasse 115, 10115 Berlin, Germany > > -- > Ververica GmbH > Registered at Amtsgericht Charlottenburg: HRB 158244 B > Managing Directors: Timothy Alexander Steinert, Yip Park Tung Jason, Ji > (Toni) Cheng >