This is a limitation of the presto version; use flink-s3-fs-hadoop-1.11.3.jar instead.

On 08/09/2021 20:39, Dhiru wrote:
I copied
FROM flink:1.11.3-scala_2.12-java11 RUN mkdir ./plugins/flink-s3-fs-presto RUN cp ./opt/flink-s3-fs-presto-1.11.3.jar   ./plugins/flink-s3-fs-presto/ then started getting this error , trying to run on aws eks and trying to access s3 bucket
2021-09-08 14:38:10
java.lang.UnsupportedOperationException: This s3 file system implementation does not support recoverable writers.     at org.apache.flink.fs.s3.common.FlinkS3FileSystem.createRecoverableWriter(FlinkS3FileSystem.java:136)     at org.apache.flink.core.fs.PluginFileSystemFactory$ClassLoaderFixingFileSystem.createRecoverableWriter(PluginFileSystemFactory.java:134)     at org.apache.flink.core.fs.SafetyNetWrapperFileSystem.createRecoverableWriter(SafetyNetWrapperFileSystem.java:69)     at org.apache.flink.streaming.api.functions.sink.filesystem.StreamingFileSink$RowFormatBuilder.createBuckets(StreamingFileSink.java:260)     at org.apache.flink.streaming.api.functions.sink.filesystem.StreamingFileSink.initializeState(StreamingFileSink.java:396)     at org.apache.flink.streaming.util.functions.StreamingFunctionUtils.tryRestoreFunction(StreamingFunctionUtils.java:185)     at org.apache.flink.streaming.util.functions.StreamingFunctionUtils.restoreFunctionState(StreamingFunctionUtils.java:167)     at org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator.initializeState(AbstractUdfStreamOperator.java:96)     at org.apache.flink.streaming.api.operators.StreamOperatorStateHandler.initializeOperatorState(StreamOperatorStateHandler.java:106)     at org.apache.flink.streaming.api.operators.AbstractStreamOperator.initializeState(AbstractStreamOperator.java:260)     at org.apache.flink.streaming.runtime.tasks.OperatorChain.initializeStateAndOpenOperators(OperatorChain.java:290)     at org.apache.flink.streaming.runtime.tasks.StreamTask.lambda$beforeInvoke$1(StreamTask.java:506)     at org.apache.flink.streaming.runtime.tasks.StreamTaskActionExecutor$SynchronizedStreamTaskActionExecutor.runThrowing(StreamTaskActionExecutor.java:92)     at org.apache.flink.streaming.runtime.tasks.StreamTask.beforeInvoke(StreamTask.java:475)     at org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:526)
    at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:721)
    at org.apache.flink.runtime.taskmanager.Task.run(Task.java:546)
    at java.base/java.lang.Thread.run(Unknown Source)


On Wednesday, September 8, 2021, 12:47:10 PM EDT, Chesnay Schepler <ches...@apache.org> wrote:


you need to put the flink-s3-fs-hadoop/presto jar into a directory within the plugins directory, for example the final path should look like this:

/opt/flink/plugins/flink-s3-fs-hadoop/flink-s3-fs-hadoop-1.13.1.jar

Furthermore, you only need either the hadoop or presto jar, _not_ both of them.

See also:
https://nightlies.apache.org/flink/flink-docs-release-1.13/docs/deployment/filesystems/s3/#hadooppresto-s3-file-systems-plugins <https://nightlies.apache.org/flink/flink-docs-release-1.13/docs/deployment/filesystems/s3/#hadooppresto-s3-file-systems-plugins> https://nightlies.apache.org/flink/flink-docs-release-1.13/docs/deployment/filesystems/plugins/ <https://nightlies.apache.org/flink/flink-docs-release-1.13/docs/deployment/filesystems/plugins/>

On 08/09/2021 17:10, Dhiru wrote:
yes I copied to plugin folder but not sure same jar I see in  /opt as well by default

root@d852f125da1f:/opt/flink/plugins# ls
README.txt  flink-s3-fs-hadoop-1.13.1.jar metrics-datadog   metrics-influx metrics-prometheus  metrics-statsd external-resource-gpu flink-s3-fs-presto-1.13.1.jar metrics-graphite  metrics-jmx  metrics-slf4j

I need  help sooner on this

On Wednesday, September 8, 2021, 09:26:46 AM EDT, Dhiru <userdh...@yahoo.com> <mailto:userdh...@yahoo.com> wrote:



yes I copied to plugin folder but not sure same jar I see in /opt as well by default

root@d852f125da1f:/opt/flink/plugins# ls
README.txt  flink-s3-fs-hadoop-1.13.1.jar metrics-datadog   metrics-influx metrics-prometheus  metrics-statsd external-resource-gpu flink-s3-fs-presto-1.13.1.jar metrics-graphite  metrics-jmx  metrics-slf4j


On Wednesday, September 8, 2021, 02:58:38 AM EDT, Martijn Visser <mart...@ververica.com> <mailto:mart...@ververica.com> wrote:


Hi,

Have you copied the correct JAR [1] to the plugins directory?

Best regards,

Martijn

[1] https://ci.apache.org/projects/flink/flink-docs-release-1.11/ops/filesystems/s3.html <https://ci.apache.org/projects/flink/flink-docs-release-1.11/ops/filesystems/s3.html>

On Wed, 8 Sept 2021 at 04:27, Dhiru <userdh...@yahoo.com <mailto:userdh...@yahoo.com>> wrote:

    Need to configure aws S3 getting this error
    org.apache.flink.core.fs.UnsupportedFileSystemSchemeException:
    Could not find a file system implementation for scheme 's3'. The
    scheme is directly supported by Flink through the following
    plugins: flink-s3-fs-hadoop, flink-s3-fs-presto. Please ensure
    that each plugin resides within its own subfolder within the
    plugins directory. See
    https://ci.apache.org/projects/flink/flink-docs-stable/ops/plugins.html
    <https://ci.apache.org/projects/flink/flink-docs-stable/ops/plugins.html>
    for more information. If you want to use a Hadoop file system for
    that scheme, please add the scheme to the configuration
    fs.allowed-fallback-filesystems. For a full list of supported file
    systems, please see
    https://ci.apache.org/projects/flink/flink-docs-stable/ops/filesystems/
    <https://ci.apache.org/projects/flink/flink-docs-stable/ops/filesystems/>.
        at
    
org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:473)
        at org.apache.flink.core.fs.FileSystem.get(FileSystem.java:389)
        at
    
org.apache.flink.streaming.api.functions.sink.filesystem.StreamingFileSink$RowFormatBuilder.createBuckets(StreamingFileSink.java:260)
        at
    
org.apache.flink.streaming.api.functions.sink.filesystem.StreamingFileSink.initializeState(StreamingFileSink.java:396)
        at
    
org.apache.flink.streaming.util.functions.StreamingFunctionUtils.tryRestoreFunction(StreamingFunctionUtils.java:185)
        at
    
org.apache.flink.streaming.util.functions.StreamingFunctionUtils.restoreFunctionState(StreamingFunctionUtils.java:167)
        at
    
org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator.initializeState(AbstractUdfStreamOperator.java:96)
        at
    
org.apache.flink.streaming.api.operators.StreamOperatorStateHandler.initializeOperatorState(StreamOperatorStateHandler.java:106)
        at
    
org.apache.flink.streaming.api.operators.AbstractStreamOperator.initializeState(AbstractStreamOperator.java:260)
        at
    
org.apache.flink.streaming.runtime.tasks.OperatorChain.initializeStateAndOpenOperators(OperatorChain.java:290)
        at
    
org.apache.flink.streaming.runtime.tasks.StreamTask.lambda$beforeInvoke$1(StreamTask.java:506)
        at
    
org.apache.flink.streaming.runtime.tasks.StreamTaskActionExecutor$SynchronizedStreamTaskActionExecutor.runThrowing(StreamTaskActionExecutor.java:92)
        at
    
org.apache.flink.streaming.runtime.tasks.StreamTask.beforeInvoke(StreamTask.java:475)
        at
    
org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:526)
        at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:721)
        at org.apache.flink.runtime.taskmanager.Task.run(Task.java:546)
        at java.base/java.lang.Thread.run(Unknown Source)
    image used flink:1.11.3-scala_2.12-java11
    some of the part which I added for flink-conf.yaml
      # REQUIRED: set storage location for job metadata in remote storage
        high-availability.storageDir: s3://msc-sandbox-flink/test
        state.backend: filesystem
        state.checkpoints.dir: s3://msc-sandbox-flink/test
        state.backend.fs.checkpointdir: s3://msc-sandbox-flink/test
        s3.path.style.access: true
    I am trying to deploy jobmanager and taskmanager using eks ,
    please let me know if I need to do any more support for adding s3



Reply via email to