Hi Lorenzo,

Since Flink 1.8 does not support plugin mechanism to load filesystem, you
need to copy flink-s3-fs-hadoop-*.jar
from opt to lib directory.

The dockerfile could be like following.

FROM flink:1.8-scala_2.11
RUN cp /opt/flink/opt/flink-s3-fs-hadoop-*.jar /opt/flink/lib

Then build you docker image and start the session cluster again.


Best,
Yang


Lorenzo Nicora <lorenzo.nic...@gmail.com> 于2020年7月2日周四 下午6:05写道:

> Hi
>
> I need to set up a dockerized *session cluster* using Flink *1.8.2* for
> development and troubleshooting. We are bound to 1.8.2 as we are deploying
> to AWS Kinesis Data Analytics for Flink.
>
> I am using an image based on the semi-official flink:1.8-scala_2.11
> I need to add to my dockerized cluster support for S3 Hadoop File System
> (s3a://), we have on KDA out of the box.
>
> Note I do not want to add dependencies to the job directly, as I want to
> deploy locally exactly the same JAR I deploy to KDA.
>
> Flink 1.8 docs [1] say  is supported out of the box but does not look to
> be the case for dockerised version.
> I am getting "Could not find a file system implementation for scheme
> 's3a'" and "Hadoop is not in the classpath/dependencies".
> I assume I need to create a customised docker image,
> extending flink:1.8-scala_2.11, but I do not understand how to add support
> for S3 Hadoop FS.
>
> Can someone please point me in the right direction? Docs or examples?
>
>
> [1]
> https://ci.apache.org/projects/flink/flink-docs-release-1.8/ops/filesystems.html
>
>
> Lorenzo
>

Reply via email to