Hey Ken,

I have flink-s3-fs-hadoop as a provided dependency in my project, and I've
configured my IDE to include provided dependencies when starting
applications. Works just fine.

On Tue, Jan 3, 2023 at 5:06 PM Ken Krugler <kkrugler_li...@transpac.com>
wrote:

> Hi all,
>
> With Flink 1.15.x, is there a way to use the S3 Presto plugin when running
> code on the MiniCluster?
>
> I can’t just add that jar as a dependency when testing, as I get:
>
> java.lang.NoClassDefFoundError: Could not initialize class
> com.facebook.presto.hive.s3.PrestoS3FileSystem
> at
> org.apache.flink.fs.s3presto.S3FileSystemFactory.createHadoopFileSystem(S3FileSystemFactory.java:88)
> ~[flink-s3-fs-presto-1.15.1.jar:1.15.1]
> at
> org.apache.flink.fs.s3.common.AbstractS3FileSystemFactory.create(AbstractS3FileSystemFactory.java:126)
> ~[flink-s3-fs-presto-1.15.1.jar:1.15.1]
>
> I assume that’s because of this warning in the Flink docs:
>
> The s3 file systems (flink-s3-fs-presto and flink-s3-fs-hadoop) can only
> be used as plugins as we already removed the relocations. Placing them in
> libs/ will result in system failures.
>
>
> In the latest Flink JavaDocs, there’s a way to specify the PluginManager
> for the MiniClusterConfiguration, but I don’t see that in Flink 1.15.x
>
> So is there a workaround to allow me to run a test from inside of my IDE,
> using the MiniCluster, that reads from S3?
>
> Thanks,
>
> — Ken
>
> --------------------------
> Ken Krugler
> http://www.scaleunlimited.com
> Custom big data solutions
> Flink, Pinot, Solr, Elasticsearch
>
>

Reply via email to