Could you try to copy the plugin jar into FLINK_HOME/lib? On Fri, Oct 21, 2022 at 4:59 AM Matt Fysh <m...@unweb.io> wrote:
> Hi Jeff, > > I created a small reproduction repo here: > https://github.com/mattfysh/zepp-flink-s3 and the error I'm seeing is: > > org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Could not > find a file system implementation for scheme 's3a' > > > I'm using version 1.13.2 to match AWS where I hope to deploy my final app, > and I've included the steps to copy the plugin into the plugins directory > in the install.sh script > > I see this same error regardless of whether I try using the Hadoop or > Presto plugins, it seems that the plugin is not being loaded by the service > loader, so there may be some additional config required? > > Thanks, > Matt > > On Thu, 20 Oct 2022 at 18:43, Jeff Zhang <zjf...@gmail.com> wrote: > >> What exact error do you mean? Even if you run flink inside zeppelin, you >> still need to configure FLINK_HOME, for any jars, you need to put under >> FLINK_HOME I believe. >> >> >> On Thu, Oct 20, 2022 at 3:52 PM Matt Fysh <m...@unweb.io> wrote: >> >>> Hi Zeppers, >>> >>> I'm running Flink inside the Zeppelin 0.10 docker image, and having >>> difficulty writing a stream to S3. I've copied the plugin inside the Flink >>> plugin directory, but I think it also requires the Service Loader to be >>> configured according to this StackOverflow question >>> <https://stackoverflow.com/questions/64115627/flink-1-11-2-cant-find-implementation-for-s3-despite-correct-plugins-being> >>> >>> I was wondering how I can get this to work when running Flink inside of >>> Zeppelin? Given that the jobs we write in notebooks don't have a permanent >>> JAR, I'm at a loss to understand where I should palace this META-INF folder >>> with the services directory inside >>> >>> Thanks, >>> Matt >>> >> >> >> -- >> Best Regards >> >> Jeff Zhang >> > -- Best Regards Jeff Zhang