Thank you very much for the help Matyas and Gyula! I just saw a video today where you were presenting the FKO. Really nice stuff!
So I'm guessing we're executing "flink run" at some point on the master and that this is when we need the jar file to be local? Am I right in assuming that this happens after the flink cluster in question was started, as part of the job execution? On the one hand I agree with the underlying idea that authentication and security should not be a responsibility of the operator. On the other hand I could add a flink-s3 driver but then I'd also have to configure it and so on and it's just hard to get that configuration to be really clean. Do we have some service running on the flink cluster which would allow us to post/copy files from the client (running kubectl) to the master? If so, could we add an option to the job specification to that effect? Just brainstorming ;-) (and forking apache/flink-kubernetes-operator) All the best, Matt On Tue, Jun 21, 2022 at 2:52 PM Őrhidi Mátyás <matyas.orh...@gmail.com> wrote: > Hi Matt, > > - In FlinkDeployments you can utilize an init container to download your > artifact onto a shared volume, then you can refer to it as local:/.. from > the main container. FlinkDeployments comes with pod template support > https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/custom-resource/pod-template/#pod-template > > - FlinkSessionJobs comes with an artifact fetcher, but it may need some > tweaking to make it work on your environment: > > https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/custom-resource/overview/#flinksessionjob-spec-overview > > I hope it helps, let us know if you have further questions. > > Cheers, > Matyas > > > > On Tue, Jun 21, 2022 at 2:35 PM Matt Casters < > matt.cast...@neotechnology.com> wrote: > >> Hi Flink team! >> >> I'm interested in getting the new Flink Kubernetes Operator to work on >> AWS EKS. Following the documentation I got pretty far. However, when >> trying to run a job I got the following error: >> >> Only "local" is supported as schema for application mode. This assumes t >>> hat the jar is located in the image, not the Flink client. An example of >>> such path is: local:///opt/flink/examples/streaming/WindowJoin.jar >> >> >> I have an Apache Hop/Beam fat jar capable of running the Flink pipeline >> in my yml file: >> >> jarURI: s3://hop-eks/hop/hop-2.1.0-fat.jar >> >> So how could I go about getting the fat jar in a desired location for the >> operator? >> >> Getting this to work would be really cool for both short and long-lived >> pipelines in the service of all sorts of data integration work. It would >> do away with the complexity of setting up and maintaining your own Flink >> cluster. >> >> Thanks in advance! >> >> All the best, >> >> Matt (mcasters, Apache Hop PMC) >> >>