Hi Matt,

Yes. There are several official Flink images with various JVMs including
Java 11.

https://hub.docker.com/_/flink

Cheers,
Matyas

On Fri, Jun 24, 2022 at 2:06 PM Matt Casters <matt.cast...@neotechnology.com>
wrote:

> Hi Mátyás & all,
>
> Thanks again for the advice so far. On a related note I noticed Java 8
> being used, indicated in the log.
>
> org.apache.flink.runtime.entrypoint.ClusterEntrypoint        [] -
>  JAVA_HOME: /usr/local/openjdk-8
>
> Is there a way to use Java 11 to start Flink with?
>
> Kind regards,
>
> Matt
>
> On Tue, Jun 21, 2022 at 4:53 PM Őrhidi Mátyás <matyas.orh...@gmail.com>
> wrote:
>
>> Hi Matt,
>>
>> I believe an artifact fetcher (e.g
>> https://hub.docker.com/r/agiledigital/s3-artifact-fetcher ) + the pod
>> template (
>> https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/custom-resource/pod-template/#pod-template)
>> is an elegant way to solve your problem.
>>
>> The operator uses K8s native integration under the hood:
>> https://nightlies.apache.org/flink/flink-docs-release-1.13/docs/deployment/resource-providers/native_kubernetes/#application-mode
>>  In
>> application mode,  the main() method of the application is executed on the
>> JobManager, hence we need the jar locally.
>>
>> You can launch a session cluster (without job spec) on the operator that
>> allows submitting jars if you would like to avoid dealing with
>> authentication, but the recommended and safe approach is to use
>> sessionjobs for this purpose.
>>
>>
>> Cheers,
>> Matyas
>>
>> On Tue, Jun 21, 2022 at 4:03 PM Matt Casters <
>> matt.cast...@neotechnology.com> wrote:
>>
>>> Thank you very much for the help Matyas and Gyula!
>>>
>>> I just saw a video today where you were presenting the FKO.  Really nice
>>> stuff!
>>>
>>> So I'm guessing we're executing "flink run" at some point on the master
>>> and that this is when we need the jar file to be local?
>>> Am I right in assuming that this happens after the flink cluster in
>>> question was started, as part of the job execution?
>>>
>>> On the one hand I agree with the underlying idea that authentication and
>>> security should not be a responsibility of the operator.   On the other
>>> hand I could add a flink-s3 driver but then I'd also have to configure it
>>> and so on and it's just hard to get that configuration to be really clean.
>>>
>>> Do we have some service running on the flink cluster which would allow
>>> us to post/copy files from the client (running kubectl) to the master?  If
>>> so, could we add an option to the job specification to that effect?  Just
>>> brainstorming ;-) (and forking apache/flink-kubernetes-operator)
>>>
>>> All the best,
>>> Matt
>>>
>>> On Tue, Jun 21, 2022 at 2:52 PM Őrhidi Mátyás <matyas.orh...@gmail.com>
>>> wrote:
>>>
>>>> Hi Matt,
>>>>
>>>> - In FlinkDeployments you can utilize an init container to download
>>>> your artifact onto a shared volume, then you can refer to it as local:/..
>>>> from the main container. FlinkDeployments comes with pod template support
>>>> https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/custom-resource/pod-template/#pod-template
>>>>
>>>> - FlinkSessionJobs comes with an artifact fetcher, but it may need some
>>>> tweaking to make it work on your environment:
>>>>
>>>> https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/custom-resource/overview/#flinksessionjob-spec-overview
>>>>
>>>> I hope it helps, let us know if you have further questions.
>>>>
>>>> Cheers,
>>>> Matyas
>>>>
>>>>
>>>>
>>>> On Tue, Jun 21, 2022 at 2:35 PM Matt Casters <
>>>> matt.cast...@neotechnology.com> wrote:
>>>>
>>>>> Hi Flink team!
>>>>>
>>>>> I'm interested in getting the new Flink Kubernetes Operator to work on
>>>>> AWS EKS.  Following the documentation I got pretty far.  However, when
>>>>> trying to run a job I got the following error:
>>>>>
>>>>> Only "local" is supported as schema for application mode. This assumes
>>>>>> t
>>>>>> hat the jar is located in the image, not the Flink client. An example
>>>>>> of such path is: local:///opt/flink/examples/streaming/WindowJoin.jar
>>>>>
>>>>>
>>>>>  I have an Apache Hop/Beam fat jar capable of running the Flink
>>>>> pipeline in my yml file:
>>>>>
>>>>> jarURI: s3://hop-eks/hop/hop-2.1.0-fat.jar
>>>>>
>>>>> So how could I go about getting the fat jar in a desired location for
>>>>> the operator?
>>>>>
>>>>> Getting this to work would be really cool for both short and
>>>>> long-lived pipelines in the service of all sorts of data integration work.
>>>>> It would do away with the complexity of setting up and maintaining your 
>>>>> own
>>>>> Flink cluster.
>>>>>
>>>>> Thanks in advance!
>>>>>
>>>>> All the best,
>>>>>
>>>>> Matt (mcasters, Apache Hop PMC)
>>>>>
>>>>>

Reply via email to