Thanks for the info and links. I had a lot of problems I am not sure what I was doing wrong.
May be conflicts with setup from apache spark. I think I may need to setup users for each development. Anyway I kept doing fresh installs about four altogether I think. Everything works fine now Including remote access of zeppelin on machines across the local area network. Next step setup remote clusters Wish me luck ! On Sun, 19 Apr 2020, 14:58 Jeff Zhang, <zjf...@gmail.com> wrote: > Hi Som, > > You can take a look at flink on zeppelin, in zeppelin you can connect to a > remote flink cluster via a few configuration, and you don't need to worry > about the jars. Flink interpreter will ship necessary jars for you. Here's > a list of tutorials. > > 1) Get started https://link.medium.com/oppqD6dIg5 > <https://t.co/PTouUYYTrv?amp=1> 2) Batch https:// > link.medium.com/3qumbwRIg5 <https://t.co/Yo9QAY0Joj?amp=1> 3) Streaming > https://link.medium.com/RBHa2lTIg5 <https://t.co/sUapN40tvI?amp=1> 4) > Advanced usage https://link.medium.com/CAekyoXIg5 > <https://t.co/MXolULmafZ?amp=1> > > > Zahid Rahman <zahidr1...@gmail.com> 于2020年4月19日周日 下午7:27写道: > >> Hi Tison, >> >> I think I may have found what I want in example 22. >> >> https://www.programcreek.com/java-api-examples/?api=org.apache.flink.configuration.Configuration >> >> I need to create Configuration object first as shown . >> >> Also I think flink-conf.yaml file may contain configuration for client >> rather than server. So before starting is irrelevant. >> I am going to play around and see but if the Configuration class allows >> me to set configuration programmatically and overrides the yaml file then >> that would be great. >> >> >> >> On Sun, 19 Apr 2020, 11:35 Som Lima, <somplastic...@gmail.com> wrote: >> >>> Thanks. >>> flink-conf.yaml does allow me to do what I need to do without making any >>> changes to client source code. >>> >>> But >>> RemoteStreamEnvironment constructor expects a jar file as the third >>> parameter also. >>> >>> RemoteStreamEnvironment >>> <https://ci.apache.org/projects/flink/flink-docs-release-1.7/api/java/org/apache/flink/streaming/api/environment/RemoteStreamEnvironment.html#RemoteStreamEnvironment-java.lang.String-int-java.lang.String...-> >>> (String >>> <http://docs.oracle.com/javase/7/docs/api/java/lang/String.html?is-external=true> >>> host, >>> int port, String >>> <http://docs.oracle.com/javase/7/docs/api/java/lang/String.html?is-external=true> >>> ... jarFiles) >>> Creates a new RemoteStreamEnvironment that points to the master >>> (JobManager) described by the given host name and port. >>> >>> On Sun, 19 Apr 2020, 11:02 tison, <wander4...@gmail.com> wrote: >>> >>>> You can change flink-conf.yaml "jobmanager.address" or >>>> "jobmanager.port" options before run the program or take a look at >>>> RemoteStreamEnvironment which enables configuring host and port. >>>> >>>> Best, >>>> tison. >>>> >>>> >>>> Som Lima <somplastic...@gmail.com> 于2020年4月19日周日 下午5:58写道: >>>> >>>>> Hi, >>>>> >>>>> After running >>>>> >>>>> $ ./bin/start-cluster.sh >>>>> >>>>> The following line of code defaults jobmanager to localhost:6123 >>>>> >>>>> final ExecutionEnvironment env = >>>>> Environment.getExecutionEnvironment(); >>>>> >>>>> which is same on spark. >>>>> >>>>> val spark = >>>>> SparkSession.builder.master(local[*]).appname("anapp").getOrCreate >>>>> >>>>> However if I wish to run the servers on a different physical computer. >>>>> Then in Spark I can do it this way using the spark URI in my IDE. >>>>> >>>>> Conf = >>>>> SparkConf().setMaster("spark://<hostip>:<port>").setAppName("anapp") >>>>> >>>>> Can you please tell me the equivalent change to make so I can run my >>>>> servers and my IDE from different physical computers. >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> > > -- > Best Regards > > Jeff Zhang >