Thanks again!

This is strange. With both Flink 1.3.3 and Flink 1.6.0-SNAPSHOT and
1) copying gradoop-demo-shaded.jar to <Flink>/lib, and
2) using RemoteEnvironment with just jmHost and jmPort (no Jarfiles)

I get the same exception [1], caused by:
*Caused by: com.typesafe.config.ConfigException$Missing: No configuration
setting found for key 'akka.remote.log-received-messages'.*

This key is not documented anywhere, so I am confused. Also, copying with
above, also JM and TM are running, the Flink dashboard on
http://localhost:8081 is *unavailable*!

With Flink 1.3.3 and Flink 1.6.0-SNAPSHOT
1) NOT copying gradoop-shaded.jar in <Flink>/lib, and
2) using RemoteEnvironment with jmHost, jmPort and jarFiles =
{<absolute-path-to-gradoop-shaded.jar>}

I get the same exception, however the Flink dashboard on
http://localhost:8081 is *available*! This makes me believe that this is
somehow an insidious classloading issue :(.
I am really perplexed with this behavior. Let me stick to Flink 1.3.3
installation as you suggested for now.

If you have any other debugging tips, please let me know. But I am running
out of ideas to make it run with non-Local Environment.

Regards,
Kedar




[1] Gradoop shaded jar in <Flink>/lib -- exception on the web-app:
org.apache.flink.client.program.ProgramInvocationException: Could not start
the ActorSystem needed to talk to the JobManager.
at org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:461)
at
org.apache.flink.client.program.StandaloneClusterClient.submitJob(StandaloneClusterClient.java:105)
at org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:442)
at org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:429)
at org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:404)
at
org.apache.flink.client.RemoteExecutor.executePlanWithJars(RemoteExecutor.java:211)
at
org.apache.flink.client.RemoteExecutor.executePlan(RemoteExecutor.java:188)
at
org.apache.flink.api.java.RemoteEnvironment.execute(RemoteEnvironment.java:172)
at
org.apache.flink.api.java.ExecutionEnvironment.execute(ExecutionEnvironment.java:926)
at
org.gradoop.demo.server.RequestHandler.getResponse(RequestHandler.java:447)
at
org.gradoop.demo.server.RequestHandler.createResponse(RequestHandler.java:430)
at
org.gradoop.demo.server.RequestHandler.executeCypher(RequestHandler.java:121)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at
com.sun.jersey.spi.container.JavaMethodInvokerFactory$1.invoke(JavaMethodInvokerFactory.java:60)
at
com.sun.jersey.server.impl.model.method.dispatch.AbstractResourceMethodDispatchProvider$ResponseOutInvoker._dispatch(AbstractResourceMethodDispatchProvider.java:205)
at
com.sun.jersey.server.impl.model.method.dispatch.ResourceJavaMethodDispatcher.dispatch(ResourceJavaMethodDispatcher.java:75)
at
com.sun.jersey.server.impl.uri.rules.HttpMethodRule.accept(HttpMethodRule.java:302)
at
com.sun.jersey.server.impl.uri.rules.RightHandPathRule.accept(RightHandPathRule.java:147)
at
com.sun.jersey.server.impl.uri.rules.ResourceClassRule.accept(ResourceClassRule.java:108)
at
com.sun.jersey.server.impl.uri.rules.RightHandPathRule.accept(RightHandPathRule.java:147)
at
com.sun.jersey.server.impl.uri.rules.RootResourceClassesRule.accept(RootResourceClassesRule.java:84)
at
com.sun.jersey.server.impl.application.WebApplicationImpl._handleRequest(WebApplicationImpl.java:1542)
at
com.sun.jersey.server.impl.application.WebApplicationImpl._handleRequest(WebApplicationImpl.java:1473)
at
com.sun.jersey.server.impl.application.WebApplicationImpl.handleRequest(WebApplicationImpl.java:1419)
at
com.sun.jersey.server.impl.application.WebApplicationImpl.handleRequest(WebApplicationImpl.java:1409)
at
com.sun.jersey.server.impl.container.grizzly2.GrizzlyContainer._service(GrizzlyContainer.java:222)
at
com.sun.jersey.server.impl.container.grizzly2.GrizzlyContainer.service(GrizzlyContainer.java:192)
at
org.glassfish.grizzly.http.server.HttpHandler.doHandle(HttpHandler.java:164)
at
org.glassfish.grizzly.http.server.HttpHandlerChain.service(HttpHandlerChain.java:196)
at
org.glassfish.grizzly.http.server.HttpHandler.doHandle(HttpHandler.java:164)
at
org.glassfish.grizzly.http.server.HttpServerFilter.handleRead(HttpServerFilter.java:175)
at
org.glassfish.grizzly.filterchain.ExecutorResolver$9.execute(ExecutorResolver.java:119)
at
org.glassfish.grizzly.filterchain.DefaultFilterChain.executeFilter(DefaultFilterChain.java:265)
at
org.glassfish.grizzly.filterchain.DefaultFilterChain.executeChainPart(DefaultFilterChain.java:200)
at
org.glassfish.grizzly.filterchain.DefaultFilterChain.execute(DefaultFilterChain.java:134)
at
org.glassfish.grizzly.filterchain.DefaultFilterChain.process(DefaultFilterChain.java:112)
at
org.glassfish.grizzly.ProcessorExecutor.execute(ProcessorExecutor.java:78)
at
org.glassfish.grizzly.nio.transport.TCPNIOTransport.fireIOEvent(TCPNIOTransport.java:815)
at
org.glassfish.grizzly.strategies.AbstractIOStrategy.fireIOEvent(AbstractIOStrategy.java:112)
at
org.glassfish.grizzly.strategies.WorkerThreadIOStrategy.run0(WorkerThreadIOStrategy.java:115)
at
org.glassfish.grizzly.strategies.WorkerThreadIOStrategy.access$100(WorkerThreadIOStrategy.java:55)
at
org.glassfish.grizzly.strategies.WorkerThreadIOStrategy$WorkerThreadRunnable.run(WorkerThreadIOStrategy.java:135)
at
org.glassfish.grizzly.threadpool.AbstractThreadPool$Worker.doWork(AbstractThreadPool.java:567)
at
org.glassfish.grizzly.threadpool.AbstractThreadPool$Worker.run(AbstractThreadPool.java:547)
at java.lang.Thread.run(Thread.java:748)
Caused by: org.apache.flink.util.FlinkException: Could not start the
ActorSystem lazily.
at
org.apache.flink.client.program.ClusterClient$LazyActorSystemLoader.get(ClusterClient.java:230)
at org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:459)
... 47 more
*Caused by: com.typesafe.config.ConfigException$Missing: No configuration
setting found for key 'akka.remote.log-received-messages'*
at com.typesafe.config.impl.SimpleConfig.findKey(SimpleConfig.java:124)
at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:145)
at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:151)
at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:151)
at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:159)
at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:164)
at com.typesafe.config.impl.SimpleConfig.getBoolean(SimpleConfig.java:174)
at akka.remote.RemoteSettings.<init>(RemoteSettings.scala:24)
at
akka.remote.RemoteActorRefProvider.<init>(RemoteActorRefProvider.scala:114)
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
at
akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$2.apply(DynamicAccess.scala:78)
at scala.util.Try$.apply(Try.scala:192)
at
akka.actor.ReflectiveDynamicAccess.createInstanceFor(DynamicAccess.scala:73)
at
akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$3.apply(DynamicAccess.scala:84)
at
akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$3.apply(DynamicAccess.scala:84)
at scala.util.Success.flatMap(Try.scala:231)
at
akka.actor.ReflectiveDynamicAccess.createInstanceFor(DynamicAccess.scala:84)
at akka.actor.ActorSystemImpl.liftedTree1$1(ActorSystem.scala:585)
at akka.actor.ActorSystemImpl.<init>(ActorSystem.scala:578)
at akka.actor.ActorSystem$.apply(ActorSystem.scala:142)
at akka.actor.ActorSystem$.apply(ActorSystem.scala:119)
at akka.actor.ActorSystem$.create(ActorSystem.scala:67)
at
org.apache.flink.runtime.akka.AkkaUtils$.createActorSystem(AkkaUtils.scala:104)
at
org.apache.flink.runtime.akka.AkkaUtils$.createActorSystem(AkkaUtils.scala:92)
at
org.apache.flink.runtime.akka.AkkaUtils.createActorSystem(AkkaUtils.scala)
at
org.apache.flink.client.program.ClusterClient$LazyActorSystemLoader.get(ClusterClient.java:226)


On Thu, Apr 26, 2018 at 11:52 PM, Chesnay Schepler <ches...@apache.org>
wrote:

> First, a small correction for my previous mail:
>
> I could reproduce your problems locally when submitting the fat-jar.
> Turns out i never submitted the far-jar, as i didn't pass the jar file
> argument to RemoteEnvironment.
>
> Now on to your questions:
>
> *What version of Flink are you trying with?*
> I got it working *once *with 1.6-SNAPSHOT, but i would recommend sticking
> with 1.3.1 since that is the version gradoop depends on. (i haven't tried
> it with this version yet, but that's the next thing on my list)
>
>
> *Are there other config changes (flink-conf.yaml) that you made in your
> cluster? *It was the standard config.
>
>
> *Is org.apache.flink.api.common.io.FileOutputFormat a good alternative to
> LocalCollectionOutputFormat? *It can be used, but if the result is small
> you could also use accumulators.
>
>
> *Do you think it is better to use jarFiles argument on
> createRemoteEnvironment? *Yes, once we get it working this is the way to
> go.
>
>
> On 26.04.2018 18:42, kedar mhaswade wrote:
>
> Thanks Chesnay for your incredible help!
>
> I will try out the suggestions again. A few questions:
> - What version of Flink are you trying with? I have had issues when I
> placed the gradoop-demo-shaded.jar in the lib folder on Flink
> installation (1.4 even refused to start!).
> - Are there other config changes (flink-conf.yaml) that you made in your
> cluster?
> - Is org.apache.flink.api.common.io.FileOutputFormat a good alternative
> to LocalCollectionOutputFormat, or should I use
> HadoopOutputFormatCommonBase (I do want to run the cluster on YARN later;
> at the moment I am trying on a standalone cluster).
> - Do you think it is better to use jarFiles argument on
> createRemoteEnvironment (which deploys the JAR only for this job and not
> mess with the entire Flink cluster) a better option than placing the JAR(s)
> in the lib folder?
>
> Thanks again,
> Regards,
> Kedar
>
>
> On Thu, Apr 26, 2018 at 3:14 AM, Chesnay Schepler <ches...@apache.org>
> wrote:
>
>> Small update:
>>
>> I could reproduce your problems locally when submitting the fat-jar.
>> I could get the job to run after placing the gradoop-demo-shaded.jar into
>> the lib folder.
>> I have not tried yet placing only the gradoop jars into lib (but my guess
>> is you missed a gradoop jar)
>>
>> Note that the job fails to run since you use
>> "LocalCollectionOutputFormat" which can only be used for local execution,
>> i.e. when the job submission and execution happen in the same JVM.
>>
>>
>> On 25.04.2018 14:23, kedar mhaswade wrote:
>>
>> Thank you for your response!
>>
>> I have not tried the flink run app.jar route because the way the app is
>> set up does not allow me to do it. Basically, the app is a web application
>> which serves the UI and also submits a Flink job for running Cypher
>> queries. It is a proof-of-concept app, but IMO, a very useful one.
>>
>> Here's how you can reproduce:
>> 1) git clone g...@github.com:kedarmhaswade/gradoop_demo.git (this is my
>> fork of gradoop_demo)
>> 2) cd gradoop_demo
>> 3) git checkout dev => dev is the branch where my changes to make gradoop
>> work with remote environment go.
>> 4) mvn clean package => should bring the gradoop JARs that this app
>> needs; these JARs should then be placed in <flink-install>/lib.
>> 5) cp 
>> ~/.m2/repository/org/gradoop/gradoop-common/0.3.2/gradoop-common-0.3.2.jar
>> <flink-install>/lib, cp ~/.m2/repository/org/gradoop/g
>> radoop-flink/0.3.2/gradoop-flink-0.3.2.jar <flink-install>/lib,
>> cp target/gradoop-demo-0.2.0.jar <flink-install>/lib.
>> 6) start the local flink cluster (I have tried with latest
>> (built-from-source) 1.6-SNAPSHOT, or 1.4) 
>> <flink-install>/bin/start-cluster.sh
>> -- note the JM host and port
>> 7) <gradoop-demo>/start.sh --jmhost <host> --jmport 6123 (adjust host and
>> port per your cluster) => this is now configured to talk to the
>> RemoteEnvironment at given JM host and port.
>> 8) open a browser at: http://localhost:2342/gradoop/html/cypher.html
>> 9) hit the query button => this would throw the exception
>> 10) Ctrl C the process in 7 and just restart it as java -cp target/
>> classes:target/gradoop-demo-shaded.jar org.gradoop.demo.server.Server =>
>> starts LocalEnvironment
>> 11) do 9 again and see the results shown nicely in the browser.
>>
>> Here is the relevant code:
>> 1) Choosing between
>> <https://github.com/kedarmhaswade/gradoop_demo/blob/dev/src/main/java/org/gradoop/demo/server/Server.java#L107>
>> a Remote or a Local Environment.
>>
>> The instructions are correct to my knowledge. Thanks for your willingness
>> to try. I have tried everything I can. With different Flink versions, I get
>> different results (I have also tried on 1.6-SNAPSHOT with class loading
>> config being parent-first, or child-first).
>>
>> Regards,
>> Kedar
>>
>>
>> On Wed, Apr 25, 2018 at 1:08 AM, Chesnay Schepler <ches...@apache.org>
>> wrote:
>>
>>> I couldn't spot any error in what you tried to do. Does the
>>> job-submission succeed if you submit the jar through the command-line
>>> client?
>>>
>>> Can you share the project, or a minimal reproducing version?
>>>
>>>
>>> On 25.04.2018 00:41, kedar mhaswade wrote:
>>>
>>> I am trying to get gradoop_demo
>>> <https://github.com/dbs-leipzig/gradoop_demo> (a gradoop based graph
>>> visualization app) working on Flink with *Remote* Execution
>>> Environment.
>>>
>>> This app, which is based on Gradoop, submits a job to the
>>> *preconfigured* execution environment, collects the results and sends
>>> it to the UI for rendering.
>>>
>>> When the execution environment is configured to be a LocalEnvironment
>>> <https://ci.apache.org/projects/flink/flink-docs-stable/api/java/org/apache/flink/api/java/LocalEnvironment.html>,
>>> everything works fine. But when I start a cluster (using <
>>> flink-install-path>/bin/start-cluster.sh), get the Job Manager endpoint
>>> (e.g. localhost:6123) and configure a RemoteEnvironment
>>> <https://ci.apache.org/projects/flink/flink-docs-stable/api/java/org/apache/flink/api/java/ExecutionEnvironment.html#createRemoteEnvironment-java.lang.String-int-org.apache.flink.configuration.Configuration-java.lang.String...->
>>>  and
>>> use that environment to run the job, I get exceptions [1].
>>>
>>> Based on the class loading doc
>>> <https://ci.apache.org/projects/flink/flink-docs-master/monitoring/debugging_classloading.html>,
>>> I copied the gradoop classes (gradoop-flink-0.3.3-SNAPSHOT.
>>> jar, gradoop-common-0.3.3-SNAPSHOT.jar) to the <flink-install-path>/lib
>>> folder (hoping that that way those classes will be available to all the
>>> executors in the cluster). I have ensured that the class that Flink fails
>>> to load is in fact available in the Gradoop jars that I copied to the
>>> /lib folder.
>>>
>>> I have tried using the RemoteEnvironment method with jarFiles argument
>>> where the passed JAR file is a fat jar containing everything (in which case
>>> there is no Gradoop JAR file in /lib folder).
>>>
>>> So, my questions are:
>>> 1) How can I use RemoteEnvironment?
>>> 2) Is there any other way of doing this *programmatically? *(That means
>>> I can't do flink run since I am interested in the job execution result as a
>>> blocking call -- which means ideally I don't want to use the submit RESTful
>>> API as well). I just want RemoteEnvironment to work as well as
>>> LocalEnvironment.
>>>
>>> Regards,
>>> Kedar
>>>
>>>
>>> [1]
>>> 2018-04-24 15:16:02,823 ERROR org.apache.flink.runtime.jobmanager.JobManager
>>>               - Failed to submit job 0c987c8704f8b7eb4d7d38efcb3d708d
>>> (Flink Java Job at Tue Apr 24 15:15:59 PDT 2018)
>>> java.lang.NoClassDefFoundError: Could not initialize class
>>> *org.gradoop.common.model.impl.id.GradoopId*
>>>   at java.io.ObjectStreamClass.hasStaticInitializer(Native Method)
>>>   at java.io.ObjectStreamClass.computeDefaultSUID(ObjectStreamCla
>>> ss.java:1887)
>>>   at java.io.ObjectStreamClass.access$100(ObjectStreamClass.java:79)
>>>   at java.io.ObjectStreamClass$1.run(ObjectStreamClass.java:263)
>>>   at java.io.ObjectStreamClass$1.run(ObjectStreamClass.java:261)
>>>   at java.security.AccessController.doPrivileged(Native Method)
>>>   at java.io.ObjectStreamClass.getSerialVersionUID(ObjectStreamCl
>>> ass.java:260)
>>>   at java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:682)
>>>   at java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream
>>> .java:1876)
>>>   at java.io.ObjectInputStream.readClassDesc(ObjectInputStream.ja
>>> va:1745)
>>>   at java.io.ObjectInputStream.readClass(ObjectInputStream.java:1710)
>>>   at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1550)
>>>   at java.io.ObjectInputStream.readObject(ObjectInputStream.java:427)
>>>   at java.util.HashSet.readObject(HashSet.java:341)
>>>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>   at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAcce
>>> ssorImpl.java:62)
>>>   at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMe
>>> thodAccessorImpl.java:43)
>>>   at java.lang.reflect.Method.invoke(Method.java:498)
>>>   at java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass
>>> .java:1158)
>>>   at java.io.ObjectInputStream.readSerialData(ObjectInputStream.j
>>> ava:2169)
>>>   at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStre
>>> am.java:2060)
>>>   at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1567)
>>>   at java.io.ObjectInputStream.defaultReadFields(ObjectInputStrea
>>> m.java:2278)
>>>   at java.io.ObjectInputStream.readSerialData(ObjectInputStream.j
>>> ava:2202)
>>>   at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStre
>>> am.java:2060)
>>>   at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1567)
>>>   at java.io.ObjectInputStream.readObject(ObjectInputStream.java:427)
>>>   at org.apache.flink.util.InstantiationUtil.deserializeObject(In
>>> stantiationUtil.java:290)....
>>>
>>>
>>>
>>
>>
>
>

Reply via email to