Re: Published test artifacts for flink streaming

2015-11-05 Thread Nick Dimiduk
Thanks Stephan, I'll check that out in the morning. Generally speaking, it would be great to have some single-jvm example tests for those of us getting started. Following the example of WindowingIntegrationTest is mostly working, though reusing my single sink instance with its static collection res

Re: Published test artifacts for flink streaming

2015-11-05 Thread Stephan Ewen
Hey! There is also a collect() sink in the "flink-streaming-contrib" project, see here: https://github.com/apache/flink/blob/master/flink-contrib/flink-streaming-contrib/src/main/java/org/apache/flink/contrib/streaming/DataStreamUtils.java It should work well locally for testing. In that case you

Re: Published test artifacts for flink streaming

2015-11-05 Thread Nick Dimiduk
Hi Robert, It seems "type" was what I needed. This it also looks like the test jar has an undeclared dependency. In the end, the following allowed me to use TestStreamEnvironment for my integration test. Thanks a lot! -n org.apache.flink flink-streaming-core ${flink.versio

Re: Published test artifacts for flink streaming

2015-11-05 Thread Robert Metzger
Hi Nick, we are usually publishing the test artifacts. Can you try and replace the tag by test-jar: org.apache.flink flink-streaming-core ${flink.version} test-jar test On Thu, Nov 5, 2015 at 8:18 PM, Nick Dimiduk wrote: > Hello, > > I'm attempting integration tests for my

Published test artifacts for flink streaming

2015-11-05 Thread Nick Dimiduk
Hello, I'm attempting integration tests for my streaming flows. I'd like to produce an input stream of java objects and sink the results into a collection for verification via JUnit asserts. StreamExecutionEnvironment provides methods for the former, however, how to achieve the latter is not evide

Re: IT's with POJO's

2015-11-05 Thread Nick Dimiduk
Thank you Robert, I'll give this a spin -- obvious now that you point it out. I'll go ahead and continue the line on inquiry on the JIRA. -n On Thu, Nov 5, 2015 at 4:18 AM, Robert Metzger wrote: > Hi Nick, > > both JIRA and the mailing list are good. In this case I'd say JIRA would be > better b

Re: Running on a firewalled Yarn cluster?

2015-11-05 Thread Niels Basjes
That is what I tried. Couldn't find that port though. On Thu, Nov 5, 2015 at 3:06 PM, Robert Metzger wrote: > Hi, > > cool, that's good news. > > The RM proxy is only for the web interface of the AM. > > I'm pretty sure that the MapReduce AM has at least two ports: > - one for the web interface

Re: Running on a firewalled Yarn cluster?

2015-11-05 Thread Robert Metzger
Hi, cool, that's good news. The RM proxy is only for the web interface of the AM. I'm pretty sure that the MapReduce AM has at least two ports: - one for the web interface (accessible through the RM proxy, so behind the firewall) - one for the AM RPC (and that port is allocated within the confi

Re: Running on a firewalled Yarn cluster?

2015-11-05 Thread Niels Basjes
Hi, I checked and this setting has been set to a limited port range of only 100 port numbers. I tried to find the actual port an AM is running on and couldn't find it (I'm not the admin on that cluster) The url to the AM that I use to access it always looks like this: http://master-001.xx.ne

Re: Running continuously on yarn with kerberos

2015-11-05 Thread Maximilian Michels
Thank you for looking into the problem, Niels. Let us know if you need anything. We would be happy to merge a pull request once you have verified the fix. On Thu, Nov 5, 2015 at 1:38 PM, Niels Basjes wrote: > I created https://issues.apache.org/jira/browse/FLINK-2977 > > On Thu, Nov 5, 2015 at 1

Re: Running on a firewalled Yarn cluster?

2015-11-05 Thread Robert Metzger
While discussing with my colleagues about the issue today, we came up with another approach to resolve the issue: d) Upload the job jar to HDFS (or another FS) and trigger the execution of the jar using an HTTP request to the web interface. We could add some tooling into the /bin/flink client to

Re: How best to deal with wide, structured tuples?

2015-11-05 Thread Aljoscha Krettek
Hi, these are some interesting Ideas. I have some thoughts, though, about the current implementation. 1. With Schema and Field you are basically re-implementing RowTypeInfo, so it should not be required. Maybe just an easier way to create a RowTypeInfo. 2. Right now, in Flink the TypeInformatio

Re: Running continuously on yarn with kerberos

2015-11-05 Thread Niels Basjes
I created https://issues.apache.org/jira/browse/FLINK-2977 On Thu, Nov 5, 2015 at 12:25 PM, Robert Metzger wrote: > Hi Niels, > thank you for analyzing the issue so properly. I agree with you. It seems > that HDFS and HBase are using their own tokes which we need to transfer > from the client to

Re: IT's with POJO's

2015-11-05 Thread Robert Metzger
Hi Nick, both JIRA and the mailing list are good. In this case I'd say JIRA would be better because then everybody has the full context of the discussion. The issue is fixed in 0.10, which is not yet released. You can work around the issue by implementing a custom SourceFunction which returns th

Re: Running continuously on yarn with kerberos

2015-11-05 Thread Robert Metzger
Hi Niels, thank you for analyzing the issue so properly. I agree with you. It seems that HDFS and HBase are using their own tokes which we need to transfer from the client to the YARN containers. We should be able to port the fix from Spark (which they got from Storm) into our YARN client. I think