This almost worked. Of the 5 build jobs four passed and one failed.
What's the acceptance criteria for a pull request? Do I need to build again
to get all 5 build jobs passing?

Thanks,
Lokesh

On Thu, May 14, 2015 at 8:50 AM, Robert Metzger <rmetz...@apache.org> wrote:

> No, you don't have to wait.
> The KafkaITCase is not always failing. If you're lucky, it will pass with
> the next run.
>
> On Thu, May 14, 2015 at 5:48 PM, Lokesh Rajaram <rajaram.lok...@gmail.com>
> wrote:
>
> > If I understand it correct, I have to wait for your pull request to be
> > merged, I can rebase and trigger build again. is that right?
> >
> > Thanks Robert, Aljoscha for super fast reply/help.
> >
> > Thanks,
> > Lokesh
> >
> > On Thu, May 14, 2015 at 8:39 AM, Robert Metzger <rmetz...@apache.org>
> > wrote:
> >
> > > However, you can only restart runs in your travis account, not on the
> > > apache account (also used for validating pull requests).
> > >
> > > I have opened a pull request a few minutes ago which will reduce the
> > number
> > > of KafakITCase failures (there is still one other unresolved issue).
> > >
> > > On Thu, May 14, 2015 at 5:37 PM, Aljoscha Krettek <aljos...@apache.org
> >
> > > wrote:
> > >
> > > > Hi,
> > > > don't worry, there are very few stupid questions. :D
> > > >
> > > > The KafkaITCase sometimes fails on Travis, this is a known problem
> > > > currently. On travis you can restart the individual runs for a commit
> > > > in the view of the failed run.
> > > >
> > > > Hope that helps.
> > > >
> > > > Cheers,
> > > > Aljoscha
> > > >
> > > > On Thu, May 14, 2015 at 5:35 PM, Lokesh Rajaram
> > > > <rajaram.lok...@gmail.com> wrote:
> > > > > Thanks Aljoscha, Robert. After adding guava dependency for
> > > flink-spargel
> > > > I
> > > > > was able to progress further but now it's failing in
> > > > > flink-streaming-connectors for the following test case:
> > > > >
> > > > > KafkaITCase.brokerFailureTest:936->tryExecute:352 Test failed with:
> > Job
> > > > > execution failed.
> > > > >
> > > > > Any pointers would help me proceed further. Sorry for a lot of
> > trivial
> > > > > questions, I am just getting started not familiar with the code
> base.
> > > > > I tried running locally, I am able to run it successfully, don't
> know
> > > why
> > > > > it's only failing in Travis build. Not sure if I am missing
> something
> > > in
> > > > my
> > > > > local environment.
> > > > >
> > > > > Thanks,
> > > > > Lokesh
> > > > >
> > > > > On Thu, May 14, 2015 at 1:39 AM, Robert Metzger <
> rmetz...@apache.org
> > >
> > > > wrote:
> > > > >
> > > > >> I think flink-spargel is missing the guava dependency.
> > > > >>
> > > > >> On Thu, May 14, 2015 at 8:18 AM, Aljoscha Krettek <
> > > aljos...@apache.org>
> > > > >> wrote:
> > > > >>
> > > > >> > @Robert, this seems like a problem with the Shading?
> > > > >> >
> > > > >> > On Thu, May 14, 2015 at 5:41 AM, Lokesh Rajaram
> > > > >> > <rajaram.lok...@gmail.com> wrote:
> > > > >> > > Thanks Aljioscha. I was able to change as recommended and able
> > to
> > > > run
> > > > >> the
> > > > >> > > entire test suite in local successfully.
> > > > >> > > However Travis build is failing for pull request:
> > > > >> > > https://github.com/apache/flink/pull/673.
> > > > >> > >
> > > > >> > > It's a compilation failure:
> > > > >> > >
> > > > >> > > [ERROR] Failed to execute goal
> > > > >> > > org.apache.maven.plugins:maven-compiler-plugin:3.1:compile
> > > > >> > > (default-compile) on project flink-spargel: Compilation
> failure:
> > > > >> > > Compilation failure:
> > > > >> > > [ERROR]
> > > > >> > >
> > > > >> >
> > > > >>
> > > >
> > >
> >
> /home/travis/build/apache/flink/flink-staging/flink-spargel/src/main/java/org/apache/flink/spargel/java/VertexCentricIteration.java:[42,30]
> > > > >> > > package com.google.common.base does not exist
> > > > >> > >
> > > > >> > > I can definitely see the package imported in the class,
> > compiling
> > > > and
> > > > >> > > passing all tests in local.
> > > > >> > > Anything I am missing here?
> > > > >> > >
> > > > >> > > Thanks,
> > > > >> > > Lokesh
> > > > >> > >
> > > > >> > > On Mon, May 11, 2015 at 1:25 AM, Aljoscha Krettek <
> > > > aljos...@apache.org
> > > > >> >
> > > > >> > > wrote:
> > > > >> > >
> > > > >> > >> I think you can replace Validate.NotNull(p) with require(p !=
> > > > null, "p
> > > > >> > >> is null (or something like this)").
> > > > >> > >>
> > > > >> > >> On Mon, May 11, 2015 at 12:27 AM, Lokesh Rajaram
> > > > >> > >> <rajaram.lok...@gmail.com> wrote:
> > > > >> > >> > 1. I think I can use require for replacing Validate.isTrue
> > > > >> > >> > 2. What about Validate.notNull? If require is used it would
> > > throw
> > > > >> > >> > IllegalArgumentException,
> > > > >> > >> > if assume or assert is used it would throw AssertionError
> > which
> > > > is
> > > > >> not
> > > > >> > >> > compatible with current implementation.
> > > > >> > >> >
> > > > >> > >> > Please let me know if my understanding is correct. Also,
> let
> > me
> > > > know
> > > > >> > your
> > > > >> > >> > thoughts.
> > > > >> > >> >
> > > > >> > >> > Thanks,
> > > > >> > >> > Lokesh
> > > > >> > >> >
> > > > >> > >> > On Sun, May 10, 2015 at 1:04 AM, Aljoscha Krettek <
> > > > >> > aljos...@apache.org>
> > > > >> > >> > wrote:
> > > > >> > >> >
> > > > >> > >> >> I would propose using the methods as Chiwan suggested. If
> > > > everyone
> > > > >> > >> >> agrees I can change the Jira issue.
> > > > >> > >> >>
> > > > >> > >> >> On Sun, May 10, 2015 at 6:47 AM, Lokesh Rajaram
> > > > >> > >> >> <rajaram.lok...@gmail.com> wrote:
> > > > >> > >> >> > Thank you for the reference links. Which approach
> should I
> > > > take,
> > > > >> > >> casting
> > > > >> > >> >> or
> > > > >> > >> >> > use scala methods.
> > > > >> > >> >> > If it's the latter option will the JIRA ticket
> FLINK-1711
> > > > >> > >> >> > <https://issues.apache.org/jira/browse/FLINK-1711> be
> > > > updated to
> > > > >> > >> >> reflect it?
> > > > >> > >> >> >
> > > > >> > >> >> > Thanks,
> > > > >> > >> >> > Lokesh
> > > > >> > >> >> >
> > > > >> > >> >> > On Sat, May 9, 2015 at 8:16 PM, Chiwan Park <
> > > > >> chiwanp...@icloud.com
> > > > >> > >
> > > > >> > >> >> wrote:
> > > > >> > >> >> >
> > > > >> > >> >> >> Hi. There is some problems using Guava’s check method
> in
> > > > Scala.
> > > > >> (
> > > > >> > >> >> >>
> > > > >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k
> > > > >> > <
> > > > >> > >> >> >>
> > > > >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k
> > > > >> > >)
> > > > >> > >> You
> > > > >> > >> >> >> can solve this error simply with casting last argument
> to
> > > > >> > >> >> java.lang.Object.
> > > > >> > >> >> >> But I think we’d better use `require`, `assume`,
> `assert`
> > > > method
> > > > >> > >> >> provided
> > > > >> > >> >> >> by Scala. (
> > > > >> > >> >> >>
> > > > >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html
> > > > >> > <
> > > > >> > >> >> >>
> > > > >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html
> > > > >> > >)
> > > > >> > >> >> >> Because this changes affects many other codes, so we
> > should
> > > > >> > discuss
> > > > >> > >> >> about
> > > > >> > >> >> >> changing Guava's method to Scala’s method.
> > > > >> > >> >> >>
> > > > >> > >> >> >> Regards.
> > > > >> > >> >> >> Chiwan Park (Sent with iPhone)
> > > > >> > >> >> >>
> > > > >> > >> >> >>
> > > > >> > >> >> >>
> > > > >> > >> >> >> > On May 10, 2015, at 11:49 AM, Lokesh Rajaram <
> > > > >> > >> >> rajaram.lok...@gmail.com>
> > > > >> > >> >> >> wrote:
> > > > >> > >> >> >> >
> > > > >> > >> >> >> > Hello All,
> > > > >> > >> >> >> >
> > > > >> > >> >> >> > I am new to Flink community and am very excited about
> > the
> > > > >> > project
> > > > >> > >> and
> > > > >> > >> >> >> work
> > > > >> > >> >> >> > you all have been doing. Kudos!!
> > > > >> > >> >> >> >
> > > > >> > >> >> >> > I was looking to pickup some starter task. Robert
> > > > recommended
> > > > >> to
> > > > >> > >> pick
> > > > >> > >> >> up
> > > > >> > >> >> >> > https://issues.apache.org/jira/browse/FLINK-1711.
> > Thanks
> > > > >> Robert
> > > > >> > >> for
> > > > >> > >> >> your
> > > > >> > >> >> >> > guidance.
> > > > >> > >> >> >> >
> > > > >> > >> >> >> > Sorry for a dumb question. I am done with code
> changes
> > > but
> > > > my
> > > > >> > "mvn
> > > > >> > >> >> >> verify"
> > > > >> > >> >> >> > failing only for the scala module as follows
> > > > >> > >> >> >> >
> > > > >> > >> >> >> >
> > > > >> > >> >> >>
> > > > >> > >> >>
> > > > >> > >>
> > > > >> >
> > > > >>
> > > >
> > >
> >
> flink/flink-scala/src/main/scala/org/apache/flink/api/scala/joinDataSet.scala:77:
> > > > >> > >> >> >> > error: ambiguous reference to overloaded definition,
> > > > >> > >> >> >> > [ERROR] both method checkNotNull in object
> > Preconditions
> > > of
> > > > >> type
> > > > >> > >> >> [T](x$1:
> > > > >> > >> >> >> > T, x$2: String, x$3: <repeated...>[Object])T
> > > > >> > >> >> >> > [ERROR] and  method checkNotNull in object
> > Preconditions
> > > of
> > > > >> type
> > > > >> > >> >> [T](x$1:
> > > > >> > >> >> >> > T, x$2: Any)T
> > > > >> > >> >> >> > [ERROR] match argument types ((L, R) => O,String)
> > > > >> > >> >> >> > [ERROR]     Preconditions.checkNotNull(fun, "Join
> > > function
> > > > >> must
> > > > >> > >> not be
> > > > >> > >> >> >> > null.")
> > > > >> > >> >> >> >
> > > > >> > >> >> >> > Same error I see for all of the Scala classes I
> > changed.
> > > > Any
> > > > >> > >> pointers
> > > > >> > >> >> >> here
> > > > >> > >> >> >> > will be very helpful for me to proceed further.
> Please
> > > let
> > > > me
> > > > >> > know
> > > > >> > >> if
> > > > >> > >> >> you
> > > > >> > >> >> >> > need more information.
> > > > >> > >> >> >> >
> > > > >> > >> >> >> > Thanks in advance for your help and support.
> > > > >> > >> >> >> >
> > > > >> > >> >> >> > Thanks,
> > > > >> > >> >> >> > Lokesh
> > > > >> > >> >> >>
> > > > >> > >> >> >>
> > > > >> > >> >>
> > > > >> > >>
> > > > >> >
> > > > >>
> > > >
> > >
> >
>

Reply via email to