Thanks, Stephan.

Everything is back to normal for us.

Cheers,

Cory

On Fri, Feb 12, 2016 at 6:54 AM, Stephan Ewen <se...@apache.org> wrote:

> Hi Cory!
>
> We found the problem. There is a development fork of Flink for Stream SQL,
> whose CI infrastructure accidentally also deployed snapshots and overwrote
> some of the proper master branch snapshots.
>
> That's why the snapshots got inconsistent. We fixed that, and newer
> snapshots should be online.
> Hope that this is resolved now.
>
> Sorry for the inconvenience,
> Stephan
>
>
> On Fri, Feb 12, 2016 at 12:51 AM, Stephan Ewen <se...@apache.org> wrote:
>
>> Hi!
>>
>> The CI system has just finished uploading an new snapshot. In that one,
>> the scalatest dependency is now correctly at 2.11 again.
>>
>>
>> https://repository.apache.org/content/repositories/snapshots/org/apache/flink/flink-test-utils_2.11/1.0-SNAPSHOT/flink-test-utils_2.11-1.0-20160211.232156-288.pom
>>
>> I am very puzzled, we did not touch any parts that seem to affect this. I
>> am wondering if it is possible that Maven had a hiccup...
>>
>> Can you retry (force dependency update), see if the dependencies are
>> correct again?
>>
>>
>> Greetings,
>> Stephan
>>
>>
>> On Fri, Feb 12, 2016 at 12:23 AM, Stephan Ewen <se...@apache.org> wrote:
>>
>>> Hi!
>>>
>>> I examined the Apache Snapshot Repository, and I could see that in the
>>> latest snapshot a "scalatest_2.10" version was introduced. I could not
>>> figure out how, yet. I could not find a "flink-core_2.10" or
>>> "flink-annotations_2.10" dependency, yet.
>>>
>>>
>>> Previous snapshot:
>>> https://repository.apache.org/content/repositories/snapshots/org/apache/flink/flink-test-utils_2.11/1.0-SNAPSHOT/flink-test-utils_2.11-1.0-20160211.162913-286.pom
>>>
>>> Latest Snapshot:
>>> https://repository.apache.org/content/repositories/snapshots/org/apache/flink/flink-test-utils_2.11/1.0-SNAPSHOT/flink-test-utils_2.11-1.0-20160211.201205-287.pom
>>>
>>>
>>> We'll try and fix this ASAP. Sorry for that, this is quite a mystery
>>> right now...
>>>
>>> Best,
>>> Stephan
>>>
>>> On Thu, Feb 11, 2016 at 11:56 PM, Cory Monty <
>>> cory.mo...@getbraintree.com> wrote:
>>>
>>>> Ufuk,
>>>>
>>>> Thanks for the link. I've double-checked everything in our dependencies
>>>> list and it's all correct.
>>>>
>>>> Stephan,
>>>>
>>>> We don't explicitly depend on "flink-java", so there should be no
>>>> suffix. It's curious, to me, that scalatest is showing in the stack trace.
>>>> I also tried clearing ~/.sbt/staging and it did not help. Our build server
>>>> (CircleCI) is also experiencing the same issue, so I don't think it's local
>>>> to my machine.
>>>>
>>>> On Thu, Feb 11, 2016 at 4:09 PM, Stephan Ewen <se...@apache.org> wrote:
>>>>
>>>>> Hi Cory!
>>>>>
>>>>> Hmmm, curios... I just double check the code, there are no more
>>>>> references to a Scala-versioned "flink-core" and "flink-annotations"
>>>>> project in the code base.
>>>>>
>>>>> The projects you use with Scala version suffix look good, actually.
>>>>> Just to be safe, can you check that the "flink-java" dependency is without
>>>>> suffix?
>>>>>
>>>>> One other thing I can imagine is a mixed up dependency cache. Can you
>>>>> try to refresh all snapshot dependencies (maybe clear "~/.sbt/staging/").
>>>>>
>>>>>
>>>>> It is high-time for a 1.0 release, so you need not work on the
>>>>> SNAPSHOT versions any more. That should really solve this version conflict
>>>>> pain.
>>>>> If we are fast tomorrow, there may be a nice surprise coming up in the
>>>>> next days...
>>>>>
>>>>> Greetings,
>>>>> Stephan
>>>>>
>>>>>
>>>>> On Thu, Feb 11, 2016 at 10:24 PM, Cory Monty <
>>>>> cory.mo...@getbraintree.com> wrote:
>>>>>
>>>>>> Hmm. We don't explicitly include "flink-annotations" and we do not
>>>>>> append the Scala suffix for "flink-core":
>>>>>>
>>>>>> `"org.apache.flink" % "flink-core" % "1.0-SNAPSHOT"`
>>>>>>
>>>>>> Here are the packages we currently include with a Scala suffix:
>>>>>>
>>>>>> flink-scala
>>>>>> flink-clients
>>>>>> flink-streaming-scala
>>>>>> flink-connector-kafka-0.8
>>>>>> flink-test-utils
>>>>>> flink-streaming-contrib
>>>>>>
>>>>>> If there is any documentation you can point to regarding when to
>>>>>> include the Scala suffix on Flink packages, let me know.
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Thu, Feb 11, 2016 at 2:55 PM, Stephan Ewen <se...@apache.org>
>>>>>> wrote:
>>>>>>
>>>>>>> Hi Cory!
>>>>>>>
>>>>>>> "flink-core" and "flink-annotations" should not have Scala suffixes,
>>>>>>> because they do not depend on Scala.
>>>>>>>
>>>>>>> So far, we mark the Scala independent projects without suffixes. Is
>>>>>>> that very confusing, or does that interfere with build tools?
>>>>>>>
>>>>>>> Greetings,
>>>>>>> Stephan
>>>>>>>
>>>>>>>
>>>>>>> On Thu, Feb 11, 2016 at 9:50 PM, Cory Monty <
>>>>>>> cory.mo...@getbraintree.com> wrote:
>>>>>>>
>>>>>>>> As of this afternoon, SBT is running into issues compiling with the
>>>>>>>> following error:
>>>>>>>>
>>>>>>>> [error] Modules were resolved with conflicting cross-version
>>>>>>>> suffixes in
>>>>>>>> [error]    org.scalatest:scalatest _2.10, _2.11
>>>>>>>> [error]    org.apache.flink:flink-core _2.11, <none>
>>>>>>>> [error]    org.apache.flink:flink-annotations _2.11, <none>
>>>>>>>> java.lang.RuntimeException: Conflicting cross-version suffixes in:
>>>>>>>> org.scalatest:scalatest, org.apache.flink:flink-core,
>>>>>>>> org.apache.flink:flink-annotations
>>>>>>>> at scala.sys.package$.error(package.scala:27)
>>>>>>>> at
>>>>>>>> sbt.ConflictWarning$.processCrossVersioned(ConflictWarning.scala:46)
>>>>>>>> at sbt.ConflictWarning$.apply(ConflictWarning.scala:32)
>>>>>>>> at sbt.Classpaths$$anonfun$66.apply(Defaults.scala:1164)
>>>>>>>> at sbt.Classpaths$$anonfun$66.apply(Defaults.scala:1161)
>>>>>>>> at scala.Function1$$anonfun$compose$1.apply(Function1.scala:47)
>>>>>>>> at
>>>>>>>> sbt.$tilde$greater$$anonfun$$u2219$1.apply(TypeFunctions.scala:40)
>>>>>>>> at sbt.std.Transform$$anon$4.work(System.scala:63)
>>>>>>>> at
>>>>>>>> sbt.Execute$$anonfun$submit$1$$anonfun$apply$1.apply(Execute.scala:226)
>>>>>>>> at
>>>>>>>> sbt.Execute$$anonfun$submit$1$$anonfun$apply$1.apply(Execute.scala:226)
>>>>>>>> at sbt.ErrorHandling$.wideConvert(ErrorHandling.scala:17)
>>>>>>>> at sbt.Execute.work(Execute.scala:235)
>>>>>>>> at sbt.Execute$$anonfun$submit$1.apply(Execute.scala:226)
>>>>>>>> at sbt.Execute$$anonfun$submit$1.apply(Execute.scala:226)
>>>>>>>> at
>>>>>>>> sbt.ConcurrentRestrictions$$anon$4$$anonfun$1.apply(ConcurrentRestrictions.scala:159)
>>>>>>>> at sbt.CompletionService$$anon$2.call(CompletionService.scala:28)
>>>>>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>>>>>>>> at
>>>>>>>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>>>>>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>>>>>>>> at
>>>>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>>>>>>>> at
>>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>>>>>>>> at java.lang.Thread.run(Thread.java:745)
>>>>>>>>
>>>>>>>> Any thoughts are greatly appreciated!
>>>>>>>>
>>>>>>>> Cheers,
>>>>>>>>
>>>>>>>> Cory
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to