Hi Gwen, I don't see any problem with this as long as we're convinced there's a good use case, which seems to be true.
Cheers, Geoff On Thu, Jun 4, 2015 at 5:20 PM, Gwen Shapira <gshap...@cloudera.com> wrote: > Not completely random places :) > People may use Cloudera / HWX distributions which include Kafka, but want > to verify that these bits match a specific upstream release. > > I think having the tests separately will be useful for this. In this case, > finding the tests are not a big issue - we'll add a download link :) > > On Thu, Jun 4, 2015 at 5:00 PM, Jiangjie Qin <j...@linkedin.com.invalid> > wrote: > > > Hey Gwen, > > > > Currently the test and code are downloaded at the same time. Supposedly > > the tests in the same repository should cover match the code. > > Are you saying people downloaded a release from some random place and > want > > to verify it? If that is the case, does that mean people still need to > > find the correct place to download the right test artifact? > > > > Thanks, > > > > Jiangjie (Becket) Qin > > > > > > > > On 6/4/15, 4:29 PM, "Gwen Shapira" <gshap...@cloudera.com> wrote: > > > > >Hi, > > > > > >Reviving the discussion a bit :) > > > > > >I think it will be nice if each Kafka version that we release will also > > >have a separate "tests" artifact that users can download, untar and > easily > > >run against a Kafka cluster of the same version. > > > > > >The idea is that if someone downloads packages that claim to contain > > >something of a specific Kafka version (i.e. Kafka 0.8.2.0 + patches), > > >users > > >can easily download the tests and verify that it indeed passes the tests > > >for this version and therefore behaves the way this version is expected > to > > >behave. > > > > > >Does it make sense? > > > > > >Gwen > > > > > >On Thu, May 21, 2015 at 3:26 PM, Geoffrey Anderson <ge...@confluent.io> > > >wrote: > > > > > >> Hi Ashish, > > >> > > >> Looks like Ewen already hit the main points, but a few additions: > > >> > > >> 1. ducktape repo is here: https://github.com/confluentinc/ducktape > > >> ducktape itself will be pip installable in the near future, and Kafka > > >> system tests will be able to depend on a particular version of > ducktape. > > >> > > >> 2. The reporting is nothing fancy. We're definitely open to feedback, > > >>but > > >> it consists of: > > >> - top level summary of the test run (simple PASS/FAIL for each test) > > >> - top level info and debug logs > > >> - per-test info and debug logs > > >> - per-test "service" logs gathered from each service used in the test. > > >>For > > >> example, if your test pulls up a Kafka cluster with 5 brokers, the end > > >> result will have the Kafka logs from each of those 5 machines. > > >> > > >> Cheers, > > >> Geoff > > >> > > >> On Thu, May 21, 2015 at 3:15 PM, Ewen Cheslack-Postava > > >><e...@confluent.io> > > >> wrote: > > >> > > >> > Ashish, > > >> > > > >> > 1. That was the plan. We put some effort into cleanly separating the > > >> > framework so it would be reusable across many projects. > > >> > 2. I think you're seeing a test in progress where the final report > > >>hasn't > > >> > been created yet. If you visit one of the older ones you'll see it > > >>has a > > >> > landing page with links: > > >> > http://testing.confluent.io/confluent_platform/2015-05-20--001/ > > >> Apparently > > >> > we need to adjust when we update the 'latest' symlink. The logs that > > >>are > > >> > collected for tests are configurable, and service implementations > > >>include > > >> > sane defaults (so, e.g., you will always get the normal log file for > > >> Kafka, > > >> > but only get the data files if the test asks for them). > > >> > 3. No code coverage support. Haven't looked into it, so I couldn't > > >> comment > > >> > on how hard it would be to add. > > >> > > > >> > -Ewen > > >> > > > >> > On Thu, May 21, 2015 at 2:38 PM, Ashish Singh <asi...@cloudera.com> > > >> wrote: > > >> > > > >> > > Geoffrey, > > >> > > > > >> > > This looks great! > > >> > > > > >> > > A few questions. > > >> > > 1. Will ducktape be maintained separately as a github repo? > > >> > > 2. How easy is viewing the test results and logs. The link in KIP, > > >> > > http://testing.confluent.io/confluent_platform/latest/, lists a > > >>bunch > > >> of > > >> > > files and dirs. Could you add to KIP how the result and logs for > the > > >> > tests > > >> > > will be organized. > > >> > > 3. Does it support code coverage? If not, how easy/ difficult > would > > >>it > > >> > be? > > >> > > > > >> > > On Thu, May 21, 2015 at 2:03 PM, Geoffrey Anderson > > >><ge...@confluent.io > > >> > > > >> > > wrote: > > >> > > > > >> > > > Great, I'll work on putting together a more detailed map of this > > >> > > > replacement process. > > >> > > > > > >> > > > On Thu, May 21, 2015 at 11:13 AM, Gwen Shapira < > > >> gshap...@cloudera.com> > > >> > > > wrote: > > >> > > > > > >> > > > > Love this idea :) > > >> > > > > > > >> > > > > I took a look at Ducktape API and it looks like a good fit - > > >>clean > > >> > API, > > >> > > > > extensible, easy to use and powerful enough for our use-case. > > >> > > > > > > >> > > > > Something I'd like to see as part of the KIP is a map of what > > >> > > system-test > > >> > > > > currently tests, which ones we want to replace and a JIRA for > > >> > replacing > > >> > > > > (possibly one for each group of tests). > > >> > > > > Basically, I know we all want to use the new system for new > test > > >> > cases > > >> > > > > (upgrades, etc), but I really want to make sure we don't get > > >>stuck > > >> > with > > >> > > > > both systems forever. > > >> > > > > > > >> > > > > Gwen > > >> > > > > > > >> > > > > On Thu, May 21, 2015 at 9:01 PM, Geoffrey Anderson < > > >> > ge...@confluent.io > > >> > > > > > >> > > > > wrote: > > >> > > > > > > >> > > > > > Hi, > > >> > > > > > > > >> > > > > > Just kicking off the discussion thread on KIP-25 > > >> > > > > > > > >> > > > > > > > >> > > > > > > > >> > > > > > > >> > > > > > >> > > > > >> > > > >> > > >> > > > https://cwiki.apache.org/confluence/display/KAFKA/KIP+-+25+System+test+im > > >>provements > > >> > > > > > > > >> > > > > > Thanks, > > >> > > > > > Geoff > > >> > > > > > > > >> > > > > > > >> > > > > > >> > > > > >> > > > > >> > > > > >> > > -- > > >> > > > > >> > > Regards, > > >> > > Ashish > > >> > > > > >> > > > >> > > > >> > > > >> > -- > > >> > Thanks, > > >> > Ewen > > >> > > > >> > > > > >