On 2013-10-08 05:03, Robert Collins wrote:
On 8 October 2013 22:44, Martyn Taylor <mtay...@redhat.com> wrote:
On 07/10/13 20:03, Robert Collins wrote:


Whilst I can see that deciding on who is Core is a difficult task, I do feel
that creating a competitive environment based on no. reviews will be
detrimental to the project.

I'm not sure how it's competitive : I'd be delighted if every
contributor was also a -core reviewer: I'm not setting, nor do I think
we need to think about setting (at this point anyhow), a cap on the
number of reviewers.

I do feel this is going to result in quantity over quality. Personally, I'd like to see every commit properly reviewed and tested before getting a vote
and I don't think these stats are promoting that.

I think thats a valid concern. However Nova has been running a (very
slightly less mechanical) form of this for well over a year, and they
are not drowning in -core reviewers. yes, reviewing is hard, and folk
should take it seriously.

Do you have an alternative mechanism to propose? The key things for me are:
 - folk who are idling are recognised as such and gc'd around about
the time their growing staleness will become an issue with review
correctness
 - folk who have been putting in consistent reading of code + changes
get given the additional responsibility of -core around about the time
that they will know enough about whats going on to review effectively.

This is a discussion that has come up in the other projects (not surprisingly), and I thought I would mention some of the criteria that are being used in those projects. The first, and simplest, is from Dolph Mathews:

'Ultimately, "core contributor" to me simply means that this person's downvotes on code reviews are consistently well thought out and meaningful, such that an upvote by the same person shows a lot of confidence in the patch.'

I personally like this definition because it requires a certain volume of review work (which benefits the project), but it also takes into account the quality of those reviews. Obviously both are important. Note that the +/- and disagreements columns in Russell's stats are intended to help with determining review quality. Nothing can replace the judgment of the current cores of course, but if someone has been +1'ing in 95% of their reviews it's probably a sign that they aren't doing quality reviews. Likewise if they're -1'ing everything but are constantly disagreeing with cores.

An expanded version of that can be found in this post to the list: http://lists.openstack.org/pipermail/openstack-dev/2013-June/009876.html

To me, that is along the same lines as what Dolph said, just a bit more specific as to how "quality" should be demonstrated and measured.

Hope this is helpful.

-Ben

_______________________________________________
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

Reply via email to