On Thu, Jun 27, 2013 at 5:51 PM, Hugo Trippaers <h...@trippaers.nl> wrote:
>
> I think Ilya offers is great, my current stance is also to see how we can 
> bring this forward.
>
> I've had the opportunity to meet with several people at the Citrix office in 
> Santa Clara, i'm actually working from their office at this moment. I think 
> it's also the responsibility of someone who put in a -1 to work with the 
> original committer to get the situation resolved. So i'll invest the time to 
> help with the review as well.
>
> It would be great if Alex or Kelven could take the time to explain how this 
> feature has been tested. That would give the community some insight as well.
>
> My main technical problem with this merge is that stuff is moving all over 
> the place without having even the slightest idea why. Now having discussed 
> this with Alex in person i get the general idea of this merge, so can 
> actually try to review it.
>
> I think that John have nicely explained what we could do to prevent 
> situations like this in advance. I fully understand that big features or 
> rewrites don't happen overnight and might show up near the end of the release 
> cycle. With the time based release cycle it's always a risk that some feature 
> might not make it in on time. Getting more people involved and chunking the 
> commits into master will greatly speed up the reviewing process.
>
> I'll get back to this after spending some time on reviewing the actual patch. 
> In fact i would like to ask more people to have a look at this patch and 
> reply to this thread with comments or remarks.
>
> Cheers,
>
> Hugo
>

So the problem in my mind, is that we don't have a way of verifying
that master isn't broken, and won't be broken by any given merge. I
look at even the minimal level of automated testing that I see today,
and ~20% of integration tests are failing[1]. The regression set of
tests (which isn't running as often) is seeing 75% of tests
failing[2]. Heaping on more change when we are demonstrably already
failing in many places is not behaving responsibly IMO.
The question I'd pose is this - running the various automated tests is
pretty cheap - whats the output of that compared to the current test
output on master? Better or worse? If it hasn't been done, why not?
I desperately want these features, but not necessarily at the cost of
further destabilizing what we have now in master - we can't continue
accruing technical debt.

--David

[1] 
http://jenkins.buildacloud.org/view/cloudstack-qa/job/test-smoke-matrix/lastCompletedBuild/testReport/
[2] 
http://jenkins.buildacloud.org/view/cloudstack-qa/job/test-regression-matrix/28/testReport/

Reply via email to