On Mon, Dec 2, 2013 at 1:55 PM, Maru Newby <ma...@redhat.com> wrote: > > On Dec 2, 2013, at 10:19 PM, Joe Gordon <joe.gord...@gmail.com> wrote: > > > > > On Dec 2, 2013 3:39 AM, "Maru Newby" <ma...@redhat.com> wrote: > > > > > > > > > On Dec 2, 2013, at 2:07 AM, Anita Kuno <ante...@anteaya.info> wrote: > > > > > > > Great initiative putting this plan together, Maru. Thanks for doing > > > > this. Thanks for volunteering to help, Salvatore (I'm thinking of > asking > > > > for you to be cloned - once that becomes available.) if you add your > > > > patch urls (as you create them) to the blueprint Maru started [0] > that > > > > would help to track the work. > > > > > > > > Armando, thanks for doing this work as well. Could you add the urls > of > > > > the patches you reference to the exceptional-conditions blueprint? > > > > > > > > For icehouse-1 to be a realistic goal for this assessment and > clean-up, > > > > patches for this would need to be up by Tuesday Dec. 3 at the latest > > > > (does 13:00 UTC sound like a reasonable target?) so that they can > make > > > > it through review and check testing, gate testing and merging prior > to > > > > the Thursday Dec. 5 deadline for icehouse-1. I would really like to > see > > > > this, I just want the timeline to be conscious. > > > > > > My mistake, getting this done by Tuesday does not seem realistic. > icehouse-2, then. > > > > > > > With icehouse-2 being the nova-network feature freeze reevaluation point > (possibly lifting it) I think gating on new stacktraces by icehouse-2 is > too late. Even a huge whitelist of errors is better then letting new > errors in. > > No question that it needs to happen asap. If we're talking about > milestones, though, and icehouse-1 patches need to be in by Tuesday, I > don't think icehouse-1 is realistic. It will have to be early in > icehouse-2. > > Yup, thanks for the clarification.
> > m. > > > > > > > m. > > > > > > > > > > > I would like to say talk to me tomorrow in -neutron to ensure you are > > > > getting the support you need to achieve this but I will be flying > (wifi > > > > uncertain). I do hope that some additional individuals come forward > to > > > > help with this. > > > > > > > > Thanks Maru, Salvatore and Armando, > > > > Anita. > > > > > > > > [0] > > > > > https://blueprints.launchpad.net/neutron/+spec/log-only-exceptional-conditions-as-error > > > > > > > > On 11/30/2013 08:24 PM, Maru Newby wrote: > > > >> > > > >> On Nov 28, 2013, at 1:08 AM, Salvatore Orlando <sorla...@nicira.com> > wrote: > > > >> > > > >>> Thanks Maru, > > > >>> > > > >>> This is something my team had on the backlog for a while. > > > >>> I will push some patches to contribute towards this effort in the > next few days. > > > >>> > > > >>> Let me know if you're already thinking of targeting the completion > of this job for a specific deadline. > > > >> > > > >> I'm thinking this could be a task for those not involved in fixing > race conditions, and be done in parallel. I guess that would be for > icehouse-1 then? My hope would be that the early signs of race conditions > would then be caught earlier. > > > >> > > > >> > > > >> m. > > > >> > > > >>> > > > >>> Salvatore > > > >>> > > > >>> > > > >>> On 27 November 2013 17:50, Maru Newby <ma...@redhat.com> wrote: > > > >>> Just a heads up, the console output for neutron gate jobs is about > to get a lot noisier. Any log output that contains 'ERROR' is going to be > dumped into the console output so that we can identify and eliminate > unnecessary error logging. Once we've cleaned things up, the presence of > unexpected (non-whitelisted) error output can be used to fail jobs, as per > the following Tempest blueprint: > > > >>> > > > >>> > https://blueprints.launchpad.net/tempest/+spec/fail-gate-on-log-errors > > > >>> > > > >>> I've filed a related Neutron blueprint for eliminating the > unnecessary error logging: > > > >>> > > > >>> > https://blueprints.launchpad.net/neutron/+spec/log-only-exceptional-conditions-as-error > > > >>> > > > >>> I'm looking for volunteers to help with this effort, please reply > in this thread if you're willing to assist. > > > >>> > > > >>> Thanks, > > > >>> > > > >>> > > > >>> Maru > > > >>> _______________________________________________ > > > >>> OpenStack-dev mailing list > > > >>> OpenStack-dev@lists.openstack.org > > > >>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev > > > >>> > > > >>> _______________________________________________ > > > >>> OpenStack-dev mailing list > > > >>> OpenStack-dev@lists.openstack.org > > > >>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev > > > >> > > > >> > > > >> _______________________________________________ > > > >> OpenStack-dev mailing list > > > >> OpenStack-dev@lists.openstack.org > > > >> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev > > > >> > > > > > > > > > > > > _______________________________________________ > > > > OpenStack-dev mailing list > > > > OpenStack-dev@lists.openstack.org > > > > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev > > > > > > > > > _______________________________________________ > > > OpenStack-dev mailing list > > > OpenStack-dev@lists.openstack.org > > > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev > > _______________________________________________ > > OpenStack-dev mailing list > > OpenStack-dev@lists.openstack.org > > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev > > > _______________________________________________ > OpenStack-dev mailing list > OpenStack-dev@lists.openstack.org > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev >
_______________________________________________ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev