On 2/15/25 9:54 AM, Zbigniew Jędrzejewski-Szmek wrote:
> On Fri, Feb 14, 2025 at 02:40:29PM -0800, Adam Williamson wrote:
>> On Fri, 2025-02-14 at 16:31 -0500, Dusty Mabe wrote:
>>> IMO the bar would only need to be that high if the user had no way to 
>>> ignore the test results.
>>> All gating does here (IIUC) is require them to do an extra step before it 
>>> automatically flows
>>> into the next rawhide compose.
>>
>> again, technically, yes, but *please* let's not train people to have a
>> pavlovian reaction to waive failures, that is not the way.
> 
> IMO, the bar for *gating* tests needs to be high. I think 95% true
> positives would be a reasonable threshold.

I can't promise 95% true positive rate. These aren't unit tests. They are 
system wide
tests that try to test real world scenarios as much as possible. That does mean 
pulling
things from github/quay/s3/Fedora infra/etc.. and thus flakes happen. Now, in 
our
tests we do collect failures and Retry them. If a retry succeeds we take it as 
success
and never report the failure at all. However there are parts of our pipeline 
that might
not be so good at retrying.

All I'm trying to say is that when you don't control everything it's hard to 
say with
confidence something will be 95%.

As I promised before, maybe just work with us on it. These tests have been 
enabled for
a while and I've only seen a handful of package maintainers look at the 
failures (you,
Zbyszek, being one of them; thank you!).

We do want them to be useful tests and I promise when a failure happens because 
of our
infra or tests themselves being flakey we try to get it fixed.

TANGENT: Maybe there's a bodhi feature request in here somewhere where there is 
a concept
         of soft gating. i.e. the update gets held for one cycle if the test 
fails, but
         if no other action is taken (i.e. negative karma or something) then it 
will go
         into the next cycle update. Could be a good middle ground.

> 
> I know this is hard to achieve, but gating failures are very disruptive
> to packagers. And false positives are even more disruptime: a person
> not intimately familiar with the CI needs to spend a considerable
> amount of time jumping through not-very-easy-to-read pages.
> *If* I can be almost certain that the report is real and with enough
> digging I'll figure out a bug, I'm ready to spend the time. But if
> there's a considerable chance of a false positive _and_ the effort to
> figure out if the report is real is also quite high, I'll just learn
> to waive any failures.
> 
> --
> 
> I cannot find it right now, but I hope it's still there somewhere…
> Bodhi updates have (had?) a message that said someting like "for failing
> update.* tests, contact <this channel>, for failing fedora-ci.* tests contact
> <that channel>". Is something like that planned for the coreos.* tests?
> 
> Zbyszek

-- 
_______________________________________________
devel mailing list -- devel@lists.fedoraproject.org
To unsubscribe send an email to devel-le...@lists.fedoraproject.org
Fedora Code of Conduct: 
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: 
https://lists.fedoraproject.org/archives/list/devel@lists.fedoraproject.org
Do not reply to spam, report it: 
https://pagure.io/fedora-infrastructure/new_issue

Reply via email to