On Mon, Nov 28, 2016 at 10:24:43AM +0100, Kornel Benko wrote: > You could make an alias ...
True. I didn't actually mean the command itself. > > When I add a pattern to invertedTests, it does not affect the unreliable > > tests. Can we change this? > > We had some discussions about this. The current logic is: > > ignored: we don't consider this test case > unreliable: we do not trust the result of the test > inverted: we know, the test fails > suspended: test fails, but we don't care for now Thanks for the summary. Let's dig deeper into the unreliable category: Sublabel: nonstandard These tests are marked as unreliable because of non-standard dependencies. I have those dependencies installed (or at least many of them, such as knitr, rjournal.sty, Farsi, aa.cls, iucr.cls, acmsiggraph.cls). So from my perspective I would like those test to be treated as normal ctests when I run the ctests (although I don't mind the label "unreliable"). I understand that other users of ctest don't want to pay attention to whether they fail or not, but they don't need to. Sublabel: varying_versions The tests in this category fail for some versions (e.g. of a LaTeX class) and pass for others. I propose a simple rule by which we set the test to fail/pass depending on the updated version of the latest TeX Live release (or if the dependency is not in TeX Live, then the latest version released). Sublabel: erratic To me this sublabel contains the most unreliable tests. These tests could depend on the phase of the moon or time of day. I would almost suggest a new label for them (or a new label for the other sublabels). Actually, I might just suggest these tests be ignored. (Note that I'm not actually convinced that the only test in this label should actually be labeled an erratic test. If fails every time for me.) Sublabel: wrong_output These tests do not fail but we would like them to. I want to know when these tests go from passing to failing. Then they could potentially be moved to inverted or LyX bugs or whatever the underlying problem for them not displaying correctly is. Occassionally we should audit them to see whether the output is now correct (e.g. with a fix in TeX Live). The reason that I want the output of the "unreliable" tests to be clean is because I want to have the choice of whether to pay attention to changing tests. Currently I have to manually compare one large list to another. This is quite annoying. If an unreliable test changes status (whether going from failing to passing or passing to failing) I would like to easily see this, so I can decide whether I think that information is useful, and whether I want to spend my time to act on it. If I were to report a "regression" because an uninverted test went from passing to failing, the burden would be on me to argue why I think this is actually a true regression, and not just a property of the unreliableness of the tests. > > Attached is the list of unreliable tests that I would like to invert. > > And what should happen if the the changed tests do not fail here but fail at > your side? We could at least invert the ones we have in common. Or, since you don't pay attention to the unreliable tests, we could just invert the ones failing for me since I do care :) Scott
signature.asc
Description: PGP signature