-
Notifications
You must be signed in to change notification settings - Fork 75
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Suggest policy of reviewing failures before accepting new tests #1871
Comments
Thanks for the report, I agree! We'll try to come up with some ideas on how to address this |
We are working on a way to separate the main test suite (all those "unit tests" etc) from the more complicated tests for which the setup is not always fair to all the tools. These are going to get their own table at the bottom and will be excluded from the main statistics. Later, once we see that the extra tests are equally good for a number of tools they can be moved back to the main report. This should make the problem less important and will allow us to keep the tests that may seem broken for some tools but are still useful for various reasons. |
I appreciate the work going into this project, so please take this as constructive suggestions. What it currently feels to me (perhaps wrongly), is currently if a new test is deemed interesting for the parser, it's imported, and the other tools that run here are ignored. I suggest the goal of sv-tests should be more inclusive. I can see the usefulness of splitting the tests. Splitting will be an improvement in that it will no longer harm the "grade" of those tools that are not applicable. However, I don't think addresses my larger point, it will still take time away from the tool teams whom will still (hopefully) see "secondary tests" false failures. I suggest it as a tool team's job to for a given tool to:
I suggest it as a test importer's job before importing a test to master to:
Again my goal is that when a new test is added, if it's failing on a tool that means a tool designer has work to do on the tool, not on even looking on a non-appropriate test. |
#1870 points out some tests that were added that are not consistent/complete designs.
I'd like to suggest a new policy that before any new tests be accepted, failures against every tool be manually reviewed, and the tests only accepted if all failures represent real tool issues versus some problem with the test (e.g. missing modules/packages).
Alternatively, before being accepted, the tests could be whitelisted or otherwise filtered to run against only those tools that don't need complete designs.
Thanks for considering.
The text was updated successfully, but these errors were encountered: