-
Notifications
You must be signed in to change notification settings - Fork 17
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
test: Add integration tests for public datasets #591
Conversation
The GitHub workers are failing the integration tests. I see three potential reasons:
Point 1. would be solved after we implement batched preprocessing, meaning that we won't keep all the dataset files in memory, but process If it's Point 2 or 3 then we have a problem. I'm running the integration tests locally on our DGX, let's see what we get as output from that. |
We have one expected fail on GazeBase and one unexpected fail with SB-Sat. Output
The error on GazeBase exactly reproduces #517. I will now merge #593 into this PR and see if we get rid of the error. The fail on SB-Sat is strange though. @prassepaul do you know why that happened? |
Codecov ReportAll modified lines are covered by tests ✅
Additional details and impacted files@@ Coverage Diff @@
## main #591 +/- ##
=========================================
Coverage 100.00% 100.00%
=========================================
Files 52 52
Lines 2337 2337
Branches 582 582
=========================================
Hits 2337 2337 ☔ View full report in Codecov by Sentry. |
Merging #593 into this PR resolves #517:
Also notice |
The problem with SBSAT should be solved in another issue. This PR is now ready for review. We will not include integration tests in our CI (yet). This single test run took 90 minutes (with one dataset failing at download start). As long as we don't solve our very high memory usage, I can do these test runs manually on our DGX via |
6560078
to
3aef774
Compare
Description
A first version to try out downloading and processing public datasets
This should fail until #517 is fixed
We have to find some solution such that integration tests are only tested very seldomly. (what about only when publishing new releases?)
For now I would just add
--ignore=^tests/integration
under[tool.pytest.ini_options]
inpyproject.toml
, because we definitely do not want to preprocess all datasets for each commit, because this would take forever (even if they are cached)