-
-
Notifications
You must be signed in to change notification settings - Fork 230
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Chore: Fix Flaky E2E Tests #3946
Comments
Thanks for the issue, our team will look into it as soon as possible! If you would like to work on this issue, please wait for us to decide if it's ready. The issue will be ready to work on once we remove the "needs triage" label. To claim an issue that does not have the "needs triage" label, please leave a comment that says ".take". If you have any questions, please comment on this issue. For full info on how to contribute, please check out our contributors guide. |
Do the e2e tests have to rely on It seems most of the flakey-ness comes from either timeouts reaching
Hypothesis: I wonder if GitHub is rate limiting the workers / runners that are hitting I know the idea behind the e2e tests was to have a re-producible login flow that we could test against. But in my eyes, it'd be better to have super solid, stable tests that don't rely too heavily on the backbone infrastructure of some other service. A good example of this would be the GitHub outage last week: this would entierly cripple our e2e validation and flow. Again, very very tricky because our auth flow does in fact rely on GitHub. |
I agree as I came across this also. I don't think we should be testing the full login/redirect flow past the point of whether the button is displayed or not. Testing what we can control from an e2e perspective could be less flakey. I'll disable the GitHub URL checks and see |
🎉 This issue has been resolved in version 2.59.0-beta.4 🎉 The release is available on GitHub release Your semantic-release bot 📦🚀 |
Re-opening so we can keep tracking flakey e2e tests. Looks like we still see occasional timeouts: https://github.com/open-sauced/app/actions/runs/10459941789/job/28965031250 I wonder if it'd be worth hosting some of our own powerful runners via self hosted runners . I wonder if the 4 core, 16 gb ram runners GitHub has are underpowered and during peek hours, we get throttled causing the heavy browser based e2e tests to time out. I could look at bootstrapping some powerful Azure runners to self host our actions. We'd likely see vast performance improvements across runners. |
🎉 This issue has been resolved in version 2.59.0 🎉 The release is available on GitHub release Your semantic-release bot 📦🚀 |
We can intercept the request without actually hitting github.com and just check that it was heading there. |
That sounds like a great plan. Seeing more frequent timeouts during peak engineering times: https://github.com/open-sauced/app/actions/runs/10579984281/job/29313738273
I'll see if I can carve out some time to drop some powerful self hosted runners |
Once the self-hosted runners are in place, if the current E2E tests are running well, we should revert the skip tests and test modifications in #3983 to see if they run well again as locally they always pass. |
I've been noticing more frequent flakiness in the E2E test suite, recently. Here is on from a couple minutes ag0, https://github.com/open-sauced/app/actions/runs/10380912089/job/28741490937?pr=3945, but I've also noticed them on releases to main.
The text was updated successfully, but these errors were encountered: