Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[PyTorch] Update timeout for inductor test suite: +4hours #3610

Merged
merged 1 commit into from
Mar 4, 2025

Conversation

anmyachev
Copy link
Contributor

12 hours isn't enough for all test suite: https://github.com/intel/intel-xpu-backend-for-triton/actions/runs/13632302494/job/38102585706

@pbchekin Is it possible/does it make sense to complicate and increase the timeout only for "all" test suite run?

@anmyachev anmyachev requested a review from pbchekin March 4, 2025 18:34
@anmyachev anmyachev marked this pull request as ready for review March 4, 2025 18:34
@alexbaden
Copy link
Contributor

12 hours is probably too long...
I don't think it takes 12 hours in upstream. But maybe they only run a subset?

@anmyachev
Copy link
Contributor Author

12 hours is probably too long... I don't think it takes 12 hours in upstream. But maybe they only run a subset?

PyTorch splits all tests into 4 parts, as per the following code:
https://github.com/pytorch/pytorch/actions/runs/13644202775/workflow?pr=147727#L35.
It's too difficult for us to duplicate and maintain this, but we can run the full test suite not very often, once a week or something like that.

@alexbaden
Copy link
Contributor

We should just use their CI then. Easy enough to make a PR there once a week.

@anmyachev
Copy link
Contributor Author

We should just use their CI then. Easy enough to make a PR there once a week.

But it's even easier to press just one button in our repo or even run it on schedule :) At least it’s great for us to have such an opportunity, for those who don’t have the rights to launch CI in PyTorch, or in the case of a pre-release there may simply not be enough resources for everyone since the number of launches IMHO always increases in this time.

@alexbaden
Copy link
Contributor

sure - it's a lot of machine time though.

@alexbaden
Copy link
Contributor

but I suppose it is nice to be able to control the pytorch version

@anmyachev anmyachev merged commit 117cf3e into main Mar 4, 2025
6 checks passed
@anmyachev anmyachev deleted the amyachev/pytorch-timeout branch March 4, 2025 20:08
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants