-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Clusterctl upgrade flake #11610
Comments
This issue is currently awaiting triage. If CAPI contributors determine this is a relevant issue, they will accept it by applying the The Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes-sigs/prow repository. |
heres the source code for where this load balancer gets provisioned in v0.3.25: and now on main for comparison: https://github.com/cahillsf/cluster-api/blob/fb2cc16480a62c1b43c0a1e3a7f48c328e283012/test/infrastructure/docker/internal/docker/manager.go#L110 |
/area e2e-testing |
This commit may be relevant, it's when the newer method of load balancer provisioning was added: 88b3662 |
I think the fix was: f726a2e When not trying to get a port, but let docker choose the port automatically by passing through 0. Maybe there's a smart way we can workaround that.
|
Which jobs are flaking?
periodic-cluster-api-e2e-release-1-9
periodic-cluster-api-e2e-mink8s-release-1-8
periodic-cluster-api-e2e-latestk8s-main
Which tests are flaking?
When testing clusterctl upgrades (v0.4=>v1.6=>current) Should create a management cluster and then upgrade all the providers
When testing clusterctl upgrades (v0.3=>v1.5=>current) Should create a management cluster and then upgrade all the providers
Since when has it been flaking?
Looks like maybe for awhile... failures from:
and can see similar patterns in the timeline back in august/september: https://storage.googleapis.com/k8s-triage/index.html?date=2024-09-04&text=Timed%20out%20waiting%20for%20Cluster&job=.*periodic-cluster-api-e2e.*&test=.*clusterctl%20upgrades
Testgrid link
https://prow.k8s.io/view/gs/kubernetes-ci-logs/logs/periodic-cluster-api-e2e-latestk8s-main/1870177294711001088
Reason for failure (if possible)
Seems like the issue is that the docker controller fails to start up the load balancer container of the workload cluster due to a port conflict when it tries to provision using v0.3 and v0.4 of clusterctl which causes the DockerCluster to get stuck:
Anything else we need to know?
No response
Label(s) to be applied
/kind flake
One or more /area label. See https://github.com/kubernetes-sigs/cluster-api/labels?q=area for the list of labels.
The text was updated successfully, but these errors were encountered: