-
Notifications
You must be signed in to change notification settings - Fork 229
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
New job for GitRepo is created and terminated every 3rd second #2853
Comments
The problem is that the GitRepo keeps the latest commit hash from the backing git repository, but that commit hash is wrong and isn't updated correctly. Initially it was blank, but after re-creating the GitRepo it first worked but then got stuck soon afterwards. Since the commit hash is wrong Rancher fleet thinks there are changes all the time and tries to trigger updates. |
It looks like there may be an issue with the
While the That doesn't explain why this issue would only happen for only one |
Yes, all GitRepos are in the same cluster and namespace.
Installed using Rancher 2.9.2 helm chart. |
We have upgraded to 2.9.2 since raising this issue, but the problem still exists, then main GitRepo shows wrong git commit hash. |
Cleaning up the backlog, we can't reproduce this. |
Is there an existing issue for this?
Current Behavior
We have a Rancher installation (version 2.9.1), this problem likely started after upgrading from 2.8.x.
We have 3 GitRepos, but only one of them are experiencing this problem. All point to the same Git repository in BitBucket but with different paths. We run on EKS 1.28 currently but plan on upgrading to EKS 1.29 soon.
For one of these GitRepos a job/pod is created roughly every 3 seconds and then it is terminated (usually), but sometimes they get stuck and we run out of IP-addresses in the subnet. The other GitRepos only see new jobs occasionally or when changes are done to the backing Git repository.
The GitRepo with problem also has this warning/error which we don't understand why it is there:
Expected Behavior
Pods are not created every 3rd second.
Steps To Reproduce
No response
Environment
Logs
Anything else?
We see a lot of logs like this even though no changes are made to the backing Git repo in Bitbucket.
{"level":"info","ts":"2024-09-16T14:07:30Z","logger":"clustergroup-cluster-handler","msg":"Cluster changed, enqueue matching cluster groups","namespace":"<namespace>","name":"cluster-8cf77d5971e8"}
The text was updated successfully, but these errors were encountered: