-
Notifications
You must be signed in to change notification settings - Fork 1.2k
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
AutoscalingRunnerSet gets stuck thinking runners exist when they do not #3821
Comments
Hello! Thank you for filing an issue. The maintainers will triage your issue shortly. In the meantime, please take a look at the troubleshooting guide for bug reports. If this is a feature request, please review our contribution guidelines. |
Having the same issue, wondering if you got this fixed by reverting to the previous version |
Yesterday solved same issue with 0.9.2. Had to delete custom resource with status failed , then it all returned to normal functioning |
I also tried fully deleting everything and creating it again. Some of my node pools are working, and most of them are not. This seems to be related to something on the Github Side, but I can't confirm what's going on. |
I had the same issue. I tried recreating the resources, including the CRD, but that didn't work. However, I found that giving a new |
Same issue from my side, deployment using helm. Logs from orchestrator pod (in loop, before issue was solved, pods of runners were not running at that time):
|
Checks
Controller Version
0.9.3
Deployment Method
Kustomize
Checks
To Reproduce
Describe the bug
One of our AutoscalingRunnerSets will often think that runners exist when they do not (the rest of our groups have not been affected). At this point, the github action runs will be pending a runner, but no pod will spin up. When it gets stuck in this state, the only way I've been able to correct is deleting the AutoscalingRunnerSet and re-creating it. Deleting the controller & listener did not help.
Here's an example of the kubernetes commands ran to diagnose and fix (argoCD re-creates the deleted resource for me).
Describe the expected behavior
The AutoscalingRunnerSet should create pods when desired, and not think pods exist that do not.
I'm curious if the pod tracking mechanism is a counter that increments/decrements rather than looking at source of truth (how many pods actually exist)
Additional Context
Controller Logs
Logs from when it was fixed by deleting the AutoscalingRunnerSet and re-creating
The text was updated successfully, but these errors were encountered: