-
Notifications
You must be signed in to change notification settings - Fork 2.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
boskos/janitor: track when cleanup fails repeatedly for the same resource #15866
Comments
/area boskos |
@ixdy would also help if we can publish the logs from boskos somewhere public. |
@dims I'm not sure where we'd publish them, and I'm also not sure we've done a great job of sanitizing the logs yet (to ensure that they're not leaking any sensitive information). The logs are visible to anyone maintaining the prow cluster, though are in some cases lacking useful information. Regarding tracking cleanup failures, I have a few potential ideas:
|
Option 2 is closely aligned with #14715. |
Issues go stale after 90d of inactivity. If this issue is safe to close now please do so with Send feedback to sig-testing, kubernetes/test-infra and/or fejta. |
/remove-lifecycle stale still want to do this. |
Moving to kubernetes-sigs/boskos#15. |
@ixdy: Closing this issue. In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
Due to programming errors, the janitor may continuously fail to clean up a resource. Two examples I just discovered:
possibly an order-of-deletion issue:
likely incorrect flags (gcloud changed but we didn't?):
It'd be good to have some way of detecting when we're repeatedly failing to clean up a resource.
Not sure yet what the best way would be to track that.
The text was updated successfully, but these errors were encountered: