-
Notifications
You must be signed in to change notification settings - Fork 1.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[backend] persistence agent don't update pipeline run status after workflow deleted #5722
Comments
Thank you @algs for flagging this bug report! My rough idea to improve would be that, persistence agent should also periodically list all runs from KFP API that are still in running state, if any of the runs do not have a corresponding argo workflow. We should update the DB record to maybe a special state called "WORKFLOW_DELETED" etc. What do you think? |
Maybe just list runs which have not been updated/reported for a certain time, it should be a few runs. |
would this because of service error rather than the corresponding workflow deleted? |
An alternative could be persistent agent watches workflow changes and if there's a delete workflow event, check running state of associated run and update DB. I prefer to choose the way which can reduce apiserver pressure. :D |
@Jeffwan great idea! That feels like the most efficient solution. |
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions. |
/lifecycle frozen |
Environment
Steps to reproduce
kubectl delete workflow xxx
to delete the workflow corresponding to the pipeline run, then the unfinished nodes in the pipeline run stay running, and never been updated.delete workflow
may happen when we delete the scheduledworkflow, and then the workflow without the parent scheduledworkflow got GCExpected result
The running nodes in the pipeline run been marked as
terminated
Materials and Reference
Impacted by this bug? Give it a 👍. We prioritise the issues with the most 👍.
cc @Jeffwan
The text was updated successfully, but these errors were encountered: