-
Notifications
You must be signed in to change notification settings - Fork 385
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Catch-up doesn't respect retries limit #369
Comments
That seemed to be a side effect of the default job abortion behavior. Aborted jobs get into the catch-up queue (if catch-up checked) by default. To avoid this no_rewind flag should be set to 1, which only happens if you manually aborting job. I guess it's neither bug or feature, just a weird scenario. I'd agree timeout abortion should come with no_rewind=1 by default |
Fixed in Cronicle v0.8.56. Thanks @mikeTWC1984 for the assist. |
Also thanks to @mbafford for the detailed issue report! |
Description
The catch-up flag does not respect the retries limit.
If you have a process with the following criteria:
The task will continuously restart after the initial event is triggered.
It appears the "ran on schedule" flag isn't cleared unless the job is successfully completed. My expectation was it would attempt to run the task to catch-up, but not retry if a failure occurred.
Details
Version 0.8.54
For example, my job below
too long
is configured to run:hour:05
1 minute
timeoutNone
retriescatch-up
selectedscript
plug-in that just runssleep 5m
The run log shows the task starting on schedule, and erroring out, but then running again, despite the retries disabled:
The job output confirms it's killed due to maximum run time:
Comments
For my case I will workaround this by disabling the time-out for this job (or the catch-up, since the server isn't likely to be down in a way that would cause problems).
I noticed this because I have multiple backup jobs in a category with a category limit of 1 job at a time. The long-running (and incorrectly configured) prune event was timing out, then re-running, seemingly delaying the entire backup schedule.
The text was updated successfully, but these errors were encountered: