-
Notifications
You must be signed in to change notification settings - Fork 1.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: Add a Prometheus metric for measuring the scalable object loop processing deviation #4703
Conversation
Signed-off-by: Jorge Turrado <jorge_turrado@hotmail.es>
Thank you for your contribution! 🙏 We will review your PR as soon as possible. While you are waiting, make sure to:
Learn more about: |
Signed-off-by: Jorge Turrado <jorge_turrado@hotmail.es>
/run-e2e sequential |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I wonder, do you see any actualy latency that could be mesured and exposed here? I mean one in the loop itself? We are already reporting latency per scaler. I am not sure if this kind of metric won't be confusion for users?
The difference here is what are we measuring. In one case, we are measuring the trigger latency and in the other case, we are measuring the deviation between loops. Both are similar, but they are different. For example: These new metrics measure the difference between the expected execution time and the real execution time, so we are getting a real picture of the overload that we have, if we should have been executed the loop at X time and we execute it later, it's because something is slowing us, it could be due to the triggers' latency but also due to an overload. If we have throttling, both metrics will increase, but the increase of |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The numbers are interesting, great stuff.
Let's procced with keda_internal_scale_loop_latency
as we discussed offline.
Signed-off-by: Jorge Turrado <jorge.turrado@scrm.lidl>
Signed-off-by: Jorge Turrado Ferrero <Jorge_turrado@hotmail.es>
/run-e2e sequential |
Signed-off-by: Jorge Turrado <jorge.turrado@scrm.lidl>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM, great stuff!
Signed-off-by: Jorge Turrado <jorge.turrado@scrm.lidl>
/run-e2e sequential |
Let's make sure we also update our documentation to ensure everyone knows what it represents and how to interpret it |
yeah, I plan to do it later on, I just did this PR before to ensure that we include it in the release (if the release would have been done today). |
Checklist
keda_internal_scale_loop_latency
metric keda-docs#1165Fixes #4702