You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This could introduce inevitable latency when scaling up from 0
what I can see from knative's design doc is like below:
Unlike the queue-proxy, the activator actively sends metrics to the autoscaler via a websocket connection to minimize scale-from-zero latencies as much as possible.
It'll greatly reduce the latency of cold start if keda HTTP addon could introduce something similar to this mechanism.
I can see from the code that the scaler will pull the metrics every 500 ms to determine whether to scale.
https://github.com/kedacore/http-add-on/blob/main/scaler/queue_pinger.go#L49
https://github.com/kedacore/http-add-on/blob/main/scaler/main.go#L40
It’s hardcoded in the http-add-on’s source code to 500ms
This could introduce inevitable latency when scaling up from 0
what I can see from knative's design doc is like below:
It'll greatly reduce the latency of cold start if keda HTTP addon could introduce something similar to this mechanism.
https://github.com/knative/serving/blob/main/docs/scaling/SYSTEM.md
The text was updated successfully, but these errors were encountered: