-
Notifications
You must be signed in to change notification settings - Fork 479
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Optimize Autoscaling Configuration #1357
Comments
All tests were with 1000 users over 30 minutes on staging against the latest prod release We currently have 300 seconds set for Scale-out cooldown period in prod Max of 20 tasks for api and each worker - 300 seconds for Scale-out cooldown period
Max of 30 tasks for api and each worker - 30 seconds for Scale-out cooldown period
http_req_failed................: 5.07% ✓ 19019 ✗ 355560 Max of 30 tasks for api and each worker - 300 seconds for Scale-out cooldown period http_req_failed................: 18.04% ✓ 38711 ✗ 175774 If we lower Scale-out cooldown period tasks are spinning up a lot quicker and everything seems to recover quicker as well. It seems like if keep a high limit on tasks we should be able to scale a good amount quicker 🚀 |
All tests were started with the following task counts: Additional tests:
Compared to # 1
Suggested Settings: In terms of scaling in, the settings should remain the same and tasks should spin down to the current minimums in a similar manner. |
Looks great |
User Story:
As a backend developer,
I want to optimize the autoscaling configuration,
So that it responds more quickly when the average CPU usage exceeds the threshold.
Acceptance Criteria
GIVEN the average CPU usage has been above the threshold for 30 seconds,
WHEN the autoscaling checks the CPU usage,
THEN it should initiate the scaling process immediately.
Product & Design Links:
N/A
Tech Details:
This task will involve adjusting the autoscaling configuration to use more responsive metrics, potentially including request volume.
Open Questions:
Is the CPU usage the most effective metric for our autoscaling needs, or should we consider alternatives like request volume?
Notes/Assumptions:
This feature assumes that faster autoscaling will better handle sudden surges in traffic and prevent system overloads.
The text was updated successfully, but these errors were encountered: