-
Notifications
You must be signed in to change notification settings - Fork 8.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Index Management] TESTING Added logger to fetch indices route #126169
Conversation
…ex_data_enrichers
Not sure how precise the Logger is, what I had in mind was to use |
Not sure if it's related to the changes here - but the Kibana instance in the cloud deployment ran out of memory and was restarted. |
I am actually thinking it's most likely related to this change, as it's happened three more times since and we have yet to experience it on any other deployments. |
Thanks a lot for checking on this deployment, @tylersmalley! |
@elasticmachine merge upstream |
Logs are mostly filled with Elasticsearch GC. Mind if we scale the cluster up? I'm hoping we can get stack monitoring working again to keep a closer eye on memory. |
Sure, that would be great, @jbudz! I currently have 1000 indices, but I'm planning to run my test with a large amount of indices: 2000, 3000 etc maybe up to 5000-10000 indices. Do you think the deployment has problems because of that? |
It's hard to say with the current logs. I see random spikes that look pretty consistent with user access from a browser(4 hours ago for example) and browser refreshes so I'm wondering if that's possible. Adding more shards could definitely help us narrow that down. Bumping ES to 4GB - will report back once it's back up. |
Okay it's back up - I'll keep a tab open to monitor. |
💔 Build Failed
Failed CI StepsTest Failures
Metrics [docs]Unknown metric groupsESLint disabled in files
ESLint disabled line counts
Total ESLint disabled count
History
To update your PR or re-run it, just comment with: |
FIY, I'm currently adding another 1000 indices into the deployment to test performance with 2000 indices. |
@jbudz I'm at about 5000 small indices in the deployment and would like to get to 10 000 to complete my testing. But I've started getting this error when creating new indices
Do you maybe know what that is related to and if it's possible to re-configure the deployment to handle this? |
I just bumped the cluster to 8gb of RAM. It's definitely the number of shards/indices that's causing things to slow down. Given we're the only client ATM - I expect Kibana isn't very friendly to heavily sharded deployments. It could be a lot of things - alerting, monitoring loading a list of all indices ( 1mb+ per xhr request, auto reloading every 10 seconds) and so on. This is probably something that should be added to our performance working group - cc @tylersmalley @danielmitterdorfer . Recap: a 4gb cluster with 1000-5000 one document indices is going OOM with ~1 active Kibana user. |
Thank you @jbudz!
|
Improving the efficiency of Elasticsearch with many indices/shards is actively tackled by the Elasticsearch's distributed team at the moment. See for example the blog post Three ways we've improved Elasticsearch scalability for recent improvements in 7.16. Note that these improvements do not change our recommended of shards/GB RAM yet though. For this targeted one-off test, I could see two options:
|
I have now tested with about 8000 indices and I think the results give us some good insights into what we can work on to improve indices list performance. I will add all the findings to #126242. |
This PR adds a logger to Index Management "list/reload indices" route to test where the loading time is spent when a large list of indices is being retrieved. The code is not intended to be merged and the main goal of this PR is to test the indices list performance on Cloud (see #126242).
The logger is added on following "checkpoints":