-
Notifications
You must be signed in to change notification settings - Fork 1.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
OpenSearch Performance Experiments Results #2461
Comments
Can you guide some steps on how to test @travisbenedict since I have no idea to run a benchmark with this: |
Hey @BlackMetalz After starting OpenSearch on a node I ran the following OpenSearch Benchmark commands depending on the cluster type. Clusters with security plugin disabled: Clusters with security plugin enabled: In my I hope this helps. Sorry for not including more details in the first place |
Is there similar result for vector search for reference ? |
This is quite old, I'm going to close this issue because it's not calling for anything actionable. Since then, we've built https://opensearch.org/benchmarks and there's a bunch of stuff there. @layavadi I don't believe we publish vector numbers there (yet). That's opensearch-project/opensearch-benchmark#103. However, AWS has a very detailed blog post on what vector search performance looks like in production and the tradeoffs that I think can help. |
Background
Using OpenSearch 1.2 build 762 (arm64/x64) I ran a set of ~20 performance tests for each of the following single node configurations:
All tests were running using OpenSearch Benchmark with an i3.8xlarge EC2 instance as the load generation host. The tests used a modified version of the default schedule for the nyc_taxis workload which runs the original schedule twice with all operations in warmup mode and three times as the standard schedule, commonly known as two warmup and three test iterations. Additional aggregations were run on the results of each test to average together metrics across different query types in order to create a set of query summary metrics.
A new load generator and new OpenSearch single node cluster were provisioned for each test.
Findings
Some random variation between tests is expected. For indexing throughput the standard deviation as a percentage of the mean of any percentile statistic, excluding p100, is about 5% across all configurations. For query latency this is about 10%.
Average latency for all queries in a workload can vary by 20% or more between any given test. Why this is will require more research. In the meantime we should avoid outright comparisons of one test to another.
Included below are some approximate statistics for index and query metrics for each configuration. This includes the expected (average) value, the standard deviation as a percentage of the mean and the percent difference between the min and max. This table is meant to give people a framework for understanding their performance test results and should not necessarily be taken as a ground truth.
Raw Data
The text was updated successfully, but these errors were encountered: