Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Change config defaults #568

Merged
merged 2 commits into from
Dec 22, 2015
Merged

Conversation

dedemorton
Copy link
Contributor

Changed rotate_every_kb default because it was incorrectly documented.

@@ -198,7 +198,7 @@ operation doesn't succeed after
===== bulk_max_size

The maximum number of events to bulk in a single Elasticsearch bulk API index request.
The default is 50.
The default is 200.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Ah, there's a confusion here. We've changed bulk_max_size' default for the Logstash output, where this setting is not documented at all (no idea why). Might be worth for me or @urso (hint) to come up with a longer explanation since this is critical for performance.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Whoops. Monica mentioned that you changed the default, and since the setting only appears in one place, I assumed I was changing the correct setting. :-) I'll revert this value to 50 and wait for you to submit a description.

Copy link

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

So we've got multiple output plugins:

  • Elasticsearch
  • logstash
  • console
  • file

The default bulk_max_size is used for all output plugins but elasticsearch, which sets the default to 50.

This options sets maximum number of events that can be combined internally into batches and will be publishable by the output plugins =>

  • if beat tries to send single events, the events are collected into batches

  • if beat tries to publish large batch of event (bigger bulk_max_size), the batch will be split.

    Bigger batch sizes can improve performance due to ammortizing per event sending overhead. On the other hand to big batch sizes can increase processing time such that queues in logstash/elasticsearch can not be processed -> APIs return errors, connections get killed or publish requests time out. This increases latency and lowers throughput for indexing events.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@urso I've created a new doc issue #586 to track the doc changes for bulk_max_size so these comments don't get lost under the outdated diff....

andrewkroh added a commit that referenced this pull request Dec 22, 2015
@andrewkroh andrewkroh merged commit 2c97653 into elastic:1.0.1 Dec 22, 2015
@dedemorton dedemorton deleted the change_config_defaults branch April 7, 2018 00:31
leweafan pushed a commit to leweafan/beats that referenced this pull request Apr 28, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants