Skip to content

Logstash output plugin to sign and export logstash events to Amazon Elasticsearch Service

License

Notifications You must be signed in to change notification settings

gcc-gso/logstash-output-amazon_es

 
 

Repository files navigation

Logstash Plugin

This is a plugin for Logstash.

License

This library is licensed under the Apache 2.0 License.

Setting Up

Installation

One command installation bin/logstash-plugin install logstash-output-amazon_es

While we are in the process of getting this plugin fully integrated within logstash to make installation simpler, if above does not work, or you would like to patch code here is a workaround to install this plugin within your logstash:

  1. Check out/clone this code from github
  2. Build plugin using - gem build logstash-output-amazon_es.gemspec ( this works with jruby and rubygem versions > 1.9)
  3. Install plugin using <logstash-home>/bin/plugin install logstash-output-amazon_es-0.2.0-java.gem (or the non java variant)
  4. For 2.3 support, please use '/bin/logstash-plugin install logstash-output-amazon_es-1.0-java.gem'
  5. For 5.2 support, please use '/bin/logstash-plugin install logstash-output-amazon_es-2.0.1-java.gem'

Configuration for Amazon Elasticsearch Output plugin

To run the Logstash output Amazon Elasticsearch plugin simply add a configuration following the below documentation.

An example configuration:

output {
    amazon_es {
        hosts => ["foo.us-east-1.es.amazonaws.com"]
        region => "us-east-1"
        # aws_access_key_id, aws_secret_access_key optional if instance profile is configured
        aws_access_key_id => 'ACCESS_KEY'
        aws_secret_access_key => 'SECRET_KEY'
        index => "production-logs-%{+YYYY.MM.dd}"
        }
    }
  • Required Parameters

    • hosts (array of string) - Amazon Elasticsearch domain endpoint. eg ["foo.us-east-1.es.amazonaws.com"]
    • region (string, :default => "us-east-1") - region where the domain is located
  • Optional Parameters

    • Credential parameters
      • aws_access_key_id, :validate => :string - Optional AWS Access key
      • aws_secret_access_key, :validate => :string - Optional AWS Secret Key
        The credential resolution logic can be described as follows:
        • User passed aws_access_key_id and aws_secret_access_key in aes configuration
        • Environment Variables - AWS_ACCESS_KEY_ID and AWS_SECRET_ACCESS_KEY (RECOMMENDED since they are recognized by all the AWS SDKs and CLI except for .NET), or AWS_ACCESS_KEY and AWS_SECRET_KEY (only recognized by Java SDK)
        • Credential profiles file at the default location (~/.aws/credentials) shared by all AWS SDKs and the AWS CLI
        • Instance profile credentials delivered through the Amazon EC2 metadata service
    • Retry Parameters
      • max_retries (number, default => 3) - Set max retry for each event
      • retry_max_items (number, default => 5000) - Set retry queue size for events that failed to send
      • retry_max_interval (number, default => 5) - Set max interval between bulk retries
    • index (string - all lowercase, default => "logstash-%{+YYYY.MM.dd}") - Elasticsearch index to write events into
    • flush_size (number , default => 500) - This setting controls how many events will be buffered before sending a batch of events in bulk API
    • idle_flush_time (number, default => 1) - The amount of time in seconds since last flush before a flush is forced. This setting helps ensure slow event rates don't get stuck in Logstash. For example, if your flush_size is 100, and you have received 10 events, and it has been more than idle_flush_time seconds since the last flush, Logstash will flush those 10 events automatically. This helps keep both fast and slow log streams moving along in near-real-time.
    • template (path) - You can set the path to your own template here, if you so desire. If not set, the included template will be used.
    • template_name (string, default => "logstash") - defines how the template is named inside Elasticsearch
    • port (string, default 443) - Amazon Elasticsearch Service listens on port 443 - https (default) and 80 - http. Tweak this for custom proxy.
    • protocol (string, default https) - The protocol used to connect to the Amazon Elasticsearch Service

Documentation

Logstash provides infrastructure to automatically generate documentation for this plugin. We use the asciidoc format to write documentation so any comments in the source code will be first converted into asciidoc and then into html. All plugin documentation are placed under one central location.

Need Help?

Need help? Try #logstash on freenode IRC or the https://discuss.elastic.co/c/logstash discussion forum.

Developing

1. Plugin Development and Testing

Code

  • To get started, you'll need JRuby with the Bundler gem installed.

  • Create a new plugin or clone and existing from the GitHub logstash-plugins organization. We also provide example plugins.

  • Install dependencies

bundle install

Test

  • Update your dependencies
bundle install
  • Run unit tests
bundle exec rspec
  • Run integration tests

Dependencies: Docker

Before the test suite is run, we will load and run an Elasticsearch instance within a docker container. This container will be cleaned up when suite has finished.

bundle exec rspec --tag integration

2. Running your unpublished Plugin in Logstash

2.1 Run in a local Logstash clone

  • Edit Logstash Gemfile and add the local plugin path, for example:
gem "logstash-filter-awesome", :path => "/your/local/logstash-filter-awesome"
  • Install plugin
bin/plugin install --no-verify
  • Run Logstash with your plugin
bin/logstash -e 'filter {awesome {}}'

At this point any modifications to the plugin code will be applied to this local Logstash setup. After modifying the plugin, simply rerun Logstash.

2.2 Run in an installed Logstash

You can use the same 2.1 method to run your plugin in an installed Logstash by editing its Gemfile and pointing the :path to your local plugin development directory or you can build the gem and install it using:

  • Build your plugin gem
gem build logstash-filter-awesome.gemspec
  • Install the plugin from the Logstash home
bin/plugin install /your/local/plugin/logstash-filter-awesome.gem
  • Start Logstash and proceed to test the plugin

Contributing

All contributions are welcome: ideas, patches, documentation, bug reports, complaints, and even something you drew up on a napkin.

Programming is not a required skill. Whatever you've seen about open source and maintainers or community members saying "send patches or die" - you will not see that here.

It is more important to the community that you are able to contribute.

For more information about contributing, see the CONTRIBUTING file.

Building the Logstash output plugin with Docker

Prerequisites:

This will result in a newly created binary inside the host-mounted volume ${PWD} named logstash-output-amazon_es-<VERSION>-java.gem. Where <VERSION> is defined as value of s.version in logstash-output-amazon_es.gemspec file.

About

Logstash output plugin to sign and export logstash events to Amazon Elasticsearch Service

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Ruby 95.7%
  • Dockerfile 4.3%