This project provides an extendable Java framework for creating serverless ETL functions on AWS Lambda. Bender handles the complex plumbing and provides the interfaces necessary to build modules for all aspects of the ETL process.
An example of what you can do is enrich your log data by appending GeoIP information. For example:
input: {"ip": "8.8.8.8"}
output: {"ip": "8.8.8.8", "geo_ip": {"location": {"lat": 37.751, "lon": -97.822}}}
This enables you to create dashboards like these in Kibana:
Bender is easily configurable with either json or yaml. The configuration guide provides documentation for option specifics and sample_configs/ contains real world examples of how Bender is configured with commonly used pipelines.
While developers are able to write their own Input Handlers, Deserializers, Operations, Wrappers, Serializers, Transporters, or Reporters, out of box the Bender contains basic functionality to read, filter, manipulate, and write JSON from Amazon Kinesis Streams or Amazon S3 files. Specially the following is supported:
Handlers interface between Amazon Lambda triggers and data sources provided to your ETL function. Events or lines are able to read from:
- Kinesis
- S3
- S3 via SNS
Modular filter support is not yet included but basic string matching and regex based filters are included as a part of Bender Core.
Included is a generic JSON deserializer to transform strings into GSON objects. This allows processing of loosely defined schemas such as those done in application logging. For schema validation the use of GSON to POJOs is encouraged.
Data sometimes needs to be transformed, fixed, sanitized, or enriched. Operations allow for these types of data manipulation.
- Geo IP lookup
- JSON Root node promoter (in the case you have nested data)
- JSON Array Dropping
- JSON Array Splitting (turning a single event into multiple)
- Appending value type information to key names (helps with writing data to ElasticSearch)
Optionally wrap data to provide additional information on where the data originated from and what processed it:
- Kinesis Wrapper
- S3 Wrapper
- Basic Wrapper
Write your transformed and wrapped data back into JSON before loading it elsewhere.
Transporters convert string payloads to serialized wire formats and send batches of data to destinations.
- Firehose
- S3 (partitioning support included)
- Elasticsearch (time based index inserts)
- Splunk
- Scalyr
- Sumo Logic
Monitor the ETL process at each phase within your function and output those metrics for easy consumption in
- Cloudwatch Metrics
- Datadog
The easiest way to deploy your function is to use Apex. A sample project is included under example_project/. The project provides an example of a function that is triggered by Kinesis, drops data matching a regex, and forwards the rest to Firehose.
Note to deploy the example you will need to create an IAM role to allow your lambda function to read from kinesis and write to firehose. Your role will need the following two policies:
arn:aws:iam::aws:policy/service-role/AWSLambdaKinesisExecutionRole
arn:aws:iam::aws:policy/AmazonKinesisFirehoseFullAccess
After creating your role edit example_project/project.json
with the role ARN.
You will also need to create the source Kinesis and destination Firehose
streams.
To deploy:
cd example_project/
make deploy
DRY=false make deploy
Features and bug fixes are welcome. Please adhere to the following guidelines:
- Use Google's Java style guide for your IDE.
- Be conscientious of dependencies you add to Bender Core.
- Help maintain unit test code coverage by adding tests for each branch in new code.