_ _ _
__| | _ _ _ _ __ _ _ __ ___ __| || |__
/ _` || || || ' \ / _` || ' \ / _ \/ _` || '_ \
\__,_| \_, ||_||_|\__,_||_|_|_|\___/\__,_||_.__/
|__/ _
___| |_ _ _ ___ __ _ _ __
(_-<| _|| '_|/ -_)/ _` || ' \
/__/ \__||_| \___|\__,_||_|_|_|
_ _ _ _
___ | | __ _ ___| |_ (_) __ ___ ___ __ _ _ _ __ | |_
/ -_)| |/ _` |(_-<| _|| |/ _|(_-</ -_)/ _` || '_|/ _|| ' \
\___||_|\__,_|/__/ \__||_|\__|/__/\___|\__,_||_| \__||_||_|
The missing blueprint for AWS Lambda, which reads stream from AWS DynamoDB and writes it to Elasticsearch.
Whenever data is changed (modified, removed or inserted) in DynamoDB one can use AWS Lambda function to capture this change and update Elasticsearch machine immediately. Further reading:
Indexing Amazon DynamoDB Content with Amazon Elasticsearch Service Using AWS Lambda
Install:
npm i dynamodb-stream-elasticsearch
Use it in your lambda:
const { pushStream } = require('dynamodb-stream-elasticsearch');
const { ES_ENDPOINT, INDEX } = process.env;
function myHandler(event, context, callback) {
console.log('Received event:', JSON.stringify(event, null, 2));
pushStream({ event, endpoint: ES_ENDPOINT, index: INDEX })
.then(() => {
callback(null, `Successfully processed ${event.Records.length} records.`);
})
.catch((e) => {
callback(`Error ${e}`, null);
});
}
exports.handler = myHandler;
Upload Lambda to AWS and star this repository if it works as expected!!
Param | Description | Required |
---|---|---|
event | Event object generated by the stream (pass it as it is and don't modify) | required |
endpoint | Exact url of Elasticsearch instance (it works with AWS ES and standard ES) (string) | required |
index | The name of Elasticsearch index (string). If not provided will set the same as DynamoDB table name | optional |
refresh | Force Elasticsearch refresh its index immediately more here (boolean). Default: true | optional |
useBulk | Enables bulk upserts and removals (boolean). Default: false | optional |
transformFunction | A function/promise to transform each record before sending them to ES. Applies to INSERT and UPDATE operations. If transformFunction returns an empty object or false the row will be skipped. This function will receive body (NewImage), oldBody (OldImage) and (record) as the whole record as arguments. |
optional |
elasticSearchOptions | Additional set of arguments passed to elasticsearch Client see here | optional |
Docker can be used to host a node of Elasticsearch
To run tests locally you need to have an Elasticsearch Docker container running. Simply type:
docker-compose up -d
Tests are written using Mocha [https://mochajs.org/]. Tests can be launched using:
npm test
If you modify src/utils/aws-es-connection
you will have to setup localstack on your local machine
You will need to:
- start localstack locally with docker
- install localstack via pip
- configure a new elastic-search domain (and wait for it to be available)
- run the tests
docker-compose up -d
pip install awscli-local
aws --endpoint-url http://localhost:4566 es create-elasticsearch-domain --domain-name domain-test
AWS_ES_ENDPOINT=http://localhost:4571 npm test-aws
One Note: there seem to be problems running localstack on macs M1, to check if the cluster has been created run:
awslocal es describe-elasticsearch-domain --domain-name domain-test | grep Created
If you want to commit changes, make sure if follow these rules:
- All code changes should go with a proper integration test;
- Code should follow Javascript Standard Guideline;
- Commit messages should be set according to this article.
- Introduce Continuous Integration;
- Add Elasticsearch bulk operation instead of index for multiple records
This project is licensed under the MIT License - see the LICENSE.md file for details
If you find this project to be useful and you would like to support the author for maintaining it, you might consider to make any donation under this link:
Compatible with node 8.10. (If for some reason you want to use it with node 6.10, then use 1.0.0 of this module)
Third version doesn't support types as they were deprecated in ES 7.0.