Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Lock-step batch processing & writing #5

Open
hackergrrl opened this issue Jun 15, 2019 · 0 comments
Open

Lock-step batch processing & writing #5

hackergrrl opened this issue Jun 15, 2019 · 0 comments
Labels
enhancement New feature or request

Comments

@hackergrrl
Copy link
Member

hackergrrl commented Jun 15, 2019

This would be an experiment to see if it speeds up indexing overall.

Right now, many instances of multifeed-index run concurrently, writing their own batches to whatever storage they use, plus writing index state changes (fetchState,storeState). Having 10 indexes over 10000 entries /w batchsize=100 means 100 * 2 * 10 = 2000 writes.

What if multifeed-index managed a set of views? If it used shared storage for storing state, and, when possible, did mapping in lock-step, it could reduce the # of writes considerably, more like 100 writes instead, from the above example. Fewer reads too, since a batch of 100 entries would be sent to all views at the same time.

These gains are only possible when all views are indexed up to the same point. Otherwise, they'd need to have separate fetch/process/write loops. It might still cost less to detect this, wipe all views to get them back to the same state, and then re-index them all together.

@hackergrrl hackergrrl added the enhancement New feature or request label Jun 15, 2019
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

1 participant