-
Notifications
You must be signed in to change notification settings - Fork 3.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Data loss when using big batches #502
Comments
Forgot to add something that might help finding the issue - if you run with replication on you can get different data dropped in different nodes, so if you run the test program with -write_data=false you get different errors every time (as queries get run on different nodes for each replicated shard). |
Thanks @ohurvitz. Moving this issue to 0.6.1 |
Anything about this? Still happens on latest, with and without setting write-batch-size as in new config file. |
We haven't got chance to take a look at it yet. It's definitely on my todo list. |
@ohurvitz I finally got to take a look at this and have a fix. Thanks for the best bug report ever, was little tricky to track it down though. |
I have the following case:
As the batch size increases, some data does not make it to the database, even though no error is reported. For my cases, a batch size of 5000 points does it almost every time, while 4000 does it most of the time.
Code to reproduce:
https://gist.github.com/ohurvitz/e5d74ae56d8ffa20e968
Note that there is a lot of hard coded numbers in that code, and also it expects the server name to contain a single digit '1' that is replaced by '2', '3' etc to get to more servers.
The text was updated successfully, but these errors were encountered: