-
Notifications
You must be signed in to change notification settings - Fork 561
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Race Condition in parallel_tests integration: an outdated report may be written #815
Comments
It's WIP because it's flaky at the moment and we don't want our CI to be flaky. See #815
It's WIP because it's flaky at the moment and we don't want our CI to be flaky. See #815
+1 here |
Does bumping up merge_timeout mitigate this? It looks like the number of tests involved is small enough that the default timeout of 10 minutes should be more than enough. Just curious whether that's a factor. I ran into the following case with some long running test suites:
and the fix, in my case, was to bump up That might be a separate issue with similar symptoms. Mentioning it here in case others googling for this fall into the same case. |
@benissimo similar symptoms, separate issue (more user configuration) In the local test case the tests finish within ~2 seconds so it's way beyond being impacted by the merge timeout. |
This might be related to us not doing correct checking which I ask/discuss in grosser/parallel_tests#772 - after that is resolved and changes are implemented this should be worth re checking. |
It seems to be better now but not fixed. CI now occasionally fails with:
which funnily enough happens on JRuby the 2-3 times I've seen it and only that spec. Needs a fix soon-ish or a wip on the cuke as not to be annoying/make people think they did anything wrong. |
Discovered/verified in #814
Basically apparently it can happen that we generate the correct report but override it with an outdated one. Discovered running the parallel tests part of the features introduced in #841 (currently tagged wip to avoid flakies, remove the wip tag if you want to test this out). It doesn't happen reliably. Might happen 3 times in a row or not at all for 20 runs (for me at least).
I gathered some examples here: #814 (comment)
What you can see there is something like this:
Notice that the second too last has the right number of loc and coverage expected in the test and also seems to take everything into account (unsure why it says n/8 though, guess: I could run 8 in parallel but parallel tests only spawned 4 as there are only 4 test files) but is seemingly later overridden by an outdated report.
If possible we should of course find and fix that race condition, and then remove the wip tags to always run these tests!
The text was updated successfully, but these errors were encountered: