-
Notifications
You must be signed in to change notification settings - Fork 39.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Failing Test] Strange ci-benchmark-scheduler-perf-master behavior #127245
Comments
cc: @pohly @sanposhiho |
/assign The first failed test case is:
That should have caused the per-test log output to be preserved, which might have had more information, but it's not under the artifacts. Something isn't quite right with output handling - will check. |
SchedulingCSIPVs/5000Nodes_5000Pods has 48 as threshold. In https://storage.googleapis.com/kubernetes-jenkins/logs/ci-benchmark-scheduler-perf-master/1833184953530060800/build-log.txt, it has:
Why is the threshold so low? ... never mind, wrong test. |
#125534 changed how I think we now hit gotestyourself/gotestsum#413 (comment). |
There are issues with gotestsum processing the JSON output of benchmarks. Collecting the original output may help with debugging this. If the files are not too large, then it may be worthwhile to keep this enabled even after fixing the current issue (kubernetes/kubernetes#127245).
There are issues with gotestsum processing the JSON output of benchmarks. Collecting the original output may help with debugging this. If the files are not too large, then it may be worthwhile to keep this enabled even after fixing the current issue (kubernetes/kubernetes#127245).
I submitted gotestyourself/gotestsum#438. If this doesn't get merged soonish, it's probably better to revert #125534 and re-apply it later with a new gotestsum. |
@macsko : how much longer are you willing to wait before we revert? There have been some discussions around how to fix this in gotestsum, but no conclusion yet. |
It's not that urgent until something bad happen with scheduling performance and we don't get an alert. |
/triage accepted |
I know, it took a while... 😅 |
Which jobs are failing?
ci-benchmark-scheduler-perf-master
Which tests are failing?
Interestingly, we get benchmark results for all of the test cases, but the job is somehow failing (see testgrid).
Since when has it been failing?
6th September. No changes in scheduler_perf since 26th August.
Testgrid link
https://testgrid.k8s.io/sig-scalability-benchmarks#scheduler-perf
Reason for failure (if possible)
I have no idea what's going on with the test. Tests give the correct results and populate the perf-dash dashboard.
Anything else we need to know?
No response
Relevant SIG(s)
/sig scheduling
The text was updated successfully, but these errors were encountered: