-
Notifications
You must be signed in to change notification settings - Fork 2.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Migrate merge-blocking jobs to dedicated cluster: pull-kubernetes-e2e-gce #18852
Comments
/remove-help |
Tangentially related, it would be nice to know if we even need to use |
Opened #18916 The main branch and 1.19 variants aren't merge-blocking anymore, but earlier branches are. Moving them all over |
#18916 merged 2020-08-19 16:40 PT https://prow.k8s.io/?job=pull-kubernetes-e2e-gce - shows a reasonable amount of traffic since there is now a push to get PR's landed in time for the final cut of kubernetes v1.16. The only failures appear to be flakes https://testgrid.k8s.io/presubmits-kubernetes-blocking#pull-kubernetes-e2e-gce&graph-metrics=test-duration-minutes - overall the job duration is less spiky and has maybe gone slightly down over time https://storage.googleapis.com/k8s-gubernator/triage/index.html?pr=1&job=pull-kubernetes-e2e-gce%24 - no real change in errors https://prow.k8s.io/job-history/gs/kubernetes-jenkins/pr-logs/directory/pull-kubernetes-e2e-gce - Seeing #19034, would like to understand whether this job caused that issue or something else cpu utilization - big spikes in the beginning for build, then nothing memory utilization - looks like that's about right So if it turn out #19034 is unrelated to this change, we're good. But need to dig into that a little more first |
@RobertKielty @spiffxp I would like to work on this issue |
So, the changes have already been made via #18916 (see #18852 (comment)) The reason this is still open is because #19034 is unexplained, and maybe happened around the same time #18916 merged? If we can either prove that #18916 didn't cause it (see #19034 (comment)), or if we can fix #19034, then this issue can be closed. To answer your question test-infra/config/tests/jobs/jobs_test.go Lines 983 to 989 in 6e1d254
#18832 set It is still merge blocking for older branches (release-1.18, release-1.17), as we generally don't backport policy or test changes back to already-released versions of kubernetes except under special circumstances. The reason this complicates things is the job wouldn't see as much traffic as jobs that always run for all branches, so it's tougher to avoid variance due to a smaller sample-set size, and thus tougher to make a judgement call on "does everything still look OK." However, I saw enough traffic in #18852 (comment) when cherry picks were being swept through in advance of upcoming patch releases. So aside from the question of #19034 I think this looks good |
Per #18852 (comment) the reason I held this open is because I'm still not certain that migration of this job did not cause #19034. But we've lived with it unresolved for about 90d now, so I guess we can live with it unexplained for longer. /close |
@spiffxp: Closing this issue. In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
What should be cleaned up or changed:
This is part of #18550
To properly monitor the outcome of this, you should be a member of k8s-infra-prow-viewers@kubernetes.io. PR yourself into https://github.com/kubernetes/k8s.io/blob/master/groups/groups.yaml#L603-L628 if you're not a member.
Migrate pull-kubernetes-e2e-gce to k8s-infra-prow-build by adding a
cluster: k8s-infra-prow-build
field to the job:NOTE: migrating this job is not as straightforward as some of the other #18550 issues, because:
Once the PR has merged, note the date/time it merged. This will allow you to compare before/after behavior.
Things to watch for the job
pull-kubernetes-e2e-gce
for 6hpull-kubernetes-e2e-gce
for 6hThings to watch for the build cluster
Keep this open for at least 24h of weekday PR traffic. If everything continues to look good, then this can be closed.
/wg k8s-infra
/sig testing
/area jobs
/help
The text was updated successfully, but these errors were encountered: