-
Notifications
You must be signed in to change notification settings - Fork 4.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Out of disk space building on Docker NanoServer #34780
Comments
Hi guys. First I want to say I'm impressed you filled the whole disk in only 24 minutes, that's impressive. Also, there's nothing about the 2nd job that says disk space full, just a 2 hour timeout? Anyways, I've had conversations with the MMS folks recently around similar issues in MacOS hosted pools, and I learned that the amount of free space goes up and down randomly with the payloads installed and no one actually knows what it is so they can't tell you how to keep your build from blowing up the machine. Your options include:
|
@dotnet/ncl Can someone look at this? It's failing every run. |
I think we should just move these builds out of the Hosted pools as they use considerable disk space. cc: @eiriktsarpalis |
Agree. This is the only way we're going to get the desired reliability here. |
@alnikola can you please help us here? |
FWIW this is happening when building the clr and libraries on the host machine. |
Are the machines supposed to handle the build? |
The failures seem to have started abruptly 10 days ago, which suggests that
one change may have broken it.
…On Tue, 14 Apr 2020, 22:01 Karel Zikmund, ***@***.***> wrote:
Are the machines supposed to handle the build?
Do we use the right design for this workflow?
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
<#34780 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AAVO3M4QKLU354DNXEMPPVDRMTFK5ANCNFSM4ME5ZZDA>
.
|
Actually the build was failing with an unknown switch for linux and the build was not marked as failed
https://dev.azure.com/dnceng/public/_build/results?buildId=587364 What changed the way we build this was: 42183b1#diff-41f10863d38cf298ee01c22c64e1b53a We normally don't use Hosted machines for our builds because of the limitation of space since our build uses around 8 GBs for the artifacts. This build is also producing docker containers which will take up space. |
|
FWIW, the stress pipeline doesn't run automatically when PRs happen. So, any changes to overall build scripts (like removing the |
Apparently, I misunderstood @safern's reply. That wrong argument issue has been fixed by now, so we definitely have to move stress test build to a different agent pool. |
HttpStress and SslStress tests moved off hosted pool to different queues. Note: HttpStress runs are failing but it's actual test code or prod code issue which will be investigated. Infrastructure-wise everything looks good now. Fixes #34780
stress-http and stress-ssl jobs are failing to build due to out of disk space, e.g.:
https://dev.azure.com/dnceng/public/_build/results?buildId=592971&view=logs&j=2d2b3007-3c5c-5840-9bb0-2b1ea49925f3&t=abae1f68-3c73-5bff-491f-f2b908580ce6
https://dev.azure.com/dnceng/public/_build/results?buildId=592970&view=logs&j=2d2b3007-3c5c-5840-9bb0-2b1ea49925f3&t=abae1f68-3c73-5bff-491f-f2b908580ce6
@dotnet/runtime-infrastructure
The text was updated successfully, but these errors were encountered: