-
Notifications
You must be signed in to change notification settings - Fork 4.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Wasm-opt can take way too long #100424
Comments
@sbomer Looks like this didn't work: I can't find the string in the test log. |
Based on the logs in one of my PRs: https://helixre107v0xdcypoyl9e7f.blob.core.windows.net/dotnet-runtime-refs-pull-100419-merge-91eb0ec8650642f7b5/WasmTestOnBrowser-System.Runtime.InteropServices.JavaScript.Tests/1/console.cb20659c.log?helixlogtype=result Maybe |
Actually I think this is a duplicate of #99888. Build analysis correctly picked that one up for my PR. |
This is not about the test failure in "browser-wasm linux Release LibraryTests_Threading" (build analysis identified that as #99888), but about the product build failure in "browser-wasm linux Release LibraryTests": |
Given the way it is failing this is probably an infrastructure issue |
Seems to show that this is a runtime specific issue. It seems like instances are consistently happening in the wasm legs (along with catching a few other stray disconnects) (Maybe you can try to change the error message to |
Thanks for the suggestion, didn't seem to work. :( |
I'll note that when I watched the stdout of the build step for one of my PRs, it was building, even after 2 hours. So it seems like the builds for this lane are extremely slow. For comparison, the windows equivalent lane seems to build in around 40 minutes. |
#99888 is during test run on helix not related to this. This is build/agent issue. Probably infra. This is before the agent was killed, something is terribly slow The running step is CPU+memory heavy, maybe the agent is also swapping to disk. This problem is blocking me |
How would it look like if the agent failed with OOM ? |
The cancellations are very strange Steve and I watched it happen in a recent run. cc @steveisok |
From the engineering services side, we don't have a lot more details on what is going on. While I try to track some information on what is happening to these agents and why these workloads would break them in such a way, I've reverted the version of the Whether that helps or not will be a useful data point. |
Possibly related:
Does this indicate OOM? |
I merged #100517 to reduce/avoid the problem |
I'm seeing this job time out during the "Build product" step, in a way that doesn't get reported to GitHub. On GitHub it looks like the job is still running forever.
Hit in multiple PRs, for example:
Build Information
Build: https://dev.azure.com/dnceng-public/public/_build/results?buildId=622511&view=logs&jobId=63c2d0c8-fec2-5788-81c8-f3ac95e8841f
Build error leg or test failing: browser-wasm linux Release LibraryTests Build product
Error Message
Fill the error message using step by step known issues guidance.
Known issue validation
Build: 🔎 https://dev.azure.com/dnceng-public/public/_build/results?buildId=622511
Error message validated:
[Agent was purged, cancelling the pipeline
]Result validation: ❌ Known issue did not match with the provided build.
Validation performed at: 4/1/2024 10:15:39 PM UTC
Report
Summary
The text was updated successfully, but these errors were encountered: