-
Notifications
You must be signed in to change notification settings - Fork 297
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fix issue where VRT tests fail before all screenshots are captured #7965
Comments
Not seeing what can be the cause, since I this seems to be CI related, locally it doesn't fail. Unassigning myself |
@techanvil pinging you here after many days of testing, tweaking, testing some more (because you are the issue creator and may have more insight). First off, there are not a ton of documented cases of this happening online. For the cases that do arise, the solutions vary, but the common suggestions and/or reasons include:
Another topic that came up a few times was that of disk availability. To test this theory and whether it was indeed a disk space issue, I updated the When I made the above changes, I saw an instant improvement in the occurrence of this error. In fact, for many runs and re-runs, I simply was not able to observe the error again. Unfortunately, it did return every now and then but far and few between (I ran many many test runs). What I did however observe is that when I run the VRT workflow before around 9am my time, which I'd argue is when most of the US and EU tech industries are still relatively inactive compared to business hours, I had a 100% success rate. This could simply be a coincidence, but the later on in the day that I (re) ran the workflow, the more likely I was to encounter the error again, although arguably still less frequently than before. Using the same Linux config in local Docker based testing, I was never able to reproduce the issue, but again this could too be a coincidence. This leaves me to conclude, and really with not much support or information, that this is likely a GitHub resource availability or usage issue, as the workers are shared. My suggestion would be to apply the following changes and see if we see a general decrease in the occurrence:
Beyond that, I would not discount the fact that maybe this is only happening on the Alpine Linux image? We could try with another distro to test this theory. Thoughts? |
Thanks @10upsimon, this sounds like a tricky one to diagnose so nice work on the R&D so far. TL;DR: Let's try what you've suggested for now, see how we get on, and refine things/diagnose further in a separate issue. As mentioned, nice work so far. The fact you've seen an apparent improvement makes me think we should take the pragmatic option of applying these changes in the hope we can immediately improve our CI reliability and thus save time in the development lifecycle. That said, digging into the specific changes, it's a bit of an assortment of seemingly unconnected tweaks. Did you do enough testing of each in isolation to be sure they are all having an effect? At present it's not clear to me whether each of them is strictly necessary, and it would be nice to be more certain that we are only applying changes with a quantifiable effect. For example, I'm not sure the GH Action VM actually has a GPU for The results are also a bit unquantified in terms of how much of a difference it makes. It's good you've seen an apparent positive change, but we could standardise our testing methods for some clearer results. What I'd suggest, as mentioned is to apply these changes for now, but introduce a followup issue where we do something along these lines:
With the above testing framework in place we'd be at liberty to test ideas that occur to us in a relatively time-efficient manner. The scripts that come out of it could be useful for future GH action debugging too. How does that sound to you? |
Removing next up label for now after discussion with the team it seems more thought and investigation is needed on this one. |
Thanks both. FWIW the site-kit-wp/tests/backstop/config.js Line 30 in 97816cc
IB LGTM, but we do need an estimate here @techanvil. Looking at the changes it should be very quick to do. The suggested testing of the workflow sounds good too but this should be tracked in an internal task.
Agreed – all the other GH workflows can be disabled to avoid unnecessary GHA activity
We can do this with an added workflow that runs using the The new workflow should be able to collect information about the run and write that to a file that could be committed back into a file since the repo is for debugging anyways. Analysis can be done later by copying the data over into a spreadsheet, etc. |
Thanks @aaemnnosttv, sorry I forgot to estimate 🤦 This one shouldn't take long, I've estimated it as a 7 to accommodate our estimation bump "experiment".
I had already created #8998 as a followup issue for this, would you prefer this to be migrated to an Asana task? |
Yeah I think so, particularly if we're going to get into testing on a fork and all. We could keep that issue as a placeholder for changes that will come out of the investigation though since it doesn't have much going on there yet. IB ✅ |
Thanks @aaemnnosttv, SGTM. I've created an Asana task for the investigative side and updated #8998 to be a placeholder for applying the findings. |
moving to approval since there is nothing to QA |
VRT now seems to run consistently although it is consistently failing one scenario in mobile. This should be fixed in #9101 . ![]() |
Bug Description
VRT runs in CI are intermittently failing before all the screenshots are captured, with the following error:
Steps to reproduce
Screenshots
This error has been seen with a couple of permutations - note there is an additional log message in the 2nd screenshot.
Do not alter or remove anything below. The following sections will be managed by moderators only.
Acceptance criteria
Implementation Brief
Dockerfile
to include manual installation of theopenrc
anddbus
APK packages.--no-sandbox
,-disable-gpu
,--disable-setuid-sandbox
,--no-zygote
.Test Coverage
QA Brief
Changelog entry
The text was updated successfully, but these errors were encountered: