-
Notifications
You must be signed in to change notification settings - Fork 244
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
component push throws error of waited 4m0s but couldn't find running pod matching selector #2877
Comments
similar error with little bit twist - #2942 (comment) |
@mik-dass This issue appears more frequently than before when the test node was 2. I am also assigning you to this issue along with me as you have fixed similar kind of issue before. |
Let's reduce the number of nodes back to 2. |
Not a bad though I suspect the failure might be due to less resources in travis CI. Is there any way to increase the resources in travis while doing |
Hitting this issue more frequently when running tests with
However there are some other consequences also of using older version of minikube like lag of latest feature implementation and I am suspecting this could be one of the reason that Right now we are running our test on travis CI with @kadel @girishramnani WDYT? |
I have raised a ticket asking for more resources on travis CI. Let see what they are replying on that. I will update the same once the reply |
Got the response from Travis CI team The provided default memory 7.5 is enough for running 4 component push in parallel. So i think we should check component push failure from odo end. WDYT @mik-dass ? |
But decreasing the test nodes to 2 has indeed reduced the amount of this failure. Also 7.5 may not be enough as we are running a cluster in the background in most of our tests which can be a expensive operation too. Also the pod initialization step can consume a lot of the time. I would suggest increasing the push timeout value by |
@mik-dass may be you are right, however iirc even on single test node we had the similar issue. Anyway we can try your suggestion to narrow down the reason for failure. |
@prietyc123 Can you please apply @mik-dass suggestion in one of your pr you mentioned in the comment #2877 (comment). You just need to overwrite the |
Sure I will definitely try it out and update the result. |
@mik-dass I have set the pushtimeout to 8 min but still getting the same failure.
More details : https://travis-ci.com/github/openshift/odo/jobs/325397186#L1861 |
There seems to be some issue with the network on travis and most probably the ImagePullBackOff happened because of that https://travis-ci.com/github/openshift/odo/jobs/325397186#L1693 Also TBH I haven't seen this error on most PRs since we switched back to 2 nodes. Even on 4 nodes it happened in 2-4 test scripts. But for your PR #2913 it's happening for all the test scripts. In fact all the jobs on Travis, which run on
Maybe there is some comparability issue with xenial or some problem on Travis side regarding xenial. |
Issues go stale after 90d of inactivity. Mark the issue as fresh by commenting If this issue is safe to close now please do so with /lifecycle stale |
Stale issues rot after 30d of inactivity. Mark the issue as fresh by commenting If this issue is safe to close now please do so with /lifecycle rotten |
/remove-lifecycle rotten |
Issues go stale after 90d of inactivity. Mark the issue as fresh by commenting If this issue is safe to close now please do so with /lifecycle stale |
Stale issues rot after 30d of inactivity. Mark the issue as fresh by commenting If this issue is safe to close now please do so with /lifecycle rotten |
Rotten issues close after 30d of inactivity. Reopen the issue by commenting /close |
@openshift-bot: Closing this issue. In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
/kind flake
What versions of software are you using?
Operating System:
All Supported
Output of
odo version
:master
How did you run odo exactly?
odo push --context context
on OpenShift ci.Actual behavior
Throwing error as:
Expected behavior
It should push the component successfully into the deployment.
Any logs, error output, etc?
For more details: https://prow.svc.ci.openshift.org/view/gcs/origin-ci-test/pr-logs/pull/openshift_odo/2875/pull-ci-openshift-odo-master-v4.1-integration-e2e-benchmark/1778#1:build-log.txt%3A710
The text was updated successfully, but these errors were encountered: