Bug 1726370 - Cluster operator machine-config is reporting a failure: Failed to resync 4.2.0-0.nightly-2019-07-01-102521
Summary: Cluster operator machine-config is reporting a failure: Failed to resync 4.2....
Keywords:
Status: CLOSED ERRATA
Alias: None
Product: OpenShift Container Platform
Classification: Red Hat
Component: Machine Config Operator
Version: 4.2.0
Hardware: Unspecified
OS: Unspecified
medium
high
Target Milestone: ---
: 4.2.0
Assignee: Antonio Murdaca
QA Contact: Micah Abbott
URL:
Whiteboard:
Depends On: 1728223 1728639
Blocks:
TreeView+ depends on / blocked
 
Reported: 2019-07-02 17:07 UTC by Anurag saxena
Modified: 2019-10-16 06:32 UTC (History)
7 users (show)

Fixed In Version:
Doc Type: If docs needed, set a value
Doc Text:
Clone Of:
Environment:
Last Closed: 2019-10-16 06:32:48 UTC
Target Upstream Version:
Embargoed:


Attachments (Terms of Use)


Links
System ID Private Priority Status Summary Last Updated
Red Hat Product Errata RHBA-2019:2922 0 None None None 2019-10-16 06:32:59 UTC

Internal Links: 1767152

Description Anurag saxena 2019-07-02 17:07:43 UTC
Description of problem:Unable to deploy latest 4.2 nightly build. I get these errors every time i try to install 4.2.0-0.nightly-2019-07-01-102521


level=debug msg="Still waiting for the cluster to initialize: Working towards 4.2.0-0.nightly-2019-07-01-102521: 91% complete"
level=debug msg="Still waiting for the cluster to initialize: Working towards 4.2.0-0.nightly-2019-07-01-102521: 92% complete"
level=debug msg="Still waiting for the cluster to initialize: Working towards 4.2.0-0.nightly-2019-07-01-102521: 94% complete"
level=debug msg="Still waiting for the cluster to initialize: Working towards 4.2.0-0.nightly-2019-07-01-102521: 95% complete"
level=debug msg="Still waiting for the cluster to initialize: Working towards 4.2.0-0.nightly-2019-07-01-102521: 98% complete"
level=debug msg="Still waiting for the cluster to initialize: Working towards 4.2.0-0.nightly-2019-07-01-102521: 98% complete, waiting on authentication, machine-config, monitoring"
level=debug msg="Still waiting for the cluster to initialize: Working towards 4.2.0-0.nightly-2019-07-01-102521: 99% complete"
level=debug msg="Still waiting for the cluster to initialize: Cluster operator machine-config is reporting a failure: Failed to resync 4.2.0-0.nightly-2019-07-01-102521 because: timed out waiting for the condition during syncRequiredMachineConfigPools: pool master has not progressed to latest configuration: configuration status for pool master is empty, retrying"
level=debug msg="Still waiting for the cluster to initialize: Working towards 4.2.0-0.nightly-2019-07-01-102521: 99% complete"
level=debug msg="Still waiting for the cluster to initialize: Cluster operator machine-config is reporting a failure: Failed to resync 4.2.0-0.nightly-2019-07-01-102521 because: timed out waiting for the condition during syncRequiredMachineConfigPools: pool master has not progressed to latest configuration: configuration status for pool master is empty, retrying"
level=fatal msg="failed to initialize the cluster: Cluster operator machine-config is reporting a failure: Failed to resync 4.2.0-0.nightly-2019-07-01-102521 because: timed out waiting for the condition during syncRequiredMachineConfigPools: pool master has not progressed to latest configuration: configuration status for pool master is empty, retrying"


Version-Release number of selected component (if applicable):4.2.0-0.nightly-2019-07-01-102521


How reproducible: Always for me


Steps to Reproduce:
1.Install 4.2.0-0.nightly-2019-07-01-102521
2.
3.

Actual results: Installation aborts due to above said error


Expected results: Installation should succeed


Additional info: Seems like routes were not setup before it froze so there seems no way to login to the cluster

$oc login xx.xx.xx.xx:6443 -u kubeadmin -p WJbvg-SRuah-UnsD8-tCJoV --insecure-skip-tls-verify=true

error: dial tcp 10.0.142.132:6443: i/o timeout - verify you have provided the correct host and port and that the server is currently running.

Unable to ping, $ ping api.qe-anusaxen-ocp42.qe.devcluster.openshift.com

Comment 1 Antonio Murdaca 2019-07-02 18:08:26 UTC
can you provide must-gather (oc adm must-gather) - that's gonna shed some light on this.

Comment 2 Anurag saxena 2019-07-02 18:38:10 UTC
(In reply to Antonio Murdaca from comment #1)
> can you provide must-gather (oc adm must-gather) - that's gonna shed some
> light on this.

I can try again for that but as i mentioned in additional info above, i am not able to go inside the cluster for now

Comment 7 Antonio Murdaca 2019-07-03 07:40:32 UTC
Likely https://bugzilla.redhat.com/show_bug.cgi?id=1725478 - can you retest with a payload newer than 07/01?

Comment 8 Anurag saxena 2019-07-03 12:56:39 UTC
(In reply to Antonio Murdaca from comment #7)
> Likely https://bugzilla.redhat.com/show_bug.cgi?id=1725478 - can you retest
> with a payload newer than 07/01?

Hi Antonia, newer nighltly payloads seems all red (Rejected)

Comment 9 Antonio Murdaca 2019-07-03 12:59:15 UTC
(In reply to Anurag saxena from comment #8)
> (In reply to Antonio Murdaca from comment #7)
> > Likely https://bugzilla.redhat.com/show_bug.cgi?id=1725478 - can you retest
> > with a payload newer than 07/01?
> 
> Hi Antonia, newer nighltly payloads seems all red (Rejected)

There are some CI issues

Comment 10 Antonio Murdaca 2019-07-03 13:02:42 UTC
Moving to modified once a new payload is green to retest

Comment 12 Anurag saxena 2019-07-03 19:09:45 UTC
Update - Cluster configured correctly and seemed healthy on Red one 4.2.0-0.nightly-2019-07-03-082520 which came after that. SO will wait for green once to verify this issue. Thanks

Comment 13 Anurag saxena 2019-07-09 09:25:30 UTC
latest Update: No green build yet on 4.2, but the cluster seems healthy on recent 4.2.0-0.nightly-2019-07-08-142835 which is a Rejected one.
Have to wait for a green build to confirm the whether issue still persist

Comment 14 Xingxing Xia 2019-07-11 01:06:48 UTC
I failed to install latest 4.2 payload env due to bug 1728639 . Then I tried to install earlier 4.2 payload env, also met this bug. Bug 1728639 blocks this bug's verification, agreeing with above comment.

Comment 15 Anurag saxena 2019-07-12 09:42:20 UTC
Seems okay on green builds, no issues on 4.2.0-0.nightly-2019-07-11-023129

Comment 16 errata-xmlrpc 2019-10-16 06:32:48 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory, and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHBA-2019:2922


Note You need to log in before you can comment on or make changes to this bug.