Bug 2117038
| Summary: | With multiple batches, if first batch timed out, then policies for second batch will not be created | ||
|---|---|---|---|
| Product: | OpenShift Container Platform | Reporter: | OpenShift BugZilla Robot <openshift-bugzilla-robot> |
| Component: | Telco Edge | Assignee: | Steven Skeard <sskeard> |
| Telco Edge sub component: | TALO | QA Contact: | yliu1 |
| Status: | CLOSED ERRATA | Docs Contact: | |
| Severity: | high | ||
| Priority: | unspecified | CC: | ijolliff, jun, keyoung |
| Version: | 4.11 | ||
| Target Milestone: | --- | ||
| Target Release: | 4.11.z | ||
| Hardware: | Unspecified | ||
| OS: | Unspecified | ||
| Whiteboard: | |||
| Fixed In Version: | Doc Type: | If docs needed, set a value | |
| Doc Text: | Story Points: | --- | |
| Clone Of: | Environment: | ||
| Last Closed: | 2022-09-27 08:10:20 UTC | Type: | --- |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
| Bug Depends On: | 2115480 | ||
| Bug Blocks: | 2108639, 2117228 | ||
|
Description
OpenShift BugZilla Robot
2022-08-09 19:57:07 UTC
I can still reproduce this issue with latest brew build. In following CGU, you can see the timeout is 18, but second batch started 24 minutes after cgu started. The brew build used is this one: https://brewweb.engineering.redhat.com/brew/buildinfo?buildID=2118287, which was built on Aug 10th. Given the PR was merged on the Aug 9th. (Unless there is a build issue...) [kni ~]$ oc get cgu -o yaml apiVersion: v1 items: - apiVersion: ran.openshift.io/v1alpha1 kind: ClusterGroupUpgrade metadata: annotations: kubectl.kubernetes.io/last-applied-configuration: | {"apiVersion":"ran.openshift.io/v1alpha1","kind":"ClusterGroupUpgrade","metadata":{"annotations":{},"name":"test","namespace":"default"},"spec":{"backup":false,"clusterSelector":["group-du-sno"],"enable":true,"managedPolicies":["common-config-policy"],"preCaching":false,"remediationStrategy":{"maxConcurrency":1,"timeout":18}}} creationTimestamp: "2022-08-11T17:11:02Z" finalizers: - ran.openshift.io/cleanup-finalizer generation: 2 name: test namespace: default resourceVersion: "72544555" uid: eb0f56e7-e8a1-4a7a-ade9-2fde7b1d4fa9 spec: actions: afterCompletion: deleteObjects: true beforeEnable: {} backup: false clusterSelector: - group-du-sno enable: true managedPolicies: - common-config-policy preCaching: false remediationStrategy: maxConcurrency: 1 timeout: 18 status: computedMaxConcurrency: 1 conditions: - lastTransitionTime: "2022-08-11T17:11:02Z" message: The ClusterGroupUpgrade CR policies are taking too long to complete reason: UpgradeTimedOut status: "False" type: Ready copiedPolicies: - test-common-config-policy-jmb4j managedPoliciesContent: common-config-policy: "null" managedPoliciesForUpgrade: - name: common-config-policy namespace: ztp-common managedPoliciesNs: common-config-policy: ztp-common placementBindings: - test-common-config-policy-placement-fr9th placementRules: - test-common-config-policy-placement-fr9th remediationPlan: - - ocp-edge87 - - ocp-edge88 safeResourceNames: test-common-config-policy: test-common-config-policy-jmb4j test-common-config-policy-config: test-common-config-policy-config-9wzxk test-common-config-policy-placement: test-common-config-policy-placement-fr9th status: currentBatch: 2 currentBatchRemediationProgress: ocp-edge88: policyIndex: 0 state: InProgress currentBatchStartedAt: "2022-08-11T17:36:02Z" startedAt: "2022-08-11T17:11:02Z" kind: List metadata: resourceVersion: "" Verified on latest 4.11 TALM. Second batch started at expected time.
spec:
actions:
afterCompletion:
deleteObjects: true
beforeEnable: {}
backup: false
clusters:
- spoke-1
- spoke-2
enable: true
managedPolicies:
- du-upgrade-cluster-version-policy1
preCaching: false
remediationStrategy:
maxConcurrency: 1
timeout: 18
status:
computedMaxConcurrency: 1
conditions:
- lastTransitionTime: "2022-09-09T18:22:34Z"
message: The ClusterGroupUpgrade CR has upgrade policies that are still non compliant
reason: UpgradeNotCompleted
status: "False"
type: Ready
copiedPolicies:
- spoke-2-du-upgrade-cluster-version-policy1-hg4s7
managedPoliciesForUpgrade:
- name: du-upgrade-cluster-version-policy1
namespace: ztp-upgrade
managedPoliciesNs:
du-upgrade-cluster-version-policy1: ztp-upgrade
placementBindings:
- spoke-2-du-upgrade-cluster-version-policy1-placement-t4625
placementRules:
- spoke-2-du-upgrade-cluster-version-policy1-placement-t4625
remediationPlan:
- - spoke-1
- - spoke-2
safeResourceNames:
spoke-2-du-upgrade-cluster-version-policy1: spoke-2-du-upgrade-cluster-version-policy1-hg4s7
spoke-2-du-upgrade-cluster-version-policy1-config: spoke-2-du-upgrade-cluster-version-policy1-config-nrls5
spoke-2-du-upgrade-cluster-version-policy1-placement: spoke-2-du-upgrade-cluster-version-policy1-placement-t4625
status:
currentBatch: 2
currentBatchRemediationProgress:
spoke-2:
policyIndex: 0
state: InProgress
currentBatchStartedAt: "2022-09-09T18:32:35Z"
startedAt: "2022-09-09T18:22:34Z"
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (OpenShift Container Platform 4.11.6 CNF vRAN extras update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHEA-2022:6669 |