Bug 2000100
| Summary: | [GSS][RFE] OCS installation Fails on ipv6 network | ||
|---|---|---|---|
| Product: | [Red Hat Storage] Red Hat OpenShift Data Foundation | Reporter: | Priya Pandey <prpandey> |
| Component: | unclassified | Assignee: | Mudit Agarwal <muagarwa> |
| Status: | CLOSED CURRENTRELEASE | QA Contact: | Vijay Avuthu <vavuthu> |
| Severity: | high | Docs Contact: | |
| Priority: | unspecified | ||
| Version: | 4.8 | CC: | abose, assingh, bkunal, bniver, gmeno, jniu, kramdoss, muagarwa, nsatsia, ocs-bugs, odf-bz-bot, pbunev, rcyriac, rsussman, sapillai, shberry, sostapov, tmuthami |
| Target Milestone: | --- | Keywords: | FutureFeature |
| Target Release: | ODF 4.12.0 | ||
| Hardware: | All | ||
| OS: | Unspecified | ||
| Whiteboard: | |||
| Fixed In Version: | Doc Type: | Enhancement | |
| Doc Text: | Story Points: | --- | |
| Clone Of: | Environment: | ||
| Last Closed: | 2023-02-08 14:06:28 UTC | Type: | Bug |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
|
Description
Priya Pandey
2021-09-01 11:53:33 UTC
Based on request from engineering, the 'installation' component has been deprecated update: ========= > Install IPv6 on BM using below storagecluster - apiVersion: ocs.openshift.io/v1 kind: StorageCluster metadata: annotations: cluster.ocs.openshift.io/local-devices: 'true' name: ocs-storagecluster namespace: openshift-storage spec: flexibleScaling: true manageNodes: false monDataDirHostPath: /var/lib/rook network: ipFamily: IPv6 storageDeviceSets: - count: 3 dataPVCTemplate: spec: accessModes: - ReadWriteOnce resources: requests: storage: 1490Gi storageClassName: localblock volumeMode: Block name: ocs-deviceset-localblock placement: {} portable: false replica: 1 resources: {} > # oc get nodes -o wide NAME STATUS ROLES AGE VERSION INTERNAL-IP EXTERNAL-IP OS-IMAGE KERNEL-VERSION CONTAINER-RUNTIME e26-h15-740xd Ready control-plane,master 10h v1.25.2+5533733 fc00:xxxx::5 <none> Red Hat Enterprise Linux CoreOS 412.86.202212030032-0 (Ootpa) 4.18.0-372.32.1.el8_6.x86_64 cri-o://1.25.1-5.rhaos4.12.git6005903.el8 e26-h17-740xd Ready control-plane,master 10h v1.25.2+5533733 fc00:xxxx::6 <none> Red Hat Enterprise Linux CoreOS 412.86.202212030032-0 (Ootpa) 4.18.0-372.32.1.el8_6.x86_64 cri-o://1.25.1-5.rhaos4.12.git6005903.el8 e26-h19-740xd Ready control-plane,master 10h v1.25.2+5533733 fc00:xxxx::7 <none> Red Hat Enterprise Linux CoreOS 412.86.202212030032-0 (Ootpa) 4.18.0-372.32.1.el8_6.x86_64 cri-o://1.25.1-5.rhaos4.12.git6005903.el8 e26-h21-740xd Ready worker 10h v1.25.2+5533733 fc00:xxxx::8 <none> Red Hat Enterprise Linux CoreOS 412.86.202212030032-0 (Ootpa) 4.18.0-372.32.1.el8_6.x86_64 cri-o://1.25.1-5.rhaos4.12.git6005903.el8 e26-h23-740xd Ready worker 10h v1.25.2+5533733 fc00:xxxx::9 <none> Red Hat Enterprise Linux CoreOS 412.86.202212030032-0 (Ootpa) 4.18.0-372.32.1.el8_6.x86_64 cri-o://1.25.1-5.rhaos4.12.git6005903.el8 e26-h25-740xd Ready worker 10h v1.25.2+5533733 fc00:xxxx::a <none> Red Hat Enterprise Linux CoreOS 412.86.202212030032-0 (Ootpa) 4.18.0-372.32.1.el8_6.x86_64 cri-o://1.25.1-5.rhaos4.12.git6005903.el8 # oc get sc NAME PROVISIONER RECLAIMPOLICY VOLUMEBINDINGMODE ALLOWVOLUMEEXPANSION AGE localblock kubernetes.io/no-provisioner Delete WaitForFirstConsumer false 26m ocs-storagecluster-ceph-rbd openshift-storage.rbd.csi.ceph.com Delete Immediate true 20m ocs-storagecluster-ceph-rgw openshift-storage.ceph.rook.io/bucket Delete Immediate false 23m ocs-storagecluster-cephfs openshift-storage.cephfs.csi.ceph.com Delete Immediate true 20m openshift-storage.noobaa.io openshift-storage.noobaa.io/obc Delete Immediate false 17m $ ceph osd tree ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF -1 4.36647 root default -3 1.45549 host e26-xxx-740xd 0 ssd 1.45549 osd.0 up 1.00000 1.00000 -5 1.45549 host e26-xxx-740xd 2 ssd 1.45549 osd.2 up 1.00000 1.00000 -7 1.45549 host e26-xxx-740xd 1 ssd 1.45549 osd.1 up 1.00000 1.00000 sh-4.4$ ceph -s cluster: id: c1e2c726-238c-4ac1-8594-a07ea3a09165 health: HEALTH_OK services: mon: 3 daemons, quorum a,b,c (age 21m) mgr: a(active, since 20m) mds: 1/1 daemons up, 1 hot standby osd: 3 osds: 3 up (since 20m), 3 in (since 20m) rgw: 1 daemon active (1 hosts, 1 zones) data: volumes: 1/1 healthy pools: 12 pools, 353 pgs objects: 675 objects, 498 MiB usage: 806 MiB used, 4.4 TiB / 4.4 TiB avail pgs: 353 active+clean io: client: 852 B/s rd, 6.0 KiB/s wr, 1 op/s rd, 1 op/s wr |