Bug 1857114
Summary: | [ROKS] PVC from OCS on IBM cloud does not mount to pod | ||
---|---|---|---|
Product: | [Red Hat Storage] Red Hat OpenShift Container Storage | Reporter: | Elvir Kuric <ekuric> |
Component: | ocs-operator | Assignee: | Jose A. Rivera <jarrpa> |
Status: | CLOSED DUPLICATE | QA Contact: | Raz Tamir <ratamir> |
Severity: | unspecified | Docs Contact: | |
Priority: | unspecified | ||
Version: | 4.3 | CC: | hchiramm, jijoy, madam, mrajanna, muagarwa, ocs-bugs, pbalogh, sostapov |
Target Milestone: | --- | ||
Target Release: | --- | ||
Hardware: | Unspecified | ||
OS: | Unspecified | ||
Whiteboard: | |||
Fixed In Version: | Doc Type: | If docs needed, set a value | |
Doc Text: | Story Points: | --- | |
Clone Of: | Environment: | ||
Last Closed: | 2020-07-16 13:35:16 UTC | Type: | Bug |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: |
Description
Elvir Kuric
2020-07-15 07:19:11 UTC
@Elvir please provide logs for ceph-csi plugins (rbd plugins) and also the definition of ibmc-block-gold storage class. (In reply to Mudit Agarwal from comment #3) > @Elvir please provide logs for ceph-csi plugins (rbd plugins) and also the > definition of ibmc-block-gold storage class. q: Is this deployment using a custom KUBELET path for ceph-csi? a: not sure, IBM team installed cluster q: --root-dir is /var/data/kubelet , check [1] below # oc get sc ibmc-block-gold -o yaml allowVolumeExpansion: true apiVersion: storage.k8s.io/v1 kind: StorageClass metadata: annotations: storageclass.kubernetes.io/is-default-class: "true" creationTimestamp: "2020-07-10T10:13:07Z" labels: app: ibmcloud-block-storage-plugin chart: ibmcloud-block-storage-plugin-1.7.1 heritage: Helm release: release-name name: ibmc-block-gold resourceVersion: "3754" selfLink: /apis/storage.k8s.io/v1/storageclasses/ibmc-block-gold uid: a13ff46d-192b-4848-a304-36b717679954 parameters: billingType: hourly classVersion: "2" fsType: ext4 iopsPerGB: "10" sizeRange: '[20-4000]Gi' type: Endurance provisioner: ibm.io/ibmc-block reclaimPolicy: Delete volumeBindingMode: Immediate --- # oc logs csi-rbdplugin-l2cpx -c driver-registrar I0713 22:18:02.263705 32109 main.go:110] Version: v4.3.27-202006211650.p0-0-g23f9061-dirty I0713 22:18:02.264000 32109 main.go:120] Attempting to open a gRPC connection with: "/csi/csi.sock" I0713 22:18:02.264036 32109 connection.go:151] Connecting to unix:///csi/csi.sock I0713 22:18:03.264721 32109 main.go:127] Calling CSI driver to discover driver name I0713 22:18:03.264761 32109 connection.go:180] GRPC call: /csi.v1.Identity/GetPluginInfo I0713 22:18:03.264767 32109 connection.go:181] GRPC request: {} I0713 22:18:03.268470 32109 connection.go:183] GRPC response: {"name":"openshift-storage.rbd.csi.ceph.com","vendor_version":"release-4.4"} I0713 22:18:03.269055 32109 connection.go:184] GRPC error: <nil> I0713 22:18:03.269080 32109 main.go:137] CSI driver name: "openshift-storage.rbd.csi.ceph.com" I0713 22:18:03.269126 32109 node_register.go:58] Starting Registration Server at: /registration/openshift-storage.rbd.csi.ceph.com-reg.sock I0713 22:18:03.269279 32109 node_register.go:67] Registration Server started at: /registration/openshift-storage.rbd.csi.ceph.com-reg.sock I0713 22:18:03.402840 32109 main.go:77] Received GetInfo call: &InfoRequest{} I0713 22:18:04.403358 32109 main.go:77] Received GetInfo call: &InfoRequest{} I0713 22:18:04.533539 32109 main.go:87] Received NotifyRegistrationStatus call: &RegistrationStatus{PluginRegistered:true,Error:,} Must gather and csi pod log : http://jmencak-pub.usersys.redhat.com/ekuric/rooks_bz/ [1] cat /usr/lib/systemd/system/kubelet.service [Unit] After=decrypt-docker.service Requires=decrypt-docker.service Description=Kubernetes Kubelet Documentation=https://github.com/kubernetes/kubernetes After=network.target auditd.service [Service] ExecStartPre=/sbin/swapoff -a ExecStartPre=/bin/systemctl stop -f haproxy.service ExecStartPre=-/usr/local/sbin/create-localproxy-netns.sh ExecStart=/usr/bin/hyperkube kubelet \ --root-dir=/var/data/kubelet \ --enable-controller-attach-detach=false \ --cgroup-driver=systemd \ --provider-id=ibm://a068244bebefc19e34e92d445a8504f3///bs43q5qd0d2ev8f1d8bg/kube-bs43q5qd0d2ev8f1d8bg-perfocs-default-000003f4 \ --cloud-provider=external \ --cluster-dns=172.21.0.10 \ --cluster-domain=cluster.local \ \ --feature-gates=ExpandInUsePersistentVolumes=true,LegacyNodeRoleBehavior=false,NodeDisruptionExclusion=false,ServiceNodeExclusion=false,SCTPSupport=false \ \ --pod-manifest-path=/etc/kubernetes/manifests \ --kubeconfig=/etc/kubernetes/kubelet-kubeconfig \ \ --max-pods=250 \ \ --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305 \ --v=2 \ --file-check-frequency=5s \ "--hostname-override=10.208.42.210" \ "--anonymous-auth=false" \ "--client-ca-file=/etc/kubernetes/cert/ca.pem" \ "--read-only-port=0" \ --network-plugin=cni --cni-conf-dir=/etc/kubernetes/cni/net.d/ --cni-bin-dir=/var/lib/cni/bin \ --tls-cert-file=/etc/kubernetes/cert/kubelet.pem \ --tls-private-key-file=/etc/kubernetes/cert/kubelet-key.pem \ --authorization-mode=Webhook \ --authentication-token-webhook \ --container-runtime=remote \ --runtime-request-timeout=15m \ --container-runtime-endpoint=/var/run/crio/crio.sock \ --feature-gates=CRIContainerLogRotation=true \ --container-log-max-size=100Mi \ --container-log-max-files=3 \ --streaming-connection-idle-timeout=30m \ --event-qps=0 \ --kube-reserved-cgroup=/podruntime.slice \ --system-reserved-cgroup=/system.slice \ --pod-max-pids=228748 \ --kube-reserved=memory=2197Mi,cpu=64m,pid=12708 \ --system-reserved=memory=3295Mi,cpu=96m,pid=12708 \ --kubelet-cgroups=/podruntime.slice \ --runtime-cgroups=/podruntime.slice \ --enforce-node-allocatable=pods \ --kube-api-qps=20 \ --kube-api-burst=40 \ --eviction-soft=memory.available<100Mi,nodefs.available<10%,imagefs.available<10%,nodefs.inodesFree<10%,imagefs.inodesFree<10% \ --eviction-soft-grace-period=memory.available=10m,nodefs.available=10m,imagefs.available=10m,nodefs.inodesFree=10m,imagefs.inodesFree=10m \ --eviction-hard=memory.available<100Mi,nodefs.available<5%,imagefs.available<5%,nodefs.inodesFree<5%,imagefs.inodesFree<5% Restart=always RestartSec=5 TimeoutStartSec=15 SyslogIdentifier=kubelet.service [Install] WantedBy=multi-user.target Elvir, the csi plugin logs are not useful because they don't capture the time when the issue occured. But looks like the issue is with the KUBELET path as Madhu mentioned, by default ceph csi uses "/var/lib/kubelet" but here we are using "/var/data/kubelet". We can confirm the same by looking at the complete plugin logs, OCS4.5 onwards this path is configurable via rook-cephcsi config file see https://github.com/openshift/ocs-operator/issues/454 For now you have to define the complete path in mountPath field of the pod. @Madhu, please add/correct if I have missed anything. So why isn't it closed as dup? *** This bug has been marked as a duplicate of bug 1801365 *** (In reply to Yaniv Kaul from comment #13) > So why isn't it closed as dup? > > *** This bug has been marked as a duplicate of bug 1801365 *** Waited for someone from ocs-operator to confirm. |