Bug 1318472
Summary: | Registry pod doesn't mount persistent volume(NFS) after restarting system | ||
---|---|---|---|
Product: | OpenShift Container Platform | Reporter: | Kenjiro Nakayama <knakayam> |
Component: | Storage | Assignee: | Paul Morie <pmorie> |
Status: | CLOSED ERRATA | QA Contact: | Jianwei Hou <jhou> |
Severity: | urgent | Docs Contact: | |
Priority: | urgent | ||
Version: | 3.1.0 | CC: | agoldste, aos-bugs, bleanhar, erjones, knakayam, mbarrett, pmorie |
Target Milestone: | --- | Keywords: | NeedsTestCase |
Target Release: | 3.1.1 | ||
Hardware: | Unspecified | ||
OS: | Unspecified | ||
Whiteboard: | |||
Fixed In Version: | atomic-openshift-3.1.1.6-4.git.32.adf8ec9.el7aos | Doc Type: | Bug Fix |
Doc Text: |
Cause: Persistent Volume Claims were added to the list of volumes to preserve rather than the actual name of the Persistent Volume associated with the Persistent Volume Claim.
Consequence: The periodic cleanup process would unmount the volume if a pod utilizing the Persistent Volume Claim had not yet entered running state.
Fix: The actual name of the Persistent Volume associated with a Persistent Volume Claim is used when determining which volumes can be cleaned up, preventing the cleanup process from considering them orphaned.
Result: Persistent Volumes are no longer unmounted while the pod requiring the volume is starting.
|
Story Points: | --- |
Clone Of: | Environment: | ||
Last Closed: | 2016-03-24 15:54:02 UTC | Type: | Bug |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: |
Description
Kenjiro Nakayama
2016-03-17 01:08:52 UTC
Exit code 32 means the mountpoint is busy or already in use -- what does the mount table look like _after_ you restart the system, but _before_ you restart openshift? This sure looks like https://github.com/kubernetes/kubernetes/issues/20734 I'm working on a backport for the fix of this to 3.1.1 Verified with: [root@openshift-143 ~]# openshift version openshift v3.1.1.6-33-g81eabcc kubernetes v1.1.0-origin-1107-g4c8e6f4 etcd 2.1.2 atomic-openshift-sdn-ovs-3.1.1.6-4.git.32.adf8ec9.el7aos.x86_64 atomic-openshift-3.1.1.6-4.git.32.adf8ec9.el7aos.x86_64 tuned-profiles-atomic-openshift-node-3.1.1.6-4.git.32.adf8ec9.el7aos.x86_64 atomic-openshift-node-3.1.1.6-4.git.32.adf8ec9.el7aos.x86_64 atomic-openshift-clients-3.1.1.6-4.git.32.adf8ec9.el7aos.x86_64 atomic-openshift-master-3.1.1.6-4.git.32.adf8ec9.el7aos.x86_64 Steps: 1. Stop atomic-openshift-node 2. On node where docker-registry is hosted: mount|grep nfs openshift-143.lab.sjc.redhat.com:/var/lib/exports/regpv on /var/lib/origin/openshift.local.volumes/pods/7ff08b81-eff0-11e5-9c0a-fa163e88dd03/volumes/kubernetes.io~nfs/regpv-volume type nfs4 (rw,relatime,vers=4.0,rsize=524288,wsize=524288,namlen=255,hard,proto=tcp,port=0,timeo=600,retrans=2,sec=sys,clientaddr=192.168.1.10,local_lock=none,addr=10.14.6.143) 3. Reboot the node 4. atomic-openshift-node starts after node reboots 5. On node where docker-registry is hosted: mount|grep nfs openshift-143.lab.sjc.redhat.com:/var/lib/exports/regpv on /var/lib/origin/openshift.local.volumes/pods/7ff08b81-eff0-11e5-9c0a-fa163e88dd03/volumes/kubernetes.io~nfs/regpv-volume type nfs4 (rw,relatime,vers=4.0,rsize=524288,wsize=524288,namlen=255,hard,proto=tcp,port=0,timeo=600,retrans=2,sec=sys,clientaddr=192.168.1.10,local_lock=none,addr=10.14.6.143) 6. oc get pods -n default The docker-registry pod is healthy and running Jianwei- Were you able to reproduce the original issue on a build from before the patch went in? Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2016:0510 |