Bug 1449097
| Summary: | Pod is terminated, but some volumes have not been cleaned up | ||
|---|---|---|---|
| Product: | OpenShift Container Platform | Reporter: | Xia Zhao <xiazhao> |
| Component: | Node | Assignee: | Seth Jennings <sjenning> |
| Status: | CLOSED WONTFIX | QA Contact: | DeShuai Ma <dma> |
| Severity: | medium | Docs Contact: | |
| Priority: | medium | ||
| Version: | 3.6.0 | CC: | aos-bugs, decarr, eparis, jokerman, jpeeler, mmccomas, nbhatt, pmorie, sjenning, vgoyal, xiazhao |
| Target Milestone: | --- | Keywords: | Reopened |
| Target Release: | --- | ||
| Hardware: | Unspecified | ||
| OS: | Unspecified | ||
| Whiteboard: | |||
| Fixed In Version: | Doc Type: | If docs needed, set a value | |
| Doc Text: | Story Points: | --- | |
| Clone Of: | Environment: | ||
| Last Closed: | 2018-05-07 16:45:01 UTC | Type: | Bug |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
|
Description
Xia Zhao
2017-05-09 08:35:36 UTC
Is this something you can reproduce? (Or still have it sticking around?) Can you run https://github.com/rhvgoyal/misc/blob/master/find-busy-mnt.sh on the node where the pod was running? For this particular case I think you could run: find-busy-mnt.sh 813cefa3-3483-11e7-b539-fa163ee9df62 But if you reproduce with another pod the string would be different. Hopefully that will tell us what process is holding this mount point... The root cause here seems to be a busy mountpoint: nestedpendingoperations.go:262] Operation for "\"kubernetes.io/secret/813cefa3-3483-11e7-b539-fa163ee9df62-aggregated-logging-fluentd-token-d4c7h\" (\"813cefa3-3483-11e7-b539-fa163ee9df62\")" failed. No retries permitted until 2017-05-09 07:52:34.252695141 +0000 UTC (durationBeforeRetry 2m0s). Error: UnmountVolume.TearDown failed for volume "kubernetes.io/secret/813cefa3-3483-11e7-b539-fa163ee9df62-aggregated-logging-fluentd-token-d4c7h" (volume.spec.Name: "aggregated-logging-fluentd-token-d4c7h") pod "813cefa3-3483-11e7-b539-fa163ee9df62" (UID: "813cefa3-3483-11e7-b539-fa163ee9df62") with: rename /var/lib/origin/openshift.local.volumes/pods/813cefa3-3483-11e7-b539-fa163ee9df62/volumes/kubernetes.io~secret/aggregated-logging-fluentd-token-d4c7h /var/lib/origin/openshift.local.volumes/pods/813cefa3-3483-11e7-b539-fa163ee9df62/volumes/kubernetes.io~secret/wrapped_aggregated-logging-fluentd-token-d4c7h.deleting~053939955: device or resource busy notice: device or resource busy I filed https://bugzilla.redhat.com/show_bug.cgi?id=1450426 as a request for the docker team to add logging to get us more data when this happens. For now, there is not much we can do here unless you can reproduce and/or collect the data requested in comment #3. I'm going to close this as WONTFIX. We know about some busy mount issues. I can't immediately identify which this likely was, but if we reproduce please reopen or open a new bug. Thank you! I'm sorry we were so slow to try to collect more data. |