Bug 1778918
| Summary: | A failure in restic does not fail the migration | ||
|---|---|---|---|
| Product: | OpenShift Container Platform | Reporter: | Erik Nelson <ernelson> |
| Component: | Migration Tooling | Assignee: | Scott Seago <sseago> |
| Status: | CLOSED ERRATA | QA Contact: | Xin jiang <xjiang> |
| Severity: | medium | Docs Contact: | |
| Priority: | medium | ||
| Version: | 4.2.0 | CC: | chezhang, dymurray, ernelson, jmatthew, rpattath, sregidor, xjiang |
| Target Milestone: | --- | ||
| Target Release: | 4.4.0 | ||
| Hardware: | Unspecified | ||
| OS: | Unspecified | ||
| Whiteboard: | |||
| Fixed In Version: | Doc Type: | If docs needed, set a value | |
| Doc Text: | Story Points: | --- | |
| Clone Of: | 1776397 | Environment: | |
| Last Closed: | 2020-05-28 11:09:55 UTC | Type: | --- |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
| Bug Depends On: | |||
| Bug Blocks: | 1776397 | ||
|
Description
Erik Nelson
2019-12-02 19:21:00 UTC
There is a PR upstream that should resolve this: https://github.com/vmware-tanzu/velero/pull/2201 Will update the bug once we bring this into our fork. The upstream PR was included in Velero 1.3.1 Verified in CAM 1.2 stage (4.1 -> 4.4)
Following the steps to reproduce the problem, we got this error in the velero Restore logs
ime="2020-05-07T16:31:08Z" level=error msg="unable to successfully complete restic restores of pod's volumes" error="pod volume restore failed: error restoring volume: error creating .velero directory for done file: mkdir /host_pods/e6803556-0d07-4094-b345-a99ba8b63e79/volumes/kubernetes.io~csi/pvc-ae1b6167-d9eb-4279-bfbc-3e42e8681c40/mount/.velero: no space left on device" logSource="pkg/restore/restore.go:1287" restore=openshift-migration/db76daa0-907e-11ea-83e3-37726d4749fc-rr4bt
And the migmigration resource failed with this status:
status:
conditions:
- category: Advisory
durable: true
lastTransitionTime: "2020-05-07T16:22:41Z"
message: '[1] Stage pods created.'
status: "True"
type: StagePodsCreated
- category: Warn
durable: true
lastTransitionTime: "2020-05-07T16:31:08Z"
message: There were errors found in 1 Restic volume restores. See restore `db76daa0-907e-11ea-83e3-37726d4749fc-rr4bt`
for details
status: "True"
type: ResticErrors
- category: Advisory
durable: true
lastTransitionTime: "2020-05-07T16:31:08Z"
message: 'The migration has failed. See: Errors.'
reason: StageRestoreFailed
status: "True"
type: Failed
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHEA-2020:2326 |