Bug 1767515
| Summary: | Failed migrations can leave stale PVCs in target namespace | ||
|---|---|---|---|
| Product: | Migration Toolkit for Containers | Reporter: | Dylan Murray <dymurray> |
| Component: | General | Assignee: | Derek Whatley <dwhatley> |
| Status: | CLOSED CURRENTRELEASE | QA Contact: | Xin jiang <xjiang> |
| Severity: | medium | Docs Contact: | Avital Pinnick <apinnick> |
| Priority: | medium | ||
| Version: | 1.3.0 | CC: | ernelson, jmatthew, sseago |
| Target Milestone: | --- | ||
| Target Release: | 1.6.0 | ||
| Hardware: | Unspecified | ||
| OS: | Unspecified | ||
| Whiteboard: | |||
| Fixed In Version: | Doc Type: | If docs needed, set a value | |
| Doc Text: | Story Points: | --- | |
| Clone Of: | Environment: | ||
| Last Closed: | 2021-08-24 18:52:57 UTC | Type: | Bug |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
|
Description
Dylan Murray
2019-10-31 15:56:53 UTC
More generally, failed restores can leave other resources around too, any of which could cause subsequent failures. Should we be adding a label to *everything* we restore in the general plugin and then, either automatically when there's an explicit failure, or manually if the migration stalls, delete all resources in the target namespaces (and all cluster resources) with this label? Let's verify this as of 1.6.0, we have cleanup steps that are run that should ensure these resources are not leaked. I am not able to reproduce this failure with the given reproducer steps. I tried: --- 1. Adjusting source cluster stage pod image. This caused migration to fail before anything was created on the target cluster, so doesn't count as reproduced. migrationcontroller.spec.migration_stage_image_fqin: migration_stage_image_fqin: quay.io/djwhatle/fake:latest --- 2. Adjusting target cluster stage pod image. This caused had no effect on migration success in my test. Presumably because the working stage pod image from the source cluster is used. Is there a way to provide a bad stage pod image to the target cluster? migrationcontroller.spec.migration_stage_image_fqin: migration_stage_image_fqin: quay.io/djwhatle/fake:latest This appears to have been fixed by the stage pod cleanup code. Was not able to reproduce. https://github.com/konveyor/mig-controller/pull/738 |