Bug 1878801 - Hooks using default image are failing
Summary: Hooks using default image are failing
Keywords:
Status: CLOSED ERRATA
Alias: None
Product: OpenShift Container Platform
Classification: Red Hat
Component: Migration Tooling
Version: 4.5
Hardware: Unspecified
OS: Unspecified
unspecified
medium
Target Milestone: ---
: 4.5.0
Assignee: Jason Montleon
QA Contact: Xin jiang
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2020-09-14 14:33 UTC by Sergio
Modified: 2020-09-30 18:43 UTC (History)
6 users (show)

Fixed In Version:
Doc Type: No Doc Update
Doc Text:
Clone Of:
Environment:
Last Closed: 2020-09-30 18:43:07 UTC
Target Upstream Version:
Embargoed:


Attachments (Terms of Use)


Links
System ID Private Priority Status Summary Last Updated
Red Hat Product Errata RHBA-2020:4148 0 None None None 2020-09-30 18:43:19 UTC

Description Sergio 2020-09-14 14:33:26 UTC
Description of problem:
When we configure a hook using the default image to execute the hooks, CMT cannot execute the hooks and the migration is stuck forever.


Version-Release number of selected component (if applicable):
CMT 1.3
SOURCE CLUSTER: azure OCP 4.2
TARGET CLUSTER: azure OCP 4.5

How reproducible:
Always

Steps to Reproduce:
1. Create a namespace
   oc new-project bztest

2. Create a migration plan for this namespace

3. Add a hook for this migration (no matter which hook, or which playbook). Use the default "Ansible runtime image".

4. Run the migration


Actual results:
In the namespace used to run the hook, we find that the pods

$ oc get pods
NAME                          READY   STATUS                 RESTARTS   AGE
hooks-prebackup-54mpd-pkx7f   0/1     CreateContainerError   0          37m
hooks-prebackup-54mpd-rr4qh   0/1     CreateContainerError   0          7m42s

Looking the those pods' logs we find this messages

  Warning  Failed  35m  kubelet, cam-src-8320-v2ksq-worker-northcentralus-67q7l  Error: container create failed: time="2020-09-14T12:05:45Z" level=error msg="container_linux.go:349: starting container process caused \"exec: \\\"/bin/entrypoint\\\": stat /bin/entrypoint: no such file or directory\""
container_linux.go:349: starting container process caused "exec: \"/bin/entrypoint\": stat /bin/entrypoint: no such file or directory"
  Warning  Failed  35m  kubelet, cam-src-8320-v2ksq-worker-northcentralus-67q7l  Error: container create failed: time="2020-09-14T12:05:58Z" level=error msg="container_linux.go:349: starting container process caused \"exec: \\\"/bin/entrypoint\\\": stat /bin/entrypoint: no such file or directory\""
container_linux.go:349: starting container process caused "exec: \"/bin/entrypoint\": stat /bin/entrypoint: no such file or directory"
  Warning  Failed  34m  kubelet, cam-src-8320-v2ksq-worker-northcentralus-67q7l  Error: container create failed: time="2020-09-14T12:06:10Z" level=error msg="container_linux.go:349: starting container process caused \"exec: \\\"/bin/entrypoint\\\": stat /bin/entrypoint: no such file or directory\""
container_linux.go:349: starting container process caused "exec: \"/bin/entrypoint\": stat /bin/entrypoint: no such file or directory"
  Warning  Failed  34m (x3 over 34m)  kubelet, cam-src-8320-v2ksq-worker-northcentralus-67q7l  (combined from similar events): Error: container create failed: time="2020-09-14T12:06:54Z" level=error msg="container_linux.go:349: starting container process caused \"exec: \\\"/bin/entrypoint\\\": stat /bin/entrypoint: no such file or directory\""
container_linux.go:349: starting container process caused "exec: \"/bin/entrypoint\": stat /bin/entrypoint: no such file or directory"
  Normal  Pulled  12m (x106 over 36m)  kubelet, cam-src-8320-v2ksq-worker-northcentralus-67q7l  Container image "quay-enterprise-quay-enterprise.apps.cam-tgt-8320.qe.azure.devcluster.openshift.com/admin/openshift-migration-hook-runner-rhel7@sha256:fda22f17e8e019e47b09c2e5dc06743bcae2f38fe0e81180850baf7eae2d5263" already present on machine



Expected results:
The hook should be executed without problems, and the migration should finish without problems too.

Additional info:

Comment 4 Sergio 2020-09-18 13:46:56 UTC
Verified using MTC 1.3

openshift-migration-rhel7-operator@sha256:233af9517407e792bbb34c58558346f2424b8b0ab54be6f12f9f97513e391a6a
    - name: HOOK_RUNNER_REPO
      value: openshift-migration-hook-runner-rhel7@sha256
    - name: HOOK_RUNNER_TAG
      value: 9e84c265df7882b1a9bdabd51e5c9927da271cd1738082466accc0b9c87f7143
    - name: MIG_CONTROLLER_REPO
      value: openshift-migration-controller-rhel8@sha256
    - name: MIG_CONTROLLER_TAG
      value: d58cccd15cc61be039cd1c8dae9584132dbd59095faf4f4f027fdb05d1860bdb
    - name: MIG_UI_REPO
      value: openshift-migration-ui-rhel8@sha256
    - name: MIG_UI_TAG
      value: f306de1051cd2029944b2aa9511626b1dce365317fd04168478f14a43ad95e44
    - name: MIGRATION_REGISTRY_REPO
      value: openshift-migration-registry-rhel8@sha256
    - name: MIGRATION_REGISTRY_TAG
      value: 3b4a26983053bccc548bc106bdfc0f651075301b90572a03d9d31d62a6c3d769
    - name: VELERO_REPO
      value: openshift-migration-velero-rhel8@sha256
    - name: VELERO_TAG
      value: f844d84dd85f8ae75dc651ca7dd206463f4a10167417f8d6c8793c01c9b72152


Configured hooks run OK. Checked in 4.5 and 3.7.

Moved to VERIFIED status.

Comment 8 errata-xmlrpc 2020-09-30 18:43:07 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory (Migration Toolkit for Containers (MTC) Tool image release advisory 1.3.0), and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHBA-2020:4148


Note You need to log in before you can comment on or make changes to this bug.