Note: This bug is displayed in read-only format because the product is no longer active in Red Hat Bugzilla.

Bug 1878801

Summary: Hooks using default image are failing
Product: OpenShift Container Platform Reporter: Sergio <sregidor>
Component: Migration ToolingAssignee: Jason Montleon <jmontleo>
Status: CLOSED ERRATA QA Contact: Xin jiang <xjiang>
Severity: medium Docs Contact:
Priority: unspecified    
Version: 4.5CC: chezhang, mberube, rjohnson, rpattath, whu, xjiang
Target Milestone: ---   
Target Release: 4.5.0   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: Doc Type: No Doc Update
Doc Text:
Story Points: ---
Clone Of: Environment:
Last Closed: 2020-09-30 18:43:07 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:

Description Sergio 2020-09-14 14:33:26 UTC
Description of problem:
When we configure a hook using the default image to execute the hooks, CMT cannot execute the hooks and the migration is stuck forever.


Version-Release number of selected component (if applicable):
CMT 1.3
SOURCE CLUSTER: azure OCP 4.2
TARGET CLUSTER: azure OCP 4.5

How reproducible:
Always

Steps to Reproduce:
1. Create a namespace
   oc new-project bztest

2. Create a migration plan for this namespace

3. Add a hook for this migration (no matter which hook, or which playbook). Use the default "Ansible runtime image".

4. Run the migration


Actual results:
In the namespace used to run the hook, we find that the pods

$ oc get pods
NAME                          READY   STATUS                 RESTARTS   AGE
hooks-prebackup-54mpd-pkx7f   0/1     CreateContainerError   0          37m
hooks-prebackup-54mpd-rr4qh   0/1     CreateContainerError   0          7m42s

Looking the those pods' logs we find this messages

  Warning  Failed  35m  kubelet, cam-src-8320-v2ksq-worker-northcentralus-67q7l  Error: container create failed: time="2020-09-14T12:05:45Z" level=error msg="container_linux.go:349: starting container process caused \"exec: \\\"/bin/entrypoint\\\": stat /bin/entrypoint: no such file or directory\""
container_linux.go:349: starting container process caused "exec: \"/bin/entrypoint\": stat /bin/entrypoint: no such file or directory"
  Warning  Failed  35m  kubelet, cam-src-8320-v2ksq-worker-northcentralus-67q7l  Error: container create failed: time="2020-09-14T12:05:58Z" level=error msg="container_linux.go:349: starting container process caused \"exec: \\\"/bin/entrypoint\\\": stat /bin/entrypoint: no such file or directory\""
container_linux.go:349: starting container process caused "exec: \"/bin/entrypoint\": stat /bin/entrypoint: no such file or directory"
  Warning  Failed  34m  kubelet, cam-src-8320-v2ksq-worker-northcentralus-67q7l  Error: container create failed: time="2020-09-14T12:06:10Z" level=error msg="container_linux.go:349: starting container process caused \"exec: \\\"/bin/entrypoint\\\": stat /bin/entrypoint: no such file or directory\""
container_linux.go:349: starting container process caused "exec: \"/bin/entrypoint\": stat /bin/entrypoint: no such file or directory"
  Warning  Failed  34m (x3 over 34m)  kubelet, cam-src-8320-v2ksq-worker-northcentralus-67q7l  (combined from similar events): Error: container create failed: time="2020-09-14T12:06:54Z" level=error msg="container_linux.go:349: starting container process caused \"exec: \\\"/bin/entrypoint\\\": stat /bin/entrypoint: no such file or directory\""
container_linux.go:349: starting container process caused "exec: \"/bin/entrypoint\": stat /bin/entrypoint: no such file or directory"
  Normal  Pulled  12m (x106 over 36m)  kubelet, cam-src-8320-v2ksq-worker-northcentralus-67q7l  Container image "quay-enterprise-quay-enterprise.apps.cam-tgt-8320.qe.azure.devcluster.openshift.com/admin/openshift-migration-hook-runner-rhel7@sha256:fda22f17e8e019e47b09c2e5dc06743bcae2f38fe0e81180850baf7eae2d5263" already present on machine



Expected results:
The hook should be executed without problems, and the migration should finish without problems too.

Additional info:

Comment 4 Sergio 2020-09-18 13:46:56 UTC
Verified using MTC 1.3

openshift-migration-rhel7-operator@sha256:233af9517407e792bbb34c58558346f2424b8b0ab54be6f12f9f97513e391a6a
    - name: HOOK_RUNNER_REPO
      value: openshift-migration-hook-runner-rhel7@sha256
    - name: HOOK_RUNNER_TAG
      value: 9e84c265df7882b1a9bdabd51e5c9927da271cd1738082466accc0b9c87f7143
    - name: MIG_CONTROLLER_REPO
      value: openshift-migration-controller-rhel8@sha256
    - name: MIG_CONTROLLER_TAG
      value: d58cccd15cc61be039cd1c8dae9584132dbd59095faf4f4f027fdb05d1860bdb
    - name: MIG_UI_REPO
      value: openshift-migration-ui-rhel8@sha256
    - name: MIG_UI_TAG
      value: f306de1051cd2029944b2aa9511626b1dce365317fd04168478f14a43ad95e44
    - name: MIGRATION_REGISTRY_REPO
      value: openshift-migration-registry-rhel8@sha256
    - name: MIGRATION_REGISTRY_TAG
      value: 3b4a26983053bccc548bc106bdfc0f651075301b90572a03d9d31d62a6c3d769
    - name: VELERO_REPO
      value: openshift-migration-velero-rhel8@sha256
    - name: VELERO_TAG
      value: f844d84dd85f8ae75dc651ca7dd206463f4a10167417f8d6c8793c01c9b72152


Configured hooks run OK. Checked in 4.5 and 3.7.

Moved to VERIFIED status.

Comment 8 errata-xmlrpc 2020-09-30 18:43:07 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory (Migration Toolkit for Containers (MTC) Tool image release advisory 1.3.0), and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHBA-2020:4148