Description of problem: When migrating sr-iov nodes that don't have OVS agent running the migration steps fail and will not continue with migration tasks on the sriov nodes. This will lead to situation where nodes are left in inconsistent state with uncompleted migration. Version-Release number of selected component (if applicable): How reproducible: 100% Steps to Reproduce: 1. Run OVN migration with SR-IOV nodes 2. 3. Actual results: Migration itself passes but sr-iov nodes are not migrated and have some leftovers. Expected results: sr-iov nodes have no leftovers Additional info: Perhaps sriov nodes don't need to run any steps during the migration. br-migration is present on sr-iov nodes OVN was not activated TASK [stop-agents : stop and disable ovs services and healthchecks] ************************************************************************************************************** task path: /home/stack/ovn_migration/playbooks/roles/stop-agents/tasks/cleanup.yml:2 Wednesday 06 October 2021 18:19:17 +0000 (0:00:00.648) 0:00:08.091 ***** META: noop META: noop META: noop failed: [computesriov-0] (item=tripleo_neutron_ovs_agent_healthcheck.timer) => {"ansible_loop_var": "item", "changed": false, "item": "tripleo_neutron_ovs_agent_healthcheck.timer ", "msg": "Could not find the requested service tripleo_neutron_ovs_agent_healthcheck.timer: host"} failed: [computesriov-1] (item=tripleo_neutron_ovs_agent_healthcheck.timer) => {"ansible_loop_var": "item", "changed": false, "item": "tripleo_neutron_ovs_agent_healthcheck.timer ", "msg": "Could not find the requested service tripleo_neutron_ovs_agent_healthcheck.timer: host"} failed: [computesriov-0] (item=tripleo_neutron_ovs_agent_healthcheck.service) => {"ansible_loop_var": "item", "changed": false, "item": "tripleo_neutron_ovs_agent_healthcheck.ser vice", "msg": "Could not find the requested service tripleo_neutron_ovs_agent_healthcheck.service: host"} failed: [computesriov-1] (item=tripleo_neutron_ovs_agent_healthcheck.service) => {"ansible_loop_var": "item", "changed": false, "item": "tripleo_neutron_ovs_agent_healthcheck.ser vice", "msg": "Could not find the requested service tripleo_neutron_ovs_agent_healthcheck.service: host"} failed: [computesriov-0] (item=tripleo_neutron_ovs_agent.service) => {"ansible_loop_var": "item", "changed": false, "item": "tripleo_neutron_ovs_agent.service", "msg": "Service i s in unknown state", "status": {}} failed: [computesriov-1] (item=tripleo_neutron_ovs_agent.service) => {"ansible_loop_var": "item", "changed": false, "item": "tripleo_neutron_ovs_agent.service", "msg": "Service i s in unknown state", "status": {}}
Verified that the issue does not happen on RHOS-17.1-RHEL-9-20230607.n.2 with openstack-neutron-ovn-migration-tool-18.6.1-1.20230518200966.el9ost.noarch
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Release of components for Red Hat OpenStack Platform 17.1 (Wallaby)), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHEA-2023:4577