Created attachment 1204228 [details] engine.log, and vdsm.log, supervdsm.log, sanlock.log, messages from 2 hosts: rebooted host and host running vms that get paused Description of problem: When any component in FC SAN changes its status, i.e. if I reboot a host, and HBA ports get online, some vms residing in FC storage domain get automatically paused and I have to resume it manually. Version-Release number of selected component (if applicable): How reproducible: Allways Steps to Reproduce: 1.Put a host in maintenance mode and reboot it 2.Wait until the host boot up for activations of HBA FC ports Actual results: Engine reports I/O problems and some vms running in other hosts get paused. Vms need to be resumed manually. Not all vms sharing same storage get paused. Expected results: Changes in port status of SAN components should not be detected as I/O problems as far the storage is accessible. Vms should not be paused. Additional info: Scenario: 8 virtualization hosts RHEL - 7 - 2.1511.el7.centos.2.10 2 IBM ds3512 storage arrays 2 paths per Host each connected to a QLogic(R) 10-port 4Gb SAN Switch Module for IBM BladeCenter(R) 4 luns connected as FC domains
The pause is done by QEMU, not by any oVirt/RHV components. It is detected as an IO error - so I assume the multipathing is not handling the port state change correctly. Have you seen anything in the logs indicating otherwise?
(In reply to Yaniv Kaul from comment #1) > The pause is done by QEMU, not by any oVirt/RHV components. It is detected > as an IO error - so I assume the multipathing is not handling the port state > change correctly. > Have you seen anything in the logs indicating otherwise? Closing - please re-open if you have more details.