Red Hat Bugzilla – Bug 1312805
Adding an iSCSI storage crashes node and sets all VM states to Unknown
Last modified: 2016-03-08 05:41:19 EST
Created attachment 1131472 [details]
Engine log at crash time
Description of problem:
We're running oVirt with Gluster as the main storage backend currently. We're planning to migrate storage to iSCSI, so we've been trying to add a iSCSI backend from the webadmin, resulting in the chosen node (by default the SPM) crashed and all its VMs set status to Unknown.
VDSM version is 4.17.18.
Version-Release number of selected component (if applicable):
oVirt-engine: v. 18.104.22.168-1
Steps to Reproduce:
1. Click on the storage tab
2. Click on the New domain button
3. Change storage type to iSCSI
4. As discover targets we set the IP address (10.X.X.80 in the logs), port is default.
5. Click on discover: The target appears correctly (iqn: iqn.2003-10.com.lefthandnetworks:p4000-kvm:67:ovirt-rojo)
6. Click on the arrow button ([->])
7. The chosen VDSM node hangs, on the Hosts tab we see 'Connecting' and timing out and setting host as non-responsive, which leads to VMs be set as Unknown and needing to be forced to be reset via DB.
To minimize impact, we migrated all VMs from one host and forced that host to add the storage and see it crash so no VMs are affected (getting the same result, though).
I'm attaching 2 logs: engine.log and vdsm.log from the crashing node.
* kvmr0X.domain.com: Nodes
* IPs 10.X.X.11 to 10.X.X.15 are nodes' service IPs of hosts 1 to 5 respectively.
* IP 10.X.X.60 and on are dedicated to storage (10.X.X.80 is the iSCSI target VIP).
* 192.168.100.X are migration IPs.
Created attachment 1131473 [details]
VDSM log at crash time
Tal, as the QE contact, can you take a look please?
Created attachment 1131747 [details]
FWIW, I found this entries in the /var/log/messages file, seems that iscsi is crashing badly
Created attachment 1131751 [details]
multipath -v10 output
I'm attaching the multipath -v10 command output. As relevant I see this line:
36000eb3a4f1acbc20000000000000043: set ACT_CREATE (map does not exist)
Still can't find out why is it caused by
I finally could find the culprit. Seems that one of the hops between hosts and storage servers had MTU 1500 when 9000 was expected. Fixed that and now it works.