Bug 1267807
| Summary: | [RFE] HostedEngine - support for multiple iscsi targets | ||
|---|---|---|---|
| Product: | Red Hat Enterprise Virtualization Manager | Reporter: | Marcus West <mwest> |
| Component: | ovirt-hosted-engine-setup | Assignee: | Simone Tiraboschi <stirabos> |
| Status: | CLOSED ERRATA | QA Contact: | Nikolai Sednev <nsednev> |
| Severity: | medium | Docs Contact: | |
| Priority: | medium | ||
| Version: | 3.5.4 | CC: | bburmest, ecohen, lsurette, mavital, michal.skrivanek, sbonazzo, ylavi |
| Target Milestone: | ovirt-4.2.0 | Keywords: | FutureFeature, Reopened, Triaged |
| Target Release: | 4.2.0 | Flags: | mavital:
testing_plan_complete+
|
| Hardware: | All | ||
| OS: | Linux | ||
| Whiteboard: | integration | ||
| Fixed In Version: | Doc Type: | Enhancement | |
| Doc Text: |
All portals of an iSCSI portal group are connected to enable iSCSI multipath, saving IP and port values in a string.
|
Story Points: | --- |
| Clone Of: | Environment: | ||
| Last Closed: | 2018-05-15 17:32:21 UTC | Type: | Bug |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | Integration | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
| Bug Depends On: | 1149579, 1512534, 1536055 | ||
| Bug Blocks: | 1458709 | ||
|
Description
Marcus West
2015-10-01 05:59:39 UTC
Thanks for reporting, issue seems to be already covered by bug #1193961. Closing as duplicate. *** This bug has been marked as a duplicate of bug 1193961 *** *** This bug has been marked as a duplicate of bug 1149579 *** I've deployed SHE on two ha-hosts, with single NFS data storage domain, over iSCSI, while was using one portal with two NICs on FreeNAS as target, deploymet was successful and once at least on IP of a target was provided, additional path (10.35.72.53:3260) was discovered automatically during deployment as appears bellow:
[ INFO ] Stage: Environment customization
--== STORAGE CONFIGURATION ==--
Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs3, nfs4)[nfs3]: iscsi
Please specify the iSCSI portal IP address: 10.35.72.52
Please specify the iSCSI portal port [3260]:
Please specify the iSCSI portal user:
The following targets have been found:
[1] iqn.2005-10.org.freenas.ctl:4he3iwscsitarget
TPGT: 1, portals:
10.35.72.52:3260
10.35.72.53:3260
Please select a target (1) [1]:
[ INFO ] Connecting to the storage server
The following luns have been found on the requested target:
[1] 36589cfc0000000c55aa7e1416b12a8fc 88GiB FreeNAS iSCSI Disk
status: free, paths: 2 active
Please select the destination LUN (1) [1]:
[ INFO ] Connecting to the storage server
Moving to verified.
Tested on these components:
rhvm-appliance-4.2-20171207.0.el7.noarch
sanlock-3.5.0-1.el7.x86_64
ovirt-host-deploy-1.7.0-1.el7ev.noarch
ovirt-hosted-engine-ha-2.2.2-1.el7ev.noarch
vdsm-4.20.9.3-1.el7ev.x86_64
ovirt-hosted-engine-setup-2.2.2-1.el7ev.noarch
mom-0.5.11-1.el7ev.noarch
libvirt-client-3.2.0-14.el7_4.5.x86_64
ovirt-host-4.2.0-1.el7ev.x86_64
qemu-kvm-rhev-2.9.0-16.el7_4.13.x86_64
Linux version 3.10.0-693.15.1.el7.x86_64 (mockbuild.eng.bos.redhat.com) (gcc version 4.8.5 20150623 (Red Hat 4.8.5-16) (GCC) ) #1 SMP Thu Dec 14 05:13:32 EST 2017
Linux 3.10.0-693.15.1.el7.x86_64 #1 SMP Thu Dec 14 05:13:32 EST 2017 x86_64 x86_64 x86_64 GNU/Linux
Red Hat Enterprise Linux Server release 7.4 (Maipo)
After deployment I've also verified proper ha behavior and it worked as expected.
puma18 ~]# hosted-engine --vm-status
--== Host 1 status ==--
conf_on_shared_storage : True
Status up-to-date : True
Hostname : puma18
Host ID : 1
Engine status : {"health": "good", "vm": "up", "detail": "Up"}
Score : 3400
stopped : False
Local maintenance : False
crc32 : 9656e9f2
local_conf_timestamp : 8092
Host timestamp : 8090
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=8090 (Mon Dec 25 14:16:57 2017)
host-id=1
score=3400
vm_conf_refresh_time=8092 (Mon Dec 25 14:16:59 2017)
conf_on_shared_storage=True
maintenance=False
state=EngineUp
stopped=False
--== Host 2 status ==--
conf_on_shared_storage : True
Status up-to-date : False
Hostname : puma19
Host ID : 2
Engine status : unknown stale-data
Score : 3400
stopped : False
Local maintenance : False
crc32 : 7151579e
local_conf_timestamp : 7994
Host timestamp : 7992
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=7992 (Mon Dec 25 14:16:46 2017)
host-id=2
score=3400
vm_conf_refresh_time=7994 (Mon Dec 25 14:16:49 2017)
conf_on_shared_storage=True
maintenance=False
state=EngineDown
stopped=False
Both paths are up and active by default on both ha-hosts:
puma18 ~]# iscsiadm -m session
tcp: [1] 10.35.72.52:3260,1 iqn.2005-10.org.freenas.ctl:4he3iwscsitarget (non-flash)
tcp: [2] 10.35.72.53:3260,1 iqn.2005-10.org.freenas.ctl:4he3iwscsitarget (non-flash)
puma18 ~]# multipath -ll
36589cfc0000000c55aa7e1416b12a8fc dm-0 FreeNAS ,iSCSI Disk
size=88G features='0' hwhandler='0' wp=rw
|-+- policy='service-time 0' prio=1 status=active
| `- 7:0:0:0 sdc 8:32 active ready running
`-+- policy='service-time 0' prio=1 status=enabled
`- 6:0:0:0 sdb 8:16 active ready running
puma19 ~]# iscsiadm -m session
tcp: [1] 10.35.72.52:3260,1 iqn.2005-10.org.freenas.ctl:4he3iwscsitarget (non-flash)
tcp: [2] 10.35.72.53:3260,1 iqn.2005-10.org.freenas.ctl:4he3iwscsitarget (non-flash)
puma19 ~]# multipath -ll
36589cfc0000000c55aa7e1416b12a8fc dm-0 FreeNAS ,iSCSI Disk
size=88G features='0' hwhandler='0' wp=rw
|-+- policy='service-time 0' prio=1 status=active
| `- 6:0:0:0 sdb 8:16 active ready running
`-+- policy='service-time 0' prio=1 status=enabled
`- 7:0:0:0 sdc 8:32 active ready running
Tested ha on puma18 (ha-host on which SHE-VM was deployed first and was running), I've tried to block active path 10.35.72.53:3260 and observed the results, fail over worked nicely, and second path 10.35.72.52:3260 became active, and engine remained up and running on puma18:
puma18 ~]# multipath -ll
36589cfc0000000c55aa7e1416b12a8fc dm-0 FreeNAS ,iSCSI Disk
size=88G features='0' hwhandler='0' wp=rw
|-+- policy='service-time 0' prio=0 status=enabled
| `- 7:0:0:0 sdc 8:32 failed faulty running
`-+- policy='service-time 0' prio=1 status=active
`- 6:0:0:0 sdb 8:16 active ready running
puma18 ~]# hosted-engine --vm-status
--== Host 1 status ==--
conf_on_shared_storage : True
Status up-to-date : True
Hostname : puma18
Host ID : 1
Engine status : {"health": "good", "vm": "up", "detail": "Up"}
Score : 3400
stopped : False
Local maintenance : False
crc32 : 45053952
local_conf_timestamp : 9507
Host timestamp : 9504
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=9504 (Mon Dec 25 14:40:30 2017)
host-id=1
score=3400
vm_conf_refresh_time=9507 (Mon Dec 25 14:40:33 2017)
conf_on_shared_storage=True
maintenance=False
state=EngineUp
stopped=False
--== Host 2 status ==--
conf_on_shared_storage : True
Status up-to-date : True
Hostname : puma19
Host ID : 2
Engine status : {"reason": "vm not running on this host", "health": "bad", "vm": "down", "detail": "unknown"}
Score : 3400
stopped : False
Local maintenance : False
crc32 : acf0b444
local_conf_timestamp : 9417
Host timestamp : 9414
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=9414 (Mon Dec 25 14:40:29 2017)
host-id=2
score=3400
vm_conf_refresh_time=9417 (Mon Dec 25 14:40:32 2017)
conf_on_shared_storage=True
maintenance=False
state=EngineDown
stopped=False
Blocked second path 10.35.72.52:3260 and first path 10.35.72.53:3260, then SHE-VM got migrated to puma19:
puma18 ~]# multipath -ll
36589cfc0000000c55aa7e1416b12a8fc dm-0 FreeNAS ,iSCSI Disk
size=88G features='0' hwhandler='0' wp=rw
|-+- policy='service-time 0' prio=0 status=enabled
| `- 7:0:0:0 sdc 8:32 failed faulty running
`-+- policy='service-time 0' prio=0 status=enabled
`- 6:0:0:0 sdb 8:16 failed faulty running
puma19 ~]# multipath -ll
36589cfc0000000c55aa7e1416b12a8fc dm-0 FreeNAS ,iSCSI Disk
size=88G features='0' hwhandler='0' wp=rw
|-+- policy='service-time 0' prio=1 status=active
| `- 6:0:0:0 sdb 8:16 active ready running
`-+- policy='service-time 0' prio=1 status=enabled
`- 7:0:0:0 sdc 8:32 active ready running
puma19 ~]# hosted-engine --vm-status
--== Host 1 status ==--
conf_on_shared_storage : True
Status up-to-date : False
Hostname : puma18
Host ID : 1
Engine status : unknown stale-data
Score : 3400
stopped : False
Local maintenance : False
crc32 : 3bf71c33
local_conf_timestamp : 9643
Host timestamp : 9640
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=9640 (Mon Dec 25 14:42:47 2017)
host-id=1
score=3400
vm_conf_refresh_time=9643 (Mon Dec 25 14:42:50 2017)
conf_on_shared_storage=True
maintenance=False
state=EngineUp
stopped=False
--== Host 2 status ==--
conf_on_shared_storage : True
Status up-to-date : True
Hostname : puma19
Host ID : 2
Engine status : {"reason": "bad vm status", "health": "bad", "vm": "up", "detail": "WaitForLaunch"}
Score : 3400
stopped : False
Local maintenance : False
crc32 : 30e86f03
local_conf_timestamp : 9660
Host timestamp : 9657
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=9657 (Mon Dec 25 14:44:31 2017)
host-id=2
score=3400
vm_conf_refresh_time=9660 (Mon Dec 25 14:44:34 2017)
conf_on_shared_storage=True
maintenance=False
state=EngineStarting
stopped=False
puma19 ~]# hosted-engine --vm-status
--== Host 1 status ==--
conf_on_shared_storage : True
Status up-to-date : False
Hostname : puma18
Host ID : 1
Engine status : unknown stale-data
Score : 3400
stopped : False
Local maintenance : False
crc32 : 3bf71c33
local_conf_timestamp : 9643
Host timestamp : 9640
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=9640 (Mon Dec 25 14:42:47 2017)
host-id=1
score=3400
vm_conf_refresh_time=9643 (Mon Dec 25 14:42:50 2017)
conf_on_shared_storage=True
maintenance=False
state=EngineUp
stopped=False
--== Host 2 status ==--
conf_on_shared_storage : True
Status up-to-date : True
Hostname : puma19
Host ID : 2
Engine status : {"reason": "failed liveliness check", "health": "bad", "vm": "up", "detail": "Up"}
Score : 3400
stopped : False
Local maintenance : False
crc32 : a383c8b0
local_conf_timestamp : 9717
Host timestamp : 9714
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=9714 (Mon Dec 25 14:45:28 2017)
host-id=2
score=3400
vm_conf_refresh_time=9717 (Mon Dec 25 14:45:32 2017)
conf_on_shared_storage=True
maintenance=False
state=EngineStarting
stopped=False
puma19 ~]# hosted-engine --vm-status
--== Host 1 status ==--
conf_on_shared_storage : True
Status up-to-date : False
Hostname : puma18
Host ID : 1
Engine status : unknown stale-data
Score : 3400
stopped : False
Local maintenance : False
crc32 : 3bf71c33
local_conf_timestamp : 9643
Host timestamp : 9640
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=9640 (Mon Dec 25 14:42:47 2017)
host-id=1
score=3400
vm_conf_refresh_time=9643 (Mon Dec 25 14:42:50 2017)
conf_on_shared_storage=True
maintenance=False
state=EngineUp
stopped=False
--== Host 2 status ==--
conf_on_shared_storage : True
Status up-to-date : True
Hostname : puma19
Host ID : 2
Engine status : {"health": "good", "vm": "up", "detail": "Up"}
Score : 3400
stopped : False
Local maintenance : False
crc32 : 02524a7e
local_conf_timestamp : 9764
Host timestamp : 9760
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=9760 (Mon Dec 25 14:46:15 2017)
host-id=2
score=3400
vm_conf_refresh_time=9764 (Mon Dec 25 14:46:18 2017)
conf_on_shared_storage=True
maintenance=False
state=EngineUp
stopped=False
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2018:1471 BZ<2>Jira Resync |