Description of problem: Installation of the hosted engine via Cockpit fails with monitor process being blocked and resulting in sanlock calling watchdog and restarting the VDSMD service on the host terminating the install process. Version-Release number of selected component (if applicable): RHVH-4.2-20171218.0-RHVH-x86_64-dvd1.iso vdsm-4.19.45-1.el7ev.x86_64 rhvm-appliance-4.1.20180125.0-1.el7.noarch How reproducible: Steps to Reproduce: 1. Install RHV on a host. 2. Install Hosted Engine via cockpit 3. Enter all info for storage/config etc and proceed with the installation. Actual results: - Stops on the step from GUI . - VDSM reports errors related to creation of hosted engine disk creation =----------------------------------------------------- 2018-02-16 15:54:06,888+1100 WARN (check/loop) [storage.check] Checker u'/rhev/data-center/mnt/_var_lib_ovirt-hosted-engine-setup_tmp2gg__IU/d42ad173-214a-4135-9195-763f585ddb30/dom_md/metadata' is blocked for 50.00 seconds (check:279) =----------------------------------------------------- - sanlock indicating -202 timeout errors and calling watchdog to restart vdsm service =---------------------------------------------------------- 2018-02-16 15:53:40 3640 [28014]: d42ad173 aio timeout RD 0x7fccbc0008c0:0x7fccbc0008d0:0x7fcccf912000 ioto 10 to_count 1 2018-02-16 15:53:40 3640 [28014]: s3 delta_renew read timeout 10 sec offset 0 /rhev/data-center/mnt/_var_lib_ovirt-hosted-engine-setup_tmp2gg__IU/d42ad173-214a-4135-9195-763f585ddb30/dom_md/ids 2018-02-16 15:53:40 3640 [28014]: s3 renewal error -202 delta_length 10 last_success 3610 2018-02-16 15:54:00 3661 [28014]: d42ad173 aio timeout RD 0x7fccbc000910:0x7fccbc000920:0x7fcccc3fd000 ioto 10 to_count 2 2018-02-16 15:54:00 3661 [28014]: s3 delta_renew read timeout 10 sec offset 0 /rhev/data-center/mnt/_var_lib_ovirt-hosted-engine-setup_tmp2gg__IU/d42ad173-214a-4135-9195-763f585ddb30/dom_md/ids 2018-02-16 15:54:00 3661 [28014]: s3 renewal error -202 delta_length 20 last_success 3610 =---------------------------------------------------------- Expected results: Hosted Engine to continue installation and complete successfully. Additional info:
Can you try to reproduce with the RHV 4.2 Beta 2? I think this is already resolved.
Yaniv, The same issue is seen again with the latest RHV 4.2 beta for this. Below are the logs for the same and I tried using NFS storage =----------------------------------------------------------------- 2018-02-20 12:40:41,328+1100 WARN (check/loop) [storage.check] Checker u'/rhev/data-center/mnt/_var_lib_ovirt-hosted-engine-setup_tmpdWJSkK/bd5f10ec-cb5c-4cfe-8816-4b7cae635be1/dom_md/metadata' is blocked for 70.01 seconds (check:278) 2018-02-20 12:40:43,197+1100 INFO (jsonrpc/0) [api.host] START getAllVmStats() from=::1,58430 (api:46) 2018-02-20 12:40:43,198+1100 INFO (jsonrpc/0) [api.host] FINISH getAllVmStats return={'status': {'message': 'Done', 'code': 0}, 'statsList': (suppressed)} from=::1,58430 (api:52) 2018-02-20 12:40:43,199+1100 INFO (jsonrpc/0) [jsonrpc.JsonRpcServer] RPC call Host.getAllVmStats succeeded in 0.00 seconds (__init__:573) 2018-02-20 12:40:44,915+1100 INFO (MainThread) [vds] Received signal 15, shutting down (vdsmd:67) 2018-02-20 12:40:44,916+1100 INFO (MainThread) [jsonrpc.JsonRpcServer] Stopping JsonRPC Server (__init__:703) 2018-02-20 12:40:44,921+1100 INFO (MainThread) [vds] Stopping http server (http:79) 2018-02-20 12:40:44,922+1100 INFO (http) [vds] Server stopped (http:69) 2018-02-20 12:40:44,922+1100 INFO (MainThread) [root] Unregistering all secrets (secret:91) 2018-02-20 12:40:44,923+1100 INFO (MainThread) [vds] Stopping QEMU-GA poller (qemuguestagent:119) 2018-02-20 12:40:44,924+1100 INFO (MainThread) [vdsm.api] START prepareForShutdown(options=None) from=internal, task_id=48ca73c6-2a83-4130-9d8f-af092d81a823 (api:46) 2018-02-20 12:40:46,568+1100 INFO (MainThread) [vds] Received signal 15, shutting down (vdsmd:67) =-------------------------------------------------------------------------- Also when I tried using ISCSI storage during installation I received the error as follows =--------------------------------------------------------------------------------------------------- 2018-02-20 11:48:37,940+1100 DEBUG otopi.plugins.gr_he_setup.storage.blockd blockd._customization:720 exception Traceback (most recent call last): File "/usr/share/ovirt-hosted-engine-setup/scripts/../plugins/gr-he-setup/storage/blockd.py", line 716, in _customization lunGUID File "/usr/share/ovirt-hosted-engine-setup/scripts/../plugins/gr-he-setup/storage/blockd.py", line 517, in _validate_domain _('The requested device is not listed by VDSM') RuntimeError: The requested device is not listed by VDSM 2018-02-20 11:48:37,943+1100 ERROR otopi.plugins.gr_he_setup.storage.blockd blockd._customization:721 The requested device is not listed by VDSM 2018-02-20 11:48:37,943+1100 DEBUG otopi.context context._executeMethod:143 method exception Traceback (most recent call last): File "/usr/lib/python2.7/site-packages/otopi/context.py", line 133, in _executeMethod method['method']() File "/usr/share/ovirt-hosted-engine-setup/scripts/../plugins/gr-he-setup/storage/blockd.py", line 723, in _customization raise RuntimeError(_('Cannot access LUN')) RuntimeError: Cannot access LUN 2018-02-20 11:48:37,945+1100 ERROR otopi.context context._executeMethod:152 Failed to execute stage 'Environment customization': Cannot access LUN =------------------------------------------------------------------------------------------------------ I am not too sure for the above error being received as the storage is accessible and I am able to login to the target successfully from the host. Ribu
Simone please check if this is a vdsm storage issue or a he issue. If vdsm, please move to vdsm storage team
At the time of the issue (around 2018-02-16 15:54 ) the system seamed to be really slow; according to the log it took 4'30" to locally extract a 1 GB tar.gz file (with a less than 3 GB inside ) and about 10 seconds just to run '/bin/qemu-img info --output json' on it while normally it's by far faster: 2018-02-16 15:52:09 INFO otopi.plugins.gr_he_common.vm.boot_disk boot_disk.prepare:216 Extracting disk image from OVF archive (could take a few minutes depending on archive size) 2018-02-16 15:56:32 INFO otopi.plugins.gr_he_common.vm.boot_disk boot_disk._validate_volume:91 Validating pre-allocated volume size 2018-02-16 15:56:32 DEBUG otopi.plugins.gr_he_common.vm.boot_disk plugin.executeRaw:813 execute: ('/bin/sudo', '-u', 'vdsm', '-g', 'kvm', '/bin/qemu-img', 'info', '--output', 'json', '/var/tmp/tmpl1ssJY'), executable='None', cwd='None', env=None 2018-02-16 15:56:42 DEBUG otopi.plugins.gr_he_common.vm.boot_disk plugin.executeRaw:863 execute-result: ('/bin/sudo', '-u', 'vdsm', '-g', 'kvm', '/bin/qemu-img', 'info', '--output', 'json', '/var/tmp/tmpl1ssJY'), rc=0 I suspect that the sanlock timeout could be due to that. Also using up to 100% of the available memory on the host for the engine VM appears really risky: 2018-02-16 15:49:28 DEBUG otopi.plugins.gr_he_setup.vm.memory dialog.queryEnvKey:90 queryEnvKey called for key OVEHOSTED_VM/vmMemSizeMB 2018-02-16 15:49:28 DEBUG otopi.plugins.otopi.dialog.machine dialog.__logString:204 DIALOG:SEND **%QStart: ovehosted_vmenv_mem 2018-02-16 15:49:28 DEBUG otopi.plugins.otopi.dialog.machine dialog.__logString:204 DIALOG:SEND ### Please specify the memory size of the VM in MB (Defaults to maximum available): [4993]: 2018-02-16 15:49:28 DEBUG otopi.plugins.otopi.dialog.machine dialog.__logString:204 DIALOG:SEND **%QDefault: 4993 2018-02-16 15:49:28 DEBUG otopi.plugins.otopi.dialog.machine dialog.__logString:204 DIALOG:SEND **%QHidden: FALSE 2018-02-16 15:49:28 DEBUG otopi.plugins.otopi.dialog.machine dialog.__logString:204 DIALOG:SEND ***Q:STRING ovehosted_vmenv_mem 2018-02-16 15:49:28 DEBUG otopi.plugins.otopi.dialog.machine dialog.__logString:204 DIALOG:SEND **%QEnd: ovehosted_vmenv_mem 2018-02-16 15:49:29 DEBUG otopi.plugins.otopi.dialog.machine dialog.__logString:204 DIALOG:RECEIVE 4993 2018-02-16 15:49:29 DEBUG otopi.context context.dumpEnvironment:760 ENVIRONMENT DUMP - BEGIN 2018-02-16 15:49:29 DEBUG otopi.context context.dumpEnvironment:770 ENV OVEHOSTED_VM/vmMemSizeMB=int:'4993' Ribu, could you please try reproducing on a system with more resources?
(In reply to Ribu Tho from comment #7) > I am not too sure for the above error being received as the storage is > accessible and I am able to login to the target successfully from the host. This is a duplicate of https://bugzilla.redhat.com/show_bug.cgi?id=1542426 See https://bugzilla.redhat.com/show_bug.cgi?id=1539598#c17
(In reply to Simone Tiraboschi from comment #10) > At the time of the issue (around 2018-02-16 15:54 ) the system seamed to be > really slow; according to the log it took 4'30" to locally extract a 1 GB > tar.gz file (with a less than 3 GB inside ) and about 10 seconds just to run > '/bin/qemu-img info --output json' on it while normally it's by far faster: Simone, I suspected hardware latency as well but sanlock timeout occured on my test systems running much faster than the one reported above. For 1.6G qemu info of the RVH Appliance image, time was <<1s. time sudo -u vdsm -g kvm qemu-img info --output json /usr/share/ovirt-engine-appliance/rhvm-appliance-4.1.20180125.0-1.el7.ova { "virtual-size": 1670002176, "filename": "/usr/share/ovirt-engine-appliance/rhvm-appliance-4.1.20180125.0-1.el7.ova", "format": "raw", "actual-size": 1670008832, "dirty-flag": false } real 0m0.069s user 0m0.057s sys 0m0.012s > Ribu, could you please try reproducing on a system with more resources? Ribu tried NFS (I suppose remote storage backend?) and iSCSI path; I've tested GlusterFS and NFS (locally backed storage) as well. sanlock is occuring in each of these scenarios for different test systems (I'm not sure of Ribu's test system). The RH doc says HE min req's are 4GiB memory over 1GiB Ether. Perhaps that isn't enough ...
(In reply to Simone Tiraboschi from comment #11) > (In reply to Ribu Tho from comment #7) > > I am not too sure for the above error being received as the storage is > > accessible and I am able to login to the target successfully from the host. > > This is a duplicate of https://bugzilla.redhat.com/show_bug.cgi?id=1542426 > See https://bugzilla.redhat.com/show_bug.cgi?id=1539598#c17 Those bugs are only about iSCSI attached storage, so how are the related to storage bexposed by NFS and Gluster for the other 2 tests? I do see this in the supervdsm.log: MainThread::DEBUG::2018-03-01 11:15:54,985::__init__::47::blivet::(register_device_format) registered device format class EFIVarFS as efivarfs MainThread::DEBUG::2018-03-01 11:15:55,102::storage_log::69::blivet::(log_exception_info) IGNORED: Caught exception, continuing. MainThread::DEBUG::2018-03-01 11:15:55,102::storage_log::72::blivet::(log_exception_info) IGNORED: Problem description: failed to get initiator name from iscsi firmware MainThread::DEBUG::2018-03-01 11:15:55,103::storage_log::73::blivet::(log_exception_info) IGNORED: Begin exception details. MainThread::DEBUG::2018-03-01 11:15:55,103::storage_log::76::blivet::(log_exception_info) IGNORED: Traceback (most recent call last): MainThread::DEBUG::2018-03-01 11:15:55,103::storage_log::76::blivet::(log_exception_info) IGNORED: File "/usr/lib/python2.7/site-packages/blivet/iscsi.py", line 146, in __init__ MainThread::DEBUG::2018-03-01 11:15:55,103::storage_log::76::blivet::(log_exception_info) IGNORED: initiatorname = libiscsi.get_firmware_initiator_name() MainThread::DEBUG::2018-03-01 11:15:55,103::storage_log::76::blivet::(log_exception_info) IGNORED: IOError: Unknown error MainThread::DEBUG::2018-03-01 11:15:55,103::storage_log::77::blivet::(log_exception_info) IGNORED: End exception details.
(In reply to Justin Pittman from comment #13) > (In reply to Simone Tiraboschi from comment #11) > > (In reply to Ribu Tho from comment #7) > > > I am not too sure for the above error being received as the storage is > > > accessible and I am able to login to the target successfully from the host. > > > > This is a duplicate of https://bugzilla.redhat.com/show_bug.cgi?id=1542426 > > See https://bugzilla.redhat.com/show_bug.cgi?id=1539598#c17 > > Those bugs are only about iSCSI attached storage, so how are the related to > storage bexposed by NFS and Gluster for the other 2 tests? I was just referring to 2018-02-20 11:48:37,940+1100 DEBUG otopi.plugins.gr_he_setup.storage.blockd blockd._customization:720 exception Traceback (most recent call last): File "/usr/share/ovirt-hosted-engine-setup/scripts/../plugins/gr-he-setup/storage/blockd.py", line 716, in _customization lunGUID File "/usr/share/ovirt-hosted-engine-setup/scripts/../plugins/gr-he-setup/storage/blockd.py", line 517, in _validate_domain _('The requested device is not listed by VDSM') RuntimeError: The requested device is not listed by VDSM on comment 7 on thkis bug. That one was on iSCSI.
(In reply to Simone Tiraboschi from comment #10) > At the time of the issue (around 2018-02-16 15:54 ) the system seamed to be > really slow; according to the log it took 4'30" to locally extract a 1 GB > tar.gz file (with a less than 3 GB inside ) and about 10 seconds just to run > '/bin/qemu-img info --output json' on it while normally it's by far faster: > > 2018-02-16 15:52:09 INFO otopi.plugins.gr_he_common.vm.boot_disk > boot_disk.prepare:216 Extracting disk image from OVF archive (could take a > few minutes depending on archive size) > 2018-02-16 15:56:32 INFO otopi.plugins.gr_he_common.vm.boot_disk > boot_disk._validate_volume:91 Validating pre-allocated volume size > 2018-02-16 15:56:32 DEBUG otopi.plugins.gr_he_common.vm.boot_disk > plugin.executeRaw:813 execute: ('/bin/sudo', '-u', 'vdsm', '-g', 'kvm', > '/bin/qemu-img', 'info', '--output', 'json', '/var/tmp/tmpl1ssJY'), > executable='None', cwd='None', env=None > 2018-02-16 15:56:42 DEBUG otopi.plugins.gr_he_common.vm.boot_disk > plugin.executeRaw:863 execute-result: ('/bin/sudo', '-u', 'vdsm', '-g', > 'kvm', '/bin/qemu-img', 'info', '--output', 'json', '/var/tmp/tmpl1ssJY'), > rc=0 > > I suspect that the sanlock timeout could be due to that. > > Also using up to 100% of the available memory on the host for the engine VM > appears really risky: > 2018-02-16 15:49:28 DEBUG otopi.plugins.gr_he_setup.vm.memory > dialog.queryEnvKey:90 queryEnvKey called for key OVEHOSTED_VM/vmMemSizeMB > 2018-02-16 15:49:28 DEBUG otopi.plugins.otopi.dialog.machine > dialog.__logString:204 DIALOG:SEND **%QStart: ovehosted_vmenv_mem > 2018-02-16 15:49:28 DEBUG otopi.plugins.otopi.dialog.machine > dialog.__logString:204 DIALOG:SEND ### Please specify the memory size > of the VM in MB (Defaults to maximum available): [4993]: > 2018-02-16 15:49:28 DEBUG otopi.plugins.otopi.dialog.machine > dialog.__logString:204 DIALOG:SEND **%QDefault: 4993 > 2018-02-16 15:49:28 DEBUG otopi.plugins.otopi.dialog.machine > dialog.__logString:204 DIALOG:SEND **%QHidden: FALSE > 2018-02-16 15:49:28 DEBUG otopi.plugins.otopi.dialog.machine > dialog.__logString:204 DIALOG:SEND ***Q:STRING ovehosted_vmenv_mem > 2018-02-16 15:49:28 DEBUG otopi.plugins.otopi.dialog.machine > dialog.__logString:204 DIALOG:SEND **%QEnd: ovehosted_vmenv_mem > 2018-02-16 15:49:29 DEBUG otopi.plugins.otopi.dialog.machine > dialog.__logString:204 DIALOG:RECEIVE 4993 > 2018-02-16 15:49:29 DEBUG otopi.context context.dumpEnvironment:760 > ENVIRONMENT DUMP - BEGIN > 2018-02-16 15:49:29 DEBUG otopi.context context.dumpEnvironment:770 ENV > OVEHOSTED_VM/vmMemSizeMB=int:'4993' > > Ribu, could you please try reproducing on a system with more resources? Simone , As per my comment #15 I have reproduced the issue with cockpit again. Please check out the comment for more info Ribu
According to the comment 15, we move it to cockpit-ovirt component for further investigation from UI side.
Moving to Phillip since Ryan is on PTO
I cannot reproduce this issue with the given versions in the description Test versions: rhvh-4.2.0.6-0.20171218.0+1 ovirt-hosted-engine-ha-2.2.2-1.el7ev.noarch ovirt-hosted-engine-setup-2.2.2-1.el7ev.noarch cockpit-ws-157-1.el7.x86_64 cockpit-storaged-157-1.el7.noarch cockpit-bridge-157-1.el7.x86_64 cockpit-system-157-1.el7.noarch cockpit-dashboard-157-1.el7.x86_64 cockpit-ovirt-dashboard-0.11.3-0.1.el7ev.noarch cockpit-157-1.el7.x86_64 rhvm-appliance-4.1.20180125.0-1.el7.noarch vdsm-4.20.9.3-1.el7ev.x86_64 Test steps: 1. Install RHVH 2. Deploy HE via cockpit Test results: [root@hp-dl388g9-05 tmp]# cat he-setup-answerfile.conf [environment:default] OVEHOSTED_CORE/rollbackProceed=none:None OVEHOSTED_CORE/screenProceed=none:None OVEHOSTED_CORE/deployProceed=bool:true OVEHOSTED_CORE/upgradeProceed=none:None OVEHOSTED_CORE/confirmSettings=bool:true OVEHOSTED_STORAGE/domainType=str:nfs3 OVEHOSTED_STORAGE/imgSizeGB=str:58 OVEHOSTED_STORAGE/storageDomain=str:hosted_storage OVEHOSTED_STORAGE/storageDomainConnection=str:10.66.148.11:/home/yzhao1/nfs8 OVEHOSTED_STORAGE/mntOptions=str: OVEHOSTED_NETWORK/bridgeIf=str:eno1 OVEHOSTED_NETWORK/bridgeName=str:ovirtmgmt OVEHOSTED_NETWORK/firewallManager=str:iptables OVEHOSTED_NETWORK/gateway=str:10.73.75.254 OVEHOSTED_NETWORK/fqdn=str:rhevh-hostedengine-vm-05.lab.eng.pek2.redhat.com OVEHOSTED_VM/bootDevice=str:disk OVEHOSTED_VM/vmVCpus=str:2 OVEHOSTED_VM/vmMACAddr=str:52:54:00:5d:21:64 OVEHOSTED_VM/vmMemSizeMB=int:4096 OVEHOSTED_VM/cloudinitVMStaticCIDR=str: OVEHOSTED_VM/cloudinitVMDNS=str: OVEHOSTED_VM/cloudinitVMTZ=str:Asia/Shanghai OVEHOSTED_VM/cloudInitISO=str:generate OVEHOSTED_VM/cloudinitInstanceHostName=str:rhevh-hostedengine-vm-05 OVEHOSTED_VM/cloudinitInstanceDomainName=str:lab.eng.pek2.redhat.com OVEHOSTED_VM/cloudinitExecuteEngineSetup=bool:true OVEHOSTED_VM/automateVMShutdown=bool:true OVEHOSTED_VM/cloudinitRootPwd=str:redhat OVEHOSTED_VM/rootSshPubkey=str: OVEHOSTED_VM/rootSshAccess=str:yes OVEHOSTED_VM/cloudinitVMETCHOSTS=bool:true OVEHOSTED_ENGINE/adminPassword=str:redhat OVEHOSTED_VDSM/consoleType=str:vnc OVEHOSTED_VDSM/cpu=str:model_Broadwell OVEHOSTED_NOTIF/smtpServer=str:localhost OVEHOSTED_NOTIF/smtpPort=str:25 OVEHOSTED_NOTIF/sourceEmail=str:root@localhost OVEHOSTED_NOTIF/destEmail=str:root@localhost [root@hp-dl388g9-05 tmp]# hosted-engine --vm-status --== Host 1 status ==-- conf_on_shared_storage : True Status up-to-date : True Hostname : hp-dl388g9-05.lab.eng.pek2.redhat.com Host ID : 1 Engine status : {"health": "good", "vm": "up", "detail": "Up"} Score : 3400 stopped : False Local maintenance : False crc32 : ff9346da local_conf_timestamp : 3043 Host timestamp : 3042 Extra metadata (valid at timestamp): metadata_parse_version=1 metadata_feature_version=1 timestamp=3042 (Wed Mar 14 18:32:47 2018) host-id=1 score=3400 vm_conf_refresh_time=3043 (Wed Mar 14 18:32:48 2018) conf_on_shared_storage=True maintenance=False state=EngineUp stopped=False Also, I have the question about that, this attachment: https://bugzilla.redhat.com/attachment.cgi?id=1396820 seems that it is a RHVH4.1 system. Why describe the version which you tested with RHVH-4.2-20171218.0-RHVH-x86_64-dvd1.iso? Then, if installed the RHVH-4.2-20171218.0-RHVH-x86_64-dvd1.iso, the vdsm version should be vdsm-4.20.9.3-1.el7ev.x86_64 , not vdsm-4.19.45-1.el7ev.x86_64. I confused about that.
Not blocking 4.2.2 on this being not reproducible. Moving to 4.2.4 waiting for needed info.
Hi Ribu, Pls see the picture which attached: https://bugzilla.redhat.com/attachment.cgi?id=1396820 From the cockpit, "The image not uploaded to the data domain", maybe it is the point. 1. Confirm to install rhvm-appliance successfully 2. Prepare the clean nfs storage domain, confirm to connect Also , you can attach the answer file to the attachment. Here is mine. [root@hp-dl388g9-05 tmp]# cat he-setup-answerfile.conf [environment:default] OVEHOSTED_CORE/rollbackProceed=none:None OVEHOSTED_CORE/screenProceed=none:None OVEHOSTED_CORE/deployProceed=bool:true OVEHOSTED_CORE/upgradeProceed=none:None OVEHOSTED_CORE/confirmSettings=bool:true OVEHOSTED_STORAGE/domainType=str:nfs3 OVEHOSTED_STORAGE/imgSizeGB=str:58 OVEHOSTED_STORAGE/storageDomain=str:hosted_storage OVEHOSTED_STORAGE/storageDomainConnection=str:10.66.148.11:/home/yzhao1/nfs8 OVEHOSTED_STORAGE/mntOptions=str: OVEHOSTED_NETWORK/bridgeIf=str:eno1 OVEHOSTED_NETWORK/bridgeName=str:ovirtmgmt OVEHOSTED_NETWORK/firewallManager=str:iptables OVEHOSTED_NETWORK/gateway=str:10.73.75.254 OVEHOSTED_NETWORK/fqdn=str:rhevh-hostedengine-vm-05.lab.eng.pek2.redhat.com OVEHOSTED_VM/bootDevice=str:disk OVEHOSTED_VM/vmVCpus=str:2 OVEHOSTED_VM/vmMACAddr=str:52:54:00:5d:21:64 OVEHOSTED_VM/vmMemSizeMB=int:4096 OVEHOSTED_VM/cloudinitVMStaticCIDR=str: OVEHOSTED_VM/cloudinitVMDNS=str: OVEHOSTED_VM/cloudinitVMTZ=str:Asia/Shanghai OVEHOSTED_VM/cloudInitISO=str:generate OVEHOSTED_VM/cloudinitInstanceHostName=str:rhevh-hostedengine-vm-05 OVEHOSTED_VM/cloudinitInstanceDomainName=str:lab.eng.pek2.redhat.com OVEHOSTED_VM/cloudinitExecuteEngineSetup=bool:true OVEHOSTED_VM/automateVMShutdown=bool:true OVEHOSTED_VM/cloudinitRootPwd=str:redhat OVEHOSTED_VM/rootSshPubkey=str: OVEHOSTED_VM/rootSshAccess=str:yes OVEHOSTED_VM/cloudinitVMETCHOSTS=bool:true OVEHOSTED_ENGINE/adminPassword=str:redhat OVEHOSTED_VDSM/consoleType=str:vnc OVEHOSTED_VDSM/cpu=str:model_Broadwell OVEHOSTED_NOTIF/smtpServer=str:localhost OVEHOSTED_NOTIF/smtpPort=str:25 OVEHOSTED_NOTIF/sourceEmail=str:root@localhost OVEHOSTED_NOTIF/destEmail=str:root@localhost
From the logs I see that this happens starting the vintage flow from cockpit but now we completely removed it supporting only the new ansible flow from cockpit and there we are not directly dealing with sanlock. According to comment 15 is not reproducible from command line so let's close it as WONTFIX since we removed the vintage flow support from cockpit and CLI works.
BZ<2>Jira Resync