Bug 1769382
Summary: | RHHI-V deployment failed using the ansible playbook execution from command line | |||
---|---|---|---|---|
Product: | [Red Hat Storage] Red Hat Gluster Storage | Reporter: | Krist van Besien <kvanbesi> | |
Component: | gluster-ansible | Assignee: | Gobinda Das <godas> | |
Status: | CLOSED ERRATA | QA Contact: | SATHEESARAN <sasundar> | |
Severity: | medium | Docs Contact: | ||
Priority: | unspecified | |||
Version: | rhhiv-1.6 | CC: | bshetty, godas, lsurette, mtessun, rhs-bugs, sabose, sasundar, stirabos | |
Target Milestone: | --- | Keywords: | ZStream | |
Target Release: | RHGS 3.5.z Batch Update 2 | |||
Hardware: | x86_64 | |||
OS: | Linux | |||
Whiteboard: | ||||
Fixed In Version: | gluster-ansible-roles-1.0.5-7.el8rhgs | Doc Type: | No Doc Update | |
Doc Text: | Story Points: | --- | ||
Clone Of: | ||||
: | 1793398 (view as bug list) | Environment: |
rhhiv, rhel8
|
|
Last Closed: | 2020-06-16 05:57:29 UTC | Type: | Bug | |
Regression: | --- | Mount Type: | --- | |
Documentation: | --- | CRM: | ||
Verified Versions: | Category: | --- | ||
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | ||
Cloudforms Team: | --- | Target Upstream Version: | ||
Embargoed: | ||||
Bug Depends On: | ||||
Bug Blocks: | 1755481, 1793398 |
Description
Krist van Besien
2019-11-06 14:01:14 UTC
Can you please provide "rpm -qa |grep ovirt" output? trying to understand which version is affected. Closing with insufficient data resolution. If you can provide requested information please reopen. (In reply to Sandro Bonazzola from comment #1) > Can you please provide "rpm -qa |grep ovirt" output? trying to understand > which version is affected. I see the exact same problem. Here are the details that was requested earlier: [root@rhsqa-grafton7-nic2 defaults]# rpm -qa | grep ovirt ovirt-ansible-repositories-1.1.5-1.el7ev.noarch ovirt-hosted-engine-ha-2.3.5-1.el7ev.noarch ovirt-vmconsole-1.0.7-3.el7ev.noarch python-ovirt-engine-sdk4-4.3.2-1.el7ev.x86_64 ovirt-host-deploy-common-1.8.2-1.el7ev.noarch ovirt-ansible-hosted-engine-setup-1.0.28-1.el7ev.noarch cockpit-ovirt-dashboard-0.13.8-1.el7ev.noarch ovirt-vmconsole-host-1.0.7-3.el7ev.noarch cockpit-machines-ovirt-195-1.el7.noarch python2-ovirt-node-ng-nodectl-4.3.6-0.20190820.0.el7ev.noarch ovirt-provider-ovn-driver-1.2.22-1.el7ev.noarch ovirt-hosted-engine-setup-2.3.12-1.el7ev.noarch ovirt-node-ng-nodectl-4.3.6-0.20190820.0.el7ev.noarch ovirt-imageio-daemon-1.5.2-0.el7ev.noarch ovirt-host-4.3.4-1.el7ev.x86_64 python2-ovirt-setup-lib-1.2.0-1.el7ev.noarch ovirt-ansible-engine-setup-1.1.9-1.el7ev.noarch ovirt-host-dependencies-4.3.4-1.el7ev.x86_64 ovirt-imageio-common-1.5.2-0.el7ev.x86_64 python2-ovirt-host-deploy-1.8.2-1.el7ev.noarch Hi, in the extra vars the following vars were configured: { "he_appliance_password": "****", "he_admin_password": "****", "he_domain_type": "glusterfs", "he_fqdn": "hostedenginesm3.lab.eng.blr.redhat.com", "he_vm_mac_addr": "00:47:55:20:49:01", "he_default_gateway": "10.70.37.254", "he_mgmt_network": "ovirtmgmt", "he_ansible_host_name": "rhsqa-grafton7-nic2.lab.eng.blr.redhat.com", "he_storage_domain_name": "HostedEngine", "he_storage_domain_path": "/engine", "he_storage_domain_addr": "rhsqa-grafton7.lab.eng.blr.redhat.com", "he_mount_options": "backup-volfile-servers=rhsqa-grafton8.lab.eng.blr.redhat.com:rhsqa-grafton9.lab.eng.blr.redhat.com", "he_bridge_if": "enp129s0f0", "he_enable_hc_gluster_service": true, "he_mem_size_MB": "16384", "he_cluster": "Default" } its appears that `he_ansible_host_name` cannot be changed to something other then default `localhost` , because it relies on the ansible facts (which are gathered for localhost) . when removing this line `"he_ansible_host_name": "rhsqa-grafton7-nic2.lab.eng.blr.redhat.com",` deployment progressed. @simone can you confirm please. (In reply to Sandro Bonazzola from comment #1) > Can you please provide "rpm -qa |grep ovirt" output? trying to understand > which version is affected. As I have provided the required information, removing the needinfo on Krist van Besien Removed 'he_ansible_host_name' from extra vars and it lead to the success. Fix needs to be done at gluster-ansible-roles package that provides this extra vars file. [root@ ~]# rpm -qf /etc/ansible/roles/gluster.ansible/playbooks/hc-ansible-deployment/he_gluster_vars.json gluster-ansible-roles-1.0.5-7.el7rhgs.noarch So I will move this bug as part of gluster-ansible component *** Bug 1755481 has been marked as a duplicate of this bug. *** Fix is required in gluster-ansible-roles to remove the param 'he_ansible_host_name' from /etc/ansible/roles/gluster.ansible/playbooks/hc-ansible-deployment/he_gluster_vars.json [root@ ~]# cat /etc/ansible/roles/gluster.ansible/playbooks/hc-ansible-deployment/he_gluster_vars.json { "he_appliance_password": "encrypt-password-using-ansible-vault", "he_admin_password": "UI-password-for-login", "he_domain_type": "glusterfs", "he_fqdn": "FQDN-for-Hosted-Engine", "he_vm_mac_addr": "Valid MAC address", "he_default_gateway": "Valid Gateway", "he_mgmt_network": "ovirtmgmt", "he_ansible_host_name": "host1", <<------------------- This needs to be removed "he_storage_domain_name": "HostedEngine", "he_storage_domain_path": "/engine", "he_storage_domain_addr": "host1", "he_mount_options": "backup-volfile-servers=host2:host3", "he_bridge_if": "interface name for bridge creation", "he_enable_hc_gluster_service": true, "he_mem_size_MB": "4096", "he_cluster": "Default" } @Gobinda, Also make sure that the HE VM is allocated with 16GB of ram, with this change in vars "he_mem_size_MB": "16384" Tested with gluster-ansible-roles-1.0.5-7.el8rhgs Contents of he_gluster_vars.json file below: [root@ ~]# cat /etc/ansible/roles/gluster.ansible/playbooks/hc-ansible-deployment/he_gluster_vars.json { "he_appliance_password": "encrypt-password-using-ansible-vault", "he_admin_password": "UI-password-for-login", "he_domain_type": "glusterfs", "he_fqdn": "FQDN-for-Hosted-Engine", "he_vm_mac_addr": "Valid MAC address", "he_default_gateway": "Valid Gateway", "he_mgmt_network": "ovirtmgmt", "he_storage_domain_name": "HostedEngine", "he_storage_domain_path": "/engine", "he_storage_domain_addr": "host1-backend-network-FQDN", "he_mount_options": "backup-volfile-servers=host2-backend-network-FQDN:host3-backend-network-FQDN", "he_bridge_if": "interface name for bridge creation", "he_enable_hc_gluster_service": true, "he_mem_size_MB": "16384", "he_cluster": "Default" } 1. he_ansible_host_name is removed 2. he_mem_size_MB is updated to 16384 Ansible based CLI deployment is successful Removing old needinfo Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHEA-2020:2575 |