Description of problem: ------------------------- The gluster_cleanup.yml playbook doesnot detach the gluster peer's probed while the gluster deployment. Version-Release number of selected component: ---------------------------------------------- gluster-ansible-repositories-1.0-1.el7rhgs.noarch gluster-ansible-maintenance-1.0.1-1.el7rhgs.noarch gluster-ansible-features-1.0.4-5.el7rhgs.noarch gluster-ansible-cluster-1.0-1.el7rhgs.noarch gluster-ansible-roles-1.0.4-4.el7rhgs.noarch gluster-ansible-infra-1.0.3-3.el7rhgs.noarch rhvh-4.3.0.5-0.20190313 glusterfs-server-3.12.2-47.el7rhgs How reproducible: ------------------- Everytime Steps to Reproduce: -------------------- 1.After the successful gluster deployment , Cleanup the cluster via gluster_cleanup.yml playbook. 2.Go to the cli and run the command "gluster peer status " and user will see the previously probed peers Actual results: ------------------ The peers do not detach . Expected results: -------------------- The peers should detach after the cleanup playbook .This sometimes leads to failure in deployment.
Moving the bug to ON_QA based on the status of the base bug.
Upadte based on Base bug ------------------------ FailQA'ing the bug since the playbook was stuck in 'TASK [Delete a node from the trusted storage pool]' for hour n so. Since there was a warning added for gluster peer detach in RHGS 3.4.2 , the ansible script should use --mode=script to run. Thanks Sas for the info. Components: ========== glusterfs-6.0-2.el7rhgs.x86_64 gluster-ansible-roles-1.0.5-1.el7rhgs.noarch gluster-ansible-infra-1.0.4-1.el7rhgs.noarch gluster-ansible-repositories-1.0.1-1.el7rhgs.noarch gluster-ansible-features-1.0.5-1.el7rhgs.noarch Steps: ===== 1. Complete the gluster deployment 2. Run the gluster_cleanup.yml 3. The ansible run gets stuck for long in 'TASK [Delete a node from the trusted storage pool]'.
Tested with RHVH 4.3.5 + RHEL 7.7 + RHGS 3.4.4 ( interim build - glusterfs-6.0-6 ) with ansible 2.8.1-1 with: gluster-ansible-features-1.0.5-2.el7rhgs.noarch gluster-ansible-roles-1.0.5-2.el7rhgs.noarch gluster-ansible-infra-1.0.4-3.el7rhgs.noarch Post running cleanup playbook, it was observed that all the peers from that node are detached <snip> TASK [Get the list of hosts to be detached] *********************************************************************************************************************************************************************** changed: [host2.lab.eng.blr.redhat.com] TASK [Delete a node from the trusted storage pool] **************************************************************************************************************************************************************** changed: [host2.lab.eng.blr.redhat.com] => (item=host1.lab.eng.blr.redhat.com) skipping: [host2.lab.eng.blr.redhat.com] => (item=host2.lab.eng.blr.redhat.com) changed: [host2.lab.eng.blr.redhat.com] => (item=host3.lab.eng.blr.redhat.com) </snip> [root@host2 ~]# gluster pe s Number of Peers: 0
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2019:2963