Description of problem: ------------------------- Unable to remove MDS by using shrink-mds.yml using fqdn, Tried 3 different commands but none seems to work # ansible-playbook -e ireallymeanit=yes infrastructure-playbooks/shrink-mds.yml -i hosts -vv -e mds_to_kill=dell-r640-013 # ansible-playbook -e ireallymeanit=yes infrastructure-playbooks/shrink-mds.yml -i hosts -vv -e mds_to_kill=dell-r640-013.dsal.lab.eng.rdu2.redhat.com ansible-playbook -e ireallymeanit=yes infrastructure-playbooks/shrink-mds.yml -i hosts -vv -e mds_to_kill=ceph-mds-dell-r640-013 Errors pasted http://pastebin.test.redhat.com/894482 Version-Release number of selected component (if applicable): [root@dell-r640-012 /]# ceph version ceph version 14.2.8-89.el8cp (9ab115d618c72e7d9227441ec25ceb1487c76fb8) nautilus (stable) [root@dell-r640-012 /]# ceph versions { "mon": { "ceph version 14.2.8-89.el8cp (9ab115d618c72e7d9227441ec25ceb1487c76fb8) nautilus (stable)": 2 }, "mgr": { "ceph version 14.2.8-89.el8cp (9ab115d618c72e7d9227441ec25ceb1487c76fb8) nautilus (stable)": 1 }, "osd": { "ceph version 14.2.8-89.el8cp (9ab115d618c72e7d9227441ec25ceb1487c76fb8) nautilus (stable)": 2 }, "mds": { "ceph version 14.2.8-89.el8cp (9ab115d618c72e7d9227441ec25ceb1487c76fb8) nautilus (stable)": 1 }, "overall": { "ceph version 14.2.8-89.el8cp (9ab115d618c72e7d9227441ec25ceb1487c76fb8) nautilus (stable)": 6 } } How reproducible: Many times Steps to Reproduce: 1. Try to shrink-mds using shrink-mds.yml playbook 2. the hosts file contained fqdn 3. Check if the mds is removed Actual results: The MDS did not get removed Expected results: The mds should have been removed from the cluster Workaround: ------------ Instead of not having FQDNs as the hostname, try having something like this in hosts file #dell-r640-013.dsal.lab.eng.rdu2.redhat.com monitor_interface=em1 dell-r640-013 monitor_interface=em1 and run ansible-playbook -e ireallymeanit=yes infrastructure-playbooks/shrink-mds.yml -i hosts -vv -e mds_to_kill=dell-r640-013 The removal of MDS should be a success Additional info:
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Red Hat Ceph Storage 4.1 Bug Fix update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2020:4144