Description of problem: On existing ceph node are deployed OSDs using ceph-ansible ceph-volume batch scenario. When adding additional devices to the host_vars/node configuration file and running add-osd.yml or site.yml, the new devices are not being populated by OSDs: Initial node deployment: ---- # cat /etc/ansible/host_vars/osds-3 devices: - /dev/vde dedicated_devices: - /dev/vdb [root@osds-3 ~]# lsblk NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT vda 253:0 0 20G 0 disk ├─vda1 253:1 0 1M 0 part ├─vda2 253:2 0 100M 0 part /boot/efi └─vda3 253:3 0 19.9G 0 part / vdb 253:16 0 30G 0 disk └─ceph--block--dbs--199c644a--8657--46f3--8136--b6679cd68153-osd--block--db--7d8b9f07--e822--43da--866b--dc202c3740bd 252:1 0 30G 0 lvm vdc 253:32 0 30G 0 disk vdd 253:48 0 30G 0 disk vde 253:64 0 30G 0 disk └─ceph--block--6171157b--47dc--426a--ba3f--5876bb49e42b-osd--block--4b812472--826e--41c4--ba00--c6c0eedc64ae 252:0 0 30G 0 lvm Adding new OSDs ---- # cat /etc/ansible/host_vars/osds-3 devices: - /dev/vde - /dev/vdd dedicated_devices: - /dev/vdb - /dev/vdc # ansible-playbook infrastructure-playbooks/add-osd.yml --limit osds-3 ... 2020-11-11 03:15:58,084 p=47055 u=root | PLAY RECAP *********************************************************************************************************************************************************************************************************************************** 2020-11-11 03:15:58,085 p=47055 u=root | osds-3 : ok=184 changed=8 unreachable=0 failed=0 skipped=299 rescued=0 ignored=1 2020-11-11 03:15:58,093 p=47055 u=root | Wednesday 11 November 2020 03:15:58 -0500 (0:00:00.056) 0:04:41.749 **** 2020-11-11 03:15:58,094 p=47055 u=root | =============================================================================== .... [root@osds-3 ~]# lsblk NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT vda 253:0 0 20G 0 disk ├─vda1 253:1 0 1M 0 part ├─vda2 253:2 0 100M 0 part /boot/efi └─vda3 253:3 0 19.9G 0 part / vdb 253:16 0 30G 0 disk └─ceph--block--dbs--199c644a--8657--46f3--8136--b6679cd68153-osd--block--db--7d8b9f07--e822--43da--866b--dc202c3740bd 252:1 0 30G 0 lvm vdc 253:32 0 30G 0 disk vdd 253:48 0 30G 0 disk vde 253:64 0 30G 0 disk └─ceph--block--6171157b--47dc--426a--ba3f--5876bb49e42b-osd--block--4b812472--826e--41c4--ba00--c6c0eedc64ae 252:0 0 30G 0 lvm ---- The playbook ends with OK, but no new devices are being deployed. The task 2020-11-11 03:03:36,685 p=45877 u=root | TASK [ceph-config : run 'ceph-volume lvm batch --report' to see how many osds are to be created] does not set any devices to be deployed When we have in host_vars/node configuration file only new devices, the OSDs will get deployed, but that brings different issue, that the 'osd memory target' value is calculated per number of OSDs in host_vars/node configuration file, so it will get calculated wrongly for the amount of OSDs the node will have in total, old+new. Version-Release number of selected component (if applicable): ansible-2.8.16-1.el8ae.noarch ceph-ansible-4.0.31-1.el8cp.noarch "ceph version 14.2.8-111.el8cp How reproducible: Always Steps to Reproduce: 1. deploy node with x amount of devices 2. add additional device into host_vars/node configuration file, keep the old ones 3. try to deploy OSDs on those Actual results: Expected results: Additional info:
*** Bug 1966934 has been marked as a duplicate of this bug. ***
*** Bug 1970031 has been marked as a duplicate of this bug. ***
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Moderate: Red Hat Ceph Storage 4.3 Security and Bug Fix update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2022:1716
Thanks to this article I can learn more. Expand my knowledge and abilities. Actually the article is very real. https://www.mybalancenow.page/