Bug 1876827
| Summary: | [cee/sd][ceph-volume] "ceph-volume lvm batch --report /dev/vdd " fails on lvm OSDs set up with dm-cache | ||
|---|---|---|---|
| Product: | [Red Hat Storage] Red Hat Ceph Storage | Reporter: | Tomas Petr <tpetr> |
| Component: | Ceph-Volume | Assignee: | Andrew Schoen <aschoen> |
| Status: | CLOSED ERRATA | QA Contact: | Ameena Suhani S H <amsyedha> |
| Severity: | high | Docs Contact: | |
| Priority: | high | ||
| Version: | 4.1 | CC: | anharris, aschoen, ceph-eng-bugs, ceph-qe-bugs, gmeno, gsitlani, knortema, mhackett, mmuench, tserlin |
| Target Milestone: | --- | ||
| Target Release: | 4.2z1 | ||
| Hardware: | Unspecified | ||
| OS: | Unspecified | ||
| Whiteboard: | |||
| Fixed In Version: | ceph-14.2.11-98.el8cp, ceph-14.2.11-98.el7cp | Doc Type: | Deprecated Functionality |
| Doc Text: |
.Deploying an OSD on a device setup for `dm-cache` is unsupported
Deploying an OSD with `dm-cache` is no longer supported. Use the Bluestore backend instead of `dm-cache`.
For more information, see the link:{install-guide}#how-to-migrate-the-object-store-from-filestore-to-bluestore[_How to migrate the object store from FileStore to BlueStore_] section in the _{storage-product} Installation Guide_.
|
Story Points: | --- |
| Clone Of: | Environment: | ||
| Last Closed: | 2021-04-28 20:12:31 UTC | Type: | Bug |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
| Bug Depends On: | |||
| Bug Blocks: | 1890121 | ||
|
Description
Tomas Petr
2020-09-08 09:34:35 UTC
Cluster deployed with: - to force vdb, vdc act like SSD (as this is virtual machine): # ansible osds -m shell -a ' echo 0 > /sys/block/vdb/queue/rotational ' # ansible osds -m shell -a ' echo 0 > /sys/block/vdc/queue/rotational ' - deploy cluster # ansible-playbook site-container.yml # cat group_vars/osds.yml osd_scenario: lvm copy_admin_key: true devices: - /dev/vdc - /dev/vdd - /dev/vde ----- # ansible osds -m shell -a 'sgdisk --new=0:0:+14G -- /dev/vdb; sgdisk --new=0:0:+10G -- /dev/vdb; pvcreate /dev/vdb1; pvcreate /dev/vdb2' [root@osds-0 ~]# pvs; vgs; lvs PV VG Fmt Attr PSize PFree /dev/vdb1 lvm2 --- 14.00g 14.00g /dev/vdb2 lvm2 --- 10.00g 10.00g /dev/vdc ceph-block-dbs-f1ad666e-7db9-4124-a25f-ace6f190b68a lvm2 a-- <30.00g 4.00m /dev/vdd ceph-block-a4b04d43-666d-4be6-b182-3adba73757fa lvm2 a-- <30.00g 0 /dev/vde ceph-block-8fe3deaa-6b03-45c9-975e-12d5107ab171 lvm2 a-- <30.00g 0 VG #PV #LV #SN Attr VSize VFree ceph-block-8fe3deaa-6b03-45c9-975e-12d5107ab171 1 1 0 wz--n- <30.00g 0 ceph-block-a4b04d43-666d-4be6-b182-3adba73757fa 1 1 0 wz--n- <30.00g 0 ceph-block-dbs-f1ad666e-7db9-4124-a25f-ace6f190b68a 1 2 0 wz--n- <30.00g 4.00m LV VG Attr LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert osd-block-de5e67e8-b562-42c9-b137-c0e28b0659f2 ceph-block-8fe3deaa-6b03-45c9-975e-12d5107ab171 -wi-ao---- <30.00g osd-block-9b97f428-992d-420a-bc56-df26204d974e ceph-block-a4b04d43-666d-4be6-b182-3adba73757fa -wi-ao---- <30.00g osd-block-db-05fdfa57-8d60-4000-bcc8-9f5c83ece9ec ceph-block-dbs-f1ad666e-7db9-4124-a25f-ace6f190b68a -wi-ao---- <15.00g osd-block-db-24857e40-d02d-42f0-83f1-f3ebc3544302 ceph-block-dbs-f1ad666e-7db9-4124-a25f-ace6f190b68a -wi-ao---- <15.00g [root@osds-0 ~]# systemctl stop ceph-osd@1 [root@osds-0 ~]# systemctl stop ceph-osd@4 [root@osds-0 ~]# lvchange -an ceph-block-8fe3deaa-6b03-45c9-975e-12d5107ab171/^C [root@osds-0 ~]# vgextend ceph-block-8fe3deaa-6b03-45c9-975e-12d5107ab171 /dev/vdb1 Volume group "ceph-block-8fe3deaa-6b03-45c9-975e-12d5107ab171" successfully extended [root@osds-0 ~]# vgextend ceph-block-a4b04d43-666d-4be6-b182-3adba73757fa /dev/vdb2 Volume group "ceph-block-a4b04d43-666d-4be6-b182-3adba73757fa" successfully extended [root@osds-0 ~]# lvcreate -L 9G -n cachedisk ceph-block-8fe3deaa-6b03-45c9-975e-12d5107ab171 /dev/vdb1 Logical volume "cachedisk" created. [root@osds-0 ~]# lvcreate -L 9G -n cachedisk ceph-block-a4b04d43-666d-4be6-b182-3adba73757fa /dev/vdb2 Logical volume "cachedisk" created. [root@osds-0 ~]# lvchange -an ceph-block-8fe3deaa-6b03-45c9-975e-12d5107ab171/cachedisk ceph-block-a4b04d43-666d-4be6-b182-3adba73757fa/cachedisk ceph-block-8fe3deaa-6b03-45c9-975e-12d5107ab171/osd-block-de5e67e8-b562-42c9-b137-c0e28b0659f2 ceph-block-a4b04d43-666d-4be6-b182-3adba73757fa/osd-block-9b97f428-992d-420a-bc56-df26204d974e [root@osds-0 ~]# lvconvert --type writecache --cachevol cachedisk ceph-block-8fe3deaa-6b03-45c9-975e-12d5107ab171/osd-block-de5e67e8-b562-42c9-b137-c0e28b0659f2 Erase all existing data on ceph-block-8fe3deaa-6b03-45c9-975e-12d5107ab171/cachedisk? [y/n]: y Logical volume ceph-block-8fe3deaa-6b03-45c9-975e-12d5107ab171/osd-block-de5e67e8-b562-42c9-b137-c0e28b0659f2 now has write cache. [root@osds-0 ~]# lvconvert --type writecache --cachevol cachedisk ceph-block-a4b04d43-666d-4be6-b182-3adba73757fa/osd-block-9b97f428-992d-420a-bc56-df26204d974e Erase all existing data on ceph-block-a4b04d43-666d-4be6-b182-3adba73757fa/cachedisk? [y/n]: y Logical volume ceph-block-a4b04d43-666d-4be6-b182-3adba73757fa/osd-block-9b97f428-992d-420a-bc56-df26204d974e now has write cache. [root@osds-0 ~]# lvchange -ay ceph-block-8fe3deaa-6b03-45c9-975e-12d5107ab171/cachedisk ceph-block-a4b04d43-666d-4be6-b182-3adba73757fa/cachedisk ceph-block-8fe3deaa-6b03-45c9-975e-12d5107ab171/osd-block-de5e67e8-b562-42c9-b137-c0e28b0659f2 ceph-block-a4b04d43-666d-4be6-b182-3adba73757fa/osd-block-9b97f428-992d-420a-bc56-df26204d974e Failed to find logical volume "ceph-block-8fe3deaa-6b03-45c9-975e-12d5107ab171/cachedisk" Failed to find logical volume "ceph-block-a4b04d43-666d-4be6-b182-3adba73757fa/cachedisk" [root@osds-0 ~]# systemctl start ceph-osd@1 [root@osds-0 ~]# systemctl start ceph-osd@4 [root@osds-0 ~]# pvs; vgs; lvs PV VG Fmt Attr PSize PFree /dev/vdb1 ceph-block-8fe3deaa-6b03-45c9-975e-12d5107ab171 lvm2 a-- <14.00g <5.00g /dev/vdb2 ceph-block-a4b04d43-666d-4be6-b182-3adba73757fa lvm2 a-- <10.00g 1020.00m /dev/vdc ceph-block-dbs-f1ad666e-7db9-4124-a25f-ace6f190b68a lvm2 a-- <30.00g 4.00m /dev/vdd ceph-block-a4b04d43-666d-4be6-b182-3adba73757fa lvm2 a-- <30.00g 0 /dev/vde ceph-block-8fe3deaa-6b03-45c9-975e-12d5107ab171 lvm2 a-- <30.00g 0 VG #PV #LV #SN Attr VSize VFree ceph-block-8fe3deaa-6b03-45c9-975e-12d5107ab171 2 1 0 wz--n- 43.99g <5.00g ceph-block-a4b04d43-666d-4be6-b182-3adba73757fa 2 1 0 wz--n- 39.99g 1020.00m ceph-block-dbs-f1ad666e-7db9-4124-a25f-ace6f190b68a 1 2 0 wz--n- <30.00g 4.00m LV VG Attr LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert osd-block-de5e67e8-b562-42c9-b137-c0e28b0659f2 ceph-block-8fe3deaa-6b03-45c9-975e-12d5107ab171 Cwi-aoC--- <30.00g [cachedisk_cvol] [osd-block-de5e67e8-b562-42c9-b137-c0e28b0659f2_wcorig] 0.01 osd-block-9b97f428-992d-420a-bc56-df26204d974e ceph-block-a4b04d43-666d-4be6-b182-3adba73757fa Cwi-aoC--- <30.00g [cachedisk_cvol] [osd-block-9b97f428-992d-420a-bc56-df26204d974e_wcorig] 0.01 osd-block-db-05fdfa57-8d60-4000-bcc8-9f5c83ece9ec ceph-block-dbs-f1ad666e-7db9-4124-a25f-ace6f190b68a -wi-ao---- <15.00g osd-block-db-24857e40-d02d-42f0-83f1-f3ebc3544302 ceph-block-dbs-f1ad666e-7db9-4124-a25f-ace6f190b68a -wi-ao---- <15.00g [root@osds-0 ~]# ceph-volume lvm batch --report /dev/vdd --> RuntimeError: Unable to use device, already a member of LVM: /dev/vdd When rerunning "ansible-playbook site-container.yml" playbook fails on following TASK for node where dm-cache is set up
TASK [ceph-config : run 'ceph-volume lvm batch --report' to see how many osds are to be created] *********************************************************************************************************************************************
task path: /usr/share/ceph-ansible/roles/ceph-config/tasks/main.yml:40
Tuesday 08 September 2020 05:32:50 -0400 (0:00:00.223) 0:05:37.147 *****
[WARNING]: The value 5120 (type int) in a string field was converted to '5120' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change.
[WARNING]: The value -1 (type int) in a string field was converted to '-1' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change.
changed: [10.10.95.125] => changed=true
cmd:
- podman
- run
- --rm
- --privileged
- --net=host
- --ipc=host
- --ulimit
- nofile=1024:4096
- -v
- /run/lock/lvm:/run/lock/lvm:z
- -v
- /var/run/udev/:/var/run/udev/:z
- -v
- /dev:/dev
- -v
- /etc/ceph:/etc/ceph:z
- -v
- /run/lvm/:/run/lvm/
- -v
- /var/lib/ceph/:/var/lib/ceph/:z
- -v
- /var/log/ceph/:/var/log/ceph/:z
- --entrypoint=ceph-volume
- registry.redhat.io/rhceph/rhceph-4-rhel8:4-32
- --cluster
- ceph
- lvm
- batch
- --bluestore
- --yes
- --prepare
- /dev/vdc
- /dev/vdd
- /dev/vde
- --report
- --format=json
delta: '0:00:01.686883'
end: '2020-09-08 05:32:52.626293'
rc: 0
start: '2020-09-08 05:32:50.939410'
stderr: |-
WARNING: The same type, major and minor should not be used for multiple devices.
WARNING: The same type, major and minor should not be used for multiple devices.
WARNING: The same type, major and minor should not be used for multiple devices.
WARNING: The same type, major and minor should not be used for multiple devices.
stderr_lines: <omitted>
stdout: |-
{
"changed": false,
"osds": [],
"vgs": []
}
stdout_lines: <omitted>
fatal: [10.10.93.192]: FAILED! => changed=true
cmd:
- podman
- run
- --rm
- --privileged
- --net=host
- --ipc=host
- --ulimit
- nofile=1024:4096
- -v
- /run/lock/lvm:/run/lock/lvm:z
- -v
- /var/run/udev/:/var/run/udev/:z
- -v
- /dev:/dev
- -v
- /etc/ceph:/etc/ceph:z
- -v
- /run/lvm/:/run/lvm/
- -v
- /var/lib/ceph/:/var/lib/ceph/:z
- -v
- /var/log/ceph/:/var/log/ceph/:z
- --entrypoint=ceph-volume
- registry.redhat.io/rhceph/rhceph-4-rhel8:4-32
- --cluster
- ceph
- lvm
- batch
- --bluestore
- --yes
- --prepare
- /dev/vdc
- /dev/vdd
- /dev/vde
- --report
- --format=json
msg: non-zero return code
rc: 1
stderr: |-
WARNING: The same type, major and minor should not be used for multiple devices.
WARNING: The same type, major and minor should not be used for multiple devices.
WARNING: The same type, major and minor should not be used for multiple devices.
WARNING: The same type, major and minor should not be used for multiple devices.
--> RuntimeError: Unable to use device, already a member of LVM: /dev/vdd
stderr_lines: <omitted>
stdout: ''
stdout_lines: <omitted>
changed: [10.10.92.15] => changed=true
cmd:
- podman
- run
- --rm
- --privileged
- --net=host
- --ipc=host
- --ulimit
- nofile=1024:4096
- -v
- /run/lock/lvm:/run/lock/lvm:z
- -v
- /var/run/udev/:/var/run/udev/:z
- -v
- /dev:/dev
- -v
- /etc/ceph:/etc/ceph:z
- -v
- /run/lvm/:/run/lvm/
- -v
- /var/lib/ceph/:/var/lib/ceph/:z
- -v
- /var/log/ceph/:/var/log/ceph/:z
- --entrypoint=ceph-volume
- registry.redhat.io/rhceph/rhceph-4-rhel8:4-32
- --cluster
- ceph
- lvm
- batch
- --bluestore
- --yes
- --prepare
- /dev/vdc
- /dev/vdd
- /dev/vde
- --report
- --format=json
delta: '0:00:02.095281'
end: '2020-09-08 05:32:53.130394'
rc: 0
start: '2020-09-08 05:32:51.035113'
stderr: |-
WARNING: The same type, major and minor should not be used for multiple devices.
WARNING: The same type, major and minor should not be used for multiple devices.
WARNING: The same type, major and minor should not be used for multiple devices.
WARNING: The same type, major and minor should not be used for multiple devices.
stderr_lines: <omitted>
stdout: |-
{
"changed": false,
"osds": [],
"vgs": []
}
stdout_lines: <omitted>
tested with same results for combination of: RHEL 8.2 (4.18.0-193.el8.x86_64) + RHCS 4.1z1a2 (14.2.8-91.el8cp) [root@osds-0 ~]# ceph-volume lvm batch --report /dev/vdd --> RuntimeError: Unable to use device, already a member of LVM: /dev/vdd RHEL 7.8 (3.10.0-1127.el7.x86_64) + RHCS 3.3z6 (12.2.12-124.el7cp) [root@osds-2 ~]# ceph-volume lvm batch --report /dev/vdd --> RuntimeError: Unable to use device, already a member of LVM: /dev/vdd Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Important: Red Hat Ceph Storage security, bug fix, and enhancement Update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2021:1452 |