Bug 1564214
Summary: | [ceph-ansible] : osd scenario -lvm : playbook failing when initiated second time | ||||||||
---|---|---|---|---|---|---|---|---|---|
Product: | [Red Hat Storage] Red Hat Ceph Storage | Reporter: | Vasishta <vashastr> | ||||||
Component: | Ceph-Ansible | Assignee: | Sébastien Han <shan> | ||||||
Status: | CLOSED ERRATA | QA Contact: | Vasishta <vashastr> | ||||||
Severity: | high | Docs Contact: | Bara Ancincova <bancinco> | ||||||
Priority: | high | ||||||||
Version: | 3.0 | CC: | adeza, aschoen, ceph-eng-bugs, edonnell, gmeno, hnallurv, kdreyer, nthomas, rperiyas, sankarshan, shan, tserlin, vakulkar | ||||||
Target Milestone: | rc | ||||||||
Target Release: | 3.2 | ||||||||
Hardware: | Unspecified | ||||||||
OS: | Unspecified | ||||||||
Whiteboard: | |||||||||
Fixed In Version: | RHEL: ceph-ansible-3.2.0-0.1.rc1.el7cp Ubuntu: ceph-ansible_3.2.0~rc1-2redhat1 | Doc Type: | Bug Fix | ||||||
Doc Text: |
.Expanding clusters deployed with `osd_scenario: lvm` works
Previously, the `ceph-ansible` utility could not expand a cluster that was deployed by using the `osd_scenario: lvm` option. The underlying source code has been modified, and clusters deployed with `osd_scenario: lvm` can be expanded as expected.
|
Story Points: | --- | ||||||
Clone Of: | Environment: | ||||||||
Last Closed: | 2019-01-03 19:01:22 UTC | Type: | Bug | ||||||
Regression: | --- | Mount Type: | --- | ||||||
Documentation: | --- | CRM: | |||||||
Verified Versions: | Category: | --- | |||||||
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |||||||
Cloudforms Team: | --- | Target Upstream Version: | |||||||
Embargoed: | |||||||||
Bug Depends On: | |||||||||
Bug Blocks: | 1557269, 1629656 | ||||||||
Attachments: |
|
The PRs that fix this have not been backported to the stable-3.0 branch. However, even if they were you could not use a partition or raw device for 'data' and expect the playbook to be idempotent until https://github.com/ceph/ceph/pull/20620 makes it into a release. That PR 20620 will be in Ceph v12.2.5 upstream. (In reply to Vasishta from comment #0) > Description of problem: > When playbook was initiated to add nodes, the task 'create filestore osds > with dedicated journal' failed trying to create OSD on lvs and disk > partitions which are being used by existing OSDs. > With this issue, user won't be able to successfully add new nodes to the cluster with OSDs having data part on logical volumes and journal on disk partitions. (In reply to Ken Dreyer (Red Hat) from comment #4) > That PR 20620 will be in Ceph v12.2.5 upstream. @Ken, it means we will not have the fix for this in z2? As per comment 5, this bug limits the ability to expand the cluster. Is there a way we can get the fix in z2? Created attachment 1419115 [details]
File contains contents ansible-playbook log
Not able to expand cluster even when data and journal were on logical volumes.
Failing while running same task which should have skipped as per my understanding.
$ cat /usr/share/ceph-ansible/group_vars/osds.yml | egrep -v ^# | grep -v ^$
---
dummy:
copy_admin_key: true
osd_scenario: lvm
lvm_volumes:
- data: data1
data_vg: d_vg
journal: journal1
journal_vg: j_vg
- data: data2
data_vg: d_vg
journal: journal2
journal_vg: j_vg
- data: data3
data_vg: d_vg
journal: journal3
journal_vg: j_vg
Not sure I fully got what happened here, Andrew has more knowledge than me on that ceph-ansible code and on the BZ itself. Andrew, could you please fill out the Doc Text field for me? Thanks We have to fix the idempotent nature of rerunning the playbook here, we use that for other add/remove operations. lgtm, thanks Working fine with lvm-batch scenario, moving to VERIFIED state. Regards, Vasishta Shastry QE, Ceph Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2019:0020 |
Created attachment 1417857 [details] File contains contents of ansible-playbook log Description of problem: When playbook was initiated to add nodes, the task 'create filestore osds with dedicated journal' failed trying to create OSD on lvs and disk partitions which are being used by existing OSDs. Version-Release number of selected component (if applicable): ceph-ansible-3.0.28-1.el7cp.noarch How reproducible: Always (1/1) Steps to Reproduce: 1.Configure ceph-ansible to initiate a ceph cluster with at least OSD with lv as data part and a disk partition as journal part 2. Once the Cluster is up, rerun playbook Actual results: TASK [ceph-osd : use ceph-volume to create filestore osds with dedicated journals] is trying to create OSD on lv and disk partition which is already being used by another OSD Expected results: Task must be skipped