Bug 1507943
| Summary: | [Ansible]: lvm osd cluster installation fails when using custom cluster name | ||
|---|---|---|---|
| Product: | [Red Hat Storage] Red Hat Ceph Storage | Reporter: | Ramakrishnan Periyasamy <rperiyas> |
| Component: | Ceph-Ansible | Assignee: | Sébastien Han <shan> |
| Status: | CLOSED WONTFIX | QA Contact: | Ramakrishnan Periyasamy <rperiyas> |
| Severity: | medium | Docs Contact: | Bara Ancincova <bancinco> |
| Priority: | high | ||
| Version: | 3.0 | CC: | adeza, anharris, aschoen, ceph-eng-bugs, flucifre, gmeno, hnallurv, kdreyer, nthomas, sankarshan, shan |
| Target Milestone: | rc | ||
| Target Release: | 3.1 | ||
| Hardware: | Unspecified | ||
| OS: | Unspecified | ||
| Whiteboard: | |||
| Fixed In Version: | Doc Type: | Known Issue | |
| Doc Text: |
.Adding LVM-based OSDs fail on clusters with other names than "ceph"
An attempt to install a new Ceph cluster or add OSDs by using the `osd_scenario: lvm` parameter fails on clusters that use other names than the default "ceph". To work around this problem on new clusters, use the default cluster name ("ceph").
|
Story Points: | --- |
| Clone Of: | Environment: | ||
| Last Closed: | 2018-05-15 23:06:15 UTC | Type: | Bug |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
| Bug Depends On: | 1266828, 1459861, 1635924 | ||
| Bug Blocks: | 1494421 | ||
Not sure why this is a 3.0 bug? This is no longer a problem in ceph-ansible master which requires the latest tip of luminous, that has this PR https://github.com/ceph/ceph/pull/18593 https://marc.info/?l=ceph-devel&m=149702444111127&w=2 says that we will be heading in a direction that custom cluster names won't be included in upstream deployment tools, this thread includes some directions on how existing cluster might migrate away from such a setup. This work that has been done doesn't cherry-pick cleanly. I'm going to recommend that we include a release note that ceph-volume isn't supported for clusters that aren't named ceph and push this out to 3.1 as not a blocker (In reply to Gregory Meno from comment #4) > https://marc.info/?l=ceph-devel&m=149702444111127&w=2 says that we will be > heading in a direction that custom cluster names won't be included in > upstream deployment tools, this thread includes some directions on how > existing cluster might migrate away from such a setup. > > This work that has been done doesn't cherry-pick cleanly. I'm going to > recommend that we include a release note that ceph-volume isn't supported > for clusters that aren't named ceph and push this out to 3.1 as not a blocker @Federico, please let us know the product's decision on this. Not fixing this bug in 3.0 will impact the use case of adding an lvm based osd to an existing cluster with custom cluster name. such users will not be able to use lvm feature. As per comment https://bugzilla.redhat.com/show_bug.cgi?id=1507943#c3 it seems that the issue is solved in 3.1, also this https://bugzilla.redhat.com/show_bug.cgi?id=1507943#c4 suggests no fix should be attempted on 3.0. So nothing to do for ceph-ansible. custom cluster names -> wontfix |
Description of problem: ceph-ansible fails to install lvm osd based cluster when using custom cluster name. Failed command: "ceph-volume lvm create --filestore --data d_vg/slow3 --journal j_vg/journal3" failed: [magna031] (item={u'data_vg': u'd_vg', u'journal': u'journal3', u'data': u'slow3', u'journal_vg': u'j_vg'}) => { "changed": true, "cmd": [ "ceph-volume", "lvm", "create", "--filestore", "--data", "d_vg/slow3", "--journal", "j_vg/journal3" ], "delta": "0:00:00.134095", "end": "2017-10-31 13:04:01.716948", "failed": true, "invocation": { "module_args": { "_raw_params": "ceph-volume lvm create --filestore --data d_vg/slow3 --journal j_vg/journal3", "_uses_shell": false, "chdir": null, "creates": null, "executable": null, "removes": null, "stdin": null, "warn": true } }, "item": { "data": "slow3", "data_vg": "d_vg", "journal": "journal3", "journal_vg": "j_vg" }, "msg": "non-zero return code", "rc": 1, "start": "2017-10-31 13:04:01.582853", "stderr": "--> NoSectionError: No section: 'global'", "stderr_lines": [ "--> NoSectionError: No section: 'global'" ], "stdout": "Running command: ceph-authtool --gen-print-key", "stdout_lines": [ "Running command: ceph-authtool --gen-print-key" ] } PLAY RECAP ****************************************************************************************************************************************************************** magna006 : ok=2 changed=0 unreachable=0 failed=0 magna028 : ok=100 changed=5 unreachable=0 failed=0 magna031 : ok=53 changed=2 unreachable=0 failed=1 magna058 : ok=2 changed=0 unreachable=0 failed=0 magna061 : ok=2 changed=0 unreachable=0 failed=0 magna063 : ok=2 changed=0 unreachable=0 failed=0 magna070 : ok=2 changed=0 unreachable=0 failed=0 magna077 : ok=2 changed=0 unreachable=0 failed=0 magna080 : ok=2 changed=0 unreachable=0 failed=0 magna086 : ok=51 changed=1 unreachable=0 failed=0 magna087 : ok=53 changed=1 unreachable=0 failed=0 Version-Release number of selected component (if applicable): ceph: ceph version 12.2.1-34.el7cp (dcc52a7ff94884920f4118fa44a24594df9475ec) luminous (stable) ceph-ansible: How reproducible: 2/2 Steps to Reproduce: 1. Configure lvm in OSD machines 2. Install cluster on lvm osd using ceph-ansible "osd_scenario='lvm'" Actual results: Expected results: Additional info: It will also fail in scenario the existing cluster is configured with custom name.