Bug 1509230 - [container][ceph-ansible] container deployment with osd-scenario as lvm fails
Summary: [container][ceph-ansible] container deployment with osd-scenario as lvm fails
Alias: None
Product: Red Hat Ceph Storage
Classification: Red Hat
Component: Ceph-Ansible
Version: 3.0
Hardware: Unspecified
OS: Unspecified
Target Milestone: rc
: 3.2
Assignee: Guillaume Abrioux
QA Contact: Ramakrishnan Periyasamy
Bara Ancincova
Depends On:
Blocks: 1494421 1553640 1629656
TreeView+ depends on / blocked
Reported: 2017-11-03 11:36 UTC by Madhavi Kasturi
Modified: 2019-01-03 19:01 UTC (History)
21 users (show)

Fixed In Version: RHEL: ceph-ansible-3.2.0-0.1.rc1.el7cp Ubuntu: ceph-ansible_3.2.0~rc1-2redhat1
Doc Type: Bug Fix
Doc Text:
.`osd_scenario: lvm` now works when deploying Ceph in containers Previously, the `lvm` installation scenario did not work when deploying a Ceph cluster in containers. With this update, the `osd_scenario: lvm` installation method is supported as expected in this situation.
Clone Of:
Last Closed: 2019-01-03 19:01:20 UTC
Target Upstream Version:

Attachments (Terms of Use)
ceph-ansible playbook log (24.41 KB, text/plain)
2017-11-03 11:36 UTC, Madhavi Kasturi
no flags Details

System ID Priority Status Summary Last Updated
Github ceph ceph-ansible pull 2144 None closed osd: skip some set_fact when osd_scenario=lvm 2020-05-06 19:34:52 UTC
Github ceph ceph-ansible pull 2866 None closed ceph-osd: ceph-volume container support 2020-05-06 19:34:53 UTC
Red Hat Product Errata RHBA-2019:0020 None None None 2019-01-03 19:01:49 UTC

Description Madhavi Kasturi 2017-11-03 11:36:26 UTC
Created attachment 1347282 [details]
ceph-ansible playbook log

Description of problem:
Deployment of container with osd scenario as "lvm" fails.

Version-Release number of selected component (if applicable):
[admin@magna003 ceph-ansible]$ rpm -qa | grep ansible
[admin@magna003 ceph-ansible]$ 

How reproducible:

Steps to Reproduce:
1. created lv cache volume on the osd nodes using the below commands.
a. pvcreate  /dev/sdb1 /dev/sdc1
b. vgcreate data_vg /dev/sdb1 /dev/sdc1
c. lvcreate -L 400G -n slowdisk data_vg /dev/sdb1 
d. lvcreate -L 100G -n cachedisk data_vg /dev/sdc1
e. lvcreate -L 2G -n metadisk data_vg /dev/sdc1
f. lvconvert --type cache-pool /dev/data_vg/cachedisk --poolmetadata /dev/data_vg/metadisk
g. lvconvert --type cache data_vg/slowdisk --cachepool data_vg/cachedisk
2. In osds.yml set the osd_scenario to "lvm"
3. The contianer deployment fails.

p.s. used /dev/sdd1 partition for journal

TASK [ceph-defaults : resolve device link(s)] *********************************************************************************************************************************************************************
fatal: [magna030]: FAILED! => {"failed": true, "msg": "'devices' is undefined"}
Actual results:
The deployment fails 

Expected results:
The deployment should succeed.

Additional info:
ubuntu@magna003 ceph-ansible]$ cat group_vars/osds.yml | egrep -v ^# | grep -v ^$
osd_scenario: lvm #"{{ 'collocated' if journal_collocation or dmcrytpt_journal_collocation else 'non-collocated' if raw_multi_journal or dmcrypt_dedicated_journal else 'dummy' }}" # backward compatibility with stable-2.2, will disappear in stable 3.1
   - data: slowdisk
     journal: /dev/sdd1
     data_vg: data_vg

[ubuntu@magna003 ceph-ansible]$ cat group_vars/all.yml | egrep -v ^# | grep -v ^$
fetch_directory: ~/ceph-ansible-keys
ceph_origin: distro
ceph_repository: rhcs
monitor_interface: eno1
ceph_docker_image: "rhcs"
ceph_docker_image_tag: "ceph-3.0-rhel-7-docker-candidate-82532-20171102231218"
ceph_docker_registry: ""brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888"
containerized_deployment: True #"{{ True if mon_containerized_deployment or osd_containerized_deployment or mds_containerized_deployment or rgw_containerized_deployment else False }}" # backward compatibility with stable-2.2, will disappear in stable 3.1
[ubuntu@magna003 ceph-ansible]$ 

[ubuntu@magna003 ceph-ansible]$ cat /etc/ansible/hosts 



[ubuntu@magna003 ceph-ansible]$

Comment 3 Alfredo Deza 2017-11-03 17:08:35 UTC
Not sure I can help with the container portion of this. I can assist with the ceph-volume implementation though.

Comment 4 Guillaume Abrioux 2017-11-07 15:06:08 UTC

there are tasks to set some facts in ceph-defaults role which should be skipped when using osd_scenario: lvm.
it's fixed upstream, waiting for the CI to merge the commit in master.

Comment 5 Guillaume Abrioux 2017-11-07 15:50:09 UTC
In any case, even if the error mentioned here is fixed, you are going to get an issue like no osd up, indeed, as far as I know, lvm scenario is not supposed to work with containerized deployments yet.

Comment 11 Ben England 2018-02-07 12:33:49 UTC
Since RHOSP 13 now depends on RHCS 3.0z2, can we get this support into that z-stream build?  

Background: we need ceph-volume support in Ceph containers, particularly for bluestore.  Bluestore is critical to RHCS performance improvement, and ceph-volume seems critical for supporting deployment of bluestore, explained here by upstream documentation.  


This is particularly true for increasingly common all-flash configurations, where you need multiple OSDs/NVM device (or lose >= 40% throughput).  Only osd_scenario=lvm allows this functionality today.


Comment 15 Guillaume Abrioux 2018-07-17 09:06:57 UTC
Hi John,

This patch has been backported in 3.0 since 3.0.11 and v3.1.0beta2 for 3.1.

Note that these backports are not going to add ceph-volume support in containers.

Comment 16 Ken Dreyer (Red Hat) 2018-10-16 18:54:35 UTC
Guillaume, what additional changes in ceph-ansible and ceph-container are needed for this BZ?

Comment 23 Ramakrishnan Periyasamy 2018-11-21 03:36:48 UTC
Moving this bug to verified state

In Container OSD_scenarios as lvm works, verified in build ceph-ansible-3.2.0-0.1.rc3.el7cp.noarch

Comment 25 errata-xmlrpc 2019-01-03 19:01:20 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory, and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.


Note You need to log in before you can comment on or make changes to this bug.