Bug 1455187 - [ceph-ansible] [ceph-container] : purge cluster fails with error 'raw_journal_devices is undefined' for collocated journal and non dmcrypt osd options
Summary: [ceph-ansible] [ceph-container] : purge cluster fails with error 'raw_journal...
Keywords:
Status: CLOSED ERRATA
Alias: None
Product: Red Hat Storage Console
Classification: Red Hat Storage
Component: ceph-ansible
Version: 2
Hardware: x86_64
OS: Linux
unspecified
high
Target Milestone: ---
: 2
Assignee: Andrew Schoen
QA Contact: Vasishta
URL:
Whiteboard:
: 1456085 (view as bug list)
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2017-05-24 12:35 UTC by Rachana Patel
Modified: 2017-06-19 13:18 UTC (History)
13 users (show)

Fixed In Version: ceph-ansible-2.2.9-1.el7scon
Doc Type: If docs needed, set a value
Doc Text:
Clone Of:
Environment:
Last Closed: 2017-06-19 13:18:27 UTC
Embargoed:


Attachments (Terms of Use)


Links
System ID Private Priority Status Summary Last Updated
Github ceph ceph-ansible pull 1568 0 None None None 2017-05-25 18:40:36 UTC
Github ceph ceph-ansible pull 1582 0 None None None 2017-06-02 14:19:44 UTC
Github ceph ceph-ansible pull 1585 0 None None None 2017-06-05 14:39:55 UTC
Red Hat Product Errata RHBA-2017:1496 0 normal SHIPPED_LIVE ceph-installer, ceph-ansible, and ceph-iscsi-ansible update 2017-06-19 17:14:02 UTC

Description Rachana Patel 2017-05-24 12:35:20 UTC
Description of problem:
=======================
 purge cluster fails with error 'raw_journal_devices is undefined'  for collocated journal and non dmcrypt  osd options


Version-Release number of selected component (if applicable):
===========================================================
ceph-ansible-2.2.6-1.el7scon.noarch


How reproducible:
================
always


Steps to Reproduce:
===================
1. create cluster(containerized) having 1 MON and 6 OSD, 1 RGW and 1 MDSS  (choose collocated journal scenario)
2. once cluster is up and healthy, purge it using below command

ansible-playbook purge-docker-cluster.yml -i /etc/ansible/temp -vv


Actual results:
===============TASK [zap ceph osd disks] ******************************************************
task path: /root/temp/purge-docker-cluster.yml:241
fatal: [magna078]: FAILED! => {"failed": true, "msg": "'raw_journal_devices' is undefined"}
fatal: [magna082]: FAILED! => {"failed": true, "msg": "'raw_journal_devices' is undefined"}
fatal: [magna084]: FAILED! => {"failed": true, "msg": "'raw_journal_devices' is undefined"}
	to retry, use: --limit @/root/temp/purge-docker-cluster.retry

PLAY RECAP *********************************************************************
localhost                  : ok=0    changed=0    unreachable=0    failed=0   
magna075                   : ok=5    changed=3    unreachable=0    failed=0   
magna078                   : ok=4    changed=2    unreachable=0    failed=1   
magna082                   : ok=4    changed=2    unreachable=0    failed=1   
magna084                   : ok=9    changed=4    unreachable=0    failed=1 


Expected results:


Additional info:

Comment 5 John Poelstra 2017-05-24 15:09:34 UTC
discussed at program meeting, definitely a blocker, will look at today and give an estimate on fixing deadline today

Comment 6 seb 2017-05-24 15:21:01 UTC
I don't see why this is a blocker, this happens when trying to purge a cluster, which is IMHO not so common. Can we get more background why this is a blocker?

Comment 7 Christina Meno 2017-05-24 23:20:11 UTC
Seb,

My experience with Ceph says that when customers go to do a POC deployment they will make mistakes and want to start "from scratch" which usually leads them to wanting something like this. Is there another way for us to satisfy that?

cheers,
G

Comment 8 Andrew Schoen 2017-05-25 14:53:53 UTC
Upstream PR: https://github.com/ceph/ceph-ansible/pull/1568

Comment 9 Andrew Schoen 2017-05-25 21:10:44 UTC
(In reply to Andrew Schoen from comment #8)
> Upstream PR: https://github.com/ceph/ceph-ansible/pull/1568

This PR allows the purge-docker-cluster.yml playbook to complete but in my testing I'm unable to redploy new OSDs to a purged node, still investigating.

Comment 10 seb 2017-05-29 09:02:07 UTC
@Greg, fair enough.

Comment 11 Ian Colle 2017-05-31 14:47:44 UTC
*** Bug 1456085 has been marked as a duplicate of this bug. ***

Comment 12 John Poelstra 2017-05-31 15:06:22 UTC
discussed at program meeting, believe we can have in build today

Comment 13 Andrew Schoen 2017-05-31 15:13:12 UTC
backport PR: https://github.com/ceph/ceph-ansible/pull/1575

Comment 14 Andrew Schoen 2017-05-31 17:10:12 UTC
backport PR merged and new 2.2.8 tag cut https://github.com/ceph/ceph-ansible/tree/v2.2.8

Comment 20 Andrew Schoen 2017-06-02 14:00:57 UTC
Upstream PR: https://github.com/ceph/ceph-ansible/pull/1582

Comment 21 Andrew Schoen 2017-06-05 14:37:33 UTC
backport PR: https://github.com/ceph/ceph-ansible/pull/1585

Comment 22 John Poelstra 2017-06-05 15:09:08 UTC
discussed at meeting. another day until build for QE

Comment 23 Andrew Schoen 2017-06-05 22:27:47 UTC
This is included in the v2.2.9 upstream tag

Comment 26 Vasishta 2017-06-06 16:09:59 UTC
Hi, 

'ansible-playbook purge-docker-cluster.yml' ran successfully without any error.
Moving BZ to VERIFIED state.

verified using :
ceph-ansible-2.2.9-1.el7
ceph-10.2.7-24.el7


Regards,
Vasishta

Comment 28 errata-xmlrpc 2017-06-19 13:18:27 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory, and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHBA-2017:1496


Note You need to log in before you can comment on or make changes to this bug.