Note: This bug is displayed in read-only format because the product is no longer active in Red Hat Bugzilla.
This project is now read‑only. Starting Monday, February 2, please use https://ibm-ceph.atlassian.net/ for all bug tracking management.

Bug 1614878

Summary: [ceph-container] : OSD containers are not coming up when there are multiple cluster conf files present in ceph directory
Product: [Red Hat Storage] Red Hat Ceph Storage Reporter: Vasishta <vashastr>
Component: ContainerAssignee: Sébastien Han <shan>
Status: CLOSED ERRATA QA Contact: Vasishta <vashastr>
Severity: high Docs Contact:
Priority: medium    
Version: 3.1CC: anharris, ceph-eng-bugs, evelu, gabrioux, hnallurv, kdreyer, shan
Target Milestone: rc   
Target Release: 3.1   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: rhceph:ceph-3.1-rhel-7-containers-candidate-59306-20180821201312 Doc Type: If docs needed, set a value
Doc Text:
Story Points: ---
Clone Of: Environment:
Last Closed: 2018-09-26 19:16:42 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:
Attachments:
Description Flags
File contains snippet of journald logs of an OSD service with verbose enabled none

Description Vasishta 2018-08-10 15:24:49 UTC
Created attachment 1475108 [details]
File contains snippet of journald logs of an OSD service with verbose enabled

Description of problem:
When multiple cluster confs are there in ceph directory (/etc/ceph) OSD containers fail to start after service restart

Version-Release number of selected component (if applicable):
ceph-3.1-rhel-7-containers-candidate-64277-20180806190637

How reproducible:
Always (1/1)

Steps to Reproduce:
1. Configure containerized cluster 
2. copy some other clusters conf (or any file with ' fsid ' in the file) file to /etc/ceph
3. Restart OSD

Actual results:
OSD containers fail to restart

Expected results:
OSD containers must be up and running

Additional info:
1) other conf file was included to configure rbd-mirroring
2) We were trying to upgrade from 3.0.z5 to 3.1

It was observed that -
"basename "$(grep -R fsid /etc/ceph/ | grep -oE '^[^.]*')" in [1] is returning cluster name of unexpected cluster (The other one)


[1] - https://github.com/ceph/ceph-container/blob/fa58105eb5d8765bf66ec1d2f52b2585cbcac7f4/src/daemon/common_functions.sh#L487

Comment 12 Sébastien Han 2018-08-20 08:27:46 UTC
The latest contains image has the fix, commit ID: 3da098568429e5c24d9924b08c819303ac6d7732

Comment 15 Vasishta 2018-08-23 05:29:58 UTC
Working fine with ceph-3.1-rhel-7-containers-candidate-79213-20180822202815

Moving to VERIFIED state.

Regards,
Vasishta shatsry
QE, Ceph

Comment 17 errata-xmlrpc 2018-09-26 19:16:42 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory, and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHBA-2018:2820