Bug 1252158
Summary: | overcloud deploy with ceph reports success but ceph is not usable because OSD/journals not created and no ceph.conf | ||
---|---|---|---|
Product: | Red Hat OpenStack | Reporter: | jliberma <jliberma> |
Component: | rhosp-director | Assignee: | John Fulton <johfulto> |
Status: | CLOSED DUPLICATE | QA Contact: | Yogev Rabl <yrabl> |
Severity: | unspecified | Docs Contact: | |
Priority: | urgent | ||
Version: | 7.0 (Kilo) | CC: | hbrock, jdonohue, jean-francois.bibeau, jefbrown, johfulto, jomurphy, jraju, mburns, mcornea, morazi, rhel-osp-director-maint, skinjo |
Target Milestone: | beta | ||
Target Release: | 11.0 (Ocata) | ||
Hardware: | Unspecified | ||
OS: | Unspecified | ||
Whiteboard: | |||
Fixed In Version: | Doc Type: | Bug Fix | |
Doc Text: | Story Points: | --- | |
Clone Of: | Environment: | ||
Last Closed: | 2017-01-17 16:32:40 UTC | Type: | Bug |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: | |||
Bug Depends On: | |||
Bug Blocks: | 1399824 |
Description
jliberma@redhat.com
2015-08-10 20:32:25 UTC
Couple of thoughts here: Is there a good way to validate ceph.conf once it is created? If so perhaps this is something we might add to puppet-ceph to make it more robust there. With regards to the existing fsid on OSD disks should we be wiping disks clean on provisioning. Or perhaps we wipe disks clean when they get deleted. Ironic does have a clean_nodes setting (which we set to false in Instack) but we could instruct users of Ceph clusters to enable if this is a concern. This bug did not make the OSP 8.0 release. It is being deferred to OSP 10. I think we should fail early by checking the syntax on the ceph.yaml. However this could mean applying this kind of checks to all the other templates. Perhaps the easiest thing to do is to validate the state of the cluster once the deployment is done? Basically if Ceph health reports HEALTH_ERR we fail the stack and start investigating. *** Bug 1312192 has been marked as a duplicate of this bug. *** - This situation has been helped tremendously by RH 1370439, fixed in OSP10 - This bug should be closed after implementing the description in comment #12 This issue is a symptom of what happens when you don't clean your disks during deployment or redeployment. That symptom is now captured and the deploy fails as requested here as a result of the outcome of RH 1370439. After that failure happens, the fix is to enable a new flag to zap as described in RH 1377867. Now that our fix for 1377867 in the works upstream, to zap the old disks as per Dan in comment #5, and as per a discussion in DFG:Ceph (including Seb who had added comment #12), our conclusion is to mark this as a duplicate of 1377867. 1377867 is on schedule to be fixed in OSP11. *** This bug has been marked as a duplicate of bug 1377867 *** |