Bug 1831105

Summary: Docs: [ceph-osd] osd failed to come up(ceph_assert(ondisk_format > 0)) on rhel 7.8 rhcs4 deployment
Product: [Red Hat Storage] Red Hat Ceph Storage Reporter: Andrew Schoen <aschoen>
Component: DocumentationAssignee: ceph-docs <ceph-docs>
Status: CLOSED NOTABUG QA Contact: Tejas <tchandra>
Severity: high Docs Contact:
Priority: high    
Version: 4.0CC: agunn, akupczyk, anharris, aschoen, bhubbard, ceph-eng-bugs, ceph-qe-bugs, dzafman, hyelloji, jdurgin, kchai, kdreyer, msekleta, nojha, rzarzyns, sseshasa, tchandra, vpoliset
Target Milestone: rcKeywords: AutomationBlocker
Target Release: 4.1   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: Doc Type: Known Issue
Doc Text:
Cause: Using partitions for the the --block.db and --block.wal arguments of the ceph-volume lvm create command. The db and wal options for the `lvm_volumes` config option in ceph-ansible is used to set those arguments during a deployment. Consequence: Occasionally the OSD will not start because udev resets the partitions permissions back to root:disk after creation by ceph-volume. Workaround (if any): Start the ceph-volume systemd unit manually for the failed OSD. For example, if the failed OSD has an ID of 8 the workaround would be running `systemctl start 'ceph-volume@lvm-8-*'`. If you know the failed OSDs UUID as well you can use the service command: `service ceph-volume@lvm-8-4c6ddc44-9037-477d-903c-63b5a789ade5 start`, where 4c6ddc44-9037-477d-903c-63b5a789ade5 is the UUID for osd.8. Result: Permissions on the partitions affected would be changed back to ceph:ceph and the OSD will be restarted and join the cluster.
Story Points: ---
Clone Of: 1822134 Environment:
Last Closed: 2020-05-08 15:07:52 UTC Type: ---
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:
Bug Depends On: 1822134    
Bug Blocks: