Red Hat Bugzilla – Bug 1304401
OSPD failed to detect and install ceph storage node
Last modified: 2016-10-13 16:06:19 EDT
Created attachment 1120790 [details]
Description of problem:
The overcloud was successfully installed without 1 Ceph storage node installed properly. The overcloud required topology is:
2 compute nodes
3 Ceph storage nodes, each with 3 hard drives, vdb, vdc ,vdd.
The ceph.yaml file is configured:
The result is the controller is installed on 1 of the servers with the 4 hard drives, 2 Ceph storage nodes are installed properly and an additional Ceph storage node is set with no OSDs.
* the installation runs on a virtual setup
Version-Release number of selected component (if applicable):
Steps to Reproduce:
1. Set ceph.yaml with additional hard drives
2. Install overcloud
The overcloud installation failed - the storage nodes are misconfigured and the OSPD says that installation finished successfully.
The OSPD should detect the server with the 4 hard drives, install and run the OSDs services
Can you paste the deploy command, attach any customized yaml and the output from 'sudo ceph status' from one of the controller nodes?
hi Yogev, is this still a bug? Can you reply to comment #2?
(In reply to Giulio Fidente from comment #3)
> hi Yogev, is this still a bug? Can you reply to comment #2?
I got a workaround this issue:
1) Create a new flavor
$openstack flavor create --id auto --ram 4096 --disk 10 --vcpus 1 cephStorage
2) Add a property to the flavor
$openstack flavor set --property 'cpu_arch'='x86_64' --property 'capabilities:boot_option'='local' --property 'capabilities:profile'='cephStorage' cephStorage
3) Add a property to the node with ironic
$ ironic node-update <ceph storage node uuid> add properties/capabilities='profile:cephStorage,boot_option:local'
And the customize yaml file is in the description of the bug
This bug did not make the OSP 8.0 release. It is being deferred to OSP 10.
If disks were previously formatted by/for a different Ceph cluster, the cluster FSID won't match and the OSP Director won't reuse them.
Before BZ #1370439, the deployment would not fail in such a scenario, but silently discard pre-owned disks.
With recent builds instead (which include the fix for BZ #1370439), the deployment should fail instead.
Can you retry formatting the Ceph disks with an empty GPT label during the deployment, as documented in: https://access.redhat.com/documentation/en/red-hat-openstack-platform/9/single/red-hat-ceph-storage-for-the-overcloud/#Formatting_Ceph_Storage_Nodes_Disks_to_GPT
(In reply to Giulio Fidente from comment #9)
> If disks were previously formatted by/for a different Ceph cluster, the
> cluster FSID won't match and the OSP Director won't reuse them.
> Before BZ #1370439, the deployment would not fail in such a scenario, but
> silently discard pre-owned disks.
> With recent builds instead (which include the fix for BZ #1370439), the
> deployment should fail instead.
> Can you retry formatting the Ceph disks with an empty GPT label during the
> deployment, as documented in:
I have tried it and it worked.
A deployment finished successfully, with Giulio's comment