Bug 1356619
Summary: | [USM] Import Cluster fails on an upgraded ceph cluster from 1.3.2 to 2.0 | ||
---|---|---|---|
Product: | [Red Hat Storage] Red Hat Storage Console | Reporter: | Tejas <tchandra> |
Component: | documentation | Assignee: | Aron Gunn <agunn> |
Status: | CLOSED CURRENTRELEASE | QA Contact: | sds-qe-bugs |
Severity: | high | Docs Contact: | |
Priority: | unspecified | ||
Version: | 2 | CC: | asriram, ceph-eng-bugs, ceph-qe-bugs, dahorak, gdrapeau, gmeno, hnallurv, kdreyer, lmiccini, nthomas, sankarshan, shtripat, tchandra |
Target Milestone: | --- | ||
Target Release: | 2 | ||
Hardware: | Unspecified | ||
OS: | Linux | ||
Whiteboard: | |||
Fixed In Version: | Doc Type: | If docs needed, set a value | |
Doc Text: | Story Points: | --- | |
Clone Of: | Environment: | ||
Last Closed: | 2018-11-19 05:31:40 UTC | Type: | Bug |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: |
Description
Tejas
2016-07-14 13:27:47 UTC
Hi Tejas, did you configure calamari user "admin" with password "admin"? See Bug 1345983. Gregory, We need some help from you to understand why CSRF token could be coming as blank for setup where it migrated to higher versions of ceph and calamari. After migration we had done "calamarictl inistailize" with admin user name, password. We suspected firewalld and selinux and tried with disabled mode as well but no luck. Request to have a look at the logs and let us know if you you can make out something. I tried simulating the issue today and find import cluster working with one work-around. The steps followed are as below - Step-1: Created a ceph 1.3.2 cluster with 1 MON and 2 OSD nodes (total 8 osds). Thanks to Tejas for getting this in place :) Step-2: Followed Red Hat Ceph Storage Installation guide, section-5.1/5.2 and upgraded the MON and OSD nodes to higher versions (2.0). Step-3: In this case calamari-1.3.3 was installed on the MON node only, so removed all calamari related rpms (diamond, graphite, salt-minion, calamari) using the command "yum remove" Step-4: After upgrade rebooted all the MON and OSD nodes and verified that "ceph -s" works fine Step-5: Created a fresh node and installed latest USM server bits and executed skyring-setup Step-6: On MON and OSD nodes added the USM agent repo and bootstrapped the nodes with ceph-installer using the ceph-installer apis /setup/ and /setup/agent Step-7: Verified that salt-minions are up and running on MON and OSD nodes and salt-master running on USM server node. Verified if salt-minions on MON and OSD nodes are properly referring USM server node's salt-master by looking at /etc/saly/minion.d/ceph_agent.conf Step-8: Opened USM UI and selected option for import cluster. All the 3 nodes (1 MON + 2 OSDs) get listed. Selected the MON node where calamari is up and running now. Continued and it listed all the participating nodes of the cluster. Submitted and cluster import started as a task. There was a glitch here while importing cluster as "rbd ls" command was hanging in MON node. Suspected default created pool "rbd" and removed the same using command "ceph osd pool delete rbd rbd --yes-i-really-really-mean-it" Step-9: Once default created pool "rbd" was removed, the import cluster task got completed immediately as it was still waiting for calamari to respond. SO, THIS WAY USM IMPORT CLUSTER FEATURE I FIND WORKING PERFECTLY FINE. @Gregory, I have a question here, why the default created pool "rbd" is causing issue and command "rbd ls" hangs. Cluster was and is in WARN state only. @Tejas, as discussed in person, this whole flow I tried with rhevm VMs and once more you would like to simulate the same with magna nodes in QE setup. Today self and Tejas tried the import cluster cluster flow with migrated cluster from 1.3.2 to 2.0 and it properly imported the cluster in USM. I suspect in old setup there was some cleanup issue with calamari old version as we need to do a "calamari-ctl clear" for old and then "calamari-ctl initialize" and it works fie. Tejas, please move/close the BZ accordingly. Thanks (In reply to Tejas from comment #7) > Thanks Shubhendu for helping me with this bug. I was not aware of 2 steps > that needed to be done for this. So I will move this bug to Doc, as we need > this upgrade process to be documented. > > Here are the list of steps for Calamari and USM setup on a ceph upgrade: > > 1. Create a Ceph 1.3.2 cluster with nodes : > - Admin + calamari server (1.3.3.1) > - MON > - OSD > > 2. Install USM packages on the Admin node ( rhscon-core rhscon-ceph > rhscon-ui) > and install the rhscon-agent on the MON's and OSD's. > > 3. Setup the USM server(skyrng setup) > > 4. Upgrade the Ceph cluster from 1.3.2 to 2.0. > > 5. Remove the Calamari Salt packages, diamond and graphite packages from Admin node. > > 6. Also remove the /etc/salt/minion.d/calamari.conf and > /etc/salt/pki/minion/* files on all the MON and OSD nodes. > > 7. Install latest calamari-server (say calamari-lite 1.4.5.1) on the first MON node, and configure /etc/salt/minion.d/ceph_agent.conf 8. Run "calamari-ctl clear --yes-i-am-sure" and "calamari-ctl initialize --admin-username admin --admin-password admin --admin-email junk" on calamari node. Then restart supervisord service once on node. > > 9. The salt-keys are still accepted by the older calamari-server, so delete > the salt-keys on the admin. Now these keys are ready to be accepted by the > USM server. > > 10. Import the cluster to USM using the GUI. > > > Thanks, > Tejas Thanks Aron. About comment 12 step 2, yes you are right. I suggest we ask the customers to delete the older packages first, before asking them to install the new RHSC. That is step 2 and 3 should come before step 1. Shubhendu, Can you let us know if our thinking is right? Thanks, Tejas Thats correct I feel. We might ask the user to manually remove older version of salt from storage nodes and then migrate cluster to higher version of ceph. What you say Tejas? Thanks Shubhendu, yes we can ask customers to delete older version before ceph upgrade. But I feel ceph upgrade has nothing to do with the Calamari and USM setup on the new admin node. We can do these steps even after a ceph upgrade. Thanks, Tejas I feel this section is clear to the customer. |