Bug 2000412
Summary: | infrastructure-playbooks/cephadm-adopt.yml fails to start iscsi daemons while converting the storage cluster daemons to run cephadm | |||
---|---|---|---|---|
Product: | [Red Hat Storage] Red Hat Ceph Storage | Reporter: | Gopi <gpatta> | |
Component: | Ceph-Ansible | Assignee: | Guillaume Abrioux <gabrioux> | |
Status: | CLOSED ERRATA | QA Contact: | Ameena Suhani S H <amsyedha> | |
Severity: | high | Docs Contact: | ||
Priority: | unspecified | |||
Version: | 5.0 | CC: | aschoen, ceph-eng-bugs, ceph-qe-bugs, gabrioux, gmeno, gsitlani, kdreyer, mgowri, mmurthy, nthomas, tserlin, vereddy, xiubli, ykaul | |
Target Milestone: | --- | Flags: | gpatta:
needinfo+
gpatta: needinfo+ gpatta: needinfo+ |
|
Target Release: | 5.0z2 | |||
Hardware: | Unspecified | |||
OS: | Unspecified | |||
Whiteboard: | ||||
Fixed In Version: | ceph-ansible-6.0.16-1.el8cp | Doc Type: | Bug Fix | |
Doc Text: |
Cause: cephadm-adopt playbook makes cephadm start iscsi service before stopping the containers that were managed by ceph-ansible, it makes the tcmu-runner process unable to open devices, there is nothing else to do than restarting the containers.
Consequence: iscsigw daemons don't work properly after the adoption is done.
Fix: Stop iscsi daemons containers that were managed by ceph-ansible before starting new container managed by cephadm. Also, iscsigws services should be migrated before OSDs (The clients should be closed before the server).
Result: iscsigw daemons work properly after the adoption is done.
|
Story Points: | --- | |
Clone Of: | ||||
: | 2007683 (view as bug list) | Environment: | ||
Last Closed: | 2021-12-08 13:57:04 UTC | Type: | Bug | |
Regression: | --- | Mount Type: | --- | |
Documentation: | --- | CRM: | ||
Verified Versions: | Category: | --- | ||
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | ||
Cloudforms Team: | --- | Target Upstream Version: | ||
Embargoed: | ||||
Bug Depends On: | 2026861 | |||
Bug Blocks: |
Description
Gopi
2021-09-02 06:23:59 UTC
Hi Xiubo, You need to enter in to root after login with cephuser (Do "sudo su"). Else use "sudo" before your command. Hi Guillaume, Can you look in to this issue? Hi Guillaume, Setup is still there. [ceph: root@ceph-gp-rbd-fqj2nr-node1-installer ~]# ceph -s cluster: id: 7c80e5b9-c8ae-4fbb-b9e5-36b4288182d0 health: HEALTH_WARN mons are allowing insecure global_id reclaim insufficient standby MDS daemons available 1 pools have too many placement groups services: mon: 3 daemons, quorum ceph-gp-rbd-fqj2nr-node2,ceph-gp-rbd-fqj2nr-node3,ceph-gp-rbd-fqj2nr-node1-installer (age 5d) mgr: ceph-gp-rbd-fqj2nr-node1-installer(active, since 5d), standbys: ceph-gp-rbd-fqj2nr-node2 mds: 1/1 daemons up osd: 12 osds: 12 up (since 5d), 12 in (since 6d) data: volumes: 1/1 healthy pools: 4 pools, 193 pgs objects: 1.09k objects, 4.1 GiB usage: 13 GiB used, 167 GiB / 180 GiB avail pgs: 193 active+clean io: client: 853 B/s rd, 0 op/s rd, 0 op/s wr [ceph: root@ceph-gp-rbd-fqj2nr-node1-installer ~]# ceph orch ls NAME RUNNING REFRESHED AGE PLACEMENT crash 0/6 - 5d label:ceph iscsi.rbd 0/2 - 5d count:2;label:iscsigws mds.cephfs 0/3 - 5d count:3;label:mdss mgr 0/2 - 5d count:2;label:mgrs mon 0/3 - 5d count:3;label:mons [ceph: root@ceph-gp-rbd-fqj2nr-node1-installer ~]# ceph orch host ls HOST ADDR LABELS STATUS ceph-gp-rbd-fqj2nr-node1-installer 10.0.210.115 mgrs mons ceph ceph-gp-rbd-fqj2nr-node2 10.0.209.21 mgrs mons osds ceph ceph-gp-rbd-fqj2nr-node3 10.0.210.56 iscsigws mons osds ceph ceph-gp-rbd-fqj2nr-node4 10.0.208.78 mdss ceph ceph-gp-rbd-fqj2nr-node5 10.0.209.206 iscsigws mdss osds ceph ceph-gp-rbd-fqj2nr-node6 10.0.210.179 grafana-server mdss monitoring ceph [ceph: root@ceph-gp-rbd-fqj2nr-node1-installer ~]# [root@ceph-gp-rbd-fqj2nr-node3 cephuser]# systemctl status tcmu-runner ● tcmu-runner.service - LIO Userspace-passthrough daemon Loaded: loaded (/usr/lib/systemd/system/tcmu-runner.service; disabled; vendor preset: disabled) Active: active (running) since Wed 2021-09-01 22:26:22 EDT; 5 days ago Docs: man:tcmu-runner(8) Main PID: 7350 (tcmu-runner) Tasks: 5 (limit: 23465) Memory: 45.6M CGroup: /system.slice/tcmu-runner.service └─7350 /usr/bin/tcmu-runner Warning: Journal has been rotated since unit was started. Log output is incomplete or unavailable. Hi Guillaume, I don't have setup right now but i will share new setup soon by reproducing the issue. Thanks, Gopi Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Red Hat Ceph Storage 5.0 Bug Fix update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2021:5020 |