Bug 2092508
Summary: | [iscsi] tcmu container crash when cluster is upgraded from 5.1Z1 to 5.2 | ||
---|---|---|---|
Product: | [Red Hat Storage] Red Hat Ceph Storage | Reporter: | Preethi <pnataraj> |
Component: | Cephadm | Assignee: | Adam King <adking> |
Status: | CLOSED ERRATA | QA Contact: | Preethi <pnataraj> |
Severity: | high | Docs Contact: | Anjana Suparna Sriram <asriram> |
Priority: | unspecified | ||
Version: | 5.0 | CC: | adking, akraj, gabrioux, tonay, tserlin, vereddy |
Target Milestone: | --- | ||
Target Release: | 5.2 | ||
Hardware: | Unspecified | ||
OS: | Unspecified | ||
Whiteboard: | |||
Fixed In Version: | ceph-16.2.8-45.el8cp | Doc Type: | No Doc Update |
Doc Text: | Story Points: | --- | |
Clone Of: | Environment: | ||
Last Closed: | 2022-08-09 17:38:27 UTC | Type: | Bug |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: | |||
Bug Depends On: | |||
Bug Blocks: | 2102272 |
Description
Preethi
2022-06-01 17:30:47 UTC
Ceph version of 5.1Z1 - ceph version 16.2.7-112.el8cp @Adam, Issue still exist with latest 5.2 builds and update in the BZ https://bugzilla.redhat.com/show_bug.cgi?id=1976128. This could be verified once we have fix for the BZ dependent as per the discussions. However, I was able to see the issue when I upgraded to the latest builds. I guess we cannot verify until we have fix i.e pids.max is set for both tcmu and iscsi containers As per the discussions, moved the issue back to assigned state, Also,We cannot perform lun test if the fix of BZ 1976128 available with PIDs to MAX limit for both ISCSI and TCMU containers and the upgrade is again dependent on 255 luns and cannot be verified until we have the fix for BZ 1976128. issue is not seen with the latest 5.2 builds where fix is implemented. Below snippet after upgrade [ceph: root@magna021 /]# ceph status cluster: id: c8ce6d50-c0a1-11ec-a99b-002590fc2a2e health: HEALTH_OK services: mon: 5 daemons, quorum magna021,magna022,magna024,magna025,magna026 (age 19m) mgr: magna022.icxgsh(active, since 22m), standbys: magna021.syfuos osd: 42 osds: 42 up (since 6m), 42 in (since 8w) rbd-mirror: 1 daemon active (1 hosts) tcmu-runner: 512 portals active (2 hosts) data: pools: 11 pools, 801 pgs objects: 891.65k objects, 3.4 TiB usage: 10 TiB used, 28 TiB / 38 TiB avail pgs: 801 active+clean io: client: 218 KiB/s rd, 231 KiB/s wr, 59 op/s rd, 56 op/s wr [ceph: root@magna021 /]# ceph version ceph version 16.2.8-47.el8cp (48087358763c55c41f590e2beabc1fd341b89226) pacific (stable) [ceph: root@magna021 /]# ceph versions { "mon": { "ceph version 16.2.8-47.el8cp (48087358763c55c41f590e2beabc1fd341b89226) pacific (stable)": 5 }, "mgr": { "ceph version 16.2.8-47.el8cp (48087358763c55c41f590e2beabc1fd341b89226) pacific (stable)": 2 }, "osd": { "ceph version 16.2.8-47.el8cp (48087358763c55c41f590e2beabc1fd341b89226) pacific (stable)": 42 }, "mds": {}, "rbd-mirror": { "ceph version 16.2.8-47.el8cp (48087358763c55c41f590e2beabc1fd341b89226) pacific (stable)": 1 }, "tcmu-runner": { "ceph version 16.2.8-47.el8cp (48087358763c55c41f590e2beabc1fd341b89226) pacific (stable)": 512 }, "overall": { "ceph version 16.2.8-47.el8cp (48087358763c55c41f590e2beabc1fd341b89226) pacific (stable)": 562 } } [ceph: root@magna021 /]# Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Moderate: Red Hat Ceph Storage Security, Bug Fix, and Enhancement Update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2022:5997 |