Bug 2280332
| Summary: | nvmeof GW exits after fully startup is not performed after brought back post failover leading to WAIT_FAILBACK_PREPARED ana_state | ||
|---|---|---|---|
| Product: | [Red Hat Storage] Red Hat Ceph Storage | Reporter: | Rahul Lepakshi <rlepaksh> |
| Component: | NVMeOF | Assignee: | Aviv Caro <aviv.caro> |
| Status: | CLOSED ERRATA | QA Contact: | Rahul Lepakshi <rlepaksh> |
| Severity: | urgent | Docs Contact: | ceph-doc-bot <ceph-doc-bugzilla> |
| Priority: | unspecified | ||
| Version: | 7.1 | CC: | acaro, aviv.caro, cephqe-warriors, mmurthy, tserlin, vdas |
| Target Milestone: | --- | Keywords: | BetaBlocker, TestBlocker |
| Target Release: | 7.1 | Flags: | rlepaksh:
needinfo-
rlepaksh: needinfo- |
| Hardware: | Unspecified | ||
| OS: | Unspecified | ||
| Whiteboard: | |||
| Fixed In Version: | ceph-18.2.1-176.el9cp | Doc Type: | If docs needed, set a value |
| Doc Text: | Story Points: | --- | |
| Clone Of: | Environment: | ||
| Last Closed: | 2024-06-13 14:32:49 UTC | Type: | Bug |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
|
Description
Rahul Lepakshi
2024-05-14 08:27:30 UTC
Please specify the severity of this bug. Severity is defined here: https://bugzilla.redhat.com/page.cgi?id=fields.html#bug_severity. Issue is understood. Leonid prepares a fix. @aviv.caro I am terming this as blocker as ina 2 GW config, if other GW also comes down for some reason, there is no one to handle IO and maintain namespaces in ceph cluster. We hit Data unavailability in this case Not seeing this issue on recent builds, but I have an Observation on scale cluster, GW takes at least 2 minutes to load NS post GW comes to ACTIVE STANDBY state. [root@argo023 ~]# nvmeof subsystem list Subsystems: ╒═══════════╤════════════════════════════╤════════════╤════════════════════╤══════════════════╤═════════════╤══════════════╕ │ Subtype │ NQN │ HA State │ Serial │ Controller IDs │ Namespace │ Max │ │ │ │ │ Number │ │ Count │ Namespaces │ ╞═══════════╪════════════════════════════╪════════════╪════════════════════╪══════════════════╪═════════════╪══════════════╡ │ NVMe │ nqn.2016-06.io.spdk:cnode1 │ enabled │ Ceph76593830561176 │ 2041-4080 │ 179 │ 400 │ ├───────────┼────────────────────────────┼────────────┼────────────────────┼──────────────────┼─────────────┼──────────────┤ │ NVMe │ nqn.2016-06.io.spdk:cnode2 │ enabled │ Ceph50770207011824 │ 2041-4080 │ 182 │ 400 │ ╘═══════════╧════════════════════════════╧════════════╧════════════════════╧══════════════════╧═════════════╧══════════════╛ [root@argo023 ~]# nvmeof subsystem list Subsystems: ╒═══════════╤════════════════════════════╤════════════╤════════════════════╤══════════════════╤═════════════╤══════════════╕ │ Subtype │ NQN │ HA State │ Serial │ Controller IDs │ Namespace │ Max │ │ │ │ │ Number │ │ Count │ Namespaces │ ╞═══════════╪════════════════════════════╪════════════╪════════════════════╪══════════════════╪═════════════╪══════════════╡ │ NVMe │ nqn.2016-06.io.spdk:cnode1 │ enabled │ Ceph76593830561176 │ 2041-4080 │ 190 │ 400 │ ├───────────┼────────────────────────────┼────────────┼────────────────────┼──────────────────┼─────────────┼──────────────┤ │ NVMe │ nqn.2016-06.io.spdk:cnode2 │ enabled │ Ceph50770207011824 │ 2041-4080 │ 191 │ 400 │ ╘═══════════╧════════════════════════════╧════════════╧════════════════════╧══════════════════╧═════════════╧══════════════╛ [root@argo023 ~]# nvmeof subsystem list Subsystems: ╒═══════════╤════════════════════════════╤════════════╤════════════════════╤══════════════════╤═════════════╤══════════════╕ │ Subtype │ NQN │ HA State │ Serial │ Controller IDs │ Namespace │ Max │ │ │ │ │ Number │ │ Count │ Namespaces │ ╞═══════════╪════════════════════════════╪════════════╪════════════════════╪══════════════════╪═════════════╪══════════════╡ │ NVMe │ nqn.2016-06.io.spdk:cnode1 │ enabled │ Ceph76593830561176 │ 2041-4080 │ 196 │ 400 │ ├───────────┼────────────────────────────┼────────────┼────────────────────┼──────────────────┼─────────────┼──────────────┤ │ NVMe │ nqn.2016-06.io.spdk:cnode2 │ enabled │ Ceph50770207011824 │ 2041-4080 │ 197 │ 400 │ ╘═══════════╧════════════════════════════╧════════════╧════════════════════╧══════════════════╧═════════════╧══════════════╛ [root@argo023 ~]# nvmeof subsystem list Subsystems: ╒═══════════╤════════════════════════════╤════════════╤════════════════════╤══════════════════╤═════════════╤══════════════╕ │ Subtype │ NQN │ HA State │ Serial │ Controller IDs │ Namespace │ Max │ │ │ │ │ Number │ │ Count │ Namespaces │ ╞═══════════╪════════════════════════════╪════════════╪════════════════════╪══════════════════╪═════════════╪══════════════╡ │ NVMe │ nqn.2016-06.io.spdk:cnode1 │ enabled │ Ceph76593830561176 │ 2041-4080 │ 200 │ 400 │ ├───────────┼────────────────────────────┼────────────┼────────────────────┼──────────────────┼─────────────┼──────────────┤ │ NVMe │ nqn.2016-06.io.spdk:cnode2 │ enabled │ Ceph50770207011824 │ 2041-4080 │ 200 │ 400 │ ╘═══════════╧════════════════════════════╧════════════╧════════════════════╧══════════════════╧═════════════╧══════════════╛ Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Critical: Red Hat Ceph Storage 7.1 security, enhancements, and bug fix update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2024:3925 |