Bug 2247616
| Summary: | NVMeof GW deployment failing with 0.0.5 version | ||
|---|---|---|---|
| Product: | [Red Hat Storage] Red Hat Ceph Storage | Reporter: | Sunil Kumar Nagaraju <sunnagar> |
| Component: | NVMeOF | Assignee: | Aviv Caro <aviv.caro> |
| Status: | CLOSED ERRATA | QA Contact: | Manohar Murthy <mmurthy> |
| Severity: | urgent | Docs Contact: | Rivka Pollack <rpollack> |
| Priority: | urgent | ||
| Version: | 7.0 | CC: | acaro, adking, akraj, aviv.caro, cephqe-warriors, idryomov, mmurthy, tserlin, vereddy |
| Target Milestone: | --- | Keywords: | Automation |
| Target Release: | 7.0 | ||
| Hardware: | Unspecified | ||
| OS: | Unspecified | ||
| Whiteboard: | |||
| Fixed In Version: | ceph-18.2.0-114.el9cp | Doc Type: | No Doc Update |
| Doc Text: | Story Points: | --- | |
| Clone Of: | Environment: | ||
| Last Closed: | 2023-12-13 15:24:49 UTC | Type: | Bug |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
|
Comment 1
Aviv Caro
2023-11-02 14:58:31 UTC
Targetting it to 7.0 as we need v5 version for TP. Deployment is successful with ceph version 18.2.0-116 and nvmeof image: 0.0.5-1
[ceph: root@ceph-1sunilkumar-uriqzt-node1-installer /]# ceph orch ps --daemon_type nvmeof --format json-pretty
[
{
"container_id": "6fdf6d9df670",
"container_image_digests": [
"registry-proxy.engineering.redhat.com/rh-osbs/ceph-nvmeof@sha256:1f1ffac74f2cd5f9748d4549872ecff5a958451edcae4bf6d0113c108607d743",
"registry-proxy.engineering.redhat.com/rh-osbs/ceph-nvmeof@sha256:5a2d64954de518ffd33af352aa6d1d752f1f92af7da202487a0f744a340f37a3"
],
"container_image_id": "756696fa8cf06dbdd4cece549353b7634c922321c0c5aeaaa9023e89ee4519a5",
"container_image_name": "registry-proxy.engineering.redhat.com/rh-osbs/ceph-nvmeof:0.0.5-1",
"cpu_percentage": "396.79%",
"created": "2023-11-06T11:29:19.215452Z",
"daemon_id": "rbd.ceph-1sunilkumar-uriqzt-node6.eoefgy",
"daemon_name": "nvmeof.rbd.ceph-1sunilkumar-uriqzt-node6.eoefgy",
"daemon_type": "nvmeof",
"events": [
"2023-11-06T11:29:19.242834Z daemon:nvmeof.rbd.ceph-1sunilkumar-uriqzt-node6.eoefgy [INFO] \"Deployed nvmeof.rbd.ceph-1sunilkumar-uriqzt-node6.eoefgy on host 'ceph-1sunilkumar-uriqzt-node6'\""
],
"hostname": "ceph-1sunilkumar-uriqzt-node6",
"is_active": false,
"last_refresh": "2023-11-06T11:58:58.508235Z",
"memory_usage": 39877345,
"ports": [
5500,
4420,
8009
],
"service_name": "nvmeof.rbd",
"started": "2023-11-06T11:48:44.884317Z",
"status": 1,
"status_desc": "running",
"version": ""
}
]
[ceph: root@ceph-1sunilkumar-uriqzt-node1-installer /]# ceph version
ceph version 18.2.0-116.el9cp (6c6dfc1f2b2ce1896bf2696daea444dff96645af) reef (stable)
[ceph: root@ceph-1sunilkumar-uriqzt-node1-installer /]#
[ceph: root@ceph-1sunilkumar-uriqzt-node1-installer /]# ceph orch ls
NAME PORTS RUNNING REFRESHED AGE PLACEMENT
alertmanager ?:9093,9094 1/1 2m ago 69m count:1
ceph-exporter 6/6 8m ago 69m *
crash 6/6 8m ago 69m *
grafana ?:3000 1/1 2m ago 69m count:1
mgr 2/2 2m ago 68m label:mgr
mon 3/3 2m ago 66m label:mon
node-exporter ?:9100 6/6 8m ago 69m *
nvmeof.rbd ?:4420,5500,8009 1/1 8m ago 38m ceph-1sunilkumar-uriqzt-node6
osd.all-available-devices 12 2m ago 62m *
prometheus ?:9095 1/1 2m ago 69m count:1
[root@ceph-1sunilkumar-uriqzt-node6 ~]# podman ps -a | grep nvmeof.rbd
6fdf6d9df670 registry-proxy.engineering.redhat.com/rh-osbs/ceph-nvmeof:0.0.5-1 -c ceph-nvmeof.co... 19 minutes ago Up 19 minutes ceph-3f361aa2-7c93-11ee-83ea-fa163e37915e-nvmeof-rbd-ceph-1sunilkumar-uriqzt-node6-eoefgy
[root@ceph-1sunilkumar-uriqzt-node6 nvmeof.rbd.ceph-1sunilkumar-uriqzt-node6.eoefgy]# cat ceph-nvmeof.conf
# This file is generated by cephadm.
[gateway]
name = client.nvmeof.rbd.ceph-1sunilkumar-uriqzt-node6.eoefgy
group = None
addr = 10.0.208.194
port = 5500
enable_auth = False
state_update_notify = True
state_update_interval_sec = 5
enable_spdk_discovery_controller = true
[ceph]
pool = rbd
config_file = /etc/ceph/ceph.conf
id = nvmeof.rbd.ceph-1sunilkumar-uriqzt-node6.eoefgy
[mtls]
server_key = ./server.key
client_key = ./client.key
server_cert = ./server.crt
client_cert = ./client.crt
[spdk]
tgt_path = /usr/local/bin/nvmf_tgt
rpc_socket = /var/tmp/spdk.sock
timeout = 60
log_level = WARN
conn_retries = 10
transports = tcp
transport_tcp_options = {"in_capsule_data_size": 8192, "max_io_qpairs_per_ctrlr": 7}
tgt_cmd_extra_args = --cpumask=0xF
Hi Aviv. Could you please confirm if this BZ needs to be added to the 7.0 release notes? If so, please provide the doc type and text. Thanks. Hi Akash, It is fixed. So if the release includes this fix, I don't see any reason to add it to RN. Aviv Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Red Hat Ceph Storage 7.0 Bug Fix update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2023:7780 The needinfo request[s] on this closed bug have been removed as they have been unresolved for 120 days |