Bug 2351823 - iscsi-tcmu container not getting upgraded when upgraded from 8.0z2(19.2.0-98.el9cp) GA build to 8.0z3(19.2.0-108.el9cp) build
Summary: iscsi-tcmu container not getting upgraded when upgraded from 8.0z2(19.2.0-98....
Keywords:
Status: CLOSED ERRATA
Alias: None
Product: Red Hat Ceph Storage
Classification: Red Hat Storage
Component: Cephadm
Version: 8.0
Hardware: Unspecified
OS: Unspecified
unspecified
high
Target Milestone: ---
: 8.0z3
Assignee: JAYA PRAKASH P
QA Contact: JAYA PRAKASH P
Rivka Pollack
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2025-03-13 08:43 UTC by JAYA PRAKASH P
Modified: 2025-04-07 15:27 UTC (History)
8 users (show)

Fixed In Version: ceph-19.2.0-118.el9cp
Doc Type: No Doc Update
Doc Text:
Clone Of:
Environment:
Last Closed: 2025-04-07 15:27:26 UTC
Embargoed:


Attachments (Terms of Use)
iscsi tcmu service logs (4.82 KB, text/plain)
2025-03-13 08:47 UTC, JAYA PRAKASH P
no flags Details
iscsi service logs (934.81 KB, text/plain)
2025-03-13 08:47 UTC, JAYA PRAKASH P
no flags Details
tcmu-runner log (2.24 KB, text/plain)
2025-03-21 04:36 UTC, JAYA PRAKASH P
no flags Details
tcmu-runner log (1.85 KB, text/plain)
2025-03-21 04:37 UTC, JAYA PRAKASH P
no flags Details
cephadm log (4.38 MB, text/plain)
2025-03-21 04:37 UTC, JAYA PRAKASH P
no flags Details


Links
System ID Private Priority Status Summary Last Updated
Red Hat Issue Tracker RHCEPH-10844 0 None None None 2025-03-13 08:44:49 UTC
Red Hat Product Errata RHSA-2025:3635 0 None None None 2025-04-07 15:27:29 UTC

Description JAYA PRAKASH P 2025-03-13 08:43:01 UTC
Description of problem:

I had deployed cluster with 8.0z2 GA build, configured nvme, iscsi and upgraded to 8.0z3 build. After upgrade when checked for "ceph versions", "tcmu-runner" shows 8.0z2 GA build details

In journactl log of iscsi service we can see it is pulling 8.0z3 image 
In journactl log of iscsi tcmu-runner service we don't see any traces of pulling 8.0z3 image

[ceph: root@ceph-80z3nvme-07q3fh-node1-installer ~]# ceph --version
ceph version 19.2.0-98.el9cp (d5c3cf625491b0bd76b4585e77aa0d907446f314) squid (stable)
[ceph: root@ceph-80z3nvme-07q3fh-node1-installer ~]# ceph orch ls
NAME                                 PORTS             RUNNING  REFRESHED  AGE   PLACEMENT
alertmanager                         ?:9093,9094           1/1  5m ago     2h    count:1
ceph-exporter                                              6/6  10m ago    2h    *
crash                                                      6/6  10m ago    2h    *
grafana                              ?:3000                1/1  5m ago     2h    count:1
iscsi.iscsi                          ?:5000                2/2  2m ago     43m   ceph-80z3nvme-07q3fh-node4;ceph-80z3nvme-07q3fh-node5
mgr                                                        2/2  9m ago     2h    label:mgr
mon                                                        3/3  9m ago     2h    label:mon
node-exporter                        ?:9100                6/6  10m ago    2h    *
nvmeof.nvmeof_pool01.nvmeof_group01  ?:4420,5500,8009      2/2  9m ago     115m  label:nvmeof_label
osd.all-available-devices                                   16  10m ago    2h    *
prometheus                           ?:9095                1/1  5m ago     2h    count:1


After upgrading to 8.0z3 build
[ceph: root@ceph-80z3nvme-07q3fh-node1-installer /]# ceph --version
ceph version 19.2.0-108.el9cp (1762f710a9f63e0304d69ed81ad964841146c93d) squid (stable)
[ceph: root@ceph-80z3nvme-07q3fh-node1-installer /]# ceph orch ls
NAME                                 PORTS             RUNNING  REFRESHED  AGE  PLACEMENT
alertmanager                         ?:9093,9094           1/1  9m ago     5h   count:1
ceph-exporter                                              6/6  9m ago     5h   *
crash                                                      6/6  9m ago     5h   *
grafana                              ?:3000                1/1  9m ago     5h   count:1
iscsi.iscsi                          ?:5000                2/2  9m ago     4h   ceph-80z3nvme-07q3fh-node4;ceph-80z3nvme-07q3fh-node5
mgr                                                        2/2  9m ago     5h   label:mgr
mon                                                        3/3  9m ago     5h   label:mon
node-exporter                        ?:9100                6/6  9m ago     5h   *
nvmeof.nvmeof_pool01.nvmeof_group01  ?:4420,5500,8009      2/2  9m ago     5h   label:nvmeof_label
osd.all-available-devices                                   16  9m ago     5h   *
prometheus                           ?:9095                1/1  9m ago     5h   count:1
[ceph: root@ceph-80z3nvme-07q3fh-node1-installer /]#
ceph versions has below
[ceph: root@ceph-80z3nvme-07q3fh-node1-installer /]# ceph versions
{
    "mon": {
        "ceph version 19.2.0-108.el9cp (1762f710a9f63e0304d69ed81ad964841146c93d) squid (stable)": 3
    },
    "mgr": {
        "ceph version 19.2.0-108.el9cp (1762f710a9f63e0304d69ed81ad964841146c93d) squid (stable)": 2
    },
    "osd": {
        "ceph version 19.2.0-108.el9cp (1762f710a9f63e0304d69ed81ad964841146c93d) squid (stable)": 16
    },
    "tcmu-runner": {
        "ceph version 19.2.0-98.el9cp (d5c3cf625491b0bd76b4585e77aa0d907446f314) squid (stable)": 2
    },
    "overall": {
        "ceph version 19.2.0-108.el9cp (1762f710a9f63e0304d69ed81ad964841146c93d) squid (stable)": 21,
        "ceph version 19.2.0-98.el9cp (d5c3cf625491b0bd76b4585e77aa0d907446f314) squid (stable)": 2
    }
}

here overall has two build details and tcmu-runner shows 8.0z2 GA build but it should be upgraded to 8.0z3
ceph orch upgrade status is not having any error
[ceph: root@ceph-80z3nvme-07q3fh-node1-installer /]# ceph orch upgrade status
{
    "target_image": null,
    "in_progress": false,
    "which": "<unknown>",
    "services_complete": [],
    "progress": null,
    "message": "",
    "is_paused": false
}
ceph orch upgrade check --image   cp.stg.icr.io/cp/ibm-ceph/ceph-8-rhel9:8-90 also everything is up to date
[ceph: root@ceph-80z3nvme-07q3fh-node1-installer /]# ceph orch upgrade check --image   cp.stg.icr.io/cp/ibm-ceph/ceph-8-rhel9:8-90
{
    "needs_update": {},
    "non_ceph_image_daemons": [
        "node-exporter.ceph-80z3nvme-07q3fh-node1-installer",
        "alertmanager.ceph-80z3nvme-07q3fh-node1-installer",
        "grafana.ceph-80z3nvme-07q3fh-node1-installer",
        "prometheus.ceph-80z3nvme-07q3fh-node1-installer",
        "node-exporter.ceph-80z3nvme-07q3fh-node2",
        "nvmeof.nvmeof_pool01.nvmeof_group01.ceph-80z3nvme-07q3fh-node2.zozyap",
        "node-exporter.ceph-80z3nvme-07q3fh-node3",
        "nvmeof.nvmeof_pool01.nvmeof_group01.ceph-80z3nvme-07q3fh-node3.rrzhmu",
        "node-exporter.ceph-80z3nvme-07q3fh-node4",
        "node-exporter.ceph-80z3nvme-07q3fh-node5",
        "node-exporter.ceph-80z3nvme-07q3fh-node6"
    ],
    "target_digest": "cp.stg.icr.io/cp/ibm-ceph/ceph-8-rhel9@sha256:693be8216de8508dffad0452f6133c855c0700009feaa7838443557e8d0a0498",
    "target_id": "bb25881faf9cf0a9379deec426702df880a7cf747ce910c1353cea6faa6ca9d8",
    "target_name": "cp.stg.icr.io/cp/ibm-ceph/ceph-8-rhel9:8-90",
    "target_version": "ceph version 19.2.0-108.el9cp (1762f710a9f63e0304d69ed81ad964841146c93d) squid (stable)",
    "up_to_date": [
        "mon.ceph-80z3nvme-07q3fh-node1-installer",
        "mgr.ceph-80z3nvme-07q3fh-node1-installer.jgvfvl",
        "ceph-exporter.ceph-80z3nvme-07q3fh-node1-installer",
        "crash.ceph-80z3nvme-07q3fh-node1-installer",
        "ceph-exporter.ceph-80z3nvme-07q3fh-node2",
        "crash.ceph-80z3nvme-07q3fh-node2",
        "mgr.ceph-80z3nvme-07q3fh-node2.mfowxn",
        "mon.ceph-80z3nvme-07q3fh-node2",
        "ceph-exporter.ceph-80z3nvme-07q3fh-node3",
        "crash.ceph-80z3nvme-07q3fh-node3",
        "mon.ceph-80z3nvme-07q3fh-node3",
        "osd.1",
        "osd.13",
        "osd.5",
        "osd.9",
        "ceph-exporter.ceph-80z3nvme-07q3fh-node4",
        "crash.ceph-80z3nvme-07q3fh-node4",
        "osd.10",
        "osd.14",
        "osd.2",
        "osd.6",
        "iscsi.iscsi.ceph-80z3nvme-07q3fh-node4.bevnsh",
        "ceph-exporter.ceph-80z3nvme-07q3fh-node5",
        "crash.ceph-80z3nvme-07q3fh-node5",
        "osd.0",
        "osd.12",
        "osd.4",
        "osd.8",
        "iscsi.iscsi.ceph-80z3nvme-07q3fh-node5.rapwrd",
        "ceph-exporter.ceph-80z3nvme-07q3fh-node6",
        "crash.ceph-80z3nvme-07q3fh-node6",
        "osd.11",
        "osd.15",
        "osd.3",
        "osd.7"
    ]
}
[ceph: root@ceph-80z3nvme-07q3fh-node1-installer /]#



How reproducible:
Always

Steps to Reproduce:
1. Deploy an cluster with 8.0z2 GA build
2. Create nvme, iscsi configs
3. Upgrade to 8.0z3 build

Actual results:
[ceph: root@ceph-80z3nvme-07q3fh-node1-installer /]# ceph versions
{
    .........
    .........
    "tcmu-runner": {
        "ceph version 19.2.0-98.el9cp (d5c3cf625491b0bd76b4585e77aa0d907446f314) squid (stable)": 2
    },
    "overall": {
        "ceph version 19.2.0-108.el9cp (1762f710a9f63e0304d69ed81ad964841146c93d) squid (stable)": 21,
        "ceph version 19.2.0-98.el9cp (d5c3cf625491b0bd76b4585e77aa0d907446f314) squid (stable)": 2
    }
}
[ceph: root@ceph-80z3nvme-07q3fh-node1-installer /]#

Expected results:
{
    .........
    .........
    "tcmu-runner": {
        "ceph version 19.2.0-108.el9cp (1762f710a9f63e0304d69ed81ad964841146c93d) squid (stable)": 2
    },
    "overall": {
        "ceph version 19.2.0-108.el9cp (1762f710a9f63e0304d69ed81ad964841146c93d) squid (stable)": 21,
    }
}

Additional info:
Attached journalctl logs

Comment 1 JAYA PRAKASH P 2025-03-13 08:47:23 UTC
Created attachment 2079974 [details]
iscsi tcmu service logs

Comment 2 JAYA PRAKASH P 2025-03-13 08:47:55 UTC
Created attachment 2079975 [details]
iscsi service logs

Comment 3 gadi 2025-03-16 10:23:53 UTC
Hi, I tested that issue.

I had deployed cluster with 8.0z2 GA build (the image: cp.stg.icr.io/cp/ibm-ceph/ceph-8-rhel9:8-79) and configured iscsi.

the outputs:

[root@cephnvme-vm9 ~]# docker ps
CONTAINER ID   IMAGE                                         COMMAND                  CREATED          STATUS          PORTS     NAMES
ee130c1f7fb7   cp.stg.icr.io/cp/ibm-ceph/ceph-8-rhel9        "/usr/bin/rbd-target…"   5 minutes ago    Up 5 minutes              ceph-c8b7baa8-0241-11f0-a1c9-02001d9b9601-iscsi-iscsi-cephnvme-vm9-ncdzwz
cabf5bba0728   cp.stg.icr.io/cp/ibm-ceph/ceph-8-rhel9        "/usr/bin/tcmu-runner"   5 minutes ago    Up 5 minutes              ceph-c8b7baa8-0241-11f0-a1c9-02001d9b9601-iscsi-iscsi-cephnvme-vm9-ncdzwz-tcmu                                                                      
25c5265a883f   cp.stg.icr.io/cp/ibm-ceph/ceph-8-rhel9        "/usr/bin/ceph-osd -…"   9 minutes ago    Up 9 minutes              ceph-c8b7baa8-0241-11f0-a1c9-02001d9b9601-osd-2
071c8650aa8b   cp.stg.icr.io/cp/ibm-ceph/ceph-8-rhel9        "/usr/bin/ceph-crash…"   14 minutes ago   Up 14 minutes             ceph-c8b7baa8-0241-11f0-a1c9-02001d9b9601-crash-cephnvme-vm9
c558908139e7   cp.stg.icr.io/cp/ibm-ceph/ceph-8-rhel9        "/usr/bin/ceph-expor…"   14 minutes ago   Up 14 minutes             ceph-c8b7baa8-0241-11f0-a1c9-02001d9b9601-ceph-exporter-cephnvme-vm9
9b8690e329f7   cp.stg.icr.io/cp/ibm-ceph/ceph-8-rhel9:8-79   "/usr/bin/ceph-mgr -…"   15 minutes ago   Up 15 minutes             ceph-c8b7baa8-0241-11f0-a1c9-02001d9b9601-mgr-cephnvme-vm9-dguovc
0531497bbb48   cp.stg.icr.io/cp/ibm-ceph/ceph-8-rhel9:8-79   "/usr/bin/ceph-mon -…"   15 minutes ago   Up 15 minutes             ceph-c8b7baa8-0241-11f0-a1c9-02001d9b9601-mon-cephnvme-vm9

[root@cephnvme-vm9 ~]# ceph orch ls
NAME                       PORTS        RUNNING  REFRESHED  AGE  PLACEMENT
alertmanager               ?:9093,9094      0/1  -          33m  count:1
ceph-exporter                               4/4  9m ago     33m  *
crash                                       4/4  9m ago     33m  *
grafana                    ?:3000           0/1  -          33m  count:1
iscsi.iscsi                ?:5000           4/4  9m ago     23m  cephnvme-vm9;cephnvme-vm8;cephnvme-vm7;cephnvme-vm6
mgr                                         4/4  9m ago     29m  label:mgr
mon                                         4/4  9m ago     29m  label:mon
node-exporter              ?:9100           0/4  9m ago     33m  *
osd.all-available-devices                     4  9m ago     28m  *
prometheus                 ?:9095           0/1  -          33m  count:1


[root@cephnvme-vm9 ~]# ceph --version
ceph version 19.2.1 (58a7fab8be0a062d730ad7da874972fd3fba59fb) squid (stable)

[root@cephnvme-vm9 ~]# ceph versions
{
    "mon": {
        "ceph version 19.2.0-92.el9cp (6e1ac6f3b4e0616542ffeaeac53a3dff6a014e4a) squid (stable)": 4
    },
    "mgr": {
        "ceph version 19.2.0-92.el9cp (6e1ac6f3b4e0616542ffeaeac53a3dff6a014e4a) squid (stable)": 4
    },
    "osd": {
        "ceph version 19.2.0-92.el9cp (6e1ac6f3b4e0616542ffeaeac53a3dff6a014e4a) squid (stable)": 4
    },
    "tcmu-runner": {
        "ceph version 19.2.0-92.el9cp (6e1ac6f3b4e0616542ffeaeac53a3dff6a014e4a) squid (stable)": 4
    },
    "overall": {
        "ceph version 19.2.0-92.el9cp (6e1ac6f3b4e0616542ffeaeac53a3dff6a014e4a) squid (stable)": 16
    }
}


then, I upgraded ceph to image: cp.stg.icr.io/cp/ibm-ceph/ceph-8-rhel9:8-90

[root@cephnvme-vm9 ~]# ceph orch upgrade start --image cp.stg.icr.io/cp/ibm-ceph/ceph-8-rhel9:8-90

After the upgrading-

[root@cephnvme-vm9 ~]# ceph --version
ceph version 19.2.1 (58a7fab8be0a062d730ad7da874972fd3fba59fb) squid (stable)

[root@cephnvme-vm9 ~]# ceph versions
{
    "mon": {
        "ceph version 19.2.0-108.el9cp (1762f710a9f63e0304d69ed81ad964841146c93d) squid (stable)": 4
    },
    "mgr": {
        "ceph version 19.2.0-108.el9cp (1762f710a9f63e0304d69ed81ad964841146c93d) squid (stable)": 4
    },
    "osd": {
        "ceph version 19.2.0-108.el9cp (1762f710a9f63e0304d69ed81ad964841146c93d) squid (stable)": 4
    },
    "tcmu-runner": {
        "ceph version 19.2.0-108.el9cp (1762f710a9f63e0304d69ed81ad964841146c93d) squid (stable)": 4
    },
    "overall": {
        "ceph version 19.2.0-108.el9cp (1762f710a9f63e0304d69ed81ad964841146c93d) squid (stable)": 16
    }
}

[root@cephnvme-vm9 ~]# ceph orch ls
NAME                       PORTS        RUNNING  REFRESHED  AGE   PLACEMENT
alertmanager               ?:9093,9094      0/1  -          101m  count:1
ceph-exporter                               4/4  2m ago     101m  *
crash                                       4/4  2m ago     101m  *
grafana                    ?:3000           0/1  -          101m  count:1
iscsi.iscsi                ?:5000           4/4  2m ago     92m   cephnvme-vm9;cephnvme-vm8;cephnvme-vm7;cephnvme-vm6
mgr                                         4/4  2m ago     98m   label:mgr
mon                                         4/4  2m ago     98m   label:mon
node-exporter              ?:9100           0/4  2m ago     101m  *
osd.all-available-devices                     4  2m ago     97m   *
prometheus                 ?:9095           0/1  -          101m  count:1


I was not able to reproduce the issue.
note- when 8.0z2 was installed, I got ceph version 19.2.0-108.el9cp.
what was your image you built ceph?

Comment 4 JAYA PRAKASH P 2025-03-17 04:48:53 UTC
Have deployed CEPH 8.0z2  GA kit

[ceph: root@ceph-80z3nvme-07q3fh-node1-installer /]# ceph --version
ceph version 19.2.0-98.el9cp (d5c3cf625491b0bd76b4585e77aa0d907446f314) squid (stable)

[ceph: root@ceph-80z3nvme-07q3fh-node1-installer ~]# ceph orch ls
NAME                                 PORTS             RUNNING  REFRESHED  AGE   PLACEMENT
alertmanager                         ?:9093,9094           1/1  5m ago     2h    count:1
ceph-exporter                                              6/6  10m ago    2h    *
crash                                                      6/6  10m ago    2h    *
grafana                              ?:3000                1/1  5m ago     2h    count:1
iscsi.iscsi                          ?:5000                2/2  2m ago     43m   ceph-80z3nvme-07q3fh-node4;ceph-80z3nvme-07q3fh-node5
mgr                                                        2/2  9m ago     2h    label:mgr
mon                                                        3/3  9m ago     2h    label:mon
node-exporter                        ?:9100                6/6  10m ago    2h    *
nvmeof.nvmeof_pool01.nvmeof_group01  ?:4420,5500,8009      2/2  9m ago     115m  label:nvmeof_label
osd.all-available-devices                                   16  10m ago    2h    *
prometheus                           ?:9095                1/1  5m ago     2h    count:1

Upgraded to cp.stg.icr.io/cp/ibm-ceph/ceph-8-rhel9:8-90

[ceph: root@ceph-80z3nvme-07q3fh-node1-installer /]# ceph orch upgrade start  cp.stg.icr.io/cp/ibm-ceph/ceph-8-rhel9:8-90
Initiating upgrade to cp.stg.icr.io/cp/ibm-ceph/ceph-8-rhel9:8-90
[ceph: root@ceph-80z3nvme-07q3fh-node1-installer /]# ceph orch upgrade status
{
    "target_image": "cp.stg.icr.io/cp/ibm-ceph/ceph-8-rhel9@sha256:693be8216de8508dffad0452f6133c855c0700009feaa7838443557e8d0a0498",
    "in_progress": true,
    "which": "Upgrading all daemon types on all hosts",
    "services_complete": [],
    "progress": "0/46 daemons upgraded",
    "message": "Pulling cp.stg.icr.io/cp/ibm-ceph/ceph-8-rhel9@sha256:693be8216de8508dffad0452f6133c855c0700009feaa7838443557e8d0a0498 image on host ceph-80z3nvme-07q3fh-node2",
    "is_paused": false
}

After upgrade

[ceph: root@ceph-80z3nvme-07q3fh-node1-installer /]# ceph -v
ceph version 19.2.0-108.el9cp (1762f710a9f63e0304d69ed81ad964841146c93d) squid (stable)

Can use cluster Installer node IP 10.0.66.93(root/passwd) for debugging

Comment 6 gadi 2025-03-18 08:00:28 UTC
I see your ceph-common version is 19.2.0.98.

[ceph: root@ceph-80z3nvme-07q3fh-node1-installer /]# ceph --version
ceph version 19.2.0-98.el9cp (d5c3cf625491b0bd76b4585e77aa0d907446f314) squid (stable)


It appears that the specific version ceph-common-19.2.0-98.el9cp.x86_64.rpm is not available in the standard Ceph repositories. 
The closest available version is ceph-common-19.2.0-0.el9.x86_64.rpm

how did you install that version?

Comment 7 JAYA PRAKASH P 2025-03-18 12:19:58 UTC
I have deployed 8.0z2 released version and ceph repo is https://public.dhe.ibm.com/ibmdl/export/pub/storage/ceph/8/

can find from below email 8.0z2 equivalent version is ceph-19.2.0-98.el9cp.

After deploying cluster with 8.0z2 build I have upgraded to 8.0z3 build


Greetings,

We have released IBM Ceph 8.0z2. This includes ceph-19.2.0-98.el9cp, matching the latest RH Ceph Storage 8.0z2.

You can find the RHEL repos here:

https://public.dhe.ibm.com/ibmdl/export/pub/storage/ceph/8

Containers in IBM's Production registry:

  cp.icr.io/cp/ibm-ceph/ceph-8-crimson-rhel9:19.2.0-98.0.crimson.26.20250228032923
  cp.icr.io/cp/ibm-ceph/ceph-8-rhel9:8-80
  cp.icr.io/cp/ibm-ceph/cephcsi:v4.16.8-1
  cp.icr.io/cp/ibm-ceph/grafana-rhel9:10.4.8-27
  cp.icr.io/cp/ibm-ceph/haproxy-rhel9:2.4.22-45
  cp.icr.io/cp/ibm-ceph/keepalived-rhel9:2.2.8-43
  cp.icr.io/cp/ibm-ceph/logging-loki-rhel8:v3.1.0-15
  cp.icr.io/cp/ibm-ceph/nginx-124-rhel9:1-25.1726696143
  cp.icr.io/cp/ibm-ceph/nvmeof-cli-rhel9:1.3.5-1
  cp.icr.io/cp/ibm-ceph/nvmeof-rhel9:1.3.5-3
  cp.icr.io/cp/ibm-ceph/oauth2-proxy-rhel9:v7.6.0-10
  cp.icr.io/cp/ibm-ceph/prometheus-alertmanager:v4.15.0-202502172235.p0.g870ade5.assembly.stream.el8
  cp.icr.io/cp/ibm-ceph/prometheus-node-exporter:v4.15.0-202502171304.p0.gaed837c.assembly.stream.el8
  cp.icr.io/cp/ibm-ceph/prometheus:v4.15.0-202502171304.p0.g1b43998.assembly.stream.el8
  cp.icr.io/cp/ibm-ceph/promtail-rhel9:v3.0.0-16
  cp.icr.io/cp/ibm-ceph/samba-metrics-rhel9:v0.5-1093.20250228150555.693bfad
  cp.icr.io/cp/ibm-ceph/samba-server-rhel9:v0.5-1092.20250228144921.412f1ad
  cp.icr.io/cp/ibm-ceph/snmp-notifier-rhel9:1.2.1-94

Reminder on how to get the cp user token for cp.icr.io:
https://github.ibm.com/alchemy-registry/image-iam/blob/master/obtaining_entitlement.md#obtaining-a-production-entitlement-key

Comment 8 JAYA PRAKASH P 2025-03-20 09:04:53 UTC
I am able to reproduce this issue again another cluster when I upgrade from 8.0z2 to 8.0z3

[ceph: root@ceph-80z2to80z3-1wef1h-node1-installer /]# ceph versions
{
    "mon": {
        "ceph version 19.2.0-114.el9cp (ded81e05f6ece06e39d5564535edd25ec76f7198) squid (stable)": 3
    },
    "mgr": {
        "ceph version 19.2.0-114.el9cp (ded81e05f6ece06e39d5564535edd25ec76f7198) squid (stable)": 2
    },
    "osd": {
        "ceph version 19.2.0-114.el9cp (ded81e05f6ece06e39d5564535edd25ec76f7198) squid (stable)": 16
    },
    "tcmu-runner": {
        "ceph version 19.2.0-98.el9cp (d5c3cf625491b0bd76b4585e77aa0d907446f314) squid (stable)": 2
    },
    "overall": {
        "ceph version 19.2.0-114.el9cp (ded81e05f6ece06e39d5564535edd25ec76f7198) squid (stable)": 21,
        "ceph version 19.2.0-98.el9cp (d5c3cf625491b0bd76b4585e77aa0d907446f314) squid (stable)": 2
    }
}
[ceph: root@ceph-80z2to80z3-1wef1h-node1-installer /]#

When I tried upgrade from 7.1 to 8.0z3, we are not observing issue

Comment 9 JAYA PRAKASH P 2025-03-20 09:23:43 UTC
Please use cluster 10.0.67.12(root/passwd) for debugging purpose

Comment 10 gadi 2025-03-20 11:16:47 UTC
Hi JAYA,
1. can you share with me the log files "/var/log/ceph/<some-fsid>/tcmu-runner.log" and "/var/log/ceph/cephadm.log" please.
2. what cephadm version did you use?
3. I cannot connect to your installer. do I need any key?

Comment 12 JAYA PRAKASH P 2025-03-21 04:36:49 UTC
Created attachment 2081193 [details]
tcmu-runner log

Comment 13 JAYA PRAKASH P 2025-03-21 04:37:31 UTC
Created attachment 2081194 [details]
tcmu-runner log

Comment 14 JAYA PRAKASH P 2025-03-21 04:37:58 UTC
Created attachment 2081195 [details]
cephadm log

Comment 15 JAYA PRAKASH P 2025-03-21 04:40:47 UTC
Hi,

I have attached tcmu-runner.log and cephadm.log

I have used ceph version 19.2.0-98.el9cp to load 8.0z2 build

I have used ceph version 19.2.0-114.el9cp to load 8.0z3 build

I am able to do ssh without any key

pjayaprakash@Ps-MacBook-Pro ~ % ssh root.67.12
root.67.12's password:
Activate the web console with: systemctl enable --now cockpit.socket

Register this system with Red Hat Insights: insights-client --register
Create an account or view all your systems at https://red.ht/insights-dashboard
Last login: Thu Mar 20 08:09:38 2025 from 10.74.16.60
[root@ceph-80z2to80z3-1wef1h-node1-installer ~]# cephadm shell
Inferring fsid 46b7f330-0550-11f0-bee0-fa163e8669db
Inferring config /var/lib/ceph/46b7f330-0550-11f0-bee0-fa163e8669db/mon.ceph-80z2to80z3-1wef1h-node1-installer/config
Using ceph image with id '5673adc7ccad' and tag '8-98' created on 2025-03-19 00:13:12 +0000 UTC
cp.stg.icr.io/cp/ibm-ceph/ceph-8-rhel9@sha256:42abc709243493cfb702d26ed141f8e183982ee850c61226655223174d3b1cc6
[ceph: root@ceph-80z2to80z3-1wef1h-node1-installer /]#

Comment 24 JAYA PRAKASH P 2025-03-26 13:18:16 UTC
verified issue and now tcmu-runner version is upgraded


CEPH versions before upgrade
[ceph: root@ceph-80z2toz3-krmzg3-node1-installer ceph]# ceph versions
{
    "mon": {
        "ceph version 19.2.0-98.el9cp (d5c3cf625491b0bd76b4585e77aa0d907446f314) squid (stable)": 3
    },
    "mgr": {
        "ceph version 19.2.0-98.el9cp (d5c3cf625491b0bd76b4585e77aa0d907446f314) squid (stable)": 2
    },
    "osd": {
        "ceph version 19.2.0-98.el9cp (d5c3cf625491b0bd76b4585e77aa0d907446f314) squid (stable)": 16
    },
    "tcmu-runner": {
        "ceph version 19.2.0-98.el9cp (d5c3cf625491b0bd76b4585e77aa0d907446f314) squid (stable)": 2
    },
    "overall": {
        "ceph version 19.2.0-98.el9cp (d5c3cf625491b0bd76b4585e77aa0d907446f314) squid (stable)": 23
    }
}

CEPH versions after upgrade
[ceph: root@ceph-80z2toz3-krmzg3-node1-installer /]# ceph versions
{
    "mon": {
        "ceph version 19.2.0-120.el9cp (7c75c2332f91a9524ad9b3bcae413f8fac4dc02c) squid (stable)": 3
    },
    "mgr": {
        "ceph version 19.2.0-120.el9cp (7c75c2332f91a9524ad9b3bcae413f8fac4dc02c) squid (stable)": 2
    },
    "osd": {
        "ceph version 19.2.0-120.el9cp (7c75c2332f91a9524ad9b3bcae413f8fac4dc02c) squid (stable)": 16
    },
    "tcmu-runner": {
        "ceph version 19.2.0-120.el9cp (7c75c2332f91a9524ad9b3bcae413f8fac4dc02c) squid (stable)": 2
    },
    "overall": {
        "ceph version 19.2.0-120.el9cp (7c75c2332f91a9524ad9b3bcae413f8fac4dc02c) squid (stable)": 23
    }
}
[ceph: root@ceph-80z2toz3-krmzg3-node1-installer /]#
[ceph: root@ceph-80z2toz3-krmzg3-node1-installer /]# ceph --version
ceph version 19.2.0-120.el9cp (7c75c2332f91a9524ad9b3bcae413f8fac4dc02c) squid (stable)
[ceph: root@ceph-80z2toz3-krmzg3-node1-installer /]# ceph -v
ceph version 19.2.0-120.el9cp (7c75c2332f91a9524ad9b3bcae413f8fac4dc02c) squid (stable)
[ceph: root@ceph-80z2toz3-krmzg3-node1-installer /]# ceph version
ceph version 19.2.0-120.el9cp (7c75c2332f91a9524ad9b3bcae413f8fac4dc02c) squid (stable)
[ceph: root@ceph-80z2toz3-krmzg3-node1-installer /]#

Comment 26 JAYA PRAKASH P 2025-03-27 04:19:50 UTC
Documentation update not required

Comment 28 errata-xmlrpc 2025-04-07 15:27:26 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory (Important: Red Hat Ceph Storage 8.0 security, bug fix, and enhancement updates), and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHSA-2025:3635


Note You need to log in before you can comment on or make changes to this bug.