Description of problem: cephadm is taking osd service name as osd, a dumping ground for all OSDs made that aren't attached to an OSD spec. So it isn't really a service, doesn't have a spec but it's listing in [root@ceph-spr-cpm3gh-node4 ~]# ceph orch ls osd NAME PORTS RUNNING REFRESHED AGE PLACEMENT osd 3 10m ago - <unmanaged> osd.all-available-devices 6 10m ago 2h <unmanaged> but while restart command it's failing as below [root@ceph-spr-cpm3gh-node4 ~]# ceph orch restart osd Error EINVAL: Invalid service name "osd". View currently running services using "ceph orch ls"it's a bug right? cephadm is taking osd service name as osd but while restart command it's failing hence requesting only to list active and running osd services in ceph orch ls osd Version-Release number of selected component (if applicable): ceph version 17.2.6-167.el9cp (5ef1496ea3e9daaa9788809a172bd5a1c3192cf7) quincy (stable) How reproducible: tried twice Steps to Reproduce: 1.deploy ceph 6.1 using cephadm 2.use provided spec file to deploy ceph cluster provided in additional info 3.even service fails to deploy with service name "osd" 4.it's getting listed in ceph orch ls osd Actual results: [root@ceph-spr-cpm3gh-node4 ~]# ceph orch ls osd NAME PORTS RUNNING REFRESHED AGE PLACEMENT osd 3 4m ago - <unmanaged> osd.all-available-devices 6 4m ago 3h <unmanaged> [root@ceph-spr-cpm3gh-node4 ~]# ceph orch restart osd Error EINVAL: Invalid service name "osd". View currently running services using "ceph orch ls" Please check log 2023-12-08T09:43:05.189114+0000 mgr.ceph-spr-cpm3gh-node1-installer.wfhbhf (mgr.14219) 5974 : cephadm [ERR] Invalid service name "osd". View currently running services using "ceph orch ls" Traceback (most recent call last): File "/usr/share/ceph/mgr/orchestrator/_interface.py", line 125, in wrapper return OrchResult(f(*args, **kwargs)) File "/usr/share/ceph/mgr/cephadm/module.py", line 2033, in service_action raise OrchestratorError(f'Invalid service name "{service_name}".' orchestrator._interface.OrchestratorError: Invalid service name "osd". View currently running services using "ceph orch ls" 2023-12-08T09:53:37.348836+0000 mgr.ceph-spr-cpm3gh-node1-installer.wfhbhf (mgr.14219) 6305 : cephadm [ERR] Invalid service name "osd". View currently running services using "ceph orch ls" [root@ceph-spr-cpm3gh-node4 ~]# ceph orch ls NAME PORTS RUNNING REFRESHED AGE PLACEMENT alertmanager ?:9093,9094 1/1 10m ago 23h count:1 ceph-exporter 3/3 10m ago 23h * crash 3/3 10m ago 23h * grafana ?:3000 1/1 10m ago 23h count:1 mgr 2/2 10m ago 23h label:mgr mon 3/3 10m ago 23h label:mon node-exporter ?:9100 3/3 10m ago 23h * osd 4 9m ago - <unmanaged> osd.all-available-devices 5 10m ago 22h <unmanaged> prometheus ?:9095 1/1 10m ago 23h count:1 Expected results: list only services that got created successfully with active and running. if cephadm doesn't want user to specify osd service_name as "osd" like specified in spec file provide that error message to users while applying spec command. Additional info: [root@ceph-spr-cpm3gh-node4 ~]# ceph orch ls --export service_type: alertmanager service_name: alertmanager placement: count: 1 --- service_type: ceph-exporter service_name: ceph-exporter placement: host_pattern: '*' spec: prio_limit: 5 stats_period: 5 --- service_type: crash service_name: crash placement: host_pattern: '*' --- service_type: grafana service_name: grafana placement: count: 1 --- service_type: mgr service_name: mgr placement: label: mgr --- service_type: mon service_name: mon placement: label: mon --- service_type: node-exporter service_name: node-exporter placement: host_pattern: '*' --- service_type: osd service_name: osd unmanaged: true spec: filter_logic: AND objectstore: bluestore --- service_type: osd service_id: all-available-devices service_name: osd.all-available-devices placement: host_pattern: '*' unmanaged: true spec: data_devices: all: true filter_logic: AND objectstore: bluestore --- service_type: prometheus service_name: prometheus placement: count: 1
*** Bug 2279839 has been marked as a duplicate of this bug. ***
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Important: Red Hat Ceph Storage 8.1 security, bug fix, and enhancement updates), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2025:9775