Bug 1898196 - [cephadm] 5.0 - osd.None an unknown service is created and displayed in ceph orch ls command
Summary: [cephadm] 5.0 - osd.None an unknown service is created and displayed in ceph ...
Keywords:
Status: CLOSED ERRATA
Alias: None
Product: Red Hat Ceph Storage
Classification: Red Hat Storage
Component: Cephadm
Version: 5.0
Hardware: x86_64
OS: Linux
medium
high
Target Milestone: ---
: 5.0
Assignee: Juan Miguel Olmo
QA Contact: Vasishta
Karen Norteman
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2020-11-16 16:05 UTC by Preethi
Modified: 2021-08-30 08:27 UTC (History)
3 users (show)

Fixed In Version: ceph-16.2.0-28.el8cp
Doc Type: No Doc Update
Doc Text:
Clone Of:
Environment:
Last Closed: 2021-08-30 08:27:12 UTC
Embargoed:


Attachments (Terms of Use)
ceph orch ls screenshot (180.88 KB, image/png)
2020-11-16 16:05 UTC, Preethi
no flags Details


Links
System ID Private Priority Status Summary Last Updated
Github ceph ceph pull 40736 0 None closed mgr/orchestrator,cephadm: clean up 'orch ls', osd services 2021-04-23 07:19:38 UTC
Github ceph ceph pull 40746 0 None open pacific: cephadm: april batch backport 2 2021-04-23 07:19:38 UTC
Red Hat Issue Tracker RHCEPH-1173 0 None None None 2021-08-30 00:15:32 UTC
Red Hat Product Errata RHBA-2021:3294 0 None None None 2021-08-30 08:27:26 UTC

Description Preethi 2020-11-16 16:05:56 UTC
Created attachment 1729808 [details]
ceph orch ls screenshot

Description of problem:[cephadm] 5.0 - osd.None an unknown service is created and displayed in ceph orch ls command 


Version-Release number of selected component (if applicable):

[root@magna094 ubuntu]# ./cephadm version
Using recent ceph image registry-proxy.engineering.redhat.com/rh-osbs/rhceph:ceph-5.0-rhel-8-containers-candidate-96803-20201013192445
ceph version 16.0.0-6275.el8cp (d1e0606106224ac333f1c245150d7484cb626841) pacific (dev)

[root@magna094 ubuntu]# rpm -qa |grep cephadm
cephadm-16.0.0-6817.el8cp.x86_64
[root@magna094 ubuntu]#Release number of selected component (if applicable):


How reproducible:

Steps to Reproduce:
1.Install 5.0 cluster with dashboard enabled
2. Perform MON, MGR, MDS, RGW, OSDs deployment with all available devices daemons using ceph orch commands
2. Perform ceph orch ls to check the services details
3. observe the behaviour

OSD.none service is created along with other services details but not sure why this is created

logs attached https://drive.google.com/drive/u/1/my-drive
file name: mgr_latest_osd.txt



[ceph: root@magna094 /]# ceph orch ls
NAME                       RUNNING  REFRESHED  AGE  PLACEMENT                           IMAGE NAME                                                                                                      IMAGE ID      
alertmanager                   1/1  7s ago     5w   count:1                             docker.io/prom/alertmanager:v0.20.0                                                                             0881eb8f169f  
crash                          9/9  10s ago    5w   *                                   registry-proxy.engineering.redhat.com/rh-osbs/rhceph:ceph-5.0-rhel-8-containers-candidate-96803-20201013192445  0158d7274861  
grafana                        1/1  7s ago     5w   count:1                             docker.io/ceph/ceph-grafana:6.6.2                                                                               a0dce381714a  
iscsi.iscsi                    0/2  -          -    magna092;magna093;count:2           <unknown>                                                                                                       <unknown>     
mds.test                       3/3  10s ago    13d  count:3                             registry-proxy.engineering.redhat.com/rh-osbs/rhceph:ceph-5.0-rhel-8-containers-candidate-96803-20201013192445  0158d7274861  
mgr                            2/2  10s ago    5w   count:2                             registry-proxy.engineering.redhat.com/rh-osbs/rhceph:ceph-5.0-rhel-8-containers-candidate-96803-20201013192445  0158d7274861  
mon                            3/3  10s ago    4w   magna094;magna067;magna073;count:3  registry-proxy.engineering.redhat.com/rh-osbs/rhceph:ceph-5.0-rhel-8-containers-candidate-96803-20201013192445  0158d7274861  
nfs.ganesha-testnfs            1/1  7s ago     13d  count:1                             registry-proxy.engineering.redhat.com/rh-osbs/rhceph:ceph-5.0-rhel-8-containers-candidate-96803-20201013192445  0158d7274861  
node-exporter                  9/9  10s ago    5w   *                                   docker.io/prom/node-exporter:v0.18.1                                                                            e5a616e4b9cf  
osd.None                      10/0  10s ago    -    <unmanaged>                         registry-proxy.engineering.redhat.com/rh-osbs/rhceph:ceph-5.0-rhel-8-containers-candidate-96803-20201013192445  0158d7274861  
osd.all-available-devices    17/17  10s ago    17h  <unmanaged>                         registry-proxy.engineering.redhat.com/rh-osbs/rhceph:ceph-5.0-rhel-8-containers-candidate-96803-20201013192445  0158d7274861  
prometheus                     1/1  7s ago     5w   count:1                             docker.io/prom/prometheus:v2.18.1                                                                               de242295e225  
rgw.myorg.us-east-1            2/2  10s ago    2w   magna092;magna093;count:2           registry-proxy.engineering.redhat.com/rh-osbs/rhceph:ceph-5.0-rhel-8-containers-candidate-96803-20201013192445  0158d7274861  
[ceph: root@magna094 /]# 


[ceph: root@magna094 /]# ceph orch ls osd.None --export
No services reported


Actual results: Unknown OSD.NOne service is created along with osd services and is displayed in ceph orch ls command


Expected results: there should be any unknown service deployed 


Additional info:
magna094 root/q - bootstrap node

Comment 1 Juan Miguel Olmo 2021-02-16 15:36:04 UTC
Waiting  backport to Pacific.

Comment 5 Preethi 2021-03-08 13:25:47 UTC
@Juan, We still see the issue in the latest alpha 


[ceph: root@ceph-sunil1adm-1614692246522-node1-mon-mgr-installer-node-expor /]# ceph orch ls
NAME                                  RUNNING  REFRESHED  AGE  PLACEMENT                                                                                                                                IMAGE NAME                                                                                                                    IMAGE ID      
alertmanager                              2/2  6m ago     5d   ceph-sunil1adm-1614692246522-node2-mon-mds-node-exporter-alertm;ceph-sunil1adm-1614692246522-node1-mon-mgr-installer-node-expor          registry.redhat.io/openshift4/ose-prometheus-alertmanager:v4.5                                                                3ea01d72d22c  
crash                                     0/8  -          -    <unmanaged>                                                                                                                              <unknown>                                                                                                                     <unknown>     
grafana                                   1/1  6m ago     5d   ceph-sunil1adm-1614692246522-node1-mon-mgr-installer-node-expor                                                                          registry.redhat.io/rhceph-alpha/rhceph-5-dashboard-rhel8:latest                                                               bd3d7748747b  
iscsi.iscsi                               1/1  6m ago     5d   ceph-sunil1adm-1614692246522-node8-client-nfs-node-exporter-isc                                                                          registry-proxy.engineering.redhat.com/rh-osbs/rhceph@sha256:687d060d91102d9317fbd8ec0305a24f53bdfb86d7ca3aaacc664955da01f03f  700feae6f592  
mds.cephfs                                2/2  6m ago     5d   ceph-sunil1adm-1614692246522-node2-mon-mds-node-exporter-alertm;ceph-sunil1adm-1614692246522-node8-client-nfs-node-exporter-isc;count:2  registry-proxy.engineering.redhat.com/rh-osbs/rhceph@sha256:687d060d91102d9317fbd8ec0305a24f53bdfb86d7ca3aaacc664955da01f03f  700feae6f592  
mgr                                       1/1  6m ago     5d   ceph-sunil1adm-1614692246522-node1-mon-mgr-installer-node-expor                                                                          registry-proxy.engineering.redhat.com/rh-osbs/rhceph:ceph-5.0-rhel-8-containers-candidate-21981-20210302003306                700feae6f592  
mon                                       3/3  6m ago     5d   label:mon                                                                                                                                mix                                                                                                                           700feae6f592  
node-exporter                             8/8  6m ago     5d   *                                                                                                                                        registry.redhat.io/openshift4/ose-prometheus-node-exporter:v4.5                                                               a6af8f87dd4a  
osd.None                                  1/0  6m ago     -    <unmanaged>                                                                                                                              registry-proxy.engineering.redhat.com/rh-osbs/rhceph@sha256:687d060d91102d9317fbd8ec0305a24f53bdfb86d7ca3aaacc664955da01f03f  700feae6f592  
osd.all-available-devices                 7/9  6m ago     3d   <unmanaged>                                                                                                                              registry-proxy.engineering.redhat.com/rh-osbs/rhceph@sha256:687d060d91102d9317fbd8ec0305a24f53bdfb86d7ca3aaacc664955da01f03f  700feae6f592  
osd.dashboard-admin123-1614929231097     2/10  6m ago     3d   *                                                                                                                                        registry-proxy.engineering.redhat.com/rh-osbs/rhceph@sha256:687d060d91102d9317fbd8ec0305a24f53bdfb86d7ca3aaacc664955da01f03f  mix           
prometheus                                1/1  6m ago     5d   ceph-sunil1adm-1614692246522-node1-mon-mgr-installer-node-expor                                                                          registry.redhat.io/openshift4/ose-prometheus:v4.6                                                                             6050e785b668  


I could see the issue in the below clusters

magna021 root/q
10.0.210.149 cephuser/cephuser

Probably, we can see the issue after setting the unmanaged flag to true after deploying the osd all available service daemon.

Comment 6 Juan Miguel Olmo 2021-03-15 10:13:11 UTC
"osd:None" services appear after creating OSDs using "ceph orch daemon add osd host:device"

"osd.None" is a simulated service created to show "something" for OSD daemons created with "daemon add". This is the reason why they can not be deleted. (these kind of services really do not exist).

As a workaround, until modification to avoid "simulated" services will be merged:

In order to remove "osd.None" service, it is needed to remove all OSDs daemons associated with the simulated service.
Use "ceph orch ls osd --format yaml" to get the list of devices that are used by the "osd.None" service.
Use "ceph device ls" to find the osd ids of these devices.
USe "ceph orch osd rm <id>"  to remove the associated daemons.

Once associated OSDs will be removed the "osd.None" service will disappear.

Comment 7 Juan Miguel Olmo 2021-04-23 07:19:38 UTC
Backport to pacific on-going: https://github.com/ceph/ceph/pull/40746

Comment 11 Preethi 2021-05-19 09:48:48 UTC
@Juan, Verified with  ceph version 16.2.0-34.el8cp. Issue is not seen. Hence, moving this to verified state.


[ceph: root@ceph-5x-doc-mgowri-1620728264715-node1-mon-mgr-installer /]# ceph orch ls
NAME                       RUNNING  REFRESHED  AGE  PLACEMENT                                                                                                                                                           
alertmanager                   1/1  4m ago     6d   ceph-5x-doc-mgowri-1620728264715-node1-mon-mgr-installer                                                                                                            
crash                          7/7  8m ago     7d   *                                                                                                                                                                   
grafana                        1/1  4m ago     6d   ceph-5x-doc-mgowri-1620728264715-node1-mon-mgr-installer                                                                                                            
mds.fs_name                    2/2  4m ago     22h  ceph-5x-doc-mgowri-1620728264715-node1-mon-mgr-installer;ceph-5x-doc-mgowri-1620728264715-node7-rgw-mgr                                                             
mgr                            3/3  6m ago     23h  ceph-5x-doc-mgowri-1620728264715-node1-mon-mgr-installer;ceph-5x-doc-mgowri-1620728264715-node2-mon-mgr-rgw;ceph-5x-doc-mgowri-1620728264715-node7-rgw-mgr          
mon                            3/3  6m ago     22h  ceph-5x-doc-mgowri-1620728264715-node1-mon-mgr-installer;ceph-5x-doc-mgowri-1620728264715-node2-mon-mgr-rgw;ceph-5x-doc-mgowri-1620728264715-node6-mon-rgw;count:3  
nfs.foo                        2/2  6m ago     4h   ceph-5x-doc-mgowri-1620728264715-node2-mon-mgr-rgw;ceph-5x-doc-mgowri-1620728264715-node7-rgw-mgr;count:2                                                           
node-exporter                  7/7  8m ago     6d   *                                                                                                                                                                   
osd.all-available-devices    12/19  8m ago     23h  <unmanaged>                                                                                                                                                         
prometheus                     1/1  4m ago     6d   ceph-5x-doc-mgowri-1620728264715-node1-mon-mgr-installer                                                                                                            
rgw.foo                        6/6  6m ago     23h  count-per-host:2;label:rgw                                                                                                                                          
rgw.test                       2/2  6m ago     23h  ceph-5x-doc-mgowri-1620728264715-node2-mon-mgr-rgw;ceph-5x-doc-mgowri-1620728264715-node6-mon-rgw;count:2                                                           
[ceph: root@ceph-5x-doc-mgowri-1620728264715-node1-mon-mgr-installer /]#

Comment 13 errata-xmlrpc 2021-08-30 08:27:12 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory (Red Hat Ceph Storage 5.0 bug fix and enhancement), and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHBA-2021:3294


Note You need to log in before you can comment on or make changes to this bug.