Yogev, I think this can be moved into VERIFIED; we should check if the systemd services for ceph-osd and ceph-mon are disabled when deploying ceph in containes
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.
For information on the advisory, and where to find the updated
files, follow the link below.
If the solution does not work for you, open a new bug report.
https://access.redhat.com/errata/RHBA-2017:3387
Comment 19Red Hat Bugzilla
2023-09-15 00:03:38 UTC
The needinfo request[s] on this closed bug have been removed as they have been unresolved for 500 days
Description of problem: ----------------------- After minor update of CephStorage nodes ceph containers fail to start, but non-containerized are started: [root@ceph-0 ~]# docker ps CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES [root@ceph-0 ~]# systemctl status ceph* ● ceph-osd - Ceph OSD Loaded: loaded (/etc/systemd/system/ceph-osd@.service; enabled; vendor preset: disabled) Active: active (running) since Fri 2017-08-25 08:20:28 UTC; 3 days ago Main PID: 22630 (ceph-osd) CGroup: /system.slice/system-ceph\x2dosd.slice/ceph-osd └─22630 /usr/bin/ceph-osd -f --cluster ceph --id 0 --setuser ceph --setgroup ceph Warning: Journal has been rotated since unit was started. Log output is incomplete or unavailable. ● ceph-osd - Ceph OSD Loaded: loaded (/etc/systemd/system/ceph-osd@.service; enabled; vendor preset: disabled) Active: activating (auto-restart) (Result: exit-code) since Mon 2017-08-28 11:07:05 UTC; 7s ago Process: 779147 ExecStop=/usr/bin/docker stop ceph-osd-ceph-0-dev%i (code=exited, status=1/FAILURE) Process: 778918 ExecStart=/usr/share/ceph-osd-run.sh %i (code=exited, status=1/FAILURE) Process: 778910 ExecStartPre=/usr/bin/docker rm -f ceph-osd-ceph-0-dev%i (code=exited, status=1/FAILURE) Process: 778904 ExecStartPre=/usr/bin/docker stop ceph-osd-ceph-0-dev%i (code=exited, status=1/FAILURE) Main PID: 778918 (code=exited, status=1/FAILURE) Aug 28 11:07:05 ceph-0 systemd[1]: ceph-osd failed. [root@ceph-0 ~]# docker ps -a CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES 6ad1e9f2279a docker.io/ceph/daemon:tag-build-master-jewel-centos-7 "/entrypoint.sh" 3 days ago Exited (0) 3 days ago ceph-osd-prepare-ceph-0-devdevvdb [root@ceph-0 ~]# ceph status cluster 755c263a-88d9-11e7-af00-5254004ae3d0 health HEALTH_OK monmap e2: 3 mons at {controller-0=172.17.3.17:6789/0,controller-1=172.17.3.13:6789/0,controller-2=172.17.3.18:6789/0} election epoch 8, quorum 0,1,2 controller-1,controller-0,controller-2 osdmap e24: 3 osds: 3 up, 3 in flags sortbitwise,require_jewel_osds pgmap v6042: 224 pgs, 6 pools, 29001 kB data, 404 objects 200 MB used, 104 GB / 104 GB avail 224 active+clean client io 17 B/s rd, 0 op/s rd, 0 op/s wr Version-Release number of selected component (if applicable): ------------------------------------------------------------- ceph-selinux-10.2.7-28.el7cp.x86_64 python-cephfs-10.2.7-28.el7cp.x86_64 libcephfs1-10.2.7-28.el7cp.x86_64 puppet-ceph-2.3.1-0.20170805094345.868e6d6.el7ost.noarch ceph-radosgw-10.2.7-28.el7cp.x86_64 ceph-common-10.2.7-28.el7cp.x86_64 ceph-mon-10.2.7-28.el7cp.x86_64 ceph-mds-10.2.7-28.el7cp.x86_64 ceph-osd-10.2.7-28.el7cp.x86_64 ceph-base-10.2.7-28.el7cp.x86_64 puppet-ceph-2.3.1-0.20170805094345.868e6d6.el7ost.noarch ceph-ansible-3.0.0-0.1.rc3.el7cp.noarch openstack-tripleo-heat-templates-7.0.0-0.20170805163048.el7ost.noarch Steps to Reproduce: ------------------- 1. Follow https://etherpad.openstack.org/p/pike-update to perform minor update 2. Check ceph nodes