Bug 1489835 - [ceph-container] - OSD not getting activated - entrypoint.sh: static: does not generate config
Summary: [ceph-container] - OSD not getting activated - entrypoint.sh: static: does no...
Keywords:
Status: CLOSED NOTABUG
Alias: None
Product: Red Hat Ceph Storage
Classification: Red Hat Storage
Component: Container
Version: 3.0
Hardware: Unspecified
OS: Unspecified
unspecified
high
Target Milestone: rc
: 3.0
Assignee: Sébastien Han
QA Contact: Harish NV Rao
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2017-09-08 12:47 UTC by Vasishta
Modified: 2017-09-15 13:12 UTC (History)
5 users (show)

Fixed In Version:
Doc Type: If docs needed, set a value
Doc Text:
Clone Of:
Environment:
Last Closed: 2017-09-15 13:12:06 UTC
Embargoed:


Attachments (Terms of Use)
File contains OSD service log snippet, docker service log snippet and ansible log (279.56 KB, text/x-vhdl)
2017-09-08 12:47 UTC, Vasishta
no flags Details
File contains journald logs of an osd serivce (11.62 KB, text/plain)
2017-09-08 14:13 UTC, Vasishta
no flags Details
File contains journald logs of an osd serivce (15.26 KB, text/plain)
2017-09-08 15:16 UTC, Vasishta
no flags Details

Description Vasishta 2017-09-08 12:47:17 UTC
Created attachment 1323733 [details]
File contains OSD service log snippet, docker service log snippet and ansible log

Description of problem:
After ansible playbook successfully completed running site-docker.yml, OSDs are not getting activated. 

Faced this issue while working on BZ 1458512
Cluster name used - 11

Version-Release number of selected component (if applicable):
ceph-3.0-rhel-7-docker-candidate-71465-20170804220045
ceph-ansible-3.0.0-0.1.rc6.el7cp.noarch

How reproducible:
Always (2/2)

Steps to Reproduce:
1.Configure ceph ansible to get a containerized ceph cluster up with container name with numbers
2.run ansible-playbook site-docker.yml

Actual results:
OSD not getting activated
(/entrypoint.sh: static: does not generate config)

Expected results:
OSD must get activated

Additional info:
1) Please let me know if I have missed anything
$ cat group_vars/osds.yml | egrep -v ^# | grep -v ^$
---
dummy:
devices:
  - /dev/sdb
  - /dev/sdc
  - /dev/sdd
ceph_osd_docker_prepare_env: -e CLUSTER={{ cluster }} -e OSD_JOURNAL_SIZE={{ journal_size }} -e OSD_FORCE_ZAP=1 -e OSD_FILESTORE=1
ceph_osd_docker_extra_env: -e CLUSTER={{ cluster }} -e CEPH_DAEMON=OSD_CEPH_DISK_ACTIVATE -e OSD_JOURNAL_SIZE={{ journal_size }} -e OSD_FILESTORE=1

2) expose_partions function from /usr/share/ceph-osd-run.sh has been copied into the attachment, where OSD_BLUESTORE_BLOCK_WAL is being set, it lead to a bit confusion.

Comment 2 seb 2017-09-08 13:53:39 UTC
please edit /usr/share/ceph-osd-run.sh and add

-e DEBUG=verbose \

Then look at journald logs and write it here, thanks.

Comment 3 Vasishta 2017-09-08 14:13:05 UTC
Created attachment 1323752 [details]
File contains journald logs of an osd serivce

Comment 4 seb 2017-09-08 14:23:47 UTC
Also could you use the latest image available? Looks like ceph-3.0-rhel-7-docker-candidate-21213-20170908140758 is the latest.

Thanks!

Comment 5 Vasishta 2017-09-08 15:16:08 UTC
Created attachment 1323772 [details]
File contains journald logs of an osd serivce

Hi Sebastien,

I upgraded only one OSD node, Here's the log snippet -
raise Error('%s does not exist' % args.path)
ceph-osd-run.sh[32644]: ceph_disk.main.Error: Error: /dev/sdb1 does not exist

Larger log snippet is in the attachment.

Regards,
Vasishata

Comment 6 seb 2017-09-13 15:40:13 UTC
Please add osd_scenario: collocated 
Also leave ceph_osd_docker_extra_env empty and set ceph_osd_docker_prepare_env: -e OSD_JOURNAL_SIZE={{ journal_size }}

Thanks.

Comment 7 seb 2017-09-15 13:12:06 UTC
I'm closing this since this was due to a misconfiguration.
We will address the doc in another BZ.


Note You need to log in before you can comment on or make changes to this bug.