Bug 1329008 - [ceph-ansible] : Installation is hung at task 'collect admin and bootstrap keys'
Summary: [ceph-ansible] : Installation is hung at task 'collect admin and bootstrap k...
Keywords:
Status: CLOSED ERRATA
Alias: None
Product: Red Hat Storage Console
Classification: Red Hat Storage
Component: ceph-ansible
Version: 2
Hardware: x86_64
OS: Linux
unspecified
urgent
Target Milestone: ---
: 2
Assignee: Christina Meno
QA Contact: sds-qe-bugs
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2016-04-20 22:16 UTC by Rachana Patel
Modified: 2016-08-23 19:49 UTC (History)
8 users (show)

Fixed In Version: ceph-ansible-1.0.5-4.el7
Doc Type: Bug Fix
Doc Text:
Clone Of:
Environment:
Last Closed: 2016-08-23 19:49:30 UTC
Embargoed:


Attachments (Terms of Use)


Links
System ID Private Priority Status Summary Last Updated
Red Hat Product Errata RHEA-2016:1754 0 normal SHIPPED_LIVE New packages: Red Hat Storage Console 2.0 2017-04-18 19:09:06 UTC

Description Rachana Patel 2016-04-20 22:16:07 UTC
Description of problem:
=======================
Ceph Intsallation using ceph-anisble got hung at task 'collect admin and bootstrap keys'


Version-Release number of selected component (if applicable):
=============================================================
ceph-ansible-1.0.5-3.el7.noarch
ceph-10.1.1-1.el7cp.x86_64


How reproducible:
=================
3 times out of 4


Steps to Reproduce:
==================
1. prepare node for ceph installation.(follow prerequisites steps.)
2. create host file having only mon and osd roles.
3.run ansible command with below values
ansible-playbook site.yml -vv -i  /etc/ansible/hosts_harish  --extra-vars '{"ceph_stable": true, "ceph_origin": "distro", "ceph_stable_rh_storage": true, "monitor_interface": "eno1", "journal_collocation": true, "devices": ["/dev/sdb", "/dev/sdc", "/dev/sdd"], "journal_size": 100, "public_network": "10.8.128.0/21", "cephx": true, "fetch_directory": "~/ceph-ansible-keys"}' -u root

Actual results:
===============
Installation is stuck at below task
TASK: [ceph-mon | collect admin and bootstrap keys] ***************************
<magna051> REMOTE_MODULE command ceph-create-keys --cluster ceph --id magna051
<magna063> REMOTE_MODULE command ceph-create-keys --cluster ceph --id magna063
<magna067> REMOTE_MODULE command ceph-create-keys --cluster ceph --id magna067
ok: [magna067] => {"changed": false, "cmd": ["ceph-create-keys", "--cluster", "ceph", "--id", "magna067"], "delta": "0:00:01.251416", "end": "2016-04-20 16:13:46.546810", "failed": false, "failed_when_result": false, "rc": 0, "start": "2016-04-20 16:13:45.295394", "stderr": "INFO:ceph-create-keys:Talking to monitor...\nINFO:ceph-create-keys:Talking to monitor...\n2016-04-20 16:13:45.860068 7fe341032700 -1 asok(0x7fe33c001680) AdminSocketConfigObs::init: failed: AdminSocket::bind_and_listen: failed to bind the UNIX domain socket to '/var/run/ceph/rbd-clients/ceph-client.admin.18881.140613940942768.asok': (2) No such file or directory\nINFO:ceph-create-keys:Talking to monitor...\n2016-04-20 16:13:46.103704 7f6725e22700 -1 asok(0x7f6720001680) AdminSocketConfigObs::init: failed: AdminSocket::bind_and_listen: failed to bind the UNIX domain socket to '/var/run/ceph/rbd-clients/ceph-client.admin.18906.140080895236016.asok': (2) No such file or directory\nINFO:ceph-create-keys:Talking to monitor...\n2016-04-20 16:13:46.372301 7febb3864700 -1 asok(0x7febac001680) AdminSocketConfigObs::init: failed: AdminSocket::bind_and_listen: failed to bind the UNIX domain socket to '/var/run/ceph/rbd-clients/ceph-client.admin.18945.140650179729328.asok': (2) No such file or directory", "stdout": "", "warnings": []}

Comment 2 Andrew Schoen 2016-04-21 19:15:11 UTC
I think this might be a duplicate of https://bugzilla.redhat.com/show_bug.cgi?id=1329013

Rachana,

When you see this issue, does the 'mon addr' section of ceph.conf for each mon contain '0.0.0.0'? I think the problem here is a failure to communicate to the monitors. You're only seeing the warning about the rbd client admin socket because of unneeded config in ceph.conf. This BZ address that: https://bugzilla.redhat.com/show_bug.cgi?id=1328594

Comment 3 Rachana Patel 2016-04-25 14:56:53 UTC
Andrew,
ceph.conf has '0.0.0.0' in mon section.

Comment 5 Christina Meno 2016-04-27 00:32:08 UTC
Andrew is is a duplicate?

Comment 6 Andrew Schoen 2016-04-27 15:42:28 UTC
(In reply to Gregory Meno from comment #5)
> Andrew is is a duplicate?

Yes, I believe it is caused by the same problem as https://bugzilla.redhat.com/show_bug.cgi?id=1329013

It should be fixed in ceph-ansible-1.0.5-4.el7

Rachana, can you confirm that this is no longer an issue with the latest version of ceph-ansible?

Comment 7 Rachana Patel 2016-05-02 21:37:34 UTC
Andrew, yes it is not coming with latest version.

Comment 8 Christina Meno 2016-05-04 21:08:43 UTC
Rachana, this sounds like a candidate for moving to Verified.
Would you please qa_ack, clear needinfo and move to Verified state?

Comment 9 Rachana Patel 2016-05-05 21:34:47 UTC
working with latest build  ceph-ansible-1.0.5-4.el7. Hence moving to verified.

Comment 13 errata-xmlrpc 2016-08-23 19:49:30 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory, and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHEA-2016:1754


Note You need to log in before you can comment on or make changes to this bug.