Description of problem: When installing OSD on directories, the ansible-playbook command exits unsuccessfully. Version-Release number of selected component (if applicable): 2.0 How reproducible: Always Steps to Reproduce: 1. Run ansible-playbook command to install a new cluster with osd's on directories 2. the command executes successfully but returns a failure result, and the osd's are up and running. Actual results: ansible command exits as a fatal failure. Expected results: The ansible-playbook command should exit successfully. Additional info: TASK: [ceph-osd | activate OSD(s)] ******************************************** ok: [magna077] => (item=/mnt/osdb/osd1) ok: [magna052] => (item=/mnt/osdb/osd1) ok: [magna058] => (item=/mnt/osdb/osd1) ok: [magna077] => (item=/mnt/osdb/osd2) ok: [magna058] => (item=/mnt/osdb/osd2) ok: [magna052] => (item=/mnt/osdb/osd2) ok: [magna077] => (item=/mnt/osdb/osd3) ok: [magna052] => (item=/mnt/osdb/osd3) ok: [magna058] => (item=/mnt/osdb/osd3) ok: [magna077] => (item=/mnt/osdb/osd4) ok: [magna052] => (item=/mnt/osdb/osd4) ok: [magna058] => (item=/mnt/osdb/osd4) ok: [magna058] => (item=/mnt/osdc/osd5) ok: [magna077] => (item=/mnt/osdc/osd5) ok: [magna052] => (item=/mnt/osdc/osd5) ok: [magna058] => (item=/mnt/osdc/osd6) ok: [magna052] => (item=/mnt/osdc/osd6) ok: [magna077] => (item=/mnt/osdc/osd6) ok: [magna052] => (item=/mnt/osdc/osd7) ok: [magna058] => (item=/mnt/osdc/osd7) ok: [magna077] => (item=/mnt/osdc/osd7) ok: [magna058] => (item=/mnt/osdc/osd8) ok: [magna052] => (item=/mnt/osdc/osd8) ok: [magna077] => (item=/mnt/osdc/osd8) ok: [magna058] => (item=/mnt/osdd/osd9) ok: [magna052] => (item=/mnt/osdd/osd9) ok: [magna077] => (item=/mnt/osdd/osd9) ok: [magna058] => (item=/mnt/osdd/osd10) ok: [magna052] => (item=/mnt/osdd/osd10) ok: [magna077] => (item=/mnt/osdd/osd10) ok: [magna052] => (item=/mnt/osdd/osd11) ok: [magna058] => (item=/mnt/osdd/osd11) ok: [magna077] => (item=/mnt/osdd/osd11) ok: [magna058] => (item=/mnt/osdd/osd12) ok: [magna052] => (item=/mnt/osdd/osd12) ok: [magna077] => (item=/mnt/osdd/osd12) TASK: [ceph-osd | start and add that the OSD service to the init sequence] **** fatal: [magna052] => error while evaluating conditional: ansible_service_mgr != "systemd" fatal: [magna058] => error while evaluating conditional: ansible_service_mgr != "systemd" fatal: [magna077] => error while evaluating conditional: ansible_service_mgr != "systemd" FATAL: all hosts have already failed -- aborting PLAY RECAP ******************************************************************** to retry, use: --limit @/root/site.sample.retry magna009 : ok=70 changed=14 unreachable=0 failed=0 magna031 : ok=70 changed=14 unreachable=0 failed=0 magna046 : ok=70 changed=14 unreachable=0 failed=0 magna052 : ok=97 changed=10 unreachable=1 failed=0 magna058 : ok=97 changed=10 unreachable=1 failed=0 magna077 : ok=97 changed=10 unreachable=1 failed=0 [root@magna009 yum.repos.d]# ceph osd tree ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY -1 32.73102 root default -2 10.91034 host magna077 0 0.90919 osd.0 up 1.00000 1.00000 3 0.90919 osd.3 up 1.00000 1.00000 6 0.90919 osd.6 up 1.00000 1.00000 9 0.90919 osd.9 up 1.00000 1.00000 12 0.90919 osd.12 up 1.00000 1.00000 16 0.90919 osd.16 up 1.00000 1.00000 20 0.90919 osd.20 up 1.00000 1.00000 22 0.90919 osd.22 up 1.00000 1.00000 26 0.90919 osd.26 up 1.00000 1.00000 29 0.90919 osd.29 up 1.00000 1.00000 32 0.90919 osd.32 up 1.00000 1.00000 35 0.90919 osd.35 up 1.00000 1.00000 -3 10.91034 host magna052 1 0.90919 osd.1 up 1.00000 1.00000 5 0.90919 osd.5 up 1.00000 1.00000 8 0.90919 osd.8 up 1.00000 1.00000 10 0.90919 osd.10 up 1.00000 1.00000 13 0.90919 osd.13 up 1.00000 1.00000 17 0.90919 osd.17 up 1.00000 1.00000 19 0.90919 osd.19 up 1.00000 1.00000 23 0.90919 osd.23 up 1.00000 1.00000 25 0.90919 osd.25 up 1.00000 1.00000 28 0.90919 osd.28 up 1.00000 1.00000 31 0.90919 osd.31 up 1.00000 1.00000 33 0.90919 osd.33 up 1.00000 1.00000 -4 10.91034 host magna058 2 0.90919 osd.2 up 1.00000 1.00000 4 0.90919 osd.4 up 1.00000 1.00000 7 0.90919 osd.7 up 1.00000 1.00000 11 0.90919 osd.11 up 1.00000 1.00000 14 0.90919 osd.14 up 1.00000 1.00000 15 0.90919 osd.15 up 1.00000 1.00000 18 0.90919 osd.18 up 1.00000 1.00000 21 0.90919 osd.21 up 1.00000 1.00000 24 0.90919 osd.24 up 1.00000 1.00000 27 0.90919 osd.27 up 1.00000 1.00000 30 0.90919 osd.30 up 1.00000 1.00000 34 0.90919 osd.34 up 1.00000 1.00000
This is not something we'll propose that customers do.
I believe this is a ceph-ansible bug. See https://github.com/ceph/ceph-ansible/issues/741 for more info.
*** Bug 1332234 has been marked as a duplicate of this bug. ***
This was fixed with commit ed2b7757d43ba7af96d97a9f345aba38e4410bd2 that was released a while ago and should be in the latest puddle.
Alfredo, Could you move this to ON_QA and not closed? Thanks, Tejas
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHEA-2016:1754