Description of problem: Failed to deploy OCS3.9 on OCP3.11.z, lvm commands fail Version-Release number of selected component (if applicable): atomic-openshift-3.11.69-1.git.0.7478b86.el7.x86_64 heketi-client-7.0.0-15.el7rhgs.x86_64 rhgs3/rhgs-server-rhel7:v3.9 How reproducible: Twice Steps to Reproduce: CI for OCP3.11.z+OCS3.9 failed twice Fails at the step: TASK [openshift_storage_glusterfs : Create heketi DB volume] fatal: [apu1-v311z-ocs-v39-master-0]: FAILED! => { "changed": true, "cmd": [ "oc", "--config=/tmp/openshift-glusterfs-ansible-exsTxk/admin.kubeconfig", "rsh", "--namespace=storage", "deploy-heketi-storage-1-5npz8", "heketi-cli", "-s", "http://localhost:8080", "--user", "admin", "--secret", "admin", "setup-openshift-heketi-storage", "--image", "rhgs3/rhgs-volmanager-rhel7:v3.9", "--listfile", "/tmp/heketi-storage.json" ], "delta": "0:01:20.772013", "end": "2019-01-22 10:51:24.019110", "invocation": { "module_args": { "_raw_params": "oc --config=/tmp/openshift-glusterfs-ansible-exsTxk/admin.kubeconfig rsh --namespace=storage deploy-heketi-storage-1-5npz8 heketi-cli -s http://localhost:8080 --user admin --secret 'admin' setup-openshift-heketi-storage --image rhgs3/rhgs-volmanager-rhel7:v3.9 --listfile /tmp/heketi-storage.json", "_uses_shell": false, "argv": null, "chdir": null, "creates": null, "executable": null, "removes": null, "stdin": null, "warn": true } }, "msg": "non-zero return code", "rc": 255, "start": "2019-01-22 10:50:03.247097", "stderr": "Error: Unable to execute command on glusterfs-storage-mjcqr: /dev/vg_50add4fedc4344a89b21b1d9dfd2981d/lvol0: not found: device not cleared\n Aborting. Failed to wipe start of new LV.\ncommand terminated with exit code 255", "stderr_lines": [ "Error: Unable to execute command on glusterfs-storage-mjcqr: /dev/vg_50add4fedc4344a89b21b1d9dfd2981d/lvol0: not found: device not cleared", " Aborting. Failed to wipe start of new LV.", "command terminated with exit code 255" ], "stdout": "", "stdout_lines": [] } Expected results: Install succeeds.
The ocs anisble log file and the output of the central CI jenkins job link can be found here - http://rhsqe-repo.lab.eng.blr.redhat.com/cns/bugs/BZ-1668316/
Additional info: openshift-ansible-3.11.70-1 sosreports from all nodes available here - http://rhsqe-repo.lab.eng.blr.redhat.com/cns/bugs/BZ-1668316/
Is there an openshift-ansible bug for this? https://github.com/openshift/openshift-ansible/pull/11068 should address the problem.
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2019:0670