Created attachment 1644872 [details] environment files and logs Description of problem: An overcloud deployment with Ceph with customized crush maps, pools and an additional pool for Cinder as a second back end failed with the errors: "failed: [ceph-5 -> 192.168.24.46] (item=[{'application': 'rbd', 'name': 'tier2', 'pg_num': 64, 'rule_name': 'fast'}, {'msg': 'non-zero return code', 'cmd': ['podman', 'exec', 'ceph-mon-controller-0', 'ceph', '--cluster', 'ceph', 'osd', 'pool', 'get', 'tier2', 'size'], 'stdout': '', 'stderr': \"Error ENOENT : unrecognized pool 'tier2'\\nError: exit status 2\", 'rc': 2, 'start': '2019-12-12 17:36:14.986390', 'end': '2019-12-12 17:36:15.728381', 'delta': '0:00:00.7 41991', 'changed': False, 'failed': False, 'invocation': {'module_args': {'_raw_params': 'podman exec ceph-mon-controller-0 ceph --cluster ceph osd pool get t ier2 size\\n', 'warn': True, '_uses_shell': False, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'crea tes': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"Error ENOENT: unrecognized pool 'tier2'\", 'Error: exit status 2'], 'fail ed_when_result': False, 'item': {'application': 'rbd', 'name': 'tier2', 'pg_num': 64, 'rule_name': 'fast'}, 'ansible_loop_var': 'item'}]) => changed=false ", Version-Release number of selected component (if applicable): ceph-ansible-4.0.5-1.el8cp.noarch How reproducible: 100% (happened 3 times) Steps to Reproduce: 1. Deploy an overcloud with customized pools and crush map Actual results: the deployment fails during the Ceph cluster deployment Expected results: The ceph cluster is deployed with the customized parameters Additional info:
Created attachment 1651066 [details] overcloud deployment log file
Tried to add the flag and it didn't work, I have attached the logs to the bug
host_vars are empty but nodes_uuid_data.json isn't , we need to make sure the uuids match
The fix is not in the latest compose
Moving to VERIFIED and tracking the new ceph-ansible issue via BZ#1791283
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHEA-2020:0283
The needinfo request[s] on this closed bug have been removed as they have been unresolved for 1000 days