Note: This bug is displayed in read-only format because the product is no longer active in Red Hat Bugzilla.

Bug 1795022

Summary: [OSP16] Deployment with TLS fails, 'Error: Could not find group qemu'
Product: Red Hat OpenStack Reporter: Roman Safronov <rsafrono>
Component: openstack-tripleo-heat-templatesAssignee: Piotr Kopec <pkopec>
Status: CLOSED DUPLICATE QA Contact: Alexander Chuzhoy <sasha>
Severity: urgent Docs Contact:
Priority: urgent    
Version: 16.0 (Train)CC: gcharot, ggrasza, mburns, michele, mschuppe, nweinber, pkopec, ramishra, ratailor, rmascena, sclewis, shrjoshi, slinaber, stephenfin, yrabl
Target Milestone: gaKeywords: Triaged, ZStream
Target Release: 16.0 (Train on RHEL 8.1)   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: Doc Type: If docs needed, set a value
Doc Text:
Story Points: ---
Clone Of: Environment:
Last Closed: 2020-01-29 11:37:56 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:

Description Roman Safronov 2020-01-26 13:44:51 UTC
Description of problem:
Deployment of OSP16 with TLS fails.
Ansible failed on all controllers

tail -n 10 ansible.log
2020-01-26 08:53:13,746 p=756 u=mistral |  NO MORE HOSTS LEFT *************************************************************
2020-01-26 08:53:13,748 p=756 u=mistral |  PLAY RECAP *********************************************************************
2020-01-26 08:53:13,748 p=756 u=mistral |  compute-0                  : ok=170  changed=95   unreachable=0    failed=0    skipped=90   rescued=0    ignored=1   
2020-01-26 08:53:13,748 p=756 u=mistral |  compute-1                  : ok=165  changed=93   unreachable=0    failed=0    skipped=90   rescued=0    ignored=1   
2020-01-26 08:53:13,748 p=756 u=mistral |  controller-0               : ok=223  changed=137  unreachable=0    failed=1    skipped=84   rescued=0    ignored=0   
2020-01-26 08:53:13,748 p=756 u=mistral |  controller-1               : ok=208  changed=137  unreachable=0    failed=1    skipped=84   rescued=0    ignored=0   
2020-01-26 08:53:13,748 p=756 u=mistral |  controller-2               : ok=208  changed=137  unreachable=0    failed=1    skipped=84   rescued=0    ignored=0   
2020-01-26 08:53:13,749 p=756 u=mistral |  undercloud                 : ok=11   changed=7    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   
2020-01-26 08:53:13,819 p=756 u=mistral |  Sunday 26 January 2020  08:53:13 +0000 (0:01:42.458)       0:10:10.548 ******** 
2020-01-26 08:53:13,819 p=756 u=mistral |  ===============================================================================


cat ansible-errors.json | sed  's/Jan/\nJan/g' | grep -i error
Jan 26 08:51:36 puppet-user: Debug: Facter: Error: unable to get cib\n<13>
Jan 26 08:51:45 puppet-user: Debug: /Stage[main]/Pacemaker::Corosync/Exec[wait-for-settle]/unless: Error: cluster is not currently running on this node\n<13>
Jan 26 08:52:41 puppet-user: Error: Could not find group qemu\n<13>
Jan 26 08:52:41 puppet-user: Error: /Stage[main]/Tripleo::Profile::Base::Certmonger_user/Tripleo::Certmonger::Libvirt_vnc[libvirt-vnc-client-cert]/File[/etc/pki/libvirt-vnc/client-key.pem]/group: change from 'root' to 'qemu' failed: Could not find group qemu\n<13>
Jan 26 08:51:36 puppet-user: Debug: Facter: Error: unable to get cib", "<13>
Jan 26 08:51:45 puppet-user: Debug: /Stage[main]/Pacemaker::Corosync/Exec[wait-for-settle]/unless: Error: cluster is not currently running on this node", "<13>
Jan 26 08:52:41 puppet-user: Error: Could not find group qemu", "<13>
Jan 26 08:52:41 puppet-user: Error: /Stage[main]/Tripleo::Profile::Base::Certmonger_user/Tripleo::Certmonger::Libvirt_vnc[libvirt-vnc-client-cert]/File[/etc/pki/libvirt-vnc/client-key.pem]/group: change from 'root' to 'qemu' failed: Could not find group qemu", "<13>
Jan 26 08:51:36 puppet-user: Debug: Facter: Error: unable to get cib\n<13>
Jan 26 08:51:38 puppet-user: error: Could not connect to cluster (is it running?)\n<13>
Jan 26 08:52:42 puppet-user: Error: Could not find group qemu\n<13>
Jan 26 08:52:42 puppet-user: Error: /Stage[main]/Tripleo::Profile::Base::Certmonger_user/Tripleo::Certmonger::Libvirt_vnc[libvirt-vnc-client-cert]/File[/etc/pki/libvirt-vnc/client-key.pem]/group: change from 'root' to 'qemu' failed: Could not find group qemu\n<13>
Jan 26 08:51:36 puppet-user: Debug: Facter: Error: unable to get cib", "<13>
Jan 26 08:51:38 puppet-user: error: Could not connect to cluster (is it running?)", "<13>
Jan 26 08:52:42 puppet-user: Error: Could not find group qemu", "<13>
Jan 26 08:52:42 puppet-user: Error: /Stage[main]/Tripleo::Profile::Base::Certmonger_user/Tripleo::Certmonger::Libvirt_vnc[libvirt-vnc-client-cert]/File[/etc/pki/libvirt-vnc/client-key.pem]/group: change from 'root' to 'qemu' failed: Could not find group qemu", "<13>
Jan 26 08:51:36 puppet-user: Debug: Facter: Error: unable to get cib\n<13>
Jan 26 08:51:45 puppet-user: Debug: /Stage[main]/Pacemaker::Corosync/Exec[wait-for-settle]/unless: Error: cluster is not currently running on this node\n<13>
Jan 26 08:52:52 puppet-user: Error: Could not find group qemu\n<13>
Jan 26 08:52:52 puppet-user: Error: /Stage[main]/Tripleo::Profile::Base::Certmonger_user/Tripleo::Certmonger::Libvirt_vnc[libvirt-vnc-client-cert]/File[/etc/pki/libvirt-vnc/client-key.pem]/group: change from 'root' to 'qemu' failed: Could not find group qemu\n<13>
Jan 26 08:51:36 puppet-user: Debug: Facter: Error: unable to get cib", "<13>
Jan 26 08:51:45 puppet-user: Debug: /Stage[main]/Pacemaker::Corosync/Exec[wait-for-settle]/unless: Error: cluster is not currently running on this node", "<13>
Jan 26 08:52:52 puppet-user: Error: Could not find group qemu", "<13>
Jan 26 08:52:52 puppet-user: Error: /Stage[main]/Tripleo::Profile::Base::Certmonger_user/Tripleo::Certmonger::Libvirt_vnc[libvirt-vnc-client-cert]/File[/etc/pki/libvirt-vnc/client-key.pem]/group: change from 'root' to 'qemu' failed: Could not find group qemu", "<13>


Feel free to change the component to a more suitable one.


Version-Release number of selected component (if applicable):
puddle RHOS_TRUNK-16.0-RHEL-8-20200124.n.1

How reproducible:
100%, failed 

Steps to Reproduce:
1.Run this CI job: https://rhos-qe-jenkins.rhev-ci-vms.eng.rdu2.redhat.com/view/DFG/view/network/view/networking-ovn/job/DFG-network-networking-ovn-16_director-rhel-virthost-3cont_2comp_1ipa-ipv4-geneve-tls/

Actual results:
Job fails on overcloud deploy stage

Expected results:
Overcloud deployed successfully

Additional info:

Comment 15 Stephen Finucane 2020-01-29 11:37:56 UTC
The revert has resolved this issue and we're going to attempt this again in z1. I'm going to close this out as a dupe of bug 1775006.

*** This bug has been marked as a duplicate of bug 1775006 ***