Note: This bug is displayed in read-only format because
the product is no longer active in Red Hat Bugzilla.
RHEL Engineering is moving the tracking of its product development work on RHEL 6 through RHEL 9 to Red Hat Jira (issues.redhat.com). If you're a Red Hat customer, please continue to file support cases via the Red Hat customer portal. If you're not, please head to the "RHEL project" in Red Hat Jira and file new tickets here. Individual Bugzilla bugs in the statuses "NEW", "ASSIGNED", and "POST" are being migrated throughout September 2023. Bugs of Red Hat partners with an assigned Engineering Partner Manager (EPM) are migrated in late September as per pre-agreed dates. Bugs against components "kernel", "kernel-rt", and "kpatch" are only migrated if still in "NEW" or "ASSIGNED". If you cannot log in to RH Jira, please consult article #7032570. That failing, please send an e-mail to the RH Jira admins at rh-issues@redhat.com to troubleshoot your issue as a user management inquiry. The email creates a ServiceNow ticket with Red Hat. Individual Bugzilla bugs that are migrated will be moved to status "CLOSED", resolution "MIGRATED", and set with "MigratedToJIRA" in "Keywords". The link to the successor Jira issue will be found under "Links", have a little "two-footprint" icon next to it, and direct you to the "RHEL project" in Red Hat Jira (issue links are of type "https://issues.redhat.com/browse/RHEL-XXXX", where "X" is a digit). This same link will be available in a blue banner at the top of the page informing you that that bug has been migrated.
DescriptionMatthew Mosesohn
2013-10-22 14:40:31 UTC
Description of problem:
libqb-0.16.0 was released accidentally, breaking corosync and pacemaker deployments
Version-Release number of selected component (if applicable):
libqb-0.16.0-1.el6_4.1.x86_64
pacemaker-1.1.8-7.el6.x86_64
corosync-1.4.1-15.el6_4.1.x86_64
Additional info:
Steps to reproduce are a bit tricky to share here, but I can post crm_report data. The fact is this may severely impact several deployments in production now if they accidentally yum update, preventing any changes to existing configuration.
Devels Fabio Di Nitto and David Vossel confirmed this release was accidental and will remain a problem until either this package is recalled from RHEL HA repository or pacemaker and corosync updates get released.
I was able to reproduce this.
This issue has to do with not properly removing the ipc server's client connections from mainloop in pacemaker. There was a race condition in pacemaker that caused mainloop dispatch a fd that libqb has already told us to remove.
Here are the related upstream patches that resolved this in pacemaker.
https://github.com/ClusterLabs/pacemaker/commit/0628021134835ea1c683d0d70cef2a4112e08404
Libqb's example server implementation received this same change.
https://github.com/ClusterLabs/libqb/commit/1e1397fb22c04e46197873b116c6798892a29ee3
I'm not entirely sure why this appeared stable with the old version of libqb. There were a couple of libqb reference leaks that were discovered between 0.14.4 and 0.16.0. My guess is that those leaks covered up the issue. When I fixed the leaks, the mainloop issue likely appeared causing me to fix that as well.
The end result here is that libqb 0.16.0 should only be used with pacemaker 1.1.10 or greater.
-- Vossel
Comment 5RHEL Program Management
2013-10-26 17:14:54 UTC
This request was not resolved in time for the current release.
Red Hat invites you to ask your support representative to
propose this request, if still desired, for consideration in
the next release of Red Hat Enterprise Linux.