Bug 1751142 - [downstream clone - 4.3.6] host activation causes RHHI nodes to lose the quorum
Summary: [downstream clone - 4.3.6] host activation causes RHHI nodes to lose the quorum
Keywords:
Status: CLOSED ERRATA
Alias: None
Product: Red Hat Enterprise Virtualization Manager
Classification: Red Hat
Component: ovirt-engine
Version: 4.3.4
Hardware: Unspecified
OS: Unspecified
urgent
urgent
Target Milestone: ovirt-4.3.6
: 4.3.6
Assignee: Gobinda Das
QA Contact: SATHEESARAN
URL:
Whiteboard:
Depends On: 1741102
Blocks: 1751717
TreeView+ depends on / blocked
 
Reported: 2019-09-11 09:24 UTC by RHV bug bot
Modified: 2023-03-24 15:25 UTC (History)
13 users (show)

Fixed In Version: ovirt-engine-4.3.6.6
Doc Type: Bug Fix
Doc Text:
During host activation, the engine did not check the glusterd status. The gluster service was restarted even though the glusterd was not stopped during maintenance. In this release, the quorum is not lost upon host activation.
Clone Of: 1741102
Environment:
Last Closed: 2019-10-10 15:36:58 UTC
oVirt Team: Gluster
Target Upstream Version:
Embargoed:


Attachments (Terms of Use)


Links
System ID Private Priority Status Summary Last Updated
Red Hat Product Errata RHEA-2019:3010 0 None None None 2019-10-10 15:37:11 UTC
oVirt gerrit 102656 0 'None' MERGED engine: Start gluster service if not running during host activation 2021-01-04 10:20:55 UTC
oVirt gerrit 103177 0 'None' MERGED engine: Start gluster service if not running during host activation 2021-01-04 10:20:55 UTC

Description RHV bug bot 2019-09-11 09:24:22 UTC
+++ This bug is a downstream clone. The original bug is: +++
+++   bug 1741102 +++
======================================================================

Description of problem:
When a host providing gluster services for RHHI is activated, the RHV-M initiates gluster services restart. This causes gluster volumes to lose quorum and stop working.


Version-Release number of selected component (if applicable):
rhvm 4.3.3

How reproducible:
100%

Steps to Reproduce:
1. Put a gluster node into maintenance
2. Set debug level
    vdsm-client Host setLogLevel level=DEBUG name=jsonrpc
#. Activate the node again

Actual results:
2019-08-14 09:30:00,866+0000 DEBUG (jsonrpc/1) [jsonrpc.JsonRpcServer] Calling 'GlusterService.action' in bridge with {u'action': u'restart', u'serviceNames': [u'glusterd']} (__init__:329)

Expected results:
No restart happens. Keep in mind that the node can be put into maintenance mode without stopping the gluster services.

(Originally by Roman Hodain)

Comment 4 RHV bug bot 2019-09-11 09:24:30 UTC
sync2jira

(Originally by Daniel Gur)

Comment 5 RHV bug bot 2019-09-11 09:24:31 UTC
sync2jira

(Originally by Daniel Gur)

Comment 11 SATHEESARAN 2019-09-25 16:48:32 UTC
Verified with RHV 4.3.6.6 with the following steps

1. Created a HC cluster
2. Move the HC node in to MAINTENANCE, without stopping gluster services
3. Note the PID of glusterd process and brick process.
4. Activate the host back

Observed that the gluster process PID remained the same post the host is activated

Comment 18 errata-xmlrpc 2019-10-10 15:36:58 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory, and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHEA-2019:3010


Note You need to log in before you can comment on or make changes to this bug.