Bug 1614430
Summary: | vdsm-gluster stuck on operations forcing vdsm to stop working | ||
---|---|---|---|
Product: | Red Hat Enterprise Virtualization Manager | Reporter: | Sahina Bose <sabose> |
Component: | vdsm | Assignee: | Kaustav Majumder <kmajumde> |
Status: | CLOSED ERRATA | QA Contact: | SATHEESARAN <sasundar> |
Severity: | high | Docs Contact: | Rolfe Dlugy-Hegwer <rdlugyhe> |
Priority: | high | ||
Version: | unspecified | CC: | adevolder, bugs, dfediuck, dlavu, fdelorey, godas, info, kmajumde, lsurette, lsvaty, mkalinin, pasik, pdhange, rbarry, rcyriac, rdlugyhe, sabose, sasundar, srevivo, ycui |
Target Milestone: | ovirt-4.3.3 | ||
Target Release: | 4.3.0 | ||
Hardware: | Unspecified | ||
OS: | Unspecified | ||
Whiteboard: | |||
Fixed In Version: | vdsm-4.30.6 | Doc Type: | Bug Fix |
Doc Text: |
Vdsm-gluster tries to run heal operations on all volumes. Previously, if the gluster commands got stuck, VDSM started waiting indefinitely for them, exhausting threads, until it timed-out. Then it stopped communicating with the Manager and went offline. The current release adds a timeout to the gluster heal info command so the command terminates within a set timeout and threads do not become exhausted. On timeout, the system issues a GlusterCommandTimeoutException, which causes the command to exit and notifies the Manager. As a result, VDSM threads are not stuck, and VDSM does not go offline.
|
Story Points: | --- |
Clone Of: | 1609792 | Environment: | |
Last Closed: | 2019-05-08 12:36:02 UTC | Type: | Bug |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | Gluster | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: | |||
Bug Depends On: | |||
Bug Blocks: | 1711830 |
Description
Sahina Bose
2018-08-09 14:49:50 UTC
Cloned this bug to vdsm to investigate the issue mentioned by Denis about stuck gluster commands exhausting vdsm threads Moving to 4.3.0 as it's not a blocker This bug has not been marked as blocker for oVirt 4.3.0. Since we are releasing it tomorrow, January 29th, this bug has been re-targeted to 4.3.1. The patches are merged in master and available in 4.3.0 Kaustav, please provide doc_text Tested with RHV 4.3.3-3 And with the mix of distribute volume and replicate volume managed by RHV Manager, there are no issues found Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2019:1077 *** Bug 1712654 has been marked as a duplicate of this bug. *** sync2jira sync2jira |