Bug 1579719
Summary: | Geo-Replication failing to kick off geo-rep session daily, when the same volume is used for two different sessions and one gets destroyed. | ||
---|---|---|---|
Product: | [oVirt] ovirt-engine | Reporter: | Sahina Bose <sabose> |
Component: | BLL.Gluster | Assignee: | Sahina Bose <sabose> |
Status: | CLOSED CURRENTRELEASE | QA Contact: | SATHEESARAN <sasundar> |
Severity: | high | Docs Contact: | |
Priority: | high | ||
Version: | 4.2.3.2 | CC: | ascerra, avishwan, bugs, rhs-bugs, sasundar |
Target Milestone: | ovirt-4.2.5 | Flags: | rule-engine:
ovirt-4.2+
rule-engine: exception+ |
Target Release: | --- | ||
Hardware: | x86_64 | ||
OS: | Linux | ||
Whiteboard: | |||
Fixed In Version: | Doc Type: | Bug Fix | |
Doc Text: |
Cause: Lock was not released on a failed geo-replication start
Consequence: Subsequent geo-replication based DR sync fails
Fix: Ensure that commands are ended and locks released when there's a failure to start geo-replication.
Result: Multiple DR sync sessions can be setup to run even when there's failure in one.
|
Story Points: | --- |
Clone Of: | 1554487 | Environment: | |
Last Closed: | 2018-07-31 15:29:17 UTC | Type: | Bug |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | Gluster | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: | |||
Bug Depends On: | |||
Bug Blocks: | 1554487 |
Description
Sahina Bose
2018-05-18 07:59:31 UTC
Unable to verify this bug, as hit with another bug - BZ 1595140 where having multiple geo-rep session with gluster volume failed. This bug is blocked because of that issue. On these grounds, moving this bug out to 4.2.5 (In reply to SATHEESARAN from comment #1) > Unable to verify this bug, as hit with another bug - BZ 1595140 where having > multiple geo-rep session with gluster volume failed. > > This bug is blocked because of that issue. > > On these grounds, moving this bug out to 4.2.5 The Bug 1595140 can be resolved if you add all the gluster hosts to setup. So I think you can continue verification on this one? (In reply to Sahina Bose from comment #2) > (In reply to SATHEESARAN from comment #1) > > Unable to verify this bug, as hit with another bug - BZ 1595140 where having > > multiple geo-rep session with gluster volume failed. > > > > This bug is blocked because of that issue. > > > > On these grounds, moving this bug out to 4.2.5 > > The Bug 1595140 can be resolved if you add all the gluster hosts to setup. > So I think you can continue verification on this one? Thanks for that information. Yes, I can now proceed with this exceptions in place Tested with ovirt-4.2.5 and glusterfs-3.8.4-54.15 with the following step. 1. Complete RHHI deployment. Treat as the primary site PrimSite1. Select any volume. This case I chose 'data' volume. 2. Create 2 geo-rep session from this volume to 2 secondary site. SecSite1 & SecSite2 3. Create a remote sync for the storage domain. 4. Start geo-rep on one session. While the session is in progress stop the volume. Geo-rep will go faulty. With this faulty geo-rep session, schedule a remote data sync to SecSite1 & SecSite2. 5. Even when one of the geo-rep session is faulty, the other session worked as expected Before the remote sync: MASTER NODE MASTER VOL MASTER BRICK SLAVE USER SLAVE SLAVE NODE STATUS CRAWL STATUS LAST_SYNCED ---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- rhsqa-grafton1-nic2.lab.eng.blr.redhat.com data /gluster_bricks/data/data root ssh://10.70.45.29::sasvol1 N/A Faulty N/A N/A rhsqa-grafton2-nic2.lab.eng.blr.redhat.com data /gluster_bricks/data/data root ssh://10.70.45.29::sasvol1 N/A Faulty N/A N/A rhsqa-grafton3-nic2.lab.eng.blr.redhat.com data /gluster_bricks/data/data root ssh://10.70.45.29::sasvol1 N/A Faulty N/A N/A rhsqa-grafton1-nic2.lab.eng.blr.redhat.com data /gluster_bricks/data/data root ssh://10.70.45.32::data N/A Stopped N/A N/A rhsqa-grafton2-nic2.lab.eng.blr.redhat.com data /gluster_bricks/data/data root ssh://10.70.45.32::data N/A Stopped N/A N/A rhsqa-grafton3-nic2.lab.eng.blr.redhat.com data /gluster_bricks/data/data root ssh://10.70.45.32::data N/A Stopped N/A N/A Remote sync worked with faulty session: MASTER NODE MASTER VOL MASTER BRICK SLAVE USER SLAVE SLAVE NODE STATUS CRAWL STATUS LAST_SYNCED -------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- rhsqa-grafton1-nic2.lab.eng.blr.redhat.com data /gluster_bricks/data/data root ssh://10.70.45.29::sasvol1 N/A Faulty N/A N/A rhsqa-grafton2-nic2.lab.eng.blr.redhat.com data /gluster_bricks/data/data root ssh://10.70.45.29::sasvol1 N/A Faulty N/A N/A rhsqa-grafton3-nic2.lab.eng.blr.redhat.com data /gluster_bricks/data/data root ssh://10.70.45.29::sasvol1 N/A Faulty N/A N/A rhsqa-grafton1-nic2.lab.eng.blr.redhat.com data /gluster_bricks/data/data root ssh://10.70.45.32::data 10.70.45.33 Passive N/A N/A rhsqa-grafton2-nic2.lab.eng.blr.redhat.com data /gluster_bricks/data/data root ssh://10.70.45.32::data 10.70.45.32 Passive N/A N/A rhsqa-grafton3-nic2.lab.eng.blr.redhat.com data /gluster_bricks/data/data root ssh://10.70.45.32::data 10.70.45.34 Active History Crawl 2018-07-20 21:38:10 This bugzilla is included in oVirt 4.2.5 release, published on July 30th 2018. Since the problem described in this bug report should be resolved in oVirt 4.2.5 release, it has been closed with a resolution of CURRENT RELEASE. If the solution does not work for you, please open a new bug report. |