Bug 989192
Summary: | Dist-geo-rep : geo-rep failover-failback is broken : special-sync-mode blind results in faulty state. | |||
---|---|---|---|---|
Product: | [Red Hat Storage] Red Hat Gluster Storage | Reporter: | Vijaykumar Koppad <vkoppad> | |
Component: | geo-replication | Assignee: | Venky Shankar <vshankar> | |
Status: | CLOSED ERRATA | QA Contact: | M S Vishwanath Bhat <vbhat> | |
Severity: | high | Docs Contact: | ||
Priority: | high | |||
Version: | 2.1 | CC: | aavati, amarts, bbandari, csaba, mzywusko, rhs-bugs, sdharane, shaines, vbhat | |
Target Milestone: | --- | Keywords: | TestBlocker | |
Target Release: | --- | |||
Hardware: | x86_64 | |||
OS: | Linux | |||
Whiteboard: | ||||
Fixed In Version: | glusterfs-3.4.0.18rhs-1 | Doc Type: | Bug Fix | |
Doc Text: | Story Points: | --- | ||
Clone Of: | ||||
: | 994462 (view as bug list) | Environment: | ||
Last Closed: | 2013-09-23 22:29:51 UTC | Type: | Bug | |
Regression: | --- | Mount Type: | --- | |
Documentation: | --- | CRM: | ||
Verified Versions: | Category: | --- | ||
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | ||
Cloudforms Team: | --- | Target Upstream Version: | ||
Embargoed: | ||||
Bug Depends On: | ||||
Bug Blocks: | 957769, 994462 |
Description
Vijaykumar Koppad
2013-07-28 07:59:01 UTC
This needs to be changed with newer geo-replication. I am not sure if this is valid anymore. Venky, would be good to look into this. Fixed now. Tested in Version: [root@mustang ~]# rpm -q glusterfs glusterfs-3.4.0.20rhs-2.el6rhs.x86_64 One point to be noted is marker (geo-replication.indexing) should be turned on from the slave before the application starts writing to the slave. If it's not turned on in the slave volume during I/O from the slave, special_sync_mode will *not* work. The steps I did for verification is 1. Create some files in master and let all the files be synced to slave. 2. Shutdown all the master nodes. 3. Turn on the indexing on the slave before any IO on slave. gluster v set slave geo-replication.indexing on 4. Now create some files from the slave. (*deletes/renames/hardlinks will have isses*) The above part where the application is moved from master to slave is failover part. 5. Now bring up all the master nodes. The status of geo-rep will be defunct. Then use stop force and delete to stop and delete the geo-rep session from master to slave. 6. Create geo-rep session from slave to master. Also make sure to setup the ssh pem from slave to master. 7. Set the special_sync_mode to recover gluster v geo slave master_node::master config special_sync_mode recover 8. Start the geo-rep session. 9. Now use status detail to monitor the status of file syncing. Make sure during this time, there should be no IO going on from slave. 10. Once all the files are synced, stop and delete the session from slave to master. 11. Re establish the session from master to slave. Move the application here back to master. Moving this bug to VERIFIED. Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. http://rhn.redhat.com/errata/RHBA-2013-1262.html |