Bug 1031687

Summary: Dist-geo-rep : while doing first xsync crawl, disconnection with slave causes the geo-rep to re-crawl the whole file system and generate XSYNC-CHANGELOGS again.
Product: Red Hat Gluster Storage Reporter: Vijaykumar Koppad <vkoppad>
Component: geo-replicationAssignee: Kotresh HR <khiremat>
Status: CLOSED ERRATA QA Contact: Vijaykumar Koppad <vkoppad>
Severity: high Docs Contact:
Priority: high    
Version: 2.1CC: aavati, asriram, bbandari, bhubbard, csaba, dblack, fharshav, khiremat, nsathyan, psriniva, vagarwal, vshankar
Target Milestone: ---Keywords: ZStream
Target Release: RHGS 2.1.2Flags: khiremat: needinfo+
Hardware: x86_64   
OS: Linux   
Whiteboard:
Fixed In Version: glusterfs-3.4.0.50rhs Doc Type: Bug Fix
Doc Text:
Previously, when the first xsync crawl was in progress, disconnection with the slave volume caused Geo-replication to re-crawl the entire file system and generate XSYNC-CHANGELOGS. With this update, xsync skips the directories which are already synced to the slave volume.
Story Points: ---
Clone Of: Environment:
Last Closed: 2014-02-25 08:04:24 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:

Description Vijaykumar Koppad 2013-11-18 14:42:18 UTC
Description of problem: while doing first xsync crawl, disconnection with slave causes the geo-rep to re-crawl the whole file system and generate XSYNC-CHANGELOGS again.


Version-Release number of selected component (if applicable): glusterfs-3.4.0.44rhs

How reproducible:Happens everytime


Steps to Reproduce:
1.create a geo-rep relationship between master and slave
2.create some 20million files on master.
3.start a geo-rep session.

Actual results: some of the sessions experience disconnections with the slave and first xsync crawl does a full crawl again generated the XSYNC-CHANGELOG with the entries which has already been crawled . 


Expected results: It shouldn't do this recrawl, it should start from where it has left off 


Additional info:

Comment 4 Kotresh HR 2013-12-11 08:17:56 UTC
*** Bug 1034238 has been marked as a duplicate of this bug. ***

Comment 8 Kotresh HR 2014-01-02 10:30:43 UTC
Added Doc Text

Comment 9 Vijaykumar Koppad 2014-01-03 13:36:11 UTC
Verified on the build glusterfs-3.4.0.53rhs-1.

Steps used to verify.

1.create a geo-rep relationship between master and slave.
2.create 500K files on master.
3.start a geo-rep session between master and slave.
4.run following on one of the master active nodes.
  while : ; do ps ax | grep "ssh " | awk '{print $1}' | xargs kill ; sleep 100 ; ps ax | grep "ssh " | awk '{print $1}' | xargs kill ; sleep 1000; done

5. Wait for it to complete the syncing.

Comment 10 Pavithra 2014-01-08 10:03:49 UTC
Can you please verify the doc text for technical accuracy?

Comment 11 Kotresh HR 2014-01-08 10:58:17 UTC
Doc text looks fine.

Comment 13 errata-xmlrpc 2014-02-25 08:04:24 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory, and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

http://rhn.redhat.com/errata/RHEA-2014-0208.html