Description of problem: Environment 4 CentOS 7.2 servers, 2 in UK, and one each in New York and Sydney A replica 2 Glusterfs src volume was created using 1 brick from each of 2 Uk servers, this worked great. Geo-replication was configured from this source volume to a destination volume in NY and another in Sydney, this also worked great. Output showed first src node as Active and second src node as Passive The first src nodes was shutdown and after a short time the second src node became Active, replication continued. The first src node was started and added its brick back into the src vol, but the geo-replication status for the first node became Faulty, the second node was Passive. Upon shutting the first node down the second node geo-replication status became Active, but when started the status became faulty. I searched the error and found that it may be related to the index being rotated and thus the first node had lost track, but no instructions on how to fix this. I had to delete the geo-replication and the destination node, then recreate both to fix the issue. Version-Release number of selected component (if applicable): 3.8.5 How reproducible: twice so far Steps to Reproduce: 1.create above environment 2.shutdown one src node 3.start the src node Actual results: geo-replication status goes faulty Expected results: src node goes Active and replication continues Additional info:
Please upload the Geo-rep logs of Faulty node from /var/log/glusterfs/geo-replication directory
Created attachment 1286989 [details] Logs from georeplication volume This file contains the logs from the georeplication directory for this volume
I know the resolution make take time, is there a workaround if this occurs again?
This bug is getting closed because the 3.8 version is marked End-Of-Life. There will be no further updates to this version. Please open a new bug against a version that still receives bugfixes if you are still facing this issue in a more current release.