Description of problem:
=======================
While upgrading from glusterfs-3.8.4-54.el7rhgs.x86_64.rpm (RHEL 7.4) to glusterfs-3.12.2-7.el7rhgs.x86_64 (RHEL 7.5) encountered a FAULTY session where there was only one worker ACTIVE.
[root@dhcp42-53 master]# gluster volume geo-replication master 10.70.42.164::slave status
MASTER NODE MASTER VOL MASTER BRICK SLAVE USER SLAVE SLAVE NODE STATUS CRAWL STATUS LAST_SYNCED
------------------------------------------------------------------------------------------------------------------------------------------
10.70.42.53 master /rhs/brick1/b1 root 10.70.42.164::slave N/A Faulty N/A N/A
10.70.42.53 master /rhs/brick2/b4 root 10.70.42.164::slave N/A Faulty N/A N/A
10.70.42.138 master /rhs/brick1/b3 root 10.70.42.164::slave 10.70.42.164 Active History Crawl N/A
10.70.42.138 master /rhs/brick2/b6 root 10.70.42.164::slave N/A Faulty N/A N/A
10.70.42.160 master /rhs/brick1/b2 root 10.70.42.164::slave N/A Faulty N/A N/A
10.70.42.160 master /rhs/brick2/b5 root 10.70.42.164::slave N/A Faulty N/A N/A
Traceback in geo-rep logs:
--------------------------------
Traceback (most recent call last):
File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 210, in main
main_i()
File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 802, in main_i
local.service_loop(*[r for r in [remote] if r])
File "/usr/libexec/glusterfs/python/syncdaemon/resource.py", line 1676, in service_loop
g3.crawlwrap(oneshot=True)
File "/usr/libexec/glusterfs/python/syncdaemon/master.py", line 597, in crawlwrap
self.crawl()
File "/usr/libexec/glusterfs/python/syncdaemon/master.py", line 1470, in crawl
self.changelogs_batch_process(changes)
File "/usr/libexec/glusterfs/python/syncdaemon/master.py", line 1370, in changelogs_batch_process
self.process(batch)
File "/usr/libexec/glusterfs/python/syncdaemon/master.py", line 1204, in process
self.process_change(change, done, retry)
File "/usr/libexec/glusterfs/python/syncdaemon/master.py", line 1123, in process_change
entry_stime_to_update[0])
File "/usr/libexec/glusterfs/python/syncdaemon/gsyncdstatus.py", line 200, in set_field
return self._update(merger)
File "/usr/libexec/glusterfs/python/syncdaemon/gsyncdstatus.py", line 161, in _update
data = mergerfunc(data)
File "/usr/libexec/glusterfs/python/syncdaemon/gsyncdstatus.py", line 194, in merger
if data[key] == value:
KeyError: 'last_synced_entry'
Version-Release number of selected component (if applicable):
=============================================================
[root@dhcp42-53 glusterfs]# rpm -qa | grep gluster
libvirt-daemon-driver-storage-gluster-3.9.0-14.el7_5.2.x86_64
glusterfs-3.12.2-7.el7rhgs.x86_64
glusterfs-fuse-3.12.2-7.el7rhgs.x86_64
glusterfs-geo-replication-3.12.2-7.el7rhgs.x86_64
gluster-nagios-common-0.2.4-1.el7rhgs.noarch
glusterfs-libs-3.12.2-7.el7rhgs.x86_64
glusterfs-client-xlators-3.12.2-7.el7rhgs.x86_64
glusterfs-server-3.12.2-7.el7rhgs.x86_64
vdsm-gluster-4.19.43-2.3.el7rhgs.noarch
glusterfs-api-3.12.2-7.el7rhgs.x86_64
glusterfs-rdma-3.12.2-7.el7rhgs.x86_64
glusterfs-events-3.12.2-7.el7rhgs.x86_64
gluster-nagios-addons-0.2.10-2.el7rhgs.x86_64
glusterfs-cli-3.12.2-7.el7rhgs.x86_64
python2-gluster-3.12.2-7.el7rhgs.x86_64
How reproducible:
=================
1/1
Actual results:
===============
Session is FAULTY.
Expected results:
=================
Session should not be FAULTY.
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.
For information on the advisory, and where to find the updated
files, follow the link below.
If the solution does not work for you, open a new bug report.
https://access.redhat.com/errata/RHSA-2018:2607