Bug 1425690 - Worker restarts on log-rsync-performance config update
Summary: Worker restarts on log-rsync-performance config update
Keywords:
Status: CLOSED ERRATA
Alias: None
Product: Red Hat Gluster Storage
Classification: Red Hat Storage
Component: geo-replication
Version: rhgs-3.2
Hardware: Unspecified
OS: Unspecified
unspecified
unspecified
Target Milestone: ---
: RHGS 3.3.0
Assignee: Kotresh HR
QA Contact: Rochelle
URL:
Whiteboard:
Depends On: 1393678
Blocks: 1417147
TreeView+ depends on / blocked
 
Reported: 2017-02-22 06:16 UTC by Kotresh HR
Modified: 2017-09-21 04:57 UTC (History)
8 users (show)

Fixed In Version: glusterfs-3.8.4-19
Doc Type: If docs needed, set a value
Doc Text:
Clone Of: 1393678
Environment:
Last Closed: 2017-09-21 04:33:25 UTC
Embargoed:


Attachments (Terms of Use)


Links
System ID Private Priority Status Summary Last Updated
Red Hat Product Errata RHBA-2017:2774 0 normal SHIPPED_LIVE glusterfs bug fix and enhancement update 2017-09-21 08:16:29 UTC

Description Kotresh HR 2017-02-22 06:16:17 UTC
+++ This bug was initially created as a clone of Bug #1393678 +++

Description of problem:
If log-rsync-performance config is set using following command, workers restarts and causes reprocessing Changelogs which are processed before the config change.

gluster volume geo-replication <MASTER> <SLAVEHOST>::<SLAVEVOL> config log-rsync-performance true

Comment 6 Rochelle 2017-07-17 11:11:42 UTC
Verified this bug on the build: glusterfs-geo-replication-3.8.4-32.el7rhgs.x86_64

Initial Log:
------------

[2017-07-17 10:53:28.284065] I [gsyncdstatus(/rhs/brick1/b1):244:set_worker_crawl_status] GeorepStatus: Crawl Status: Changelog Crawl
[2017-07-17 10:53:31.302864] I [master(/rhs/brick1/b1):1125:crawl] _GMaster: slave's time: (1500288808, 0)
[2017-07-17 11:02:31.789875] I [master(/rhs/brick1/b1):1125:crawl] _GMaster: slave's time: (1500288808, 0)

Worker Restarts once the options like use-tarssh is set to true:
----------------------------------------------------------------

[2017-07-17 11:02:31.789875] I [master(/rhs/brick1/b1):1125:crawl] _GMaster: slave's time: (1500288808, 0)
[2017-07-17 11:05:32.703613] I [gsyncd(conf):688:main_i] <top>: Config Set: use-tarssh = true
[2017-07-17 11:05:32.970976] I [syncdutils(/rhs/brick1/b1):237:finalize] <top>: exiting.
[2017-07-17 11:05:32.975008] I [repce(/rhs/brick1/b1):92:service_loop] RepceServer: terminating on reaching EOF.
[2017-07-17 11:05:32.975292] I [syncdutils(/rhs/brick1/b1):237:finalize] <top>: exiting.
[2017-07-17 11:05:32.989492] I [gsyncdstatus(monitor):240:set_worker_status] GeorepStatus: Worker Status: Faulty
[2017-07-17 11:05:34.252108] I [gsyncd(conf):688:main_i] <top>: Config Set: session-owner = 96a8c224-9a1a-4069-8a3c-bfd3d132c3ef
[2017-07-17 11:05:34.818747] I [gsyncdstatus(monitor):240:set_worker_status] GeorepStatus: Worker Status: Initializing...
[2017-07-17 11:05:34.998934] I [monitor(monitor):275:monitor] Monitor: starting gsyncd worker(/rhs/brick1/b1). Slave node: ssh://root.37.105:gluster://localhost:slave
[2017-07-17 11:05:35.133186] I [changelogagent(/rhs/brick1/b1):73:__init__] ChangelogAgent: Agent listining...
[2017-07-17 11:05:35.133666] I [resource(/rhs/brick1/b1):1676:connect_remote] SSH: Initializing SSH connection between master and slave...
[2017-07-17 11:05:40.981836] I [resource(/rhs/brick1/b1):1683:connect_remote] SSH: SSH connection between master and slave established. Time taken: 5.8478 secs
[2017-07-17 11:05:40.982347] I [resource(/rhs/brick1/b1):1416:connect] GLUSTER: Mounting gluster volume locally...
[2017-07-17 11:05:46.97822] I [resource(/rhs/brick1/b1):1429:connect] GLUSTER: Mounted gluster volume. Time taken: 5.1151 secs
[2017-07-17 11:05:46.98491] I [gsyncd(/rhs/brick1/b1):778:main_i] <top>: Closing feedback fd, waking up the monitor
[2017-07-17 11:05:48.131329] I [master(/rhs/brick1/b1):1249:register] _GMaster: Working dir: /var/lib/misc/glusterfsd/master/ssh%3A%2F%2Froot%4010.70.37.105%3Agluster%3A%2F%2F127.0.0.1%3Aslave/c19b89ac45352ab8c894d210d136dd56
[2017-07-17 11:05:48.131927] I [resource(/rhs/brick1/b1):1560:service_loop] GLUSTER: Register time: 1500289548
[2017-07-17 11:05:48.159312] I [master(/rhs/brick1/b1):466:mgmt_lock] _GMaster: Got lock : /rhs/brick1/b1 : Becoming ACTIVE
[2017-07-17 11:05:48.167395] I [gsyncdstatus(/rhs/brick1/b1):271:set_active] GeorepStatus: Worker Status: Active
[2017-07-17 11:05:48.171622] I [gsyncdstatus(/rhs/brick1/b1):244:set_worker_crawl_status] GeorepStatus: Crawl Status: History Crawl



Worker do not restart when log-rsync-performance is set to true:
----------------------------------------------------------------

[2017-07-17 11:06:03.266887] I [gsyncdstatus(/rhs/brick1/b1):244:set_worker_crawl_status] GeorepStatus: Crawl Status: Changelog Crawl
[2017-07-17 11:08:28.918246] I [gsyncd(conf):688:main_i] <top>: Config Set: log-rsync-performance = true

Moving this bug to verified state.

Comment 8 errata-xmlrpc 2017-09-21 04:33:25 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory, and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHBA-2017:2774

Comment 9 errata-xmlrpc 2017-09-21 04:57:48 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory, and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHBA-2017:2774


Note You need to log in before you can comment on or make changes to this bug.