rsync --stats --progress This would give enough details to display in our log files on a timely basis or on a user request basis to check the bandwidth usage. This could be a special configuration option.
Accidentally closed the wrong BUG
Feature requests make most sense against the 'mainline' release, there is no ETA for an implementation and requests might get forgotten when filed against a particular version.
REVIEW: http://review.gluster.org/10070 (geo-rep: Log Rsync performance) posted (#1) for review on master by Aravinda VK (avishwan)
Added basic functionality to record rsync performance after the sync.
COMMIT: http://review.gluster.org/10070 committed in master by Vijay Bellur (vbellur) ------ commit fea8d9701291ed5ebdbd655c916a866b5f40a34f Author: Aravinda VK <avishwan> Date: Tue Mar 31 17:03:42 2015 +0530 geo-rep: Log Rsync performance Introducing configurable option to log the rsync performance. gluster volume geo-replication <MASTERVOL> <SLAVEHOST>::<SLAVEVOL> \ config log-rsync-performance true Default value is False. Example log: [2015-03-31 16:48:34.572022] I [resource(/bricks/b1):857:rsync] SSH: rsync performance: Number of files: 2 (reg: 1, dir: 1), Number of regular files transferred: 1, Total file size: 178 bytes, Total transferred file size: 178 bytes, Literal data: 178 bytes, Matched data: 0 bytes, Total bytes sent: 294, Total bytes received: 32, sent 294 bytes received 32 bytes 652.00 bytes/sec Change-Id: If11467e29e6ac502fa114bd5742a8434b7084f98 Signed-off-by: Aravinda VK <avishwan> BUG: 764827 Reviewed-on: http://review.gluster.org/10070 Tested-by: Gluster Build System <jenkins.com> Reviewed-by: Vijay Bellur <vbellur>
This bug is getting closed because a release has been made available that should address the reported issue. In case the problem is still not fixed with glusterfs-3.7.0, please open a new bug report. glusterfs-3.7.0 has been announced on the Gluster mailinglists [1], packages for several distributions should become available in the near future. Keep an eye on the Gluster Users mailinglist [2] and the update infrastructure for your distribution. [1] http://thread.gmane.org/gmane.comp.file-systems.gluster.devel/10939 [2] http://thread.gmane.org/gmane.comp.file-systems.gluster.user