Same case and configuration as the previous bug but upgrading from 3.7.5 to 3.7.6 one node (Brick2). Volume Name: smbshare Type: Replicate Volume ID: 40bfc10d-6f7a-45cf-81ba-0e4d531da890 Status: Started Number of Bricks: 1 x 2 = 2 Transport-type: tcp Bricks: Brick1: catsserver-node1:/srv/gluster/bricks/smbshare Brick2: catsserver-node2:/srv/gluster/bricks/smbshare Options Reconfigured: nfs.disable: on server.allow-insecure: on On Brick1: ---------- tail /var/log/glusterfs/etc-glusterfs-glusterd.vol.log [2015-11-27 20:13:35.247775] E [MSGID: 106524] [glusterd-op-sm.c:1794:glusterd_op_stage_stats_volume] 0-glusterd: Volume name get failed [2015-11-27 20:13:35.269306] E [MSGID: 106301] [glusterd-op-sm.c:5197:glusterd_op_ac_stage_op] 0-management: Stage failed on operation 'Volume Profile', Status : -2 On Brick2: ---------- tail /var/log/glusterfs/etc-glusterfs-glusterd.vol.log [2015-11-27 20:13:35.273939] E [MSGID: 106153] [glusterd-syncop.c:113:gd_collate_errors] 0-glusterd: Staging failed on catsserver-node1. Error: Volume name get failed +++ This bug was initially created as a clone of Bug #1276029 +++ This is extract from mail that a user(David Robinson) sent on gluster-users. Description of problem: I have a replica pair setup that I was trying to upgrade from 3.7.4 to 3.7.5. After upgrading the rpm packages (rpm -Uvh *.rpm) and rebooting one of the nodes, I am now receiving the following: [root@frick01 log]# gluster volume status Staging failed on frackib01.corvidtec.com. Please check log file for details. [root@frick01 log]# gluster volume info Volume Name: gfs Type: Replicate Volume ID: abc63b5c-bed7-4e3d-9057-00930a2d85d3 Status: Started Number of Bricks: 1 x 2 = 2 Transport-type: tcp,rdma Bricks: Brick1: frickib01.corvidtec.com:/data/brick01/gfs Brick2: frackib01.corvidtec.com:/data/brick01/gfs Options Reconfigured: storage.owner-gid: 100 server.allow-insecure: on performance.readdir-ahead: on server.event-threads: 4 client.event-threads: 4 How reproducible: Reported by multiple users. Logs have been attached. --- Additional comment from Raghavendra Talur on 2015-10-28 08:47 EDT --- --- Additional comment from Anand Nekkunti on 2015-10-30 10:27:08 EDT --- master branch patch link: http://review.gluster.org/#/c/12473/ --- Additional comment from Vijay Bellur on 2015-11-02 04:39:27 EST --- REVIEW: http://review.gluster.org/12486 (glusterd: move new feature (tiering) enum op to the last of the array) posted (#1) for review on release-3.7 by Gaurav Kumar Garg (ggarg) --- Additional comment from Raghavendra Talur on 2015-11-17 01:01:43 EST --- This bug is getting closed because a release has been made available that should address the reported issue. In case the problem is still not fixed with glusterfs-3.7.6, please open a new bug report. glusterfs-3.7.6 has been announced on the Gluster mailinglists [1], packages for several distributions should become available in the near future. Keep an eye on the Gluster Users mailinglist [2] and the update infrastructure for your distribution. [1] http://www.gluster.org/pipermail/gluster-users/2015-November/024359.html [2] http://thread.gmane.org/gmane.comp.file-systems.gluster.user
Works fine after both servers are upgraded to 3.7.6. Problem is linked to nodes not being at the same version.
Please refer http://www.gluster.org/pipermail/gluster-users/2015-November/024178.html. I've mentioned that the fix actually doesn't solve the immediate upgrade path as well until the whole cluster is upgraded. However once you are in 3.7.6 the subsequent upgrades will not experience it. Let me know whether we can close this bug based on the same point.
OK. So I think we can close it. Let's review it when we'll upgrade to 3.7.7. Thansk for the support.
Closing this bug since the fix was done with an assumption that we will hit the same issue with next upgrade path until and unless the whole cluster is upgraded.