Bug 1290653

Summary: [GlusterD]: GlusterD log is filled with error messages - " Failed to aggregate response from node/brick"
Product: [Red Hat Storage] Red Hat Gluster Storage Reporter: Byreddy <bsrirama>
Component: glusterdAssignee: Atin Mukherjee <amukherj>
Status: CLOSED ERRATA QA Contact: Byreddy <bsrirama>
Severity: high Docs Contact:
Priority: unspecified    
Version: rhgs-3.1CC: amukherj, asrivast, bsrirama, nicolas, nlevinki, rhinduja, sasundar, vbellur
Target Milestone: ---Keywords: ZStream
Target Release: RHGS 3.1.3   
Hardware: x86_64   
OS: Linux   
Whiteboard: glusterd
Fixed In Version: glusterfs-3.7.9-1 Doc Type: Bug Fix
Doc Text:
Story Points: ---
Clone Of:
: 1290734 (view as bug list) Environment:
Last Closed: 2016-06-23 04:59:02 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:
Bug Depends On:    
Bug Blocks: 1268895, 1290734, 1299184, 1310999    

Description Byreddy 2015-12-11 04:40:26 UTC
Description of problem:
=======================
Created a VM in RHEVM Env using gluster volume as a storage and observed below  errors in the  glusterd logs

<<<<<<<<<<<<<<<<<<<<GlusterD Log>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>

[2015-12-11 04:41:40.000010] E [MSGID: 106108] [glusterd-syncop.c:1072:_gd_syncop_commit_op_cbk] 0-management: Failed to aggregate response from  node/brick
[2015-12-11 04:41:40.004306] I [MSGID: 106499] [glusterd-handler.c:4267:__glusterd_handle_status_volume] 0-management: Received status volume req for volume vol_brs
[2015-12-11 04:42:40.657420] E [MSGID: 106108] [glusterd-syncop.c:1072:_gd_syncop_commit_op_cbk] 0-management: Failed to aggregate response from  node/brick
[2015-12-11 04:42:40.660675] I [MSGID: 106499] [glusterd-handler.c:4267:__glusterd_handle_status_volume] 0-management: Received status volume req for volume vol_brs
[2015-12-11 04:43:37.847553] I [MSGID: 106499] [glusterd-handler.c:4267:__glusterd_handle_status_volume] 0-management: Received status volume req for volume vol_brs
[2015-12-11 04:43:41.302213] E [MSGID: 106108] [glusterd-syncop.c:1072:_gd_syncop_commit_op_cbk] 0-management: Failed to aggregate response from  node/brick
[2015-12-11 04:44:41.960021] E [MSGID: 106108] [glusterd-syncop.c:1072:_gd_syncop_commit_op_cbk] 0-management: Failed to aggregate response from  node/brick
The message "I [MSGID: 106499] [glusterd-handler.c:4267:__glusterd_handle_status_volume] 0-management: Received status volume req for volume vol_brs" repeated 3 times between [2015-12-11 04:43:37.847553] and [2015-12-11 04:44:41.963542]
[2015-12-11 04:45:42.634852] E [MSGID: 106108] [glusterd-syncop.c:1072:_gd_syncop_commit_op_cbk] 0-management: Failed to aggregate response from  node/brick
[2015-12-11 04:45:42.640099] I [MSGID: 106499] [glusterd-handler.c:4267:__glusterd_handle_status_volume] 0-management: Received status volume req for volume vol_brs
[2015-12-11 04:46:43.297719] E [MSGID: 106108] [glusterd-syncop.c:1072:_gd_syncop_commit_op_cbk] 0-management: Failed to aggregate response from  node/brick
[2015-12-11 04:46:43.301371] I [MSGID: 106499] [glusterd-handler.c:4267:__glusterd_handle_status_volume] 0-management: Received status volume req for volume vol_brs
[2015-12-11 04:47:43.956339] E [MSGID: 106108] [glusterd-syncop.c:1072:_gd_syncop_commit_op_cbk] 0-management: Failed to aggregate response from  node/brick
[2015-12-11 04:47:43.959903] I [MSGID: 106499] [glusterd-handler.c:4267:__glusterd_handle_status_volume] 0-management: Received status volume req for volume vol_brs
[2015-12-11 04:51:46.542435] E [MSGID: 106108] [glusterd-syncop.c:1072:_gd_syncop_commit_op_cbk] 0-management: Failed to aggregate response from  node/brick
[2015-12-11 04:51:46.546767] I [MSGID: 106499] [glusterd-handler.c:4267:__glusterd_handle_status_volume] 0-management: Received status volume req for volume vol_brs

<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<End>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>


Version-Release number of selected component (if applicable):
glusterfs-3.7.5-10


How reproducible:
=================
One time observed in 3.1.2.


Steps to Reproduce:
===================
1. Have two RHGS nodes with version  - glusterfs-3.7.5-10
2. Add both the nodes to RHEVM
3. Create a Distributed-Replica volume and do fuse mount on client
4. Create a VM  using the fuse mount point as storage
5. Check the glusterd logs in both RHGS nodes..

Actual results:
===============
Errors in glusterd logs 
[2015-12-11 04:51:46.542435] E [MSGID: 106108] [glusterd-syncop.c:1072:_gd_syncop_commit_op_cbk] 0-management: Failed to aggregate response from  node/brick


Expected results:
=================
Above mentioned error should not come in glusterd log.


Additional info:

Comment 2 Atin Mukherjee 2015-12-11 05:34:31 UTC
I don't think this has anything to do with RHEVM setup. I remember seeing this log in some set up and on further analysis I figured out that we have inadequate logs for this path to get to the actual reason of the failure, we need to improve the logging part here and you can expect a patch to be coming soon in upstream. However, if I try to set up a two node cluster and create a volume and run volume status I don't see this log. 

Do you have a reproducer for this?

Comment 3 SATHEESARAN 2015-12-11 06:13:03 UTC
I have seen this issue also in RHGS 3.1.1 ( glusterfs-3.7.1-16.el7rhgs ) too

Comment 4 Byreddy 2015-12-11 06:39:29 UTC
(In reply to Atin Mukherjee from comment #2)
> I don't think this has anything to do with RHEVM setup. I remember seeing
> this log in some set up and on further analysis I figured out that we have
> inadequate logs for this path to get to the actual reason of the failure, we
> need to improve the logging part here and you can expect a patch to be
> coming soon in upstream. However, if I try to set up a two node cluster and
> create a volume and run volume status I don't see this log. 
> 
> Do you have a reproducer for this?

Atin, 

You are right this is not happening with RHEVM only, just with below steps we can consistently reproduce the issue.
1. Have two node cluster
2. Create a Sample (Distributed volume) volume using both the nodes and start it
3. Issue "gluster volume status all tasks" command on one of the node
4. Check the glusterd logs.

Comment 5 Atin Mukherjee 2015-12-11 09:50:09 UTC
An upstream patch http://review.gluster.org/#/c/12950/ is posted for review.

Comment 6 Nicolas Ecarnot 2015-12-18 09:04:09 UTC
Hi,

I'm witnessing the same repeated "Failed to aggregate response
from  node/brick" messages in a replica-3 centos 7.2 nodes with 3.7.6-1 gluster, approx. every 10 seconds.

Comment 7 SATHEESARAN 2015-12-18 14:32:26 UTC
(In reply to Nicolas Ecarnot from comment #6)
> Hi,
> 
> I'm witnessing the same repeated "Failed to aggregate response
> from  node/brick" messages in a replica-3 centos 7.2 nodes with 3.7.6-1
> gluster, approx. every 10 seconds.

Nicolas,

FYI
There was already a patch sent upstream master which is in "needs-code-review" state.
And that change was tracked as part of bug - https://bugzilla.redhat.com/show_bug.cgi?id=1290734 and hope the issue will be fixed in glusterfs-3.7.7

This bug is to track the issue for the product - "Red Hat Gluster Storage" 

Thanks

Comment 9 Atin Mukherjee 2016-01-25 05:36:01 UTC
Looks good to me

Comment 11 Atin Mukherjee 2016-03-22 12:09:02 UTC
The fix is now available in rhgs-3.1.3 branch, hence moving the state to Modified.

Comment 13 Byreddy 2016-04-04 05:19:44 UTC
Verified this bug using the build "glusterfs-3.7.9-1".

Checked glusterd log after executing the "volume status all tasks", No error message like "Failed to aggregate response from  node/brick" in the log.


Moving to verified state with above details.

Comment 15 errata-xmlrpc 2016-06-23 04:59:02 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory, and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHBA-2016:1240