Bug 1686255

Summary: glusterd leaking memory when issued gluster vol status all tasks continuosly
Product: [Red Hat Storage] Red Hat Gluster Storage Reporter: Bala Konda Reddy M <bmekala>
Component: glusterdAssignee: Sanju <srakonde>
Status: CLOSED ERRATA QA Contact: Kshithij Iyer <kiyer>
Severity: high Docs Contact:
Priority: medium    
Version: rhgs-3.4CC: amukherj, bmekala, kiyer, nchilaka, pasik, rhinduja, rhs-bugs, sheggodu, srakonde, storage-qa-internal, vbellur
Target Milestone: ---   
Target Release: RHGS 3.5.0   
Hardware: x86_64   
OS: Linux   
Whiteboard:
Fixed In Version: glusterfs-6.0-2 Doc Type: Bug Fix
Doc Text:
A small memory leak that occurred when viewing the status of all volumes has been fixed.
Story Points: ---
Clone Of:
: 1691164 (view as bug list) Environment:
Last Closed: 2019-10-30 12:20:22 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:
Bug Depends On: 1691164, 1694610, 1694612    
Bug Blocks: 1696807    
Attachments:
Description Flags
Top output of glusterd for all six nodes of the cluster none

Description Bala Konda Reddy M 2019-03-07 05:48:42 UTC
Created attachment 1541678 [details]
Top output of glusterd for all six nodes of the cluster

Description of problem:
glusterd is leaking memory when issused "gluster vol status tasks" continuosly for 12 hours. The memory increase is from 250MB to 1.1GB. The increase have been 750 MB.


Version-Release number of selected component (if applicable):
glusterfs-3.12.2-45.el7rhgs.x86_64

How reproducible:
1/1

Steps to Reproduce:
1. On a six node cluster with brick-multiplexing enabled
2. Created 150 disperse volumes and 250 replica volumes and started them
3. Taken memory footprint from all the nodes
4. Issued "while true; do gluster volume status all tasks; sleep 2; done" with a time gap of 2 seconds 

Actual results:
Seen a memory increase of glusterd on Node N1 from 260MB to 1.1GB

Expected results:
glusterd memory shouldn't leak

Additional info:
Attaching the screenshot of the top output before and after the command has been executed.

The setup in same state for further debugging.

Comment 9 Atin Mukherjee 2019-03-12 09:11:54 UTC
Sanju,

Looks like there's a leak on the remote glusterd i.e. in the op-sm framework based on the periodic statedump I captured while testing this.

The impacted data types are:

gf_common_mt_gf_timer_t
gf_common_mt_asprintf
gf_common_mt_strdup
gf_common_mt_char
gf_common_mt_txn_opinfo_obj_t

Please check if we're not cleaning up txn_opinfo in some place in this transaction, fixing that might implicitly fix the other leaks too.

Comment 29 errata-xmlrpc 2019-10-30 12:20:22 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory, and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHEA-2019:3249