Bug 1600790
Summary: | Segmentation fault while using gfapi while getting volume utilization | ||||||
---|---|---|---|---|---|---|---|
Product: | [Red Hat Storage] Red Hat Gluster Storage | Reporter: | Shubhendu Tripathi <shtripat> | ||||
Component: | rpc | Assignee: | Mohit Agrawal <moagrawa> | ||||
Status: | CLOSED ERRATA | QA Contact: | Upasana <ubansal> | ||||
Severity: | high | Docs Contact: | |||||
Priority: | unspecified | ||||||
Version: | rhgs-3.4 | CC: | amukherj, apaladug, dahorak, jthottan, moagrawa, rhs-bugs, sankarshan, sheggodu, shtripat, skoduri, storage-qa-internal, ubansal | ||||
Target Milestone: | --- | ||||||
Target Release: | RHGS 3.4.0 | ||||||
Hardware: | Unspecified | ||||||
OS: | Unspecified | ||||||
Whiteboard: | |||||||
Fixed In Version: | glusterfs-3.12.2-15 | Doc Type: | If docs needed, set a value | ||||
Doc Text: | Story Points: | --- | |||||
Clone Of: | |||||||
: | 1607783 (view as bug list) | Environment: | |||||
Last Closed: | 2018-09-04 06:50:20 UTC | Type: | Bug | ||||
Regression: | --- | Mount Type: | --- | ||||
Documentation: | --- | CRM: | |||||
Verified Versions: | Category: | --- | |||||
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |||||
Cloudforms Team: | --- | Target Upstream Version: | |||||
Embargoed: | |||||||
Bug Depends On: | 1607783 | ||||||
Bug Blocks: | 1503137, 1600092 | ||||||
Attachments: |
|
Description
Shubhendu Tripathi
2018-07-13 03:58:49 UTC
Created attachment 1458625 [details]
gfapi-segfault.txt
If I create a distribute-disperse volume with smaller no of bricks and from few nodes, the volume utilization details are shown properly as show below # gluster v info test-disp Volume Name: test-disp Type: Distributed-Disperse Volume ID: b2d2d004-34be-4448-9320-6a952b562447 Status: Started Snapshot Count: 0 Number of Bricks: 2 x (2 + 1) = 6 Transport-type: tcp Bricks: Brick1: dahorak-usm3-gl01.usmqe.lab.eng.blr.redhat.com:/root/gluster_bricks/test-disp_b1 Brick2: dahorak-usm3-gl01.usmqe.lab.eng.blr.redhat.com:/root/gluster_bricks/test-disp_b2 Brick3: dahorak-usm3-gl01.usmqe.lab.eng.blr.redhat.com:/root/gluster_bricks/test-disp_b3 Brick4: dahorak-usm3-gl01.usmqe.lab.eng.blr.redhat.com:/root/gluster_bricks/test-disp_b4 Brick5: dahorak-usm3-gl01.usmqe.lab.eng.blr.redhat.com:/root/gluster_bricks/test-disp_b5 Brick6: dahorak-usm3-gl01.usmqe.lab.eng.blr.redhat.com:/root/gluster_bricks/test-disp_b6 Options Reconfigured: transport.address-family: inet nfs.disable: on # tendrl-gluster-vol-utilization test-disp {"test-disp": {"pcnt_used": 16.85206375567077, "used": 2130144.0, "used_inode": 22788, "free": 10510112.0, "pcnt_inode_used": 0.7207355373730537, "total_inode": 3161770, "total": 12640256.0}} Can you provide the core dump to debug this further? Without core its not possible to analyse what caused the crash. Also, installing debuginfo and pasting the backtrace of core is more helpful. Hi, RCA: A gf_client program is getting crash in rpc_clnt_connection_cleanup at the time of destroying saved frames on connection because saved frames are already destroyed by rpc_clnt_destroy.To avoid this race set NULL to saved_frames in critical section in rpc_clnt_destroy. I have tried to execute client program with valgrind and found "O bytes inside a block" at the time of destroying frame like below =9735== Address 0x18abbe70 is 0 bytes inside a block of size 272 free'd ==9735== at 0x4C2ACBD: free (vg_replace_malloc.c:530) ==9735== by 0x5645B9D: rpc_clnt_destroy (rpc-clnt.c:1777) ==9735== by 0x5645B9D: rpc_clnt_notify (rpc-clnt.c:950) ==9735== by 0x56419AB: rpc_transport_unref (rpc-transport.c:517) ==9735== by 0x5644A38: rpc_clnt_trigger_destroy (rpc-clnt.c:1766) ==9735== by 0x5644A38: rpc_clnt_unref (rpc-clnt.c:1803) ==9735== by 0x5644E3F: call_bail (rpc-clnt.c:197) ==9735== by 0x5AA6981: gf_timer_proc (timer.c:165) ==9735== by 0x689DDD4: start_thread (pthread_create.c:308) ==9735== by 0x515DB3C: clone (clone.S:113) Regards Mohit Agrawal Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2018:2607 |