Bug 1580361 - rpc: ABRT, SEGV in rpcsvc_handle_disconnect->glusterd_rpcsvc_notify
Summary: rpc: ABRT, SEGV in rpcsvc_handle_disconnect->glusterd_rpcsvc_notify
Keywords:
Status: CLOSED INSUFFICIENT_DATA
Alias: None
Product: GlusterFS
Classification: Community
Component: glusterd
Version: mainline
Hardware: x86_64
OS: Linux
unspecified
unspecified
Target Milestone: ---
Assignee: bugs@gluster.org
QA Contact:
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2018-05-21 11:07 UTC by Kaleb KEITHLEY
Modified: 2019-11-26 18:19 UTC (History)
4 users (show)

Fixed In Version:
Doc Type: If docs needed, set a value
Doc Text:
Clone Of:
Environment:
Last Closed: 2018-12-05 06:55:05 UTC
Regression: ---
Mount Type: ---
Documentation: ---
CRM:
Verified Versions:
Embargoed:


Attachments (Terms of Use)

Description Kaleb KEITHLEY 2018-05-21 11:07:19 UTC
Description of problem:


Version-Release number of selected component (if applicable):


How reproducible:


Steps to Reproduce:
1.
2.
3.

Actual results:


Expected results:


Additional info:

Comment 1 Kaleb KEITHLEY 2018-05-21 11:09:46 UTC
See https://retrace.fedoraproject.org/faf/reports/2170288/

includes a rudimentary BT

Comment 2 Jeff Darcy 2018-05-21 12:11:20 UTC
I see there's some RDMA code in the stack trace. Any chance this is RDMA-specific, e.g. some kind of ordering/serialization assumption fulfilled by the socket code but not by the RDMA code? That wouldn't necessarily mean it's a flaw in the RDMA code or should be fixed there, but might provide a useful hint. Or maybe it's pure coincidence.

Comment 3 Atin Mukherjee 2018-06-18 03:40:29 UTC
Do we have a reproducer for this?

Comment 4 Kaleb KEITHLEY 2018-06-18 11:34:07 UTC
No, _I_ don't have a reproducer.

This is an automated ABRT report coming from boxes running community gluster installed from the CentOS Storage SIG repos.

I'm simply relaying the report that gets forwarded to me as the Gluster packager in Fedora. (CentOS ABRTs get sent to Fedora.)

Although as you can see at the link posted in Comment 1 it has occurred 67 times, so it seems like it should be easy to reproduce.

Comment 5 Atin Mukherjee 2018-10-05 02:38:48 UTC
Kaleb - do you still see this happening? I'm looking for a core file (specifically the complete backtrace) and the test which is run to get to the crash by which we can begin the investigation.

Comment 6 Shyamsundar 2018-10-23 14:54:00 UTC
Release 3.12 has been EOLd and this bug was still found to be in the NEW state, hence moving the version to mainline, to triage the same and take appropriate actions.

Comment 7 Atin Mukherjee 2018-12-05 06:55:05 UTC
Doesn't have sufficient data to debug this. Closing it.


Note You need to log in before you can comment on or make changes to this bug.