Bug 691606 - Panic due to oom on 32bit RHEL6.1 with FCoE MPIO of 256LUNs
Summary: Panic due to oom on 32bit RHEL6.1 with FCoE MPIO of 256LUNs
Keywords:
Status: CLOSED INSUFFICIENT_DATA
Alias: None
Product: Red Hat Enterprise Linux 6
Classification: Red Hat
Component: device-mapper-multipath
Version: 6.1
Hardware: Unspecified
OS: Linux
unspecified
high
Target Milestone: rc
: 6.1
Assignee: Ben Marzinski
QA Contact: Red Hat Kernel QE team
URL:
Whiteboard:
Depends On:
Blocks: 635490
TreeView+ depends on / blocked
 
Reported: 2011-03-28 23:50 UTC by Yi Zou
Modified: 2011-04-26 13:38 UTC (History)
18 users (show)

Fixed In Version:
Doc Type: Bug Fix
Doc Text:
Clone Of:
Environment:
Last Closed: 2011-04-26 13:38:38 UTC
Target Upstream Version:


Attachments (Terms of Use)
oom panic in mpio 32bit RH (85.74 KB, application/x-compressed-tar)
2011-03-28 23:50 UTC, Yi Zou
no flags Details

Description Yi Zou 2011-03-28 23:50:18 UTC
Created attachment 488304 [details]
oom panic in mpio 32bit RH

FCoE MPIO on 32bit with 256LUNs ends up oom panic:

This has been observed on both 32bit 6.0 and 6.1 with the following setup on both Intel 82599 + Open-FCoE.org FCoE protocol stack as well as Qlogic CNAs. kernel is 2.6.32-122.el6 and device-mapper-multipath-0.4.9-39.el6

Steps to Reproduce:
1. Have 256LUNs each on the dual port HBA using EMC V-MAX
2. Use default MPIO config
3. Start multipathd service
  
Actual results: panic on oom

Expected results: at least gracefully fail

Additional info: This seems to be related to a pretty big mem footprint from multipathd, while spawning per lun udevd (see syslog attached). as this is 32bit, even there is enough highmem, multipathd is able to make use of it. 
I have noticed that multipathd by default sets the oom score to be -17, i.e., not killable, this causes the oom panic every time boots up since mpio gets started right after boot since it kills everything else. When changing the oom score (after disabling multipathd on boot) of multipathd to be much higher to allow it to be killed when oom kicks in, at least the system does not panic.

Please see the attached syslog for the roughly same oom behavior on both qlogic and intel CNAs.

Comment 2 Milan Broz 2011-03-30 12:55:45 UTC
There seems to be other bugs covering multipath memory footprint, maybe it is just dup of bug #623644 ?

Comment 3 Ben Marzinski 2011-03-31 16:43:16 UTC
the fix for bug #623644 should already be in that package.  I just setup 512 iscsi paths to a x86_64 bit system, and multipathd is using less than one tenth of the memory that it is using in your logs.

I haven't done much testing on 32bit systems and none with FCoE, so that's where I'm guessing the problem is.

Comment 4 Yi Zou 2011-03-31 16:58:04 UTC
(In reply to comment #3)
> the fix for bug #623644 should already be in that package.  I just setup 512
> iscsi paths to a x86_64 bit system, and multipathd is using less than one tenth
> of the memory that it is using in your logs.
> I haven't done much testing on 32bit systems and none with FCoE, so that's
> where I'm guessing the problem is.

I believe this is already device-mapper-multipath-0.4.9-39.el6 from rhel6.1 beta, which as you mentioned has the fix from bug #623644 but still happens on 32bit. Let me double-check on the rpm pkg on the testing box where this was reported. Meanwhile, is there any similiar issue found in 32bit? 

As I have reproduced this using Qlogic CNA HBA, so this is not an issue of Open-FCoE's FCoE protocol stack.

Comment 5 Yi Zou 2011-04-01 01:07:18 UTC
(In reply to comment #4)
> (In reply to comment #3)
> > the fix for bug #623644 should already be in that package.  I just setup 512
> > iscsi paths to a x86_64 bit system, and multipathd is using less than one tenth
> > of the memory that it is using in your logs.
> > I haven't done much testing on 32bit systems and none with FCoE, so that's
> > where I'm guessing the problem is.
> 
> I believe this is already device-mapper-multipath-0.4.9-39.el6 from rhel6.1
> beta, which as you mentioned has the fix from bug #623644 but still happens on
> 32bit. Let me double-check on the rpm pkg on the testing box where this was
ok, I have doubled-checked, 0.4.9-39 has the fix from #623644 but this problem is still there on 32bit.

Comment 6 RHEL Program Management 2011-04-04 01:48:18 UTC
Since RHEL 6.1 External Beta has begun, and this bug remains
unresolved, it has been rejected as it is not proposed as
exception or blocker.

Red Hat invites you to ask your support representative to
propose this request, if appropriate and relevant, in the
next release of Red Hat Enterprise Linux.

Comment 7 Ben Marzinski 2011-04-04 16:30:02 UTC
I'll try to setup a x86 system to test this on.  I'm not sure what sort of resources you have available, bu if you would be able to test this on an x86_64 bit system, so that we can confirm that this is a x86 problem that would be really helpful.

Comment 8 Yi Zou 2011-04-04 20:12:47 UTC
(In reply to comment #7)
> I'll try to setup a x86 system to test this on.  I'm not sure what sort of
> resources you have available, bu if you would be able to test this on an x86_64
> bit system, so that we can confirm that this is a x86 problem that would be
> really helpful.
Ok, I will ask our validation folks to run the same test on x86_64 and get back to you. Also, at this point, I am not sure if this is gonna be a storage qualification blocker, I doubt that as it's unlikely people would use 32bit for MPIO w/ that many LUNs, but, if this is, we will have to have this fix targetted for rhel61.

Comment 9 Jack Morgan 2011-04-13 19:53:38 UTC
I cannot reproduce this bug on RHEL6.1-Snapshot2 x86_64 or i386. The multipath service starts up properly. I can run multipath -v2 ; multipath -ll and see all 256 dm-XXX devices and their corresponding paths. Suggest closing.

Comment 10 Russell Doty 2011-04-20 21:03:42 UTC
From comment 9, this bug has been verified by Intel as working in Snapshot 2.


Note You need to log in before you can comment on or make changes to this bug.