| Summary: | [abrt] kernel: [1136667.935710] kernel BUG at mm/huge_memory.c:1368!: TAINTED -------D | ||
|---|---|---|---|
| Product: | [Fedora] Fedora | Reporter: | Elliott Sales de Andrade <quantum.analyst> |
| Component: | kernel | Assignee: | Kernel Maintainer List <kernel-maint> |
| Status: | CLOSED ERRATA | QA Contact: | Fedora Extras Quality Assurance <extras-qa> |
| Severity: | unspecified | Docs Contact: | |
| Priority: | unspecified | ||
| Version: | 15 | CC: | gansalmon, itamar, jlmagee, jonathan, kernel-maint, madhu.chinakonda |
| Target Milestone: | --- | ||
| Target Release: | --- | ||
| Hardware: | x86_64 | ||
| OS: | Unspecified | ||
| Whiteboard: | abrt_hash:b883c5068f48831ca166b8e8dd4be016977c77b0 | ||
| Fixed In Version: | Doc Type: | Bug Fix | |
| Doc Text: | Story Points: | --- | |
| Clone Of: | Environment: | ||
| Last Closed: | 2012-04-11 16:30:42 UTC | Type: | --- |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
|
Description
Elliott Sales de Andrade
2011-10-20 21:29:36 UTC
Appears to be same issue. The system degraded to the point of being unuseable over the following 2 hours.
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.391322] mapcount 0 page_mapcount 1
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.391536] ------------[ cut here ]------------
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.391730] kernel BUG at mm/huge_memory.c:1368!
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.391923] invalid opcode: 0000 [#1] SMP
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.392117] CPU 0
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.392125] Modules linked in: ipt_MASQUERADE iptable_mangle iptable_nat nf_nat iptable_raw nf_conntrack_ipv4 nf_defr
ag_ipv4 fuse ebtable_nat ebtables xt_CHECKSUM vhost_net macvtap macvlan tun bridge stp llc bonding ip6t_REJECT nf_conntrack_ipv6 nf_defrag_ipv6 xt_state nf_co
nntrack ip6table_filter ip6_tables ses enclosure dcdbas microcode serio_raw ghes hed joydev i2c_i801 i2c_core iTCO_wdt iTCO_vendor_support i7core_edac edac_co
re igb dca virtio_net kvm_intel kvm ipv6 raid1 megaraid_sas [last unloaded: nf_defrag_ipv4]
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.393764]
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.393946] Pid: 4583, comm: postgres Not tainted 2.6.40.4-5.fc15.x86_64 #1 Dell PowerEdge C2100
/0P19C9
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.394340] RIP: 0010:[<ffffffff8111d662>] [<ffffffff8111d662>] split_huge_page+0x181/0x5ac
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.394726] RSP: 0000:ffff88234d035788 EFLAGS: 00010297
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.394921] RAX: 0000000000000001 RBX: ffffea00011db000 RCX: 0000000000003dfe
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.395296] RDX: 0000000000000000 RSI: 0000000000000046 RDI: 0000000000000246
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.395669] RBP: ffff88234d035818 R08: 0000000000000000 R09: 0000000000000000
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.396044] R10: 0000ffff00066c0a R11: 0000000000000003 R12: ffff880fce11cfd0
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.396417] R13: fffffffffffffff2 R14: ffff88234f9402d0 R15: ffff88234f9402d0
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.396790] FS: 00007f74dd5a87e0(0000) GS:ffff88243f200000(0000) knlGS:0000000000000000
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.397167] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.397363] CR2: 0000000006e00000 CR3: 0000000e25634000 CR4: 00000000000026e0
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.397736] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.398109] DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.398484] Process postgres (pid: 4583, threadinfo ffff88234d034000, task ffff881450c81730)
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.398862] Stack:
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.399046] ffff88234d035798 ffffffff81101034 ffff88234d035848 ffff88234f940300
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.399428] 0000000000000003 4000000000000001 ffff88234f940300 000001004a8e2020
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.399810] 000000004d035910 ffff88234f9402e0 0000000000000000 ffff88234d035b78
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.400194] Call Trace:
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.400384] [<ffffffff81101034>] ? page_unlock_anon_vma+0x15/0x17
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.400585] [<ffffffff81106653>] add_to_swap+0x3f/0x88
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.400784] [<ffffffff810e9b4d>] shrink_page_list+0x22a/0x6f4
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.400985] [<ffffffff81120b3a>] ? mem_cgroup_del_lru+0x1d/0x21
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.401186] [<ffffffff810f24ff>] ? __mod_zone_page_state+0x45/0x4f
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.401387] [<ffffffff810e8c4a>] ? update_isolated_counts+0x139/0x157
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.401589] [<ffffffff810ea42f>] shrink_inactive_list+0x230/0x399
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.401789] [<ffffffff810e3b37>] ? determine_dirtyable_memory+0x1a/0x23
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.401992] [<ffffffff810eac17>] shrink_zone+0x3cf/0x50c
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.402190] [<ffffffff810eb0bf>] do_try_to_free_pages+0x10c/0x34e
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.402393] [<ffffffff810e2594>] ? get_page_from_freelist+0x60b/0x64e
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.402595] [<ffffffff810eb582>] try_to_free_pages+0xad/0x100
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.402794] [<ffffffff810e2c16>] __alloc_pages_nodemask+0x4d2/0x736
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.402997] [<ffffffff8110f8bb>] alloc_pages_vma+0xf5/0xfa
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.403194] [<ffffffff8111dcac>] do_huge_pmd_anonymous_page+0xbf/0x26c
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.403396] [<ffffffff810f654f>] ? pmd_offset+0x19/0x3f
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.403592] [<ffffffff810f98ae>] handle_mm_fault+0x120/0x1db
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.403792] [<ffffffff8148b3cd>] do_page_fault+0x354/0x39b
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.403990] [<ffffffff81042cb3>] ? set_next_entity+0x45/0x97
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.404190] [<ffffffff81008842>] ? __switch_to+0x20e/0x220
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.404390] [<ffffffff8123bce0>] ? rb_insert_color+0xb8/0xe1
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.404588] [<ffffffff8104522f>] ? finish_task_switch+0x49/0xb7
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.404803] [<ffffffffa0085a48>] ? kvm_on_user_return+0x65/0x6d [kvm]
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.405003] [<ffffffff810d9d7c>] ? fire_user_return_notifiers+0x2d/0x39
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.405204] [<ffffffff81488755>] page_fault+0x25/0x30
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.405399] Code: 0c 4d 89 fe ff c0 39 45 b4 74 16 8b 53 0c 8b 75 b4 48 c7 c7 a9 55 7b 81 31 c0 ff c2 e8 92 1e 36 00
8b 43 0c ff c0 39 45 b4 74 02 <0f> 0b 4c 8b 2b 4c 8b 7b 20 4c 89 e8 49 c1 ed 35 41 83 e5 03 48
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.406107] RIP [<ffffffff8111d662>] split_huge_page+0x181/0x5ac
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.406306] RSP <ffff88234d035788>
Nov 16 21:08:06 tebreckvm10 kernel: [4321072.406937] ---[ end trace dc6935e840e8825a ]---
please raise the priority and severity to urgent Please try the 2.6.41.1-1 kernel update. Will close this bug in another week if there's no response to the request to try a 2.6.41 kernel. Is there a specific update in 2.6.41 that you believe addresses the issue? Or are we all shooting in the dark? We cannot upgrade until we position some new hardware in a couple of weeks. This occurred once in 3 months on one of six servers supporting similar workloads. so even after we upgrade, it will be difficult to declare victory. (In reply to comment #5) > Is there a specific update in 2.6.41 that you believe addresses the issue? Or > are we all shooting in the dark? We cannot upgrade until we position some new > hardware in a couple of weeks. This occurred once in 3 months on one of six > servers supporting similar workloads. so even after we upgrade, it will be > difficult to declare victory. It's an entirely new kernel release that contains many fixes across the board, including the MM subsystems. While it is indeed a small shot in the dark, it is not an unreasonable one. More importantly, F15 has move to 2.6.41 so there will be no more 2.6.40.x updates. If this still needs fixing, it needs fixing in 2.6.41. I think this was fixed in 1c641e84719429bbfe62a95ed3545ee7fe24408f upstream. 2.6.42.9-2.fc15 and newer should have this fixed. |