Bug 446284
Summary: | RHEL can't boot after installation on a Intel sata array | ||
---|---|---|---|
Product: | Red Hat Enterprise Linux 5 | Reporter: | Fernando Lozano <fernando> |
Component: | mkinitrd | Assignee: | Peter Jones <pjones> |
Status: | CLOSED DUPLICATE | QA Contact: | |
Severity: | high | Docs Contact: | |
Priority: | high | ||
Version: | 5.2 | CC: | bmr, ddumas, heinzm, jukka.lehtonen, syeghiay, tao, wmealing |
Target Milestone: | rc | ||
Target Release: | --- | ||
Hardware: | All | ||
OS: | Linux | ||
Whiteboard: | |||
Fixed In Version: | Doc Type: | Bug Fix | |
Doc Text: | Story Points: | --- | |
Clone Of: | Environment: | ||
Last Closed: | 2008-12-02 20:45:56 UTC | Type: | --- |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: | |||
Attachments: |
Description
Fernando Lozano
2008-05-13 21:52:43 UTC
Just to let you know my motherborard and raid controller are certified for RHEL: https://hardware.redhat.com/show.cgi?id=238494 OOps, my mistake, I was refering to bug #349161 (https://bugzilla.redhat.com/show_bug.cgi?id=349161) and not bug #399161; Both refere to incorrect initrd that prevents access to root partition on hardware raid. After updating my server to RHEL5.2 though RHN (just updating packages running yum) same problem, just a little bit different: the update generated a initrd containing a init script with the following commands, that were not included by previous updates: mkblkdevs echo Scanning and configuring dmraid supported devices dmraid -ay -i -p "ddf1_4c5349202020202080862682000000003547905e00000a28" kpartx -a -p p "/dev/mapper/ddf1_4c5349202020202080862682000000003547905e00000a28" echo Scanning logical volumes (Commandos "mkblkdevs" and "echo Scanning..." are there just to mark the positio n where the new commands were inserted) My server could not find the root partition, so I deleted those new commands and replaced them by the commandos that were on the original initrd generated by RHEL5.0 installer but not on any of the subsequent updates: mkblkdevs rmparts sdc rmparts sdb rmparts sda dm create ddf1_4c5349202020202080862682000000003547905e00000a28 0 2923825152 striped 3 128 8:0 0 8:16 0 8:32 0 dm partadd ddf1_4c5349202020202080862682000000003547905e00000a28 echo Scanning logical volumes And now my server boots again. There's a couple of weird things here. First, the device name generated indicates DDF1 disk metadata. That's weird, because it's not what any Intel raid firmware I've ever seen uses. I'd expect something like "isw_Volume_0". So the first thing we need to figure out is if the data is being probed wrong. Can you run "dmraid -ay -t" and post the output? Here's the output you asked for: # dmraid -ay -t ddf1_4c534920202020208086268200000000356841f000000a28: 0 2923825152 striped 3 128 /dev/sda 0 /dev/sdb 0 /dev/sdc 0 I am also giving you a sysreport of the server after updating all packages to RHEL5.2. Created attachment 306550 [details]
Sysreport of bug server after update to 5.2 (and my fixes to initrd)
just to let you know updating to kernel 2.6.18-92.1.1 (from RHN) requires me to make the same fixes to initrd. Created attachment 311984 [details]
Dmesg, raid fail to boot after intall redhat 5.2, on motherboard with ICH9R
Having the "dm create"/"dm addpart" commands replaced with dmraid/kpartx is expected - RHEL5 has moved from using nash support for activating dmraid devices to including dmraid itself in the initrd images so this seems like a problem with dmraid being able to activate the device itself - adding the maintainer on CC. I just fought a week with this/similar during install of CentOS 5.2 into ICH10R RAID1. I do get RAID-set of type isw_*, and my boot did "succeed", because the commands after 'dmraid -ay *' did spot the partitions on /dev/sda (and duplicates on /dev/sdb). The convenience of mirrors compared to stripes. The solution that does work for me is to use: dmraid -ay -i -p without the RAID-set string. As with others, that manual edit of initrd*.img is required after each kernel update. Comment #15 on bug #349161 appears to explain where dmraid fails with the explicit RAID-set string. This seems to affect all platforms, not just 64 bit.. changing to all. Created attachment 324554 [details]
Remove the dmname, as the raid set is set incorrectly.
Ive only done very basic testing of this, and I'm not sure if activating all raid sets on the system is a problem, but this seems to workforme.
The patch in comment #13 is a workaround - our strategy for device activation in the initramfs is to only activate those devices needed for booting (i.e. the ones that contain the root file system). This is true for dmraid, mpath, LVM, and MD devices. If we're setting the wrong dmname for this array then that should be fixed, rather than trying to activate every dmraid device on the system. Agreeing to Bryn's comment #14. Activating *all* RAID sets in an eg. DDF1 environment may be a long list to work during an initrd run and should be postponed to rc processing. *** This bug has been marked as a duplicate of bug 471689 *** |