Hide Forgot
Description of problem: Can not get raid detail message though the raid create successful Version-Release number of selected component (if applicable): 3.10.0-514.el7.x86_64 How reproducible: sometimes Steps to Reproduce: 1.create raid with some disks mdadm --create --run /dev/md0 --level 1 --metadata 1.2 --raid-devices 5 /dev/sdb1 /dev/sdc1 /dev/sdd1 /dev/sde1 /dev/sdf1 --spare-devices 1 /dev/sdg1--chunk 512 2. mdadm --detail /dev/md0 INFO: Found free iSCSI disks: sdb sdc sdd sde sdf sdg sdh sdi INFO: iSCSI is setuped on storageqe-55.rhts.eng.pek2.redhat.com for mdadm INFO: Executing MD_Create_RAID() to create raid 1 INFO: Created md raid with these raid devices " /dev/sdb1 /dev/sdc1 /dev/sdd1 /dev/sde1 /dev/sdf1". INFO: Created md raid with these spare disks " /dev/sdg1". mdadm: array /dev/md0 started. [INFO][23:42:51]INFO: Successfully created md raid /dev/md0 mdadm: cannot open /dev/md0: No such file or directory state is mdadm: cannot open /dev/md0: No such file or directory mdadm: cannot open /dev/md0: No such file or directory mdadm: cannot open /dev/md0: No such file or directory mdadm: cannot open /dev/md0: No such file or directory mdadm: cannot open /dev/md0: No such file or directory mdadm: cannot open /dev/md0: No such file or directory mdadm: cannot open /dev/md0: No such file or directory mdadm: cannot open /dev/md0: No such file or directory mdadm: cannot open /dev/md0: No such file or directory Actual results: can not get the raid message Expected results: get md raid message Additional info: [root@storageqe-55 grow-add-disk]# cat /proc/mdstat Personalities : [raid0] [raid1] [raid6] [raid5] [raid4] [raid10] md127 : inactive sdi1[6](S) 10477568 blocks super 1.2 md0 : active raid1 sdg1[5](S) sdf1[4] sde1[3] sdd1[2] sdc1[1] sdb1[0] 10477568 blocks super 1.2 [5/5] [UUUUU] [====>................] resync = 24.2% (2545792/10477568) finish=8.3min speed=15795K/sec unused devices: <none> [root@storageqe-55 grow-add-disk]# mdadm --detail /dev/md127 /dev/md127: Version : 1.2 Raid Level : raid0 Total Devices : 1 Persistence : Superblock is persistent State : inactive Name : 0 UUID : 3da95f68:1c79c02e:4f0fbd2f:1d751f45 Events : 259 Number Major Minor RaidDevice - 8 129 - /dev/sdi1 [root@storageqe-55 grow-add-disk]# lsblk NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT sda 8:0 0 279.4G 0 disk ├─sda1 8:1 0 1G 0 part /boot └─sda2 8:2 0 278.4G 0 part ├─rhel_storageqe--55-root 253:0 0 50G 0 lvm / ├─rhel_storageqe--55-swap 253:1 0 28G 0 lvm [SWAP] └─rhel_storageqe--55-home 253:2 0 200.4G 0 lvm /home sdb 8:16 0 20G 0 disk └─sdb1 8:17 0 10G 0 part └─md0 9:0 0 10G 0 raid1 sdc 8:32 0 20G 0 disk └─sdc1 8:33 0 10G 0 part └─md0 9:0 0 10G 0 raid1 sdd 8:48 0 20G 0 disk └─sdd1 8:49 0 10G 0 part └─md0 9:0 0 10G 0 raid1 sde 8:64 0 20G 0 disk └─sde1 8:65 0 10G 0 part └─md0 9:0 0 10G 0 raid1 sdf 8:80 0 20G 0 disk └─sdf1 8:81 0 10G 0 part └─md0 9:0 0 10G 0 raid1 sdg 8:96 0 20G 0 disk └─sdg1 8:97 0 10G 0 part └─md0 9:0 0 10G 0 raid1 sdh 8:112 0 20G 0 disk └─sdh1 8:113 0 10G 0 part sdi 8:128 0 20G 0 disk └─sdi1 8:129 0 10G 0 part job: https://beaker.engineering.redhat.com/recipes/3244084#task47921946
Created attachment 1221541 [details] sosreport
Could you please clarify your bug report: This doesn't make any sense - what happens as a result of 1? The INFO: messages do not come from the mdadm command as your report indicates. Could you please provide the /proc/mdstat output inbetween command 1 and command 2? Is this run in a script or on the command line? If you run command 1 and 2 in a script you may simply be racing the creation of the device node in /dev Jes Steps to Reproduce: 1.create raid with some disks mdadm --create --run /dev/md0 --level 1 --metadata 1.2 --raid-devices 5 /dev/sdb1 /dev/sdc1 /dev/sdd1 /dev/sde1 /dev/sdf1 --spare-devices 1 /dev/sdg1--chunk 512 2. mdadm --detail /dev/md0 INFO: Found free iSCSI disks: sdb sdc sdd sde sdf sdg sdh sdi INFO: iSCSI is setuped on storageqe-55.rhts.eng.pek2.redhat.com for mdadm INFO: Executing MD_Create_RAID() to create raid 1 INFO: Created md raid with these raid devices " /dev/sdb1 /dev/sdc1 /dev/sdd1 /dev/sde1 /dev/sdf1". INFO: Created md raid with these spare disks " /dev/sdg1". mdadm: array /dev/md0 started. [INFO][23:42:51]INFO: Successfully created md raid /dev/md0 mdadm: cannot open /dev/md0: No such file or directory state is mdadm: cannot open /dev/md0: No such file or directory mdadm: cannot open /dev/md0: No such file or directory
1.create raid 1 or raid 10 successful. I met this question on raid1 and raid10 while running /kernel/storage/mdadm/grow-add-disk test case mdadm --create --run /dev/md0 --level 1 --metadata 1.2 --raid-devices 5 /dev/sdb1 /dev/sdc1 /dev/sdd1 /dev/sde1 /dev/sdf1 --spare-devices 1 /dev/sdg1--chunk 512 2.get message about the raid command "mdadm --detail /dev/md0 ",but failed like this [root@storageqe-55 grow-add-disk]# mdadm --detail /dev/md0 mdadm: cannot open /dev/md0: No such file or directory [root@storageqe-55 grow-add-disk]# ls /dev/md* /dev/md127 but we can get the /dev/md0 from mdstat. [root@storageqe-55 grow-add-disk]# cat /proc/mdstat Personalities : [raid0] [raid1] [raid6] [raid5] [raid4] [raid10] md127 : inactive sdi1[6](S) 10477568 blocks super 1.2 md0 : active raid1 sdg1[5](S) sdf1[4] sde1[3] sdd1[2] sdc1[1] sdb1[0] 10477568 blocks super 1.2 [5/5] [UUUUU] [===========>.........] resync = 56.7% (5949248/10477568) finish=4.8min speed=15547K/sec unused devices: <none> the INFO message was my test case output when the issue occur. this issue was hard to reproduce,so i'm not sure i can get the output between command 1 and command 2.
I see, There is a possibility that mdadm --create completes before the /dev/md0 device node is created. If you add a small delay after the first command, are you still able to reproduce the problem? Thanks, Jes
i have added 30s delay time after the first time get raid status failed ,then get the raid status ,but all can not get the raid name