Bug 1330714
Summary: | docker-storage-setup frequently fails when lvm2 not initially installed in system | ||
---|---|---|---|
Product: | Red Hat Enterprise Linux 7 | Reporter: | Lokesh Mandvekar <lsm5> |
Component: | docker-latest | Assignee: | Lokesh Mandvekar <lsm5> |
Status: | CLOSED ERRATA | QA Contact: | atomic-bugs <atomic-bugs> |
Severity: | unspecified | Docs Contact: | Yoana Ruseva <yruseva> |
Priority: | unspecified | ||
Version: | 7.2 | CC: | agk, atomic-bugs, dwalsh, ejacobs, lsm5, lsu, prajnoha, teigland, vgoyal, zkabelac |
Target Milestone: | rc | Keywords: | Extras, ZStream |
Target Release: | --- | ||
Hardware: | Unspecified | ||
OS: | Unspecified | ||
Whiteboard: | |||
Fixed In Version: | docker-latest-1.10.3-20.el7 | Doc Type: | Bug Fix |
Doc Text: | Story Points: | --- | |
Clone Of: | 1330290 | Environment: | |
Last Closed: | 2016-05-12 14:56:18 UTC | Type: | Bug |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: | |||
Bug Depends On: | 1330290 | ||
Bug Blocks: | 1330706 |
Comment 4
Lokesh Mandvekar
2016-05-04 19:37:07 UTC
(In reply to Lokesh Mandvekar from comment #4) > Luwen, > > The patch at > https://github.com/projectatomic/docker-storage-setup/commit/ > df2af9439577cedc2c502512d887c8df10a33cbf should be fixing this along with > the change that we replace docker with docker-latest everywhere in > docker-latest-storage-setup except we retain the local variable > docker_devmapper_meta_dir as-is. > > Vivek, please correct me if i'm wrong. Thanks, then the wipe signatures works fine for me.. move to verified The step is just for reference, run docker-latest-storage-setup twice, the second time will trigger the "Fatal "Failed to wipe signatures on device ${dev}1" to exit. # cat /etc/sysconfig/docker-latest-storage-setup WIPE_SIGNATURES=true DEVS=vdb VG=docker-latest # docker-latest-storage-setup INFO: Volume group backing root filesystem could not be determined INFO: Wipe Signatures is set to true. Any signatures on /dev/vdb will be wiped. /dev/vdb: 2 bytes were erased at offset 0x00000438 (ext2): 53 ef create_disk_partions Checking that no-one is using this disk right now ... OK Disk /dev/vdb: 16644 cylinders, 16 heads, 63 sectors/track sfdisk: /dev/vdb: unrecognized partition table type Old situation: sfdisk: No partitions found New situation: Units: sectors of 512 bytes, counting from 0 Device Boot Start End #sectors Id System /dev/vdb1 2048 16777215 16775168 8e Linux LVM /dev/vdb2 0 - 0 0 Empty /dev/vdb3 0 - 0 0 Empty /dev/vdb4 0 - 0 0 Empty Warning: partition 1 does not start at a cylinder boundary Warning: partition 1 does not end at a cylinder boundary Warning: no primary partition is marked bootable (active) This does not matter for LILO, but the DOS MBR will not boot this disk. Successfully wrote the new partition table Re-reading the partition table ... If you created or changed a DOS partition, /dev/foo7, say, then use dd(1) to zero the first 512 bytes: dd if=/dev/zero of=/dev/foo7 bs=512 count=1 (See fdisk(8).) wipefs executing....... result.... Physical volume "/dev/vdb1" successfully created Volume group "docker-latest" successfully created Rounding up size to full physical extent 12.00 MiB Logical volume "docker-latest-poolmeta" created. Logical volume "docker-latest-pool" created. WARNING: Converting logical volume docker-latest/docker-latest-pool and docker-latest/docker-latest-poolmeta to pool's data and metadata volumes. THIS WILL DESTROY CONTENT OF LOGICAL VOLUME (filesystem etc.) Converted docker-latest/docker-latest-pool to thin pool. Logical volume "docker-latest-pool" changed. [root@myregistrydomain ~]# docker-latest-storage-setup INFO: Volume group backing root filesystem could not be determined INFO: Wipe Signatures is set to true. Any signatures on /dev/vdb will be wiped. wipefs: error: /dev/vdb: probing initialization failed: Device or resource busy ERROR: Failed to wipe signatures on device /dev/vdb Ok, this issue is solved by using udevadm settle patch. https://github.com/projectatomic/docker-storage-setup/commit/80648727909370d394b6e72af76e9c88fe8e1f4c Also, I think WIPE_SIGNATURES should take affect only when disk is not already part of volume group. I need to look at the code closely and see why it is hitting for you again when you run docker storage setup again. Ok, found out. If DEVS=/dev/vdb is used, this problem does not happen. Please use full device path name instead of just disk name. I will fix it anyway but this is not a good practice. Created a PR to fix the issue of wipefs being called on a disk which is already part of volume group. https://github.com/projectatomic/docker-storage-setup/pull/131 We will need to live with this problem with DEVS=vdb until the next release 7.2.5 Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://rhn.redhat.com/errata/RHEA-2016-1057.html |