Bug 1585207
| Summary: | oVirt node upgrade fail in %post script | ||
|---|---|---|---|
| Product: | [oVirt] ovirt-node | Reporter: | Rob Sanders <rob> |
| Component: | Installation & Update | Assignee: | Ryan Barry <rbarry> |
| Status: | CLOSED WORKSFORME | QA Contact: | Yaning Wang <yaniwang> |
| Severity: | unspecified | Docs Contact: | |
| Priority: | unspecified | ||
| Version: | 4.2 | CC: | bugs, cshao |
| Target Milestone: | --- | ||
| Target Release: | --- | ||
| Hardware: | x86_64 | ||
| OS: | Linux | ||
| Whiteboard: | |||
| Fixed In Version: | Doc Type: | If docs needed, set a value | |
| Doc Text: | Story Points: | --- | |
| Clone Of: | Environment: | ||
| Last Closed: | 2018-06-01 19:59:20 UTC | Type: | Bug |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | Node | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
|
Description
Rob Sanders
2018-06-01 14:00:57 UTC
I managed to recover my system by analysing debug log from imgbase and re-using it's actions to recreate missing LV. Pasting here for reference only if someone else has similar problem. Chances are not all the steps are needed, but this worked for me:
# lvcreate --thin --virtualsize 155508015104B --name ovirt-node-ng-4.1.8-0.20171211.0 onn/pool00
Using default stripesize 64.00 KiB.
WARNING: Sum of all thin volume sizes (<1.02 TiB) exceeds the size of thin pool onn/pool00 and the size of whole volume group (220.00 GiB)!
For thin pool auto extension activation/thin_pool_autoextend_threshold should be below 100.
Logical volume "ovirt-node-ng-4.1.8-0.20171211.0" created.
# lvchange --addtag imgbased:base onn/ovirt-node-ng-4.1.8-0.20171211.0
Logical volume onn/ovirt-node-ng-4.1.8-0.20171211.0 changed.
# lvchange --permission r onn/ovirt-node-ng-4.1.8-0.20171211.0
Logical volume onn/ovirt-node-ng-4.1.8-0.20171211.0 changed.
# lvchange --setactivationskip y onn/ovirt-node-ng-4.1.8-0.20171211.0
Logical volume onn/ovirt-node-ng-4.1.8-0.20171211.0 changed.
# lvchange --activate n onn/ovirt-node-ng-4.1.8-0.20171211.0
# lvchange --permission rw onn/ovirt-node-ng-4.1.8-0.20171211.0
Logical volume onn/ovirt-node-ng-4.1.8-0.20171211.0 changed.
# lvchange --activate y onn/ovirt-node-ng-4.1.8-0.20171211.0 --ignoreactivationskip
# mkfs.ext4 -E discard /dev/onn/ovirt-node-ng-4.1.8-0.20171211.0
mke2fs 1.42.9 (28-Dec-2013)
Discarding device blocks: done
Filesystem label=
OS type: Linux
Block size=4096 (log=2)
Fragment size=4096 (log=2)
Stride=16 blocks, Stripe width=16 blocks
9494528 inodes, 37965824 blocks
1898291 blocks (5.00%) reserved for the super user
First data block=0
Maximum filesystem blocks=2187329536
1159 block groups
32768 blocks per group, 32768 fragments per group
8192 inodes per group
Superblock backups stored on blocks:
32768, 98304, 163840, 229376, 294912, 819200, 884736, 1605632, 2654208,
4096000, 7962624, 11239424, 20480000, 23887872
Allocating group tables: done
Writing inode tables: done
Creating journal (32768 blocks): done
Writing superblocks and filesystem accounting information: done
# lvchange --permission r onn/ovirt-node-ng-4.1.8-0.20171211.0
Logical volume onn/ovirt-node-ng-4.1.8-0.20171211.0 changed.
# lvchange --setactivationskip y onn/ovirt-node-ng-4.1.8-0.20171211.0
Logical volume onn/ovirt-node-ng-4.1.8-0.20171211.0 changed.
# lvchange --activate n onn/ovirt-node-ng-4.1.8-0.20171211.0
# imgbase layout
ovirt-node-ng-4.1.8-0.20171211.0
+- ovirt-node-ng-4.1.8-0.20171211.0+1
ovirt-node-ng-4.1.9-0.20180124.0
+- ovirt-node-ng-4.1.9-0.20180124.0+1
ovirt-node-ng-4.2.3.1-0.20180530.0
+- ovirt-node-ng-4.2.3.1-0.20180530.0+1
# lvremove onn/ovirt-node-ng-4.2.3.1-0.20180530.0+1
Do you really want to remove active logical volume onn/ovirt-node-ng-4.2.3.1-0.20180530.0+1? [y/n]: y
Logical volume "ovirt-node-ng-4.2.3.1-0.20180530.0+1" successfully removed
# lvremove onn/ovirt-node-ng-4.2.3.1-0.20180530.0
Do you really want to remove active logical volume onn/ovirt-node-ng-4.2.3.1-0.20180530.0? [y/n]: y
Logical volume "ovirt-node-ng-4.2.3.1-0.20180530.0" successfully removed
# yum reinstall ovirt-node-ng-image-update -y
I'm glad you were able to resolve, but any idea how the system got in this state in the first place? This is a totally new report to me, and I've never seen anything like it... It looks like one of the LVs was removed but LVM still had it cached somewhere. Closing for now, since you worked around it, but still responding to comments... |