Bug 993119
Summary: | [RHS-RHOS] DHT hex is not reset on the glance image post remove-brick and it goes missing from filesystem_store_datadir specified in glance-api.conf | ||
---|---|---|---|
Product: | [Red Hat Storage] Red Hat Gluster Storage | Reporter: | Gowrishankar Rajaiyan <grajaiya> |
Component: | glusterfs | Assignee: | shishir gowda <sgowda> |
Status: | CLOSED ERRATA | QA Contact: | shilpa <smanjara> |
Severity: | high | Docs Contact: | |
Priority: | high | ||
Version: | 2.1 | CC: | amarts, grajaiya, nsathyan, rhs-bugs, sdharane, shaines, tkatarki, vagarwal, vbellur |
Target Milestone: | --- | ||
Target Release: | --- | ||
Hardware: | Unspecified | ||
OS: | Unspecified | ||
Whiteboard: | |||
Fixed In Version: | glusterfs-3.4.0.19rhs-1 | Doc Type: | Bug Fix |
Doc Text: | Story Points: | --- | |
Clone Of: | Environment: |
virt rhos cinder rhs integration
|
|
Last Closed: | 2013-09-23 22:35:59 UTC | Type: | Bug |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: |
Description
Gowrishankar Rajaiyan
2013-08-05 15:58:08 UTC
Tried to reproduce this on a new volume with eager-lock turned off as suggested by Shishir. The issue persists. With eager-lock enabled: Volume Name: vol-glance Type: Distributed-Replicate Volume ID: 93bd85a9-1621-444e-8f0d-3122cfa86723 Status: Started Number of Bricks: 6 x 2 = 12 Transport-type: tcp Bricks: Brick1: 10.70.37.168:/rhs/brick3/g1 Brick2: 10.70.37.74:/rhs/brick3/g2 Brick3: 10.70.37.220:/rhs/brick3/g3 Brick4: 10.70.37.203:/rhs/brick3/g4 Brick5: 10.70.37.168:/rhs/brick3/g5 Brick6: 10.70.37.74:/rhs/brick3/g6 Brick7: 10.70.37.220:/rhs/brick3/g7 Brick8: 10.70.37.203:/rhs/brick3/g8 Brick9: 10.70.37.168:/rhs/brick3/g9 Brick10: 10.70.37.74:/rhs/brick3/g10 Brick11: 10.70.37.220:/rhs/brick3/g11 Brick12: 10.70.37.203:/rhs/brick3/g12 Options Reconfigured: storage.owner-uid: 161 storage.owner-gid: 161 network.remote-dio: enable cluster.eager-lock: enable performance.stat-prefetch: off performance.io-cache: off performance.read-ahead: off performance.quick-read: off Remove-brick. File: c4efd768-3b9a-44ab-9b91-5dcfc2989fc0 [root@rhs-vm1 home]# find / -name c4efd768-3b9a-44ab-9b91-5dcfc2989fc0 /rhs/brick3/g9/glance/images/c4efd768-3b9a-44ab-9b91-5dcfc2989fc0 [root@rhs-vm1 home]# gluster v remove-brick vol-glance 10.70.37.168:/rhs/brick3/g9 10.70.37.74:/rhs/brick3/g10 start volume remove-brick start: success ID: 99dea640-3b2d-4c20-92da-0759fa860af6 [root@rhs-vm1 home]# gluster v remove-brick vol-glance 10.70.37.168:/rhs/brick3/g9 10.70.37.74:/rhs/brick3/g10 status Node Rebalanced-files size scanned failures skipped status run-time in secs --------- ----------- ----------- ----------- ----------- ----------- ------------ -------------- localhost 0 0Bytes 0 0 0 completed 0.00 10.70.37.220 0 0Bytes 0 0 0 not started 0.00 10.70.37.203 0 0Bytes 0 0 0 not started 0.00 10.70.37.74 0 0Bytes 4 0 0 completed 0.00 After remove brick: [root@rhs-vm1 home]# find / -name c4efd768-3b9a-44ab-9b91-5dcfc2989fc0 /rhs/brick3/g9/glance/images/c4efd768-3b9a-44ab-9b91-5dcfc2989fc0 [root@rhs-vm3 home]# find / -name c4efd768-3b9a-44ab-9b91-5dcfc2989fc0 /rhs/brick3/g11/glance/images/c4efd768-3b9a-44ab-9b91-5dcfc2989fc0 [root@rhs-vm1 home]# ll /rhs/brick3/g9/glance/images/c4efd768-3b9a-44ab-9b91-5dcfc2989fc0 -rw-r----- 2 161 161 251985920 Aug 6 16:35 /rhs/brick3/g9/glance/images/c4efd768-3b9a-44ab-9b91-5dcfc2989fc0 [root@rhs-vm3 home]# ll /rhs/brick3/g11/glance/images/7ece2be5-2cd5-41c7-a9bc-be44eadb84b3 ---------T 2 161 161 0 Aug 6 16:56 /rhs/brick3/g11/glance/images/7ece2be5-2cd5-41c7-a9bc-be44eadb84b3 With eager-lock off: Volume Name: vol-glance Type: Distributed-Replicate Volume ID: 93bd85a9-1621-444e-8f0d-3122cfa86723 Status: Started Number of Bricks: 5 x 2 = 10 Transport-type: tcp Bricks: Brick1: 10.70.37.168:/rhs/brick3/g1 Brick2: 10.70.37.74:/rhs/brick3/g2 Brick3: 10.70.37.220:/rhs/brick3/g3 Brick4: 10.70.37.203:/rhs/brick3/g4 Brick5: 10.70.37.168:/rhs/brick3/g5 Brick6: 10.70.37.74:/rhs/brick3/g6 Brick7: 10.70.37.220:/rhs/brick3/g7 Brick8: 10.70.37.203:/rhs/brick3/g8 Brick9: 10.70.37.220:/rhs/brick3/g11 Brick10: 10.70.37.203:/rhs/brick3/g12 Options Reconfigured: performance.quick-read: off performance.read-ahead: off performance.io-cache: off performance.stat-prefetch: off cluster.eager-lock: off network.remote-dio: enable storage.owner-gid: 161 storage.owner-uid: 161 Tested remove-brick on file : 7ece2be5-2cd5-41c7-a9bc-be44eadb84b3 [root@rhs-vm1 home]# find / -name 7ece2be5-2cd5-41c7-a9bc-be44eadb84b3 /rhs/brick3/g1/glance/images/7ece2be5-2cd5-41c7-a9bc-be44eadb84b3 [root@rhs-vm1 home]# gluster v remove-brick vol-glance 10.70.37.168:/rhs/brick3/g1 10.70.37.74:/rhs/brick3/g2 start volume remove-brick start: success ID: f63be36c-1c45-4835-b10c-bc4d784c5001 [root@rhs-vm1 home]# gluster v remove-brick vol-glance 10.70.37.168:/rhs/brick3/g1 10.70.37.74:/rhs/brick3/g2 status Node Rebalanced-files size scanned failures skipped status run-time in secs --------- ----------- ----------- ----------- ----------- ----------- ------------ -------------- localhost 0 0Bytes 0 0 0 completed 0.00 10.70.37.220 0 0Bytes 0 0 0 not started 0.00 10.70.37.203 0 0Bytes 0 0 0 not started 0.00 10.70.37.74 0 0Bytes 2 0 0 completed 0.00 [root@rhs-vm1 home]# ll /rhs/brick3/g1/glance/images/7ece2be5-2cd5-41c7-a9bc-be44eadb84b3 -rw-r----- 2 161 161 251985920 Aug 6 16:35 /rhs/brick3/g1/glance/images/7ece2be5-2cd5-41c7-a9bc-be44eadb84b3 [root@rhs-vm3 home]# ll /rhs/brick3/g11/glance/images/7ece2be5-2cd5-41c7-a9bc-be44eadb84b3 ---------T 2 161 161 0 Aug 6 16:56 /rhs/brick3/g11/glance/images/7ece2be5-2cd5-41c7-a9bc-be44eadb84b3 As seen above, both the tests yield the same result. The file is still found in the orginal brick that is removed. Tested on distribute volume 6X2. The files are successfully migrated after rebalance. Issue found only on distribute-replicate volumes. (In reply to shilpa from comment #3) > Tested on a distribute volume. The files are successfully migrated after > rebalance. Issue found only on distribute-replicate volumes. Continuing tests on distribute-replicate volume. With the gluster volume for glance unmounted on the Openstack client, rebalance seems to work. [root@rhs-client40 cinder(keystone_admin)]# umount /mnt/gluster [root@rhs-vm1 brick1]# gluster v remove-brick glance-vol 10.70.37.220:/rhs/brick1/g7 10.70.37.203:/rhs/brick1/g8 start volume remove-brick start: success ID: cb4a59ad-8888-45ea-9740-a0079c1a8efa [root@rhs-vm1 brick1]# gluster v remove-brick glance-vol 10.70.37.220:/rhs/brick1/g7 10.70.37.203:/rhs/brick1/g8 status Node Rebalanced-files size scanned failures skipped status run-time in secs --------- ----------- ----------- ----------- ----------- ----------- ------------ -------------- localhost 0 0Bytes 0 0 0 not started 0.00 10.70.37.220 3 1.4GB 5 0 0 completed 51.00 10.70.37.203 0 0Bytes 4 0 0 completed 1.00 10.70.37.74 0 0Bytes 0 0 0 not started 0.00 sosreports of RHS nodes and openstack node in http://rhsqe-repo.lab.eng.blr.redhat.com/sosreports/993119/. Verified in glusterfs-3.4.0.19rhs-1. Tested with a new volume vol-glance on file d31aca1c-d462-4241-9c4f-e0550466cedb. Volume Name: vol-glance Type: Distributed-Replicate Volume ID: 1f37f298-1563-4df7-844d-6953685ae3ff Status: Started Number of Bricks: 6 x 2 = 12 Transport-type: tcp Bricks: Brick1: 10.70.37.168:/rhs/brick3/g1 Brick2: 10.70.37.74:/rhs/brick3/g2 Brick3: 10.70.37.220:/rhs/brick3/g3 Brick4: 10.70.37.203:/rhs/brick3/g4 Brick5: 10.70.37.168:/rhs/brick3/g5 Brick6: 10.70.37.74:/rhs/brick3/g6 Brick7: 10.70.37.220:/rhs/brick3/g7 Brick8: 10.70.37.203:/rhs/brick3/g8 Brick9: 10.70.37.168:/rhs/brick3/g9 Brick10: 10.70.37.74:/rhs/brick3/g10 Brick11: 10.70.37.220:/rhs/brick3/g11 Brick12: 10.70.37.203:/rhs/brick3/g12 Options Reconfigured: storage.owner-uid: 161 storage.owner-gid: 161 network.remote-dio: enable cluster.eager-lock: enable performance.stat-prefetch: off performance.io-cache: off performance.read-ahead: off performance.quick-read: off [root@rhs-vm3 brick3]# find /rhs -name d31aca1c-d462-4241-9c4f-e0550466cedb /rhs/brick3/g7/glance/images/d31aca1c-d462-4241-9c4f-e0550466cedb [root@rhs-vm3 brick3]# gluster v remove-brick vol-glance 10.70.37.220:/rhs/brick3/g7 10.70.37.203:/rhs/brick3/g8 start volume remove-brick start: success ID: c0218192-d977-4984-bd4f-cf672eda3089 [root@rhs-vm3 brick3]# gluster v remove-brick vol-glance 10.70.37.220:/rhs/brick3/g7 10.70.37.203:/rhs/brick3/g8 stat Node Rebalanced-files size scanned failures skipped status run-time in secs --------- ----------- ----------- ----------- ----------- ----------- ------------ -------------- localhost 1 892.6MB 2 0 0 completed 17.00 10.70.37.203 0 0Bytes 1 0 0 completed 0.00 10.70.37.168 0 0Bytes 0 0 0 not started 0.00 10.70.37.74 0 0Bytes 0 0 0 not started 0.00 File successfully migrated to brick g11: [root@rhs-vm3 brick3]# find /rhs -name d31aca1c-d462-4241-9c4f-e0550466cedb /rhs/brick3/g11/glance/images/d31aca1c-d462-4241-9c4f-e0550466cedb [root@rhs-vm3 brick3]# gluster v remove-brick vol-glance 10.70.37.220:/rhs/brick3/g7 10.70.37.203:/rhs/brick3/g8 commit Removing brick(s) can result in data loss. Do you want to Continue? (y/n) y volume remove-brick commit: success Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. http://rhn.redhat.com/errata/RHBA-2013-1262.html |