Bug 1396166
Summary: | self-heal info command hangs after triggering self-heal | ||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Product: | [Red Hat Storage] Red Hat Gluster Storage | Reporter: | SATHEESARAN <sasundar> | ||||||||||||||
Component: | replicate | Assignee: | Krutika Dhananjay <kdhananj> | ||||||||||||||
Status: | CLOSED ERRATA | QA Contact: | SATHEESARAN <sasundar> | ||||||||||||||
Severity: | high | Docs Contact: | |||||||||||||||
Priority: | unspecified | ||||||||||||||||
Version: | rhgs-3.2 | CC: | amukherj, nchilaka, rhs-bugs, sabose, sasundar, storage-qa-internal | ||||||||||||||
Target Milestone: | --- | ||||||||||||||||
Target Release: | RHGS 3.2.0 | ||||||||||||||||
Hardware: | x86_64 | ||||||||||||||||
OS: | Linux | ||||||||||||||||
Whiteboard: | |||||||||||||||||
Fixed In Version: | glusterfs-3.8.4-6 | Doc Type: | If docs needed, set a value | ||||||||||||||
Doc Text: | Story Points: | --- | |||||||||||||||
Clone Of: | |||||||||||||||||
: | 1398566 (view as bug list) | Environment: |
RHV-RHGS HCI
|
||||||||||||||
Last Closed: | 2017-03-23 06:19:44 UTC | Type: | Bug | ||||||||||||||
Regression: | --- | Mount Type: | --- | ||||||||||||||
Documentation: | --- | CRM: | |||||||||||||||
Verified Versions: | Category: | --- | |||||||||||||||
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |||||||||||||||
Cloudforms Team: | --- | Target Upstream Version: | |||||||||||||||
Embargoed: | |||||||||||||||||
Bug Depends On: | |||||||||||||||||
Bug Blocks: | 1277939, 1351528, 1398566, 1398888 | ||||||||||||||||
Attachments: |
|
Description
SATHEESARAN
2016-11-17 16:11:45 UTC
I have tested this with qemu's native driver for glusterfs ( which uses gfapi ) Created attachment 1221739 [details]
Client statedump taken from qemu process of VM1 using gdb
Created attachment 1221740 [details]
Client statedump taken from qemu process of VM2 using gdb
Created attachment 1221741 [details]
clients logs from VM1
Created attachment 1221742 [details]
client logs from VM2
You do have the brick statedump too, don't you? Could you please attach those as well? -Krutika (In reply to Krutika Dhananjay from comment #7) > You do have the brick statedump too, don't you? Could you please attach > those as well? > > -Krutika Hi Krutika, I have mistakenly re-provisioned my third server in the cluster to simulate failed node scenario. But I have brick statedump from server1 and server2. I will attach them Created attachment 1223015 [details]
brick1-statedump
Created attachment 1223016 [details]
brick2-statedump
As per the triaging we all have the agreement that this BZ has to be fixed in rhgs-3.2.0. Providing devel_ack. patch on master posted for review at http://review.gluster.org/15929 Moving this bug to POST state. https://code.engineering.redhat.com/gerrit/#/c/91332/1 <-- that's the downstream patch. Waiting on QE and PM ack before asking for it to be merged. Tested with glusterfs-3.8.4-10.el7rhgs with the following steps: 1. Created replica 3 sharded volume with compound-fops enabled 2. Optimized the volume for VM store usecase and fuse mounted the volume on the hypervisor 3. Created a sparse image file on the VM and started the OS installation. 4. While VM installation is in progress, killed the first brick 5. After VM installation is completed, brought back the brick and initiated heal on that volume 'gluster volume heal <vol>' 6. Checked for heal status using 'gluster volume heal <vol> info' 'gluster volume heal info' listed the entries that were pending to heal Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://rhn.redhat.com/errata/RHSA-2017-0486.html |