| Summary: | Detach tier fails to complete, on non local hosts | ||
|---|---|---|---|
| Product: | Red Hat Gluster Storage | Reporter: | Nag Pavan Chilakam <nchilaka> |
| Component: | tier | Assignee: | Dan Lambright <dlambrig> |
| Status: | CLOSED NOTABUG | QA Contact: | Bala Konda Reddy M <bmekala> |
| Severity: | urgent | Docs Contact: | |
| Priority: | urgent | ||
| Version: | rhgs-3.1 | CC: | kramdoss, nbalacha, nchilaka, rcyriac, rhs-bugs, rkavunga, sankarshan, smohan |
| Target Milestone: | --- | Keywords: | ZStream |
| Target Release: | --- | ||
| Hardware: | Unspecified | ||
| OS: | Unspecified | ||
| Whiteboard: | tier-attach-detach | ||
| Fixed In Version: | Doc Type: | Bug Fix | |
| Doc Text: | Story Points: | --- | |
| Clone Of: | Environment: | ||
| Last Closed: | 2016-08-18 12:44:33 UTC | Type: | Bug |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
|
Description
Nag Pavan Chilakam
2016-01-18 06:23:19 UTC
Please provide gluster vol info details and sos-reports for the systems on which this was seen. Need steps to reproduce this problem. Hi Dan, I have seen this atleast another 2 time on 3.5.7-16 steps: 1)create a disperse- distribute (ec) volume 2)Now mount volume on two or more fuse clients 3) enabled quota 4) start IOs, on one untar lin-kernel on another dd create command for about 300MB files in loop for about 50 files 5) attach tier 6)now if previous IOs were complete, then re-start the same on different dir 7)while IOs are going on do a detach tier start 8)it can be seen that the detach tier status shows as complete on local node, but in progrss on other nodes. Also, the vol status tasks, show detach as in progresss sosreports @ [nchilaka@rhsqe-repo nchilaka]$ pwd /home/repo/sosreports/nchilaka/bug.1300301 I tried to reproduce the issue using the steps given in comment4. I couldn't reproduce it on my system for 2 times. My setup. 4 node server 2 clients Volume Name: patchy Type: Tier Volume ID: 5641dc88-58eb-44c6-b848-54795b32ed9c Status: Started Number of Bricks: 10 Transport-type: tcp Hot Tier : Hot Tier Type : Distributed-Replicate Number of Bricks: 2 x 2 = 4 Brick1: 10.70.42.212:/home/brick2/h2 Brick2: 10.70.43.110:/home/brick2/h1 Brick3: 10.70.43.100:/home/brick2/h1 Brick4: 10.70.42.212:/home/brick2/h1 Cold Tier: Cold Tier Type : Disperse Number of Bricks: 1 x (4 + 2) = 6 Brick5: 10.70.43.148:/home/brick1/c1 Brick6: 10.70.42.212:/home/brick1/c1 Brick7: 10.70.43.100:/home/brick1/c1 Brick8: 10.70.43.110:/home/brick1/c1 Brick9: 10.70.43.148:/home/brick1/c2 Brick10: 10.70.42.212:/home/brick1/c2 I will try couple of times more. Additional info here: The crash reported in this BZ is the same as the one tracked by BZ# 1294774. Based on comment 8 , changing the description. karthick can you kindly check if this is happening The issue reported is not seen in 3.1.3 builds. validated this in build - glusterfs-3.7.9-5.el7rhgs.x86_64 Per comment 12, Can we close this? clearing stale needinfos. |