Hide Forgot
Description of problem: ======================= On a tiered volume having in total about 100GB data, detach tier was issued. The detach tier status shows as completed on local host node, after about 2 hours, but doesnt complete at all on other nodes. Infact it doesnt even show a single file as migrated on the other nodes. After even completion of a day, it was still in same state and crash was found Version-Release number of selected component (if applicable): ==== 3.7.5-14
Please provide gluster vol info details and sos-reports for the systems on which this was seen.
Need steps to reproduce this problem.
Hi Dan, I have seen this atleast another 2 time on 3.5.7-16 steps: 1)create a disperse- distribute (ec) volume 2)Now mount volume on two or more fuse clients 3) enabled quota 4) start IOs, on one untar lin-kernel on another dd create command for about 300MB files in loop for about 50 files 5) attach tier 6)now if previous IOs were complete, then re-start the same on different dir 7)while IOs are going on do a detach tier start 8)it can be seen that the detach tier status shows as complete on local node, but in progrss on other nodes. Also, the vol status tasks, show detach as in progresss
sosreports @ [nchilaka@rhsqe-repo nchilaka]$ pwd /home/repo/sosreports/nchilaka/bug.1300301
I tried to reproduce the issue using the steps given in comment4. I couldn't reproduce it on my system for 2 times. My setup. 4 node server 2 clients Volume Name: patchy Type: Tier Volume ID: 5641dc88-58eb-44c6-b848-54795b32ed9c Status: Started Number of Bricks: 10 Transport-type: tcp Hot Tier : Hot Tier Type : Distributed-Replicate Number of Bricks: 2 x 2 = 4 Brick1: 10.70.42.212:/home/brick2/h2 Brick2: 10.70.43.110:/home/brick2/h1 Brick3: 10.70.43.100:/home/brick2/h1 Brick4: 10.70.42.212:/home/brick2/h1 Cold Tier: Cold Tier Type : Disperse Number of Bricks: 1 x (4 + 2) = 6 Brick5: 10.70.43.148:/home/brick1/c1 Brick6: 10.70.42.212:/home/brick1/c1 Brick7: 10.70.43.100:/home/brick1/c1 Brick8: 10.70.43.110:/home/brick1/c1 Brick9: 10.70.43.148:/home/brick1/c2 Brick10: 10.70.42.212:/home/brick1/c2 I will try couple of times more.
Additional info here: The crash reported in this BZ is the same as the one tracked by BZ# 1294774.
Based on comment 8 , changing the description.
karthick can you kindly check if this is happening
The issue reported is not seen in 3.1.3 builds. validated this in build - glusterfs-3.7.9-5.el7rhgs.x86_64
Per comment 12, Can we close this?
clearing stale needinfos.