Bug 893378
| Summary: | DHT - User is able to modify file when cached sub-volume is down and hashed sub-volume is up it results in data lost and at same level multiple files can be created having same name | ||
|---|---|---|---|
| Product: | [Red Hat Storage] Red Hat Gluster Storage | Reporter: | Rachana Patel <racpatel> |
| Component: | glusterfs | Assignee: | Kaushal <kaushal> |
| Status: | CLOSED ERRATA | QA Contact: | amainkar |
| Severity: | high | Docs Contact: | |
| Priority: | high | ||
| Version: | 2.0 | CC: | amarts, nsathyan, racpatel, rhs-bugs, sdharane, shaines, vbellur |
| Target Milestone: | --- | ||
| Target Release: | --- | ||
| Hardware: | x86_64 | ||
| OS: | Linux | ||
| Whiteboard: | |||
| Fixed In Version: | glusterfs-3.4.0qa8, glusterfs-3.3.0.5rhs-42 | Doc Type: | Bug Fix |
| Doc Text: | Story Points: | --- | |
| Clone Of: | Environment: | ||
| Last Closed: | 2013-09-23 22:34:48 UTC | Type: | Bug |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
|
Description
Rachana Patel
2013-01-09 09:41:58 UTC
Upstream fix http://review.gluster.org/#change,4383 in review CHANGE: http://review.gluster.org/4383 (cluster/distribute: If cached_subvol is down, return ENOTCONN in lookup) merged in master by Anand Avati (avati) *** Bug 903917 has been marked as a duplicate of this bug. *** *** Bug 903476 has been marked as a duplicate of this bug. *** verified with 3.3.0.6rhs-4.el6rhs.x86_64 bug 893378 and bug 903917 working as per expectation but Bug 903476 - not working as per expectation e.g. [root@rhsauto037 new]# cat renamefile18 cat: renamefile18: No such file or directory [root@rhsauto037 new]# cp renamefile18 abc cp: cannot stat `renamefile18': No such file or directory [root@rhsauto037 new]# ls -l renamefile18 ls: cannot access renamefile18: No such file or directory [root@rhsauto037 new]# chmod 777 f1 chmod: cannot access `f1': No such file or directory hence moving back to assigned sorry for the inconvenience caused, logs are attached to the bug. After investigating the logs, it looked like a issue fixed in bug 884379. Updated the release to [root@localhost ~]# rpm -qa |grep glusterfs glusterfs-devel-3.3.0.6rhs-6.el6rhs.x86_64 glusterfs-3.3.0.6rhs-6.el6rhs.x86_64 glusterfs-server-3.3.0.6rhs-6.el6rhs.x86_64 glusterfs-geo-replication-3.3.0.6rhs-6.el6rhs.x86_64 glusterfs-debuginfo-3.3.0.6rhs-6.el6rhs.x86_64 glusterfs-fuse-3.3.0.6rhs-6.el6rhs.x86_64 glusterfs-rdma-3.3.0.6rhs-6.el6rhs.x86_64 1. Create a 3 - dht volume, mount and rename file till we get a linkfile [root@localhost export]# mount -t glusterfs localhost:/test /mnt/dht/ [root@localhost export]# cd /mnt/dht/ [root@localhost dht]# ls [root@localhost dht]# touch 1 [root@localhost dht]# mv 1 2 [root@localhost dht]# ls -l /export/* /export/sub1: total 0 ---------T. 2 root root 0 Mar 21 03:28 2 /export/sub2: total 0 -rw-r--r--. 2 root root 0 Mar 21 03:28 2 /export/sub3: total 0 2. kill brick, and check status [root@localhost ~]# gluster volume status Status of volume: test Gluster process Port Online Pid ------------------------------------------------------------------------------ Brick vm1:/export/sub1 24012 Y 29771 Brick vm1:/export/sub2 24013 N N/A Brick vm1:/export/sub3 24014 Y 29784 NFS Server on localhost 38467 Y 29790 3. Try to perform ops on the file. [root@localhost dht]# cat 2 cat: 2: Transport endpoint is not connected [root@localhost dht]# rm 2 rm: cannot remove `2': Transport endpoint is not connected [root@localhost dht]# ls -l 2 ls: cannot access 2: Transport endpoint is not connected [root@localhost dht]# mv 2 3 mv: cannot stat `2': Transport endpoint is not connected Can you please rerun the test and check if the issue is fixed. Per 04-10-2013 Storage bug triage meeting, targeting for Big Bend. I'm not able to reproduce this bug with the glusterfs-3.4.0.9rhs-1.el6rhs. [root@localhost mnt]# gluster volume info test Volume Name: test Type: Distribute Volume ID: 713ad1ed-96ca-459b-8728-0209439b972f Status: Started Number of Bricks: 2 Transport-type: tcp Bricks: Brick1: 10.70.42.223:/brick/test1 Brick2: 10.70.42.223:/brick/test2 [root@localhost mnt]# mount | grep glusterfs localhost:test on /mnt type fuse.glusterfs (rw,default_permissions,allow_other,max_read=131072) [root@localhost mnt]# touch file [root@localhost mnt]# ls -l /brick/* /brick/test1: total 0 /brick/test2: total 0 -rw-r--r-- 2 root root 0 Jun 12 13:07 file [root@localhost mnt]# mv file file.rename [root@localhost mnt]# ls -l /brick/* /brick/test1: total 0 ---------T 2 root root 0 Jun 12 13:07 file.rename /brick/test2: total 0 -rw-r--r-- 2 root root 0 Jun 12 13:07 file.rename [root@localhost mnt]# gluster volume status test Status of volume: test Gluster process Port Online Pid ------------------------------------------------------------------------------ Brick 10.70.42.223:/brick/test1 49152 Y 28403 Brick 10.70.42.223:/brick/test2 49153 Y 29028 NFS Server on localhost 2049 Y 29039 There are no active volume tasks [root@localhost mnt]# kill 29028 [root@localhost mnt]# gluster volume status test Status of volume: test Gluster process Port Online Pid ------------------------------------------------------------------------------ Brick 10.70.42.223:/brick/test1 49152 Y 28403 Brick 10.70.42.223:/brick/test2 N/A N N/A NFS Server on localhost 2049 Y 29039 There are no active volume tasks [root@localhost mnt]# ls [root@localhost mnt]# touch file.rename touch: cannot touch `file.rename': Transport endpoint is not connected [root@localhost mnt]# cat file.rename cat: file.rename: Transport endpoint is not connected [root@localhost mnt]# echo "hello" > file.rename -bash: file.rename: Transport endpoint is not connected Can you rerun this test and confirm the same? Requesting QE to run the test as mentioned by Kaushal above. As mentioned in comment #7 of this bug " verified with 3.3.0.6rhs-4.el6rhs.x86_64 bug 893378 and bug 903917 working as per expectation but Bug 903476 - not working as per expectation" this bug was fixed but it got opened as one of its duplicate Bug 903476 was not working as per expectation. As we have removed Bug 903476 from duplicate. we can mark this as verified.(bug 893378 and bug 903917 working as per expectation) Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. http://rhn.redhat.com/errata/RHBA-2013-1262.html |