Bug 903476
| Summary: | DHT: - If brick/Sub-volume is down then any attempt to create/access/modify file which is hashed and cached on down sub-volume should give error “ Transport endpoint is not connected” (rather than “No such file or directory” or "Input/output error") | ||||||
|---|---|---|---|---|---|---|---|
| Product: | [Red Hat Storage] Red Hat Gluster Storage | Reporter: | Rachana Patel <racpatel> | ||||
| Component: | glusterfs | Assignee: | shishir gowda <sgowda> | ||||
| Status: | CLOSED ERRATA | QA Contact: | amainkar | ||||
| Severity: | medium | Docs Contact: | |||||
| Priority: | medium | ||||||
| Version: | 2.0 | CC: | nsathyan, rhs-bugs, vbellur | ||||
| Target Milestone: | --- | Keywords: | Reopened | ||||
| Target Release: | --- | ||||||
| Hardware: | x86_64 | ||||||
| OS: | Linux | ||||||
| Whiteboard: | |||||||
| Fixed In Version: | Doc Type: | Bug Fix | |||||
| Doc Text: | Story Points: | --- | |||||
| Clone Of: | Environment: | ||||||
| Last Closed: | 2013-09-23 22:29:52 UTC | Type: | Bug | ||||
| Regression: | --- | Mount Type: | --- | ||||
| Documentation: | --- | CRM: | |||||
| Verified Versions: | Category: | --- | |||||
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |||||
| Cloudforms Team: | --- | Target Upstream Version: | |||||
| Embargoed: | |||||||
| Attachments: |
|
||||||
|
Description
Rachana Patel
2013-01-24 05:07:18 UTC
Created attachment 686454 [details]
log
Per Feb-06 bug triage meeting, targeting for 2.1.0. Per Feb-06 bug triage meeting, targeting for 2.1.0. Fixed as part of bug 893378 (http://review.gluster.org/#change,4383). *** This bug has been marked as a duplicate of bug 893378 *** able to reproduce in :- 3.4.0.8rhs-1.el6rhs.x86_64 server:- [root@mia ~]# gluster v status test1 Status of volume: test1 Gluster process Port Online Pid ------------------------------------------------------------------------------ Brick fred.lab.eng.blr.redhat.com:/rhs/brick1/t1 49154 Y32380 Brick mia.lab.eng.blr.redhat.com:/rhs/brick1/t1 N/A N11173 Brick cutlass.lab.eng.blr.redhat.com:/rhs/brick1/t1 49154 Y8989 NFS Server on localhost 2049 Y11183 NFS Server on c5154da1-be15-40e2-b5f3-9be6dadafd43 2049 Y8999 NFS Server on a37ff566-da82-4ae4-90c6-17763466fd36 2049 Y15188 NFS Server on 292b158a-7650-4e09-9bc0-71e392f0d0c1 2049 Y32390 There are no active volume tasks [root@cutlass ~]# ls -l /rhs/brick1/t1/newf52 ls: cannot access /rhs/brick1/t1/newf52: No such file or directory [root@mia ~]# ls -l /rhs/brick1/t1/newf52 -rw-r--r-- 2 root root 0 Jun 4 02:14 /rhs/brick1/t1/newf52 [root@fred ~]# ls -l /rhs/brick1/t1/newf52 ls: cannot access /rhs/brick1/t1/newf52: No such file or directory for newf52 file, hashed and cached sub-vol is down on mount point:- [root@rhsauto037 test1nfs]# touch newf52 touch: cannot touch `newf52': Input/output error [root@rhsauto037 test1nfs]# cp file109 newf52 cp: cannot create regular file `newf52': Input/output error Expected results: If brick/Sub-volume is down then any attempt to access/modify file which is hashed and cached on down sub-volume should give error “ Transport endpoint is not connected” Looks like NFS is converting ENOTCONN error to EIO. The failures are not seen on fuse clients:
Create related:
[2013-06-10 06:33:06.002267] W [client-rpc-fops.c:2058:client3_3_create_cbk] 0-sng-client-0: remote oper
ation failed: Transport endpoint is not connected. Path: /new3
[2013-06-10 06:33:06.002307] W [nfs3.c:2354:nfs3svc_create_cbk] 0-nfs: 7f08d02a: /new3 => -1 (Transport
endpoint is not connected) <========ENOTCONN error
[2013-06-10 06:33:06.002356] W [nfs3-helpers.c:3460:nfs3_log_newfh_res] 0-nfs-nfsv3: XID: 7f08d02a, CREA
TE: NFS: 5(I/O error), POSIX: 107(Transport endpoint is not connected), FH: exportid 00000000-0000-0000-
0000-000000000000, gfid 00000000-0000-0000-0000-000000000000 <=====EIO error
Rename related:
[2013-06-10 06:37:01.918575] W [nfs3.c:3663:nfs3svc_rename_cbk] 0-nfs: a108d02a: rename /new1 -> /new3 =
> -1 (Transport endpoint is not connected) <==========ENOTCONN error
[2013-06-10 06:37:01.918615] W [nfs3-helpers.c:3391:nfs3_log_common_res] 0-nfs-nfsv3: XID: a108d02a, REN
AME: NFS: 5(I/O error), POSIX: 14(Bad address) <=========EIO error
BZ 903476 and BZ 860915 are similar, so marking this as duplicate of 860915. Will update the analysis in BZ 860915. *** This bug has been marked as a duplicate of bug 860915 *** I must say, detailed information with necessary links helped a lot in understanding the root cause(comment #11 of bug 860915). Thanks, Santosh. Note:- 1) removing duplicate - bug 860915. reason - agreed that root cause is same but as steps are different. one is about Dir creation and other is file acess/modify/creation. (as per Amar's mail to storage-eng 'Important: Steps to marking a bug as duplicate' , date - May 13, 2013 ) 2) opening new bug for documentation and resigning this defect to Shishir reason - This problem has been fixed in DHT - FUSE mount. So this defect can be used to track that problem in future and for NFS mount would open new defect and assigned it to Doc team 3) As this defect is not reproducible with latest build(3.4.0.9rhs-1.el6.x86_64) on Fuse mount marking it as verified. Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. http://rhn.redhat.com/errata/RHBA-2013-1262.html |