Bug 1185173 - AFR: 3-way-replication: Transport point not connected error messaged not displayed when one of the replica pair is down
Summary: AFR: 3-way-replication: Transport point not connected error messaged not disp...
Keywords:
Status: CLOSED CURRENTRELEASE
Alias: None
Product: Red Hat Gluster Storage
Classification: Red Hat Storage
Component: distribute
Version: rhgs-3.0
Hardware: x86_64
OS: Linux
unspecified
high
Target Milestone: ---
: ---
Assignee: Satish Mohan
QA Contact: Anil Shah
URL:
Whiteboard:
Depends On:
Blocks: 1279705 1285829 1308413
TreeView+ depends on / blocked
 
Reported: 2015-01-23 06:44 UTC by Anil Shah
Modified: 2017-01-02 15:21 UTC (History)
5 users (show)

Fixed In Version:
Doc Type: Bug Fix
Doc Text:
Clone Of:
: 1279705 1285829 (view as bug list)
Environment:
Last Closed: 2017-01-02 15:21:42 UTC
Embargoed:


Attachments (Terms of Use)

Description Anil Shah 2015-01-23 06:44:12 UTC
Description of problem:

When one of the replica pair is down, creating file from mount point should report transport point not connected error message 

Version-Release number of selected component (if applicable):

glusterfs 3.6.0.42

How reproducible:

100%

Steps to Reproduce:
1.create 2x3 distribute replicate volume
2.kill one of the replica pair of the volume
3.create files from mount point

Actual results:

[root@client glusterfs]# touch file{1..100}
touch: cannot touch `file3': No such file or directory
touch: cannot touch `file4': No such file or directory
touch: cannot touch `file7': No such file or directory

===============================
Mount logs /var/log/glusterfs/mnt-glusterfs-.log

[2015-01-23 11:52:17.053399] W [dht-layout.c:180:dht_layout_search] 0-testvol-dht: no subvolume for hash (value) = 1401870546
[2015-01-23 11:52:17.055207] W [dht-layout.c:180:dht_layout_search] 0-testvol-dht: no subvolume for hash (value) = 1401870546
[2015-01-23 11:52:17.055322] W [fuse-bridge.c:1937:fuse_create_cbk] 0-glusterfs-fuse: 13025: /file91 => -1 (No such file or directory)
[2015-01-23 11:52:17.055600] W [dht-layout.c:180:dht_layout_search] 0-testvol-dht: no subvolume for hash (value) = 1401870546
[2015-01-23 11:52:17.057228] W [dht-layout.c:180:dht_layout_search] 0-testvol-dht: no subvolume for hash (value) = 1401870546
[2015-01-23 11:52:17.080706] W [dht-layout.c:180:dht_layout_search] 0-testvol-dht: no subvolume for hash (value) = 1742334698
[2015-01-23 11:52:17.082387] W [dht-layout.c:180:dht_layout_search] 0-testvol-dht: no subvolume for hash (value) = 1742334698
[2015-01-23 11:52:17.084435] W [dht-layout.c:180:dht_layout_search] 0-testvol-dht: no subvolume for hash (value) = 1742334698
[2015-01-23 11:52:17.086271] W [dht-layout.c:180:dht_layout_search] 0-testvol-dht: no subvolume for hash (value) = 1742334698
[2015-01-23 11:52:17.086388] W [fuse-bridge.c:1937:fuse_create_cbk] 0-glusterfs-fuse: 13034: /file93 => -1 (No such file or directory)
[2015-01-23 11:52:17.086648] W [dht-layout.c:180:dht_layout_search] 0-testvol-dht: no subvolume for hash (value) = 1742334698
[2015-01-23 11:52:17.088220] W [dht-layout.c:180:dht_layout_search] 0-testvol-dht: no subvolume for hash (value) = 1742334698
[2015-01-23 11:52:17.156335] W [dht-layout.c:180:dht_layout_search] 0-testvol-dht: no subvolume for hash (value) = 1752000159
[2015-01-23 11:52:17.158209] W [dht-layout.c:180:dht_layout_search] 0-testvol-dht: no subvolume for hash (value) = 1752000159
[2015-01-23 11:52:17.160406] W [dht-layout.c:180:dht_layout_search] 0-testvol-dht: no subvolume for hash (value) = 1752000159
[2015-01-23 11:52:17.162200] W [dht-layout.c:180:dht_layout_search] 0-testvol-dht: no subvolume for hash (value) = 1752000159
[2015-01-23 11:52:17.162299] W [fuse-bridge.c:1937:fuse_create_cbk] 0-glusterfs-fuse: 13055: /file97 => -1 (No such file or directory)
[2015-01-23 11:52:17.162625] W [dht-layout.c:180:dht_layout_search] 0-testvol-dht: no subvolume for hash (value) = 1752000159
[2015-01-23 11:52:17.164215] W [dht-layout.c:180:dht_layout_search] 0-testvol-dht: no subvolume for hash (value) = 1752000159
[2015-01-23 11:52:17.209093] W [dht-layout.c:180:dht_layout_search] 0-testvol-dht: no subvolume for hash (value) = 1194566581
[2015-01-23 11:52:17.211148] W [dht-layout.c:180:dht_layout_search] 0-testvol-dht: no subvolume for hash (value) = 1194566581
[2015-01-23 11:52:17.213408] W [dht-layout.c:180:dht_layout_search] 0-testvol-dht: no subvolume for hash (value) = 1194566581
[2015-01-23 11:52:17.215112] W [dht-layout.c:180:dht_layout_search] 0-testvol-dht: no subvolume for hash (value) = 1194566581
[2015-01-23 11:52:17.215210] W [fuse-bridge.c:1937:fuse_create_cbk] 0-glusterfs-fuse: 13070: /file100 => -1 (No such file or directory)

=============================================================

Expected results:

Transport point not connected message should be displayed for the down replica pair 

Additional info:

Volume Name: testvol
Type: Distributed-Replicate
Volume ID: 7e154f4b-7ded-42f0-9a02-2d0ff734414b
Status: Started
Snap Volume: no
Number of Bricks: 2 x 3 = 6
Transport-type: tcp
Bricks:
Brick1: 10.70.47.143:/rhs/brick1/b1
Brick2: 10.70.47.145:/rhs/brick1/b2
Brick3: 10.70.47.150:/rhs/brick1/b3
Brick4: 10.70.47.151:/rhs/brick1/b4
Brick5: 10.70.47.143:/rhs/brick2/b5
Brick6: 10.70.47.145:/rhs/brick2/b6
Options Reconfigured:
performance.readdir-ahead: on
auto-delete: disable
snap-max-soft-limit: 90
snap-max-hard-limit: 256

==================================================

[root@node1 ~]# gluster v status testvol
Status of volume: testvol
Gluster process						Port	Online	Pid
------------------------------------------------------------------------------
Brick 10.70.47.143:/rhs/brick1/b1			N/A	N	N/A
Brick 10.70.47.145:/rhs/brick1/b2			N/A	N	N/A
Brick 10.70.47.150:/rhs/brick1/b3			N/A	N	N/A
Brick 10.70.47.151:/rhs/brick1/b4			49152	Y	7687
Brick 10.70.47.143:/rhs/brick2/b5			49153	Y	2463
Brick 10.70.47.145:/rhs/brick2/b6			49153	Y	3275
NFS Server on localhost					2049	Y	3047
Self-heal Daemon on localhost				N/A	Y	3004
NFS Server on 10.70.47.145				2049	Y	3762
Self-heal Daemon on 10.70.47.145			N/A	Y	3710
NFS Server on 10.70.47.151				2049	Y	8169
Self-heal Daemon on 10.70.47.151			N/A	Y	8108
NFS Server on 10.70.47.150				2049	Y	23171
Self-heal Daemon on 10.70.47.150			N/A	Y	23093
 
Task Status of Volume testvol
------------------------------------------------------------------------------
There are no active volume tasks

Comment 4 Anil Shah 2016-07-22 10:02:32 UTC
[root@dhcp47-158 .unsupported]# gluster  v status testvol
Status of volume: testvol
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick 10.70.47.158:/rhs/brick1/b1           N/A       N/A        N       N/A  
Brick 10.70.47.188:/rhs/brick1/b2           N/A       N/A        N       N/A  
Brick 10.70.47.170:/rhs/brick1/b3           N/A       N/A        N       N/A  
Brick 10.70.47.162:/rhs/brick1/b4           49152     0          Y       13517
Brick 10.70.47.158:/rhs/brick2/b5           49153     0          Y       15420
Brick 10.70.47.188:/rhs/brick2/b6           49153     0          Y       14248
NFS Server on localhost                     2049      0          Y       16451
NFS Server on 10.70.47.170                  2049      0          Y       16191
NFS Server on 10.70.47.162                  2049      0          Y       14327
NFS Server on 10.70.47.188                  2049      0          Y       15021
 


[root@dhcp47-190 fuse]# touch files{1..100}
touch: cannot touch ‘files1’: Transport endpoint is not connected
touch: cannot touch ‘files2’: Transport endpoint is not connected
touch: cannot touch ‘files5’: Transport endpoint is not connected
touch: cannot touch ‘files6’: Transport endpoint is not connected
touch: cannot touch ‘files10’: Transport endpoint is not connected


Bug verified on build glusterfs-3.7.9-10.el7rhgs.x86_64


Note You need to log in before you can comment on or make changes to this bug.