Bug 1311460 - unable to mount a glusterfs volume on clients
Summary: unable to mount a glusterfs volume on clients
Keywords:
Status: CLOSED EOL
Alias: None
Product: GlusterFS
Classification: Community
Component: quick-read
Version: 3.7.8
Hardware: x86_64
OS: Linux
unspecified
unspecified
Target Milestone: ---
Assignee: Atin Mukherjee
QA Contact:
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2016-02-24 10:03 UTC by bitchecker
Modified: 2017-03-08 10:50 UTC (History)
10 users (show)

Fixed In Version:
Doc Type: Bug Fix
Doc Text:
Clone Of:
Environment:
Last Closed: 2017-03-08 10:50:48 UTC
Regression: ---
Mount Type: ---
Documentation: ---
CRM:
Verified Versions:
Embargoed:


Attachments (Terms of Use)
/var/log/glusterfs/mnt-.log (14.24 KB, text/plain)
2016-02-24 10:03 UTC, bitchecker
no flags Details

Description bitchecker 2016-02-24 10:03:29 UTC
Created attachment 1130160 [details]
/var/log/glusterfs/mnt-.log

hi,
i've a glusterfs storage with two nodes that have a replicated volume.

server side it works properly:

# gluster volume heal volume info
Brick gluster01:/data/brick/volume
Number of entries: 0

Brick gluster02:/data/brick/volume
Number of entries: 0

# gluster volume info
 
Volume Name: volume
Type: Replicate
Volume ID: 3c6c72bf-f85d-4c4c-a536-19a75e2ae71b
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: gluster01:/data/brick/volume
Brick2: gluster02:/data/brick/volume
Options Reconfigured:
auth.allow: {list of ip}
performance.readdir-ahead: on


# gluster volume status                                                                                                            
Status of volume: volume
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick gluster01:/data/brick/volume      49152     0          Y       2116 
Brick gluster02:/data/brick/volume      49152     0          Y       1356 
NFS Server on localhost                     2049      0          Y       2254 
Self-heal Daemon on localhost               N/A       N/A        Y       2262 
NFS Server on gluster02                     2049      0          Y       1296 
Self-heal Daemon on gluster02               N/A       N/A        Y       1316 
 
Task Status of Volume volume
------------------------------------------------------------------------------
There are no active volume tasks


on clients i've the problem because i can't mount the volume:

# mount.glusterfs glusterfs02:/volume /mnt/
Mount failed. Please check the log file for more details.

the only log that i've is: /var/log/glusterfs/mnt-.log

i've attached the log.

Comment 2 Atin Mukherjee 2016-02-24 10:07:39 UTC
(In reply to bitchecker from comment #0)
> Created attachment 1130160 [details]
> /var/log/glusterfs/mnt-.log
> 
> hi,
> i've a glusterfs storage with two nodes that have a replicated volume.
> 
> server side it works properly:
> 
> # gluster volume heal volume info
> Brick gluster01:/data/brick/volume
> Number of entries: 0
> 
> Brick gluster02:/data/brick/volume
> Number of entries: 0
> 
> # gluster volume info
>  
> Volume Name: volume
> Type: Replicate
> Volume ID: 3c6c72bf-f85d-4c4c-a536-19a75e2ae71b
> Status: Started
> Number of Bricks: 1 x 2 = 2
> Transport-type: tcp
> Bricks:
> Brick1: gluster01:/data/brick/volume
> Brick2: gluster02:/data/brick/volume
> Options Reconfigured:
> auth.allow: {list of ip}
> performance.readdir-ahead: on
> 
> 
> # gluster volume status                                                     
> 
> Status of volume: volume
> Gluster process                             TCP Port  RDMA Port  Online  Pid
> -----------------------------------------------------------------------------
> -
> Brick gluster01:/data/brick/volume      49152     0          Y       2116 
> Brick gluster02:/data/brick/volume      49152     0          Y       1356 
> NFS Server on localhost                     2049      0          Y      
> 2254 
> Self-heal Daemon on localhost               N/A       N/A        Y      
> 2262 
> NFS Server on gluster02                     2049      0          Y      
> 1296 
> Self-heal Daemon on gluster02               N/A       N/A        Y      
> 1316 
>  
> Task Status of Volume volume
> -----------------------------------------------------------------------------
> -
> There are no active volume tasks
> 
> 
> on clients i've the problem because i can't mount the volume:
> 
> # mount.glusterfs glusterfs02:/volume /mnt/
shouldn't the address be gluster02?

> Mount failed. Please check the log file for more details.
> 
> the only log that i've is: /var/log/glusterfs/mnt-.log
> 
> i've attached the log.

Comment 3 RajeshReddy 2016-02-24 10:12:53 UTC
From the log it looks like all sub-volumes are down so mount failed,

[MSGID: 108006] [afr-common.c:4015:afr_notify] 0-volume-replicate-0: All subvolumes are down. Going offline until atleast one of them comes back up.

Comment 4 bitchecker 2016-02-24 11:02:35 UTC
(In reply to Atin Mukherjee from comment #2)
> shouldn't the address be gluster02?

addresses are fully resolvable.

Comment 5 bitchecker 2016-02-24 11:03:46 UTC
(In reply to RajeshReddy from comment #3)
> From the log it looks like all sub-volumes are down so mount failed,
> 
> [MSGID: 108006] [afr-common.c:4015:afr_notify] 0-volume-replicate-0: All
> subvolumes are down. Going offline until atleast one of them comes back up.

i've seen that, but with STATUS and INFO command i don't see any problem so i don't know how to solve, i'm searching online but nothing.
you know how solve this?

Comment 6 bitchecker 2016-02-24 11:16:33 UTC
(In reply to bitchecker from comment #5)
> (In reply to RajeshReddy from comment #3)
> > From the log it looks like all sub-volumes are down so mount failed,
> > 
> > [MSGID: 108006] [afr-common.c:4015:afr_notify] 0-volume-replicate-0: All
> > subvolumes are down. Going offline until atleast one of them comes back up.
> 
> i've seen that, but with STATUS and INFO command i don't see any problem so
> i don't know how to solve, i'm searching online but nothing.
> you know how solve this?

this problem is only on clients, servers can mount volume properly!

Comment 7 Atin Mukherjee 2016-02-24 11:29:27 UTC
[2016-02-23 16:33:23.282375] E [name.c:247:af_inet_client_get_remote_sockaddr] 0-volume-client-0: DNS resolution failed on host gluster01

[2016-02-23 16:33:23.431491] E [name.c:247:af_inet_client_get_remote_sockaddr] 0-volume-client-1: DNS resolution failed on host gluster02

Certainly they are not resolvable, atleast the log indicates the same.

Comment 8 bitchecker 2016-02-24 11:35:02 UTC
(In reply to Atin Mukherjee from comment #7)
> [2016-02-23 16:33:23.282375] E
> [name.c:247:af_inet_client_get_remote_sockaddr] 0-volume-client-0: DNS
> resolution failed on host gluster01
> 
> [2016-02-23 16:33:23.431491] E
> [name.c:247:af_inet_client_get_remote_sockaddr] 0-volume-client-1: DNS
> resolution failed on host gluster02
> 
> Certainly they are not resolvable, atleast the log indicates the same.

they are specified on hosts file, but if i try with ip address i can't even mount volume.

Comment 9 bitchecker 2016-02-24 15:00:04 UTC
i solved the problem!

the mistake was on name of server, glusterfs servers are known as gluster01 and gluster02 and client need to know them with the same name.


i don't know if is normal that if i use ip address to try mount it doesn't works properly.

Comment 10 bitchecker 2016-02-24 15:27:15 UTC
after a reboot, clients and servers are unable to mount the volume.

i've two mnt log:
/var/log/glusterfs/mnt.log:

[2016-02-24 15:08:38.101109] I [MSGID: 100030] [glusterfsd.c:2318:main] 0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version 3.7.8 (args: /usr/sbin/glusterfs --volfile-server=gluster01 --volfile-id=/volume /mnt)
[2016-02-24 15:08:38.241023] I [MSGID: 101190] [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread with index 1
[2016-02-24 15:08:38.430844] E [MSGID: 127002] [quick-read.c:835:check_cache_size_ok] 0-volume-quick-read: Cache size 1073741824 is greater than the max size of 1016430592
[2016-02-24 15:08:38.430890] E [MSGID: 101019] [xlator.c:433:xlator_init] 0-volume-quick-read: Initialization of volume 'volume-quick-read' failed, review your volfile again
[2016-02-24 15:08:38.430907] E [graph.c:322:glusterfs_graph_init] 0-volume-quick-read: initializing translator failed
[2016-02-24 15:08:38.430914] E [graph.c:661:glusterfs_graph_activate] 0-graph: init failed
pending frames:
frame : type(0) op(0)
patchset: git://git.gluster.com/glusterfs.git
signal received: 11
time of crash: 
2016-02-24 15:08:38
configuration details:
argp 1
backtrace 1
dlfcn 1
libpthread 1
llistxattr 1
setfsid 1
spinlock 1
epoll.h 1
xattr.h 1
st_atim.tv_nsec 1
package-string: glusterfs 3.7.8
[2016-02-24 15:08:38.443995] W [glusterfsd.c:1236:cleanup_and_exit] (-->/usr/sbin/glusterfs(mgmt_getspec_cbk+0x331) [0x7f17b2ab3801] -->/usr/sbin/glusterfs(glusterfs_process_volfp+0x126) [0x7f17b2aae1a6] -->/usr/sbin/glusterfs(cleanup_and_exit+0x69) [0x7f17b2aad789] ) 0-: received signum (0), shutting down
[2016-02-24 15:08:38.444110] I [fuse-bridge.c:5685:fini] 0-fuse: Unmounting '/mnt'.


/var/log/glusterfs/mnt-.log:

[2016-02-23 15:58:22.732798] I [MSGID: 100030] [glusterfsd.c:2318:main] 0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version 3.7.8 (args: /usr/sbin/glusterfs --volfile-server=gluster02 --volfile-id=/volume /mnt/)
[2016-02-23 15:58:22.758758] I [MSGID: 101190] [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread with index 1
[2016-02-23 15:58:22.773890] I [MSGID: 114020] [client.c:2106:notify] 0-volume-client-0: parent translators are ready, attempting connect on transport
[2016-02-23 15:58:22.774685] I [MSGID: 101190] [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread with index 2
[2016-02-23 15:58:22.901192] E [MSGID: 101075] [common-utils.c:306:gf_resolve_ip6] 0-resolver: getaddrinfo failed (Nome o servizio sconosciuto)
[2016-02-23 15:58:22.901258] E [name.c:247:af_inet_client_get_remote_sockaddr] 0-volume-client-0: DNS resolution failed on host gluster01
[2016-02-23 15:58:22.901318] I [MSGID: 114020] [client.c:2106:notify] 0-volume-client-1: parent translators are ready, attempting connect on transport
[2016-02-23 15:58:23.044273] E [MSGID: 101075] [common-utils.c:306:gf_resolve_ip6] 0-resolver: getaddrinfo failed (Nome o servizio sconosciuto)
[2016-02-23 15:58:23.044317] E [name.c:247:af_inet_client_get_remote_sockaddr] 0-volume-client-1: DNS resolution failed on host gluster02
Final graph:
+------------------------------------------------------------------------------+
  1: volume volume-client-0
  2:     type protocol/client
  3:     option ping-timeout 42
  4:     option remote-host gluster01
  5:     option remote-subvolume /data/brick/volume
  6:     option transport-type socket
  7:     option send-gids true
  8: end-volume
  9:  
 10: volume volume-client-1
 11:     type protocol/client
 12:     option ping-timeout 42
 13:     option remote-host gluster02
 14:     option remote-subvolume /data/brick/volume
 15:     option transport-type socket
 16:     option send-gids true
 17: end-volume
 18:  
 19: volume volume-replicate-0
 20:     type cluster/replicate
 21:     subvolumes volume-client-0 volume-client-1
 22: end-volume
 23:  
 24: volume volume-dht
 25:     type cluster/distribute
 26:     subvolumes volume-replicate-0
 27: end-volume
 28:  
 29: volume volume-write-behind
 30:     type performance/write-behind
 31:     subvolumes volume-dht
 32: end-volume
 33:  
 34: volume volume-read-ahead
 35:     type performance/read-ahead
 36:     subvolumes volume-write-behind
 37: end-volume
 38:  
 39: volume volume-readdir-ahead
 40:     type performance/readdir-ahead
 41:     subvolumes volume-read-ahead
 42: end-volume
 43:  
 44: volume volume-io-cache
 45:     type performance/io-cache
 46:     subvolumes volume-readdir-ahead
 47: end-volume
 48:  
 49: volume volume-quick-read
 50:     type performance/quick-read
 51:     subvolumes volume-io-cache
 52: end-volume
 53:  
 54: volume volume-open-behind
 55:     type performance/open-behind
 56:     subvolumes volume-quick-read
 57: end-volume
 58:  
 59: volume volume-md-cache
 60:     type performance/md-cache
 61:     subvolumes volume-open-behind
 62: end-volume
 63:  
 64: volume volume
 65:     type debug/io-stats
 66:     option latency-measurement off
 67:     option count-fop-hits off
 68:     subvolumes volume-md-cache
 69: end-volume
 70:  
 71: volume meta-autoload
 72:     type meta
 73:     subvolumes volume
 74: end-volume
 75:  
+------------------------------------------------------------------------------+
[2016-02-23 15:58:23.048949] E [MSGID: 108006] [afr-common.c:4015:afr_notify] 0-volume-replicate-0: All subvolumes are down. Going offline until atleast one of them comes back up.
[2016-02-23 15:58:23.069369] I [fuse-bridge.c:5139:fuse_graph_setup] 0-fuse: switched to graph 0
[2016-02-23 15:58:23.077380] I [fuse-bridge.c:4060:fuse_init] 0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs 7.22 kernel 7.22
[2016-02-23 15:58:23.077637] I [MSGID: 108006] [afr-common.c:4143:afr_local_init] 0-volume-replicate-0: no subvolumes up
[2016-02-23 15:58:23.077925] W [fuse-bridge.c:758:fuse_attr_cbk] 0-glusterfs-fuse: 2: LOOKUP() / => -1 (Il socket di destinazione non è connesso)
[2016-02-23 15:58:23.092700] I [fuse-bridge.c:4986:fuse_thread_proc] 0-fuse: unmounting /mnt/
The message "I [MSGID: 108006] [afr-common.c:4143:afr_local_init] 0-volume-replicate-0: no subvolumes up" repeated 2 times between [2016-02-23 15:58:23.077637] and [2016-02-23 15:58:23.091271]
[2016-02-23 15:58:23.093014] W [glusterfsd.c:1236:cleanup_and_exit] (-->/lib64/libpthread.so.0(+0x7dc5) [0x7f82ca93edc5] -->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xe5) [0x7f82cbfa9905] -->/usr/sbin/glusterfs(cleanup_and_exit+0x69) [0x7f82cbfa9789] ) 0-: received signum (15), shutting down
[2016-02-23 15:58:23.093029] I [fuse-bridge.c:5685:fini] 0-fuse: Unmounting '/mnt/'.
[2016-02-23 16:16:52.150695] I [MSGID: 100030] [glusterfsd.c:2318:main] 0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version 3.7.8 (args: /usr/sbin/glusterfs --volfile-server=gluster02 --volfile-id=/volume /mnt/)
[2016-02-23 16:16:52.181144] I [MSGID: 101190] [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread with index 1
[2016-02-23 16:16:52.199373] I [MSGID: 114020] [client.c:2106:notify] 0-volume-client-0: parent translators are ready, attempting connect on transport
[2016-02-23 16:16:52.200228] I [MSGID: 101190] [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread with index 2
[2016-02-23 16:16:52.320766] E [MSGID: 101075] [common-utils.c:306:gf_resolve_ip6] 0-resolver: getaddrinfo failed (Nome o servizio sconosciuto)
[2016-02-23 16:16:52.320809] E [name.c:247:af_inet_client_get_remote_sockaddr] 0-volume-client-0: DNS resolution failed on host gluster01
[2016-02-23 16:16:52.320888] I [MSGID: 114020] [client.c:2106:notify] 0-volume-client-1: parent translators are ready, attempting connect on transport
[2016-02-23 16:16:52.437595] E [MSGID: 101075] [common-utils.c:306:gf_resolve_ip6] 0-resolver: getaddrinfo failed (Nome o servizio sconosciuto)
[2016-02-23 16:16:52.437644] E [name.c:247:af_inet_client_get_remote_sockaddr] 0-volume-client-1: DNS resolution failed on host gluster02
Final graph:
+------------------------------------------------------------------------------+
  1: volume volume-client-0
  2:     type protocol/client
  3:     option ping-timeout 42
  4:     option remote-host gluster01
  5:     option remote-subvolume /data/brick/volume
  6:     option transport-type socket
  7:     option send-gids true
  8: end-volume
  9:  
 10: volume volume-client-1
 11:     type protocol/client
 12:     option ping-timeout 42
 13:     option remote-host gluster02
 14:     option remote-subvolume /data/brick/volume
 15:     option transport-type socket
 16:     option send-gids true
 17: end-volume
 18:  
 19: volume volume-replicate-0
 20:     type cluster/replicate
 21:     subvolumes volume-client-0 volume-client-1
 22: end-volume
 23:  
 24: volume volume-dht
 25:     type cluster/distribute
 26:     subvolumes volume-replicate-0
 27: end-volume
 28:  
 29: volume volume-write-behind
 30:     type performance/write-behind
 31:     subvolumes volume-dht
 32: end-volume
 33:  
 34: volume volume-read-ahead
 35:     type performance/read-ahead
 36:     subvolumes volume-write-behind
 37: end-volume
 38:  
 39: volume volume-readdir-ahead
 40:     type performance/readdir-ahead
 41:     subvolumes volume-read-ahead
 42: end-volume
 43:  
 44: volume volume-io-cache
 45:     type performance/io-cache
 46:     subvolumes volume-readdir-ahead
 47: end-volume
 48:  
 49: volume volume-quick-read
 50:     type performance/quick-read
 51:     subvolumes volume-io-cache
 52: end-volume
 53:  
 54: volume volume-open-behind
 55:     type performance/open-behind
 56:     subvolumes volume-quick-read
 57: end-volume
 58:  
 59: volume volume-md-cache
 60:     type performance/md-cache
 61:     subvolumes volume-open-behind
 62: end-volume
 63:  
 64: volume volume
 65:     type debug/io-stats
 66:     option latency-measurement off
 67:     option count-fop-hits off
 68:     subvolumes volume-md-cache
 69: end-volume
 70:  
 71: volume meta-autoload
 72:     type meta
 73:     subvolumes volume
 74: end-volume
 75:  
+------------------------------------------------------------------------------+
[2016-02-23 16:16:52.441886] E [MSGID: 108006] [afr-common.c:4015:afr_notify] 0-volume-replicate-0: All subvolumes are down. Going offline until atleast one of them comes back up.
[2016-02-23 16:16:52.460512] I [fuse-bridge.c:5139:fuse_graph_setup] 0-fuse: switched to graph 0
[2016-02-23 16:16:52.460789] I [fuse-bridge.c:4060:fuse_init] 0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs 7.22 kernel 7.22
[2016-02-23 16:16:52.460979] I [MSGID: 108006] [afr-common.c:4143:afr_local_init] 0-volume-replicate-0: no subvolumes up
[2016-02-23 16:16:52.461257] W [fuse-bridge.c:758:fuse_attr_cbk] 0-glusterfs-fuse: 2: LOOKUP() / => -1 (Il socket di destinazione non è connesso)
[2016-02-23 16:16:52.468348] I [fuse-bridge.c:4986:fuse_thread_proc] 0-fuse: unmounting /mnt/
The message "I [MSGID: 108006] [afr-common.c:4143:afr_local_init] 0-volume-replicate-0: no subvolumes up" repeated 2 times between [2016-02-23 16:16:52.460979] and [2016-02-23 16:16:52.466340]
[2016-02-23 16:16:52.468730] W [glusterfsd.c:1236:cleanup_and_exit] (-->/lib64/libpthread.so.0(+0x7dc5) [0x7fefc4d73dc5] -->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xe5) [0x7fefc63de905] -->/usr/sbin/glusterfs(cleanup_and_exit+0x69) [0x7fefc63de789] ) 0-: received signum (15), shutting down
[2016-02-23 16:16:52.468748] I [fuse-bridge.c:5685:fini] 0-fuse: Unmounting '/mnt/'.
[2016-02-23 16:28:45.534853] I [MSGID: 100030] [glusterfsd.c:2318:main] 0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version 3.7.8 (args: /usr/sbin/glusterfs --volfile-server=glusterfs01 --volfile-id=/volume /mnt/)
[2016-02-23 16:28:45.564469] I [MSGID: 101190] [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread with index 1
[2016-02-23 16:28:45.582692] I [MSGID: 114020] [client.c:2106:notify] 0-volume-client-0: parent translators are ready, attempting connect on transport
[2016-02-23 16:28:45.583405] I [MSGID: 101190] [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread with index 2
[2016-02-23 16:28:45.763697] E [MSGID: 101075] [common-utils.c:306:gf_resolve_ip6] 0-resolver: getaddrinfo failed (Nome o servizio sconosciuto)
[2016-02-23 16:28:45.763738] E [name.c:247:af_inet_client_get_remote_sockaddr] 0-volume-client-0: DNS resolution failed on host gluster01
[2016-02-23 16:28:45.763812] I [MSGID: 114020] [client.c:2106:notify] 0-volume-client-1: parent translators are ready, attempting connect on transport
[2016-02-23 16:28:45.940000] E [MSGID: 101075] [common-utils.c:306:gf_resolve_ip6] 0-resolver: getaddrinfo failed (Nome o servizio sconosciuto)
[2016-02-23 16:28:45.940040] E [name.c:247:af_inet_client_get_remote_sockaddr] 0-volume-client-1: DNS resolution failed on host gluster02
Final graph:
+------------------------------------------------------------------------------+
  1: volume volume-client-0
  2:     type protocol/client
  3:     option ping-timeout 42
  4:     option remote-host gluster01
  5:     option remote-subvolume /data/brick/volume
  6:     option transport-type socket
  7:     option send-gids true
  8: end-volume
  9:  
 10: volume volume-client-1
 11:     type protocol/client
 12:     option ping-timeout 42
 13:     option remote-host gluster02
 14:     option remote-subvolume /data/brick/volume
 15:     option transport-type socket
 16:     option send-gids true
 17: end-volume
 18:  
 19: volume volume-replicate-0
 20:     type cluster/replicate
 21:     subvolumes volume-client-0 volume-client-1
 22: end-volume
 23:  
 24: volume volume-dht
 25:     type cluster/distribute
 26:     subvolumes volume-replicate-0
 27: end-volume
 28:  
 29: volume volume-write-behind
 30:     type performance/write-behind
 31:     subvolumes volume-dht
 32: end-volume
 33:  
 34: volume volume-read-ahead
 35:     type performance/read-ahead
 36:     subvolumes volume-write-behind
 37: end-volume
 38:  
 39: volume volume-readdir-ahead
 40:     type performance/readdir-ahead
 41:     subvolumes volume-read-ahead
 42: end-volume
 43:  
 44: volume volume-io-cache
 45:     type performance/io-cache
 46:     subvolumes volume-readdir-ahead
 47: end-volume
 48:  
 49: volume volume-quick-read
 50:     type performance/quick-read
 51:     subvolumes volume-io-cache
 52: end-volume
 53:  
 54: volume volume-open-behind
 55:     type performance/open-behind
 56:     subvolumes volume-quick-read
 57: end-volume
 58:  
 59: volume volume-md-cache
 60:     type performance/md-cache
 61:     subvolumes volume-open-behind
 62: end-volume
 63:  
 64: volume volume
 65:     type debug/io-stats
 66:     option latency-measurement off
 67:     option count-fop-hits off
 68:     subvolumes volume-md-cache
 69: end-volume
 70:  
 71: volume meta-autoload
 72:     type meta
 73:     subvolumes volume
 74: end-volume
 75:  
+------------------------------------------------------------------------------+
[2016-02-23 16:28:45.951727] E [MSGID: 108006] [afr-common.c:4015:afr_notify] 0-volume-replicate-0: All subvolumes are down. Going offline until atleast one of them comes back up.
[2016-02-23 16:28:45.978554] I [fuse-bridge.c:5139:fuse_graph_setup] 0-fuse: switched to graph 0
[2016-02-23 16:28:45.979002] I [fuse-bridge.c:4060:fuse_init] 0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs 7.22 kernel 7.22
[2016-02-23 16:28:45.979204] I [MSGID: 108006] [afr-common.c:4143:afr_local_init] 0-volume-replicate-0: no subvolumes up
[2016-02-23 16:28:45.979502] W [fuse-bridge.c:758:fuse_attr_cbk] 0-glusterfs-fuse: 2: LOOKUP() / => -1 (Il socket di destinazione non è connesso)
[2016-02-23 16:28:45.986834] I [fuse-bridge.c:4986:fuse_thread_proc] 0-fuse: unmounting /mnt/
The message "I [MSGID: 108006] [afr-common.c:4143:afr_local_init] 0-volume-replicate-0: no subvolumes up" repeated 2 times between [2016-02-23 16:28:45.979204] and [2016-02-23 16:28:45.984908]
[2016-02-23 16:28:45.987251] W [glusterfsd.c:1236:cleanup_and_exit] (-->/lib64/libpthread.so.0(+0x7dc5) [0x7f098a6dbdc5] -->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xe5) [0x7f098bd46905] -->/usr/sbin/glusterfs(cleanup_and_exit+0x69) [0x7f098bd46789] ) 0-: received signum (15), shutting down
[2016-02-23 16:28:45.987269] I [fuse-bridge.c:5685:fini] 0-fuse: Unmounting '/mnt/'.

Comment 11 bitchecker 2016-02-24 15:56:36 UTC
i've found the problem, is on amount of cache!

Maximum cache will change after reboots and the volume config can be bigger than max cache.

Comment 12 Atin Mukherjee 2016-02-25 04:12:13 UTC
(In reply to bitchecker from comment #11)
> i've found the problem, is on amount of cache!
> 
> Maximum cache will change after reboots and the volume config can be bigger
> than max cache.

Can you please elaborate the problem?

Comment 13 Atin Mukherjee 2016-02-25 04:45:28 UTC
Ignore my previous comment as now I get what you meant!

Also since you are testing upstream bits the bug should be raised with upstream product. I've corrected it.

Following information would be useful to debug this issue further:

1. output of gluster volume info
2. cmd_history.log file of all nodes

Comment 14 bitchecker 2016-02-25 08:45:48 UTC
# gluster --version
glusterfs 3.7.8 built on Feb  9 2016 06:29:54
Repository revision: git://git.gluster.com/glusterfs.git
Copyright (c) 2006-2011 Gluster Inc. <http://www.gluster.com>
GlusterFS comes with ABSOLUTELY NO WARRANTY.
You may redistribute copies of GlusterFS under the terms of the GNU General Public License

# gluster volume info 
 
Volume Name: volume
Type: Replicate
Volume ID: 3c6c72bf-f85d-4c4c-a536-19a75e2ae71b
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: gluster01:/data/brick/volume
Brick2: gluster02:/data/brick/volume
Options Reconfigured:
performance.readdir-ahead: enable
auth.allow: *
cluster.readdir-optimize: on
cluster.lookup-optimize: on
server.allow-insecure: on
features.trash-max-filesize: 1GB
features.trash: on
performance.write-behind-window-size: 10MB
nfs.disable: on
server.root-squash: ON
performance.io-thread-count: 64
performance.io-cache: on
performance.client-io-threads: on
nfs.export-volumes: off
nfs.addr-namelookup: off
performance.cache-size: 500MB


# cat /var/log/glusterfs/cmd_history.log (NODE01)                                                                                                                           
[2016-02-23 16:09:10.183385]  : volume set help : SUCCESS
[2016-02-23 16:09:12.674174]  : volume status : SUCCESS
[2016-02-23 16:09:12.690612]  : volume status : SUCCESS
[2016-02-24 10:28:14.039661]  : volume set volume auth.allow * : SUCCESS
[2016-02-24 10:30:23.813961]  : volume set help : SUCCESS
[2016-02-24 10:30:40.667330]  : volume set volume cluster.readdir-optimize on : SUCCESS
[2016-02-24 10:31:00.538373]  : volume set volume cluster.lookup-optimize on : SUCCESS
[2016-02-24 10:31:09.257735]  : volume set volume server.allow-insecure on : SUCCESS
[2016-02-24 10:31:17.845762]  : volume set volume features.trash-max-filesize 1GB : SUCCESS
[2016-02-24 10:31:27.195250]  : volume set volume features.trash on : SUCCESS
[2016-02-24 10:31:34.849999]  : volume set volume performance.write-behind-window-size 10MB : SUCCESS
[2016-02-24 10:31:53.612553]  : volume set volume nfs.disable on : SUCCESS
[2016-02-24 10:32:01.464254]  : volume set volume server.root-squash ON : SUCCESS
[2016-02-24 10:32:12.944064]  : volume set volume performance.io-thread-count 64 : SUCCESS
[2016-02-24 10:32:21.522073]  : volume set volume performance.io-cache on : SUCCESS
[2016-02-24 10:32:29.544960]  : volume set volume performance.client-io-threads on : SUCCESS
[2016-02-24 10:32:36.873471]  : volume set volume nfs.export-volumes off : SUCCESS
[2016-02-24 10:32:44.317885]  : volume set volume nfs.addr-namelookup off : SUCCESS
[2016-02-24 10:32:51.650657]  : volume set volume performance.readdir-ahead enable : SUCCESS
[2016-02-24 10:33:00.808064]  : volume set volume performance.cache-size 2GB : SUCCESS
[2016-02-24 10:35:57.357085]  : volume set help : SUCCESS
[2016-02-24 10:38:34.864842]  : volume log rotate volume : SUCCESS
[2016-02-24 11:26:22.066994]  : volume set help : SUCCESS
[2016-02-24 11:45:28.202550]  : volume set help : SUCCESS
[2016-02-24 13:47:10.827338]  : volume set help : SUCCESS
[2016-02-24 13:48:02.221481]  : volume set volume performance.cache-size 1GB : SUCCESS
[2016-02-24 14:08:26.806732]  : volume set help : SUCCESS
[2016-02-24 14:09:47.976273]  : volume set help : SUCCESS
[2016-02-24 14:10:51.429318]  : volume status volume : SUCCESS
[2016-02-24 14:28:58.228384]  : volume set help : SUCCESS
[2016-02-24 14:53:51.402021]  : volume set help : SUCCESS
[2016-02-24 15:03:22.970201]  : volume set help : SUCCESS
[2016-02-24 15:11:10.536295]  : volume set help : SUCCESS
[2016-02-24 15:11:21.776935]  : volume status : SUCCESS
[2016-02-24 15:11:21.795461]  : volume status : SUCCESS
[2016-02-24 15:24:25.965889]  : volume set help : SUCCESS
[2016-02-24 15:24:38.642638]  : volume set volume performance.cache-size 0 : FAILED : '0' in 'option cache-size 0' is out of range [4194304 - 34359738368]
[2016-02-24 15:25:21.803863]  : volume set volume performance.cache-size 500MB : SUCCESS
[2016-02-24 15:25:43.293993]  : volume status : SUCCESS
[2016-02-24 15:25:43.299855]  : volume status : SUCCESS
[2016-02-24 15:56:26.525493]  : volume set help : SUCCESS
[2016-02-25 08:27:51.329361]  : volume set help : SUCCESS


# cat /var/log/glusterfs/cmd_history.log (NODE02) 
[2016-02-23 16:10:04.874219]  : volume set help : SUCCESS
[2016-02-23 16:10:17.948524]  : volume status : SUCCESS
[2016-02-23 16:10:17.950331]  : volume status : SUCCESS
[2016-02-23 16:15:53.717774]  : volume set help : SUCCESS
[2016-02-23 16:18:17.108678]  : volume set help : SUCCESS
[2016-02-24 10:34:56.804196]  : volume set help : SUCCESS
[2016-02-24 14:29:03.936580]  : volume set help : SUCCESS
[2016-02-24 14:53:45.494411]  : volume set help : SUCCESS
[2016-02-25 08:30:42.306042]  : volume set help : SUCCESS

Comment 15 Kvasir 2016-10-13 15:32:14 UTC
Hi Everyone,

I wanted to further detail my findings since it does not appear it has been mentioned yet.

This issue where the IP does not seem to be working for a glusterfs client mount is more related to reverse name resolution.  I tested this on an updated CentOS7 system pulling directly from the gluster 3.8 repo.  The moment I added the servers in the /etc/hosts file, I could mount by hostanme or IP with two client systems.  My logic tells me this is why the servers do not seem to be affected: They already have an updated /etc/hosts file.

Comment 16 denmat 2016-12-13 17:03:42 UTC
I get a similar problem. We only use IP addresses.

[2016-12-13 17:01:15.908466] I [MSGID: 100030] [glusterfsd.c:2454:main] 0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version 3.8.6 (args: /usr/sbin/glusterfs --volfile-server=10.90.5.105 --volfile-id=/marketplace_nfs /data)
[2016-12-13 17:01:15.912204] I [MSGID: 101190] [event-epoll.c:628:event_dispatch_epoll_worker] 0-epoll: Started thread with index 1
[2016-12-13 17:01:15.918784] E [MSGID: 127002] [quick-read.c:835:check_cache_size_ok] 0-marketplace_nfs-quick-read: Cache size 1073741824 is greater than the max size of 1040695296
[2016-12-13 17:01:15.918799] E [MSGID: 101019] [xlator.c:433:xlator_init] 0-marketplace_nfs-quick-read: Initialization of volume 'marketplace_nfs-quick-read' failed, review your volfile again
[2016-12-13 17:01:15.918805] E [MSGID: 101066] [graph.c:324:glusterfs_graph_init] 0-marketplace_nfs-quick-read: initializing translator failed
[2016-12-13 17:01:15.918813] E [MSGID: 101176] [graph.c:673:glusterfs_graph_activate] 0-graph: init failed
[2016-12-13 17:01:15.919085] W [glusterfsd.c:1327:cleanup_and_exit] (-->/usr/sbin/glusterfs(mgmt_getspec_cbk+0x3a1) [0x7f11b21aeae1] -->/usr/sbin/glusterfs(glusterfs_process_volfp+0x14a) [0x7f11b21a94da] -->/usr/sbin/glusterfs(cleanup_and_exit+0x6b) [0x7f11b21a8acb] ) 0-: received signum (1), shutting down
[2016-12-13 17:01:15.919106] I [fuse-bridge.c:5793:fini] 0-fuse: Unmounting '/data'.

# glusterfs --version
glusterfs 3.8.6

Comment 17 denmat 2016-12-13 17:44:51 UTC
Actually, my problem above is most likely caused my the instance memory.

Comment 18 Kaushal 2017-03-08 10:50:48 UTC
This bug is getting closed because GlusteFS-3.7 has reached its end-of-life.

Note: This bug is being closed using a script. No verification has been performed to check if it still exists on newer releases of GlusterFS.
If this bug still exists in newer GlusterFS releases, please reopen this bug against the newer release.


Note You need to log in before you can comment on or make changes to this bug.