Created attachment 836326 [details] sos report Description of problem: glusterd crashes and when rebalancing a volume with the name StartMigrationDuringRebalanceTest. glusterd crashes again shortly after restarting the process. Version-Release number of selected component (if applicable): glusterfs-server-3.4.0.49rhs-1.el6rhs.x86_64 How reproducible: 100% reproducible Steps to Reproduce: 1. create a simple dist volume with the name StartMigrationDuringRebalanceTest 2. start the volume 3. start rebalance on the volume Actual results: glusterd crashes, rebalance failed to start Expected results: rebalance starts successfully Additional info:
Created attachment 836327 [details] etc-glusterfs-glusterd.vol.log
Created attachment 836328 [details] glusterd.log
Hi Dusmant, The bug is reproducible by using gluster cli, rhsc gui and the rhsc rest api. Please let me know what other information I can provide. >Steps to Reproduce: >1. create a simple dist volume with the name StartMigrationDuringRebalanceTest >2. start the volume >[Dusmant] From CLI or through RHSC or through REST API? >3. start rebalance on the volume >[Dusmant] From CLI or through RHSC or through REST API?
Here is the input/ouput from the gluster cli: [root@rhs-21u2-20131208-c ~]# gluster peer probe latest-d peer probe: success. [root@rhs-21u2-20131208-c ~]# gluster vol create StartMigrationDuringRebalanceTest Usage: volume create <NEW-VOLNAME> [stripe <COUNT>] [replica <COUNT>] [device vg] [transport <tcp|rdma|tcp,rdma>] <NEW-BRICK> ... [force] [root@rhs-21u2-20131208-c ~]# gluster vol create StartMigrationDuringRebalanceTest latest-c:/bricks/foo latest-d:/bricks/foo volume create: StartMigrationDuringRebalanceTest: success: please start the volume to access data [root@rhs-21u2-20131208-c ~]# gluster vol start StartMigrationDuringRebalanceTest volume start: StartMigrationDuringRebalanceTest: success [root@rhs-21u2-20131208-c ~]# gluster vol rebalance StartMigrationDuringRebalanceTest start Connection failed. Please check if gluster daemon is operational.
Dustin, Could you attach the core file of the glusterd crash you are seeing? I couldn't see any backtrace corresponding to a crash in the log files nor a core file in sosreport that is attached to this bug.
krishnan, I wasn't able to locate the core dump either. If you run the steps listed in comment#5, glusterd goes down while starting reblance on both nodes in the cluster.
The information provided is insufficient. There is no core file. Log files don't provide any backtrace to make any conclusion. The issue is not reproducible. We have asked Dustin Tsang to provide more information in the bug, with no response.
Created attachment 838985 [details] core dump
Hi Vivek, core dump attached.
KP, I've made minor edits. Can you please verify the doc text for technical accuracy?
in glusterfs-3.4.0.55rhs-1.el6rhs.x86_64: However rebalance command does not cause a crash, rebalance is unsuccessful for volumes with names greater than equal 33 characters steps: 1. create a volume with name greater than equal to 33 characters in length 2. start rebalance => start rebalance succeeds 3. poll rebalance status `gluster rebalance $VOLNAME status` => [root@rhs-21u2-20131223-errata-a ~]# gluster vol rebalance aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa status volume rebalance: aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa: failed: error
kirshnan, setting to verified.
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. http://rhn.redhat.com/errata/RHEA-2014-0208.html