Bug 962266 - core: volume start fails on latest build
Summary: core: volume start fails on latest build
Keywords:
Status: CLOSED ERRATA
Alias: None
Product: Red Hat Gluster Storage
Classification: Red Hat Storage
Component: glusterfs
Version: 2.1
Hardware: x86_64
OS: Linux
high
urgent
Target Milestone: ---
: ---
Assignee: Krutika Dhananjay
QA Contact: Saurabh
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2013-05-13 03:49 UTC by Saurabh
Modified: 2016-01-19 06:11 UTC (History)
5 users (show)

Fixed In Version: glusterfs-3.4.0.8rhs-1
Doc Type: Bug Fix
Doc Text:
Clone Of:
Environment:
Last Closed: 2013-09-23 22:35:31 UTC
Embargoed:


Attachments (Terms of Use)

Description Saurabh 2013-05-13 03:49:15 UTC
Description of problem:

gluster volume start fails on latest build
[root@bigbend1 ~]# gluster peer status
Number of Peers: 4

Hostname: 10.70.37.121
Uuid: cd19b8d0-bb97-42a8-9860-6cbc4bda1601
State: Peer in Cluster (Connected)

Hostname: 10.70.37.211
Uuid: baed4623-ca4c-4028-8674-f5e0c285e367
State: Peer in Cluster (Connected)

Hostname: 10.70.37.100
Uuid: ccdb3e04-5889-46dc-ab35-6e624d9ef85c
State: Peer in Cluster (Connected)

Hostname: 10.70.37.155
Uuid: 15316a44-a8df-4c4d-97a6-346a51d267d2
State: Peer in Cluster (Connected)
[root@bigbend1 ~]# gluster volume info
 
Volume Name: dist-rep
Type: Distributed-Replicate
Volume ID: 5912c35e-ebff-4d5d-9258-a5426b14a6e3
Status: Stopped
Number of Bricks: 6 x 2 = 12
Transport-type: tcp
Bricks:
Brick1: 10.70.37.155:/rhs/brick1/d1r1
Brick2: 10.70.37.100:/rhs/brick1/d1r2
Brick3: 10.70.37.121:/rhs/brick1/d2r1
Brick4: 10.70.37.211:/rhs/brick1/d2r2
Brick5: 10.70.37.155:/rhs/brick1/d3r1
Brick6: 10.70.37.100:/rhs/brick1/d3r2
Brick7: 10.70.37.121:/rhs/brick1/d4r1
Brick8: 10.70.37.211:/rhs/brick1/d4r2
Brick9: 10.70.37.155:/rhs/brick1/d5r1
Brick10: 10.70.37.100:/rhs/brick1/d5r2
Brick11: 10.70.37.121:/rhs/brick1/d6r1
Brick12: 10.70.37.211:/rhs/brick1/d6r2
Options Reconfigured:
server.root-squash: enable



Version-Release number of selected component (if applicable):
glusterfs-3.4.0.6rhs-1.el6rhs.x86_64

How reproducible:
always

Steps to Reproduce:
1. volume stop and glusterd stop on all nodes
2. remove rpm of the older build
3. install the latest rpms, glusterd start on all nodes of the cluster
3. gluster volume start <volume-name>
  
Actual results:
[root@bigbend1 ~]# gluster volume start dist-rep
volume start: dist-rep: failed: Another transaction could be in progress. Please try again after sometime.
[root@bigbend1 ~]# gluster volume info
 
Volume Name: dist-rep
Type: Distributed-Replicate
Volume ID: 5912c35e-ebff-4d5d-9258-a5426b14a6e3
Status: Stopped
Number of Bricks: 6 x 2 = 12
Transport-type: tcp
Bricks:
Brick1: 10.70.37.155:/rhs/brick1/d1r1
Brick2: 10.70.37.100:/rhs/brick1/d1r2
Brick3: 10.70.37.121:/rhs/brick1/d2r1
Brick4: 10.70.37.211:/rhs/brick1/d2r2
Brick5: 10.70.37.155:/rhs/brick1/d3r1
Brick6: 10.70.37.100:/rhs/brick1/d3r2
Brick7: 10.70.37.121:/rhs/brick1/d4r1
Brick8: 10.70.37.211:/rhs/brick1/d4r2
Brick9: 10.70.37.155:/rhs/brick1/d5r1
Brick10: 10.70.37.100:/rhs/brick1/d5r2
Brick11: 10.70.37.121:/rhs/brick1/d6r1
Brick12: 10.70.37.211:/rhs/brick1/d6r2
Options Reconfigured:
server.root-squash: enable
[root@bigbend1 ~]# gluster volume status
Another transaction could be in progress. Please try again after sometime.

Interesting thing is this behaviour is not seen all nodes, 

on two nodes it does this, 

i.e. 
10.70.37.155 --- affected
10.70.37.121 --- affected
10.70.37.100 --- unaffected
10.70.37.211 --- unaffected
 


Expected results:
the gluster commands should show same result from all nodes

Additional info:

Comment 1 Saurabh 2013-05-13 03:54:31 UTC
sosreports:-
http://rhsqe-repo.lab.eng.blr.redhat.com/sosreports/962266/

Comment 3 Amar Tumballi 2013-05-17 06:30:13 UTC
Looks same issue as described in mail to storage-qa (about RPM upgrade issue). Can you please verify with latest ISO + build, without any upgrade in picture?

Comment 5 Scott Haines 2013-09-23 22:35:31 UTC
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. 

For information on the advisory, and where to find the updated files, follow the link below.

If the solution does not work for you, open a new bug report.

http://rhn.redhat.com/errata/RHBA-2013-1262.html


Note You need to log in before you can comment on or make changes to this bug.