Description of problem:
The IO's hung and mount point became inaccessible after killing starting a brick. This bug is a quite similar to bug 1385605 by seeing at the logs
Version-Release number of selected component (if applicable):
Logs are placed at
Steps to Reproduce:
1. Create 3 x (2+1) arbiter volume
2. Mount the volume on gnfs and fuse protocol
3. Create small files using small-file tool (multi-client)on gNFS
4. now kill a brick ; and Start a small file cleanup
5. Force start the volume to start the volume.
6. and start a large file from FIO tool.
7. trigger heal info on server
Heal info hung
Mount point not accessible
IO tool reports I/O error
IO's should run smoothly
No errors should be reported
My gut feeling is that its the same as bug .  was hit when protocol/client received events in the order,
However, in this bz I think protocol/client received events in the order,
Though we need to think such an ordering is possible (since there can be only one event from socket due to EPOLL_ONESHOT, but the events can be on different sockets since for every new connection transport/socket uses a new socket). Another point to note that  fixes , by making:
1. making priv->connected=0
2. notifying higher layers a DISCONNECT event
as atomic in rpc-client. However, if indeed there are racing events, what about a CONNECT and DISCONNECT racing b/w transport/socket and rpc-client and changing the order. Is it possible? Something to ponder about.
This is very intermittently reproducible but when this issue gets hit it makes the whole system in a hanged state. I have the statedump taken at that time when the issue got hit. Its placed at the location itself. pstack output is not taken.
Thanks & regards
Patches  are merged in rhgs-3.3.0. Should we close this bug as fixed?
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.
For information on the advisory, and where to find the updated
files, follow the link below.
If the solution does not work for you, open a new bug report.