Is the fencing of nodes causing quorum loss? Can you ensure that customer has set the fencing policies related to gluster at the cluster level (i.e not fencing if brick is online or if it could lead to quorum loss) Also can you confirm if these are gluster snapshots or qemu snapshots on gluster volume?
vmstore1 and vmstore2 are distributed-replica volumes. whenever concurrent delete of VM snapshots occurs, there's an issue with I/O latency , sanlock logs - "2019-05-07 00:19:43 2301165 [25167]: s10 delta_renew long write time 43 sec" Krutika, could you check the logs to see if there are any gluster issues causing this high latency?
The needinfo request[s] on this closed bug have been removed as they have been unresolved for 1000 days