Bug 1245333
| Summary: | rare leaked connection->session->message->connection cycle on client restart with blocked requests | ||
|---|---|---|---|
| Product: | [Red Hat Storage] Red Hat Ceph Storage | Reporter: | Samuel Just <sjust> |
| Component: | RADOS | Assignee: | Samuel Just <sjust> |
| Status: | CLOSED ERRATA | QA Contact: | ceph-qe-bugs <ceph-qe-bugs> |
| Severity: | low | Docs Contact: | |
| Priority: | unspecified | ||
| Version: | 1.3.0 | CC: | ceph-eng-bugs, dzafman, flucifre, hnallurv, kchai, kdreyer, shmohan |
| Target Milestone: | rc | ||
| Target Release: | 1.3.2 | ||
| Hardware: | All | ||
| OS: | All | ||
| Whiteboard: | |||
| Fixed In Version: | RHEL: ceph-0.94.5-2.el7cp Ubuntu: ceph_0.94.5-2redhat1 | Doc Type: | Bug Fix |
| Doc Text: | Story Points: | --- | |
| Clone Of: | Environment: | ||
| Last Closed: | 2016-02-29 14:42:30 UTC | Type: | Bug |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
|
Description
Samuel Just
2015-07-21 19:10:07 UTC
Shipped in v0.94.4 - will be in RHCS 1.3.2 This is a race condition which I was not able to reproduce on the older builds. After talking to sam we concluded that we have done enough automated+manual regression testing in the surrounding areas of the fix , hence marking this bug as verified.
Following tests are also run specific to this bug.
As per sam's instruction I started doing map changing commands from different terminals like following ..
T1:
===
set noout
unset noout
T2:
===
set noin
unset noin
T3:
===
ceph osd scrub 1
ceph osd deep-scrub 1
T4:
===
for i in {1..1000}; do sudo ceph osd pool create pool$i 1 1 replicated replicated_ruleset; sudo ceph osd pool mksnap pool$i snappy$i; sudo ceph osd pool rmsnap pool$i snappy$i; done
T5:
===
for i in {101..110}; do for j in {1..100}; do sudo ceph osd pool mksnap p$i s$j; sudo ceph osd pool rmsnap p$i s$j; done; done
[ubuntu@magna028 ~]$ cat snap.sh
#!/bin/bash
val=$RANDOM
for i in {1..100}
do
for j in {1..100}
do
sudo ceph osd pool mksnap p$i sna$i$val
sudo ceph osd pool rmsnap p$i sna$i$val
done
done
above script run from 4 different terminals(= 4 different clients , so 400 ops ) like below
for i in {1..100}
do
./snap.sh &
done
simultaneously ceph-radosgw process has been restarted continuously.
But still I am not able to see the blocked messages in ceph -w.
verified on ceph-0.94.5-8.el7cp.x86_64
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2016:0313 |