Bug 2257733
Summary: | [5.3 backport][GSS] "MDS_CLIENT_OLDEST_TID" Errors | ||
---|---|---|---|
Product: | [Red Hat Storage] Red Hat Ceph Storage | Reporter: | Bipin Kunal <bkunal> |
Component: | CephFS | Assignee: | Venky Shankar <vshankar> |
Status: | CLOSED ERRATA | QA Contact: | Hemanth Kumar <hyelloji> |
Severity: | medium | Docs Contact: | |
Priority: | unspecified | ||
Version: | 6.1 | CC: | bhull, bkunal, bmcmurra, ceph-eng-bugs, cephqe-warriors, dwalveka, gfarnum, hyelloji, kdreyer, mcaldeir, mchangir, tserlin, vdas, vereddy, vshankar, xiubli |
Target Milestone: | --- | Flags: | hyelloji:
needinfo-
mcaldeir: needinfo- |
Target Release: | 5.3z7 | ||
Hardware: | x86_64 | ||
OS: | Linux | ||
Whiteboard: | |||
Fixed In Version: | ceph-16.2.10-264.el8cp | Doc Type: | Bug Fix |
Doc Text: |
Previously, MDS_CLIENT_OLDEST_TID client would fail to advance the oldest client/flush tid. Due to this the MDS would pile up the completed request list in large size resulting in the MDS going read-only.
With this fix, the oldest_client_tid is updated through the session renew caps message to make sure that MDS would not pile up the completed request and now MDS_CLIENT_OLDEST_TID does not fail to advance the oldest client/flush tid.
|
Story Points: | --- |
Clone Of: | 2257730 | Environment: | |
Last Closed: | 2024-06-26 10:01:41 UTC | Type: | --- |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: | |||
Bug Depends On: | 2239951, 2257730, 2263219 | ||
Bug Blocks: |
Description
Bipin Kunal
2024-01-10 16:18:23 UTC
See KCS #6971376, (https://access.redhat.com/solutions/6971376). /MC Waiting on bkunal to get feedback from customer. (In reply to Bipin Kunal from comment #33) > Venky, > I have lost my permission to access internal comment, so I am unable to > see comments between 22-32 so no clue what is happening. I am working on > getting the access back but in the meanwhile If there a hotfix request from > support please suggest them to follow the hotfix process and send a formal > hotfix request. There is not hotfix request, just that we need to check with the customer if they are fine to upgrade to RHCS6 (where the fix is available) instead of pulling in the stripped down fix in z7. (I'm unable to mark comments as private, IDK why - can someone please do the needful) (In reply to Bipin Kunal from comment #35) > (In reply to Venky Shankar from comment #34) > > (In reply to Bipin Kunal from comment #33) > > > Venky, > > > I have lost my permission to access internal comment, so I am unable to > > > see comments between 22-32 so no clue what is happening. I am working on > > > getting the access back but in the meanwhile If there a hotfix request from > > > support please suggest them to follow the hotfix process and send a formal > > > hotfix request. > > > > There is not hotfix request, just that we need to check with the customer if > > they are fine to upgrade to RHCS6 (where the fix is available) instead of > > pulling in the stripped down fix in z7. > > I had discussion with the TAM, and got to know that customer is waiting for > it. > They might be on 5.3z till mid of next year. So it is better to ship the fix > in z7. ACK. cc Xiubo who did the hotfix backport (stripped down version) for the customer. Xiubo, we need to get in the same set of patches for z7. (In reply to Venky Shankar from comment #36) > (In reply to Bipin Kunal from comment #35) > > (In reply to Venky Shankar from comment #34) > > > (In reply to Bipin Kunal from comment #33) > > > > Venky, > > > > I have lost my permission to access internal comment, so I am unable to > > > > see comments between 22-32 so no clue what is happening. I am working on > > > > getting the access back but in the meanwhile If there a hotfix request from > > > > support please suggest them to follow the hotfix process and send a formal > > > > hotfix request. > > > > > > There is not hotfix request, just that we need to check with the customer if > > > they are fine to upgrade to RHCS6 (where the fix is available) instead of > > > pulling in the stripped down fix in z7. > > > > I had discussion with the TAM, and got to know that customer is waiting for > > it. > > They might be on 5.3z till mid of next year. So it is better to ship the fix > > in z7. > > ACK. cc Xiubo who did the hotfix backport (stripped down version) for the > customer. > > Xiubo, we need to get in the same set of patches for z7. Sure, will do it today. (In reply to Xiubo Li from comment #37) > (In reply to Venky Shankar from comment #36) > > (In reply to Bipin Kunal from comment #35) > > > (In reply to Venky Shankar from comment #34) > > > > (In reply to Bipin Kunal from comment #33) > > > > > Venky, > > > > > I have lost my permission to access internal comment, so I am unable to > > > > > see comments between 22-32 so no clue what is happening. I am working on > > > > > getting the access back but in the meanwhile If there a hotfix request from > > > > > support please suggest them to follow the hotfix process and send a formal > > > > > hotfix request. > > > > > > > > There is not hotfix request, just that we need to check with the customer if > > > > they are fine to upgrade to RHCS6 (where the fix is available) instead of > > > > pulling in the stripped down fix in z7. > > > > > > I had discussion with the TAM, and got to know that customer is waiting for > > > it. > > > They might be on 5.3z till mid of next year. So it is better to ship the fix > > > in z7. > > > > ACK. cc Xiubo who did the hotfix backport (stripped down version) for the > > customer. > > > > Xiubo, we need to get in the same set of patches for z7. > > Sure, will do it today. Done, please see https://gitlab.cee.redhat.com/ceph/ceph/-/merge_requests/642. (In reply to Xiubo Li from comment #38) > (In reply to Xiubo Li from comment #37) > > (In reply to Venky Shankar from comment #36) > > > (In reply to Bipin Kunal from comment #35) > > > > (In reply to Venky Shankar from comment #34) > > > > > (In reply to Bipin Kunal from comment #33) > > > > > > Venky, > > > > > > I have lost my permission to access internal comment, so I am unable to > > > > > > see comments between 22-32 so no clue what is happening. I am working on > > > > > > getting the access back but in the meanwhile If there a hotfix request from > > > > > > support please suggest them to follow the hotfix process and send a formal > > > > > > hotfix request. > > > > > > > > > > There is not hotfix request, just that we need to check with the customer if > > > > > they are fine to upgrade to RHCS6 (where the fix is available) instead of > > > > > pulling in the stripped down fix in z7. > > > > > > > > I had discussion with the TAM, and got to know that customer is waiting for > > > > it. > > > > They might be on 5.3z till mid of next year. So it is better to ship the fix > > > > in z7. > > > > > > ACK. cc Xiubo who did the hotfix backport (stripped down version) for the > > > customer. > > > > > > Xiubo, we need to get in the same set of patches for z7. > > > > Sure, will do it today. > > Done, please see > https://gitlab.cee.redhat.com/ceph/ceph/-/merge_requests/642. Can we merge the above into ceph-5.3-rhel-patches? Or is some review needed? Thomas Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Moderate: Red Hat Ceph Storage 5.3 security, bug fix, and enhancement update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2024:4118 |