Bug 1814057 - Ceph Monitor heartbeat grace period does not reset
Summary: Ceph Monitor heartbeat grace period does not reset
Keywords:
Status: CLOSED ERRATA
Alias: None
Product: Red Hat Ceph Storage
Classification: Red Hat Storage
Component: RADOS
Version: 3.2
Hardware: x86_64
OS: Linux
medium
high
Target Milestone: z2
: 4.1
Assignee: Sridhar Seshasayee
QA Contact: Pawan
Aron Gunn
URL:
Whiteboard:
Depends On:
Blocks: 1816167 1855472
TreeView+ depends on / blocked
 
Reported: 2020-03-16 21:55 UTC by Steve Baldwin
Modified: 2023-10-06 19:28 UTC (History)
14 users (show)

Fixed In Version: ceph-14.2.8-100.el8cp, ceph-14.2.8-100.el7cp
Doc Type: Enhancement
Doc Text:
.Update to the heartbeat grace period Previously, when there were no Ceph OSD failures for more than 48 hours, there was no mechanism to reset the grace timer back to the default value. With this release, the heartbeat grace timer is reset to the default value of 20 seconds, if there have been no failures on a Ceph OSD for 48 hours. When the failure interval between the last failure and the latest failure exceeds 48 hours, the grace timer is reset to the default value of 20 seconds. The grace time is the interval in which a Ceph storage cluster considers a Ceph OSD as down by the absence of a heartbeat. The grace time is scaled based on lag estimations or on how frequently a Ceph ODS is experiencing failures.
Clone Of:
: 1855472 (view as bug list)
Environment:
Last Closed: 2020-09-30 17:24:49 UTC
Embargoed:


Attachments (Terms of Use)


Links
System ID Private Priority Status Summary Last Updated
Github ceph ceph pull 35798 0 None closed nautilus: mon/OSDMonitor: Reset grace period if failure interval exceeds a threshold. 2021-02-12 08:50:05 UTC
Red Hat Issue Tracker RHCEPH-7646 0 None None None 2023-10-06 19:28:31 UTC
Red Hat Product Errata RHBA-2020:4144 0 None None None 2020-09-30 17:25:27 UTC

Comment 30 errata-xmlrpc 2020-09-30 17:24:49 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory (Red Hat Ceph Storage 4.1 Bug Fix update), and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHBA-2020:4144


Note You need to log in before you can comment on or make changes to this bug.