Bug 1849478 - [RADOS] Backport changes related to bluefs log not being compacted and possibly getting corrupted after growing to extreme size
Summary: [RADOS] Backport changes related to bluefs log not being compacted and possib...
Keywords:
Status: CLOSED ERRATA
Alias: None
Product: Red Hat Ceph Storage
Classification: Red Hat Storage
Component: RADOS
Version: 3.3
Hardware: x86_64
OS: Linux
high
high
Target Milestone: z6
: 3.3
Assignee: Neha Ojha
QA Contact: Manohar Murthy
Amrita
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2020-06-22 03:02 UTC by Prashant Dhange
Modified: 2020-08-25 02:46 UTC (History)
13 users (show)

Fixed In Version: RHEL: ceph-12.2.12-120.el7cp Ubuntu: ceph_12.2.12-108redhat1
Doc Type: Bug Fix
Doc Text:
Previously, BlueFS failed to replay log which was corrupted because of a previous written replay log. The BlueFS log corruption was caused by the BlueFS log growing exponentially because of OSD functioning in a way where sync_metadata was never invoked in some situations and if sync_metadata was invoked the BlueFS log was not getting compacted even though there is no new log data to flush. This log corruption caused Bluestore not to be mounted and data loss in multiple OSDs. With this update, sync_metadata is now invoked, BlueFS log is getting compacted even though there is no new log data to flush in sync_metadata and avoid log corruption if log is expanded. This prevents OSDs getting full due to BlueFS log growing exponentially and also BlueFS getting corrupted. As a result, logs replay and there is no data loss.
Clone Of:
Environment:
Last Closed: 2020-08-18 18:05:58 UTC
Embargoed:
pdhange: automate_bug?


Attachments (Terms of Use)


Links
System ID Private Priority Status Summary Last Updated
Github ceph ceph pull 34876 0 None closed luminous: os/bluestore/BlueFS: compact log even when sync_metadata sees no work 2021-02-11 20:13:58 UTC
Github ceph ceph pull 35473 0 None closed os/bluestore: fix bluefs log growth. 2021-02-11 20:13:59 UTC
Red Hat Product Errata RHSA-2020:3504 0 None None None 2020-08-18 18:06:29 UTC

Comment 14 errata-xmlrpc 2020-08-18 18:05:58 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory (Moderate: Red Hat Ceph Storage 3.3 security and bug fix update), and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHSA-2020:3504

Comment 15 Neha Ojha 2020-08-18 18:15:11 UTC
Requested info has already been provided.


Note You need to log in before you can comment on or make changes to this bug.