Bug 2275222

Summary: In Replica-1 data always goes to one particular osd and never goes to the additional osds present for a failure domain
Product: [Red Hat Storage] Red Hat OpenShift Data Foundation Reporter: Malay Kumar parida <mparida>
Component: ocs-operatorAssignee: Malay Kumar parida <mparida>
Status: CLOSED ERRATA QA Contact: Aviad Polak <apolak>
Severity: unspecified Docs Contact:
Priority: unspecified    
Version: 4.16CC: nberry, odf-bz-bot
Target Milestone: ---   
Target Release: ODF 4.16.0   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: 4.16.0-81 Doc Type: No Doc Update
Doc Text:
Story Points: ---
Clone Of:
: 2276339 (view as bug list) Environment:
Last Closed: 2024-07-17 13:19:33 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:
Bug Depends On:    
Bug Blocks: 2276339    

Description Malay Kumar parida 2024-04-16 05:55:06 UTC
In Replica-1 we support increasing the number of osds per failure domain. But even after the number of osds per failure domain is increased the data always goes to one particular osd. This results in a large imbalance of data among the osds in a failure domain.

This happens because the PG & PGP number stays at 1 always for the replica-1 pools

pool 5 'ocs-storagecluster-cephblockpool-us-east-1b' replicated size 1 min_size 1 crush_rule 8 object_hash rjenkins pg_num 1 pgp_num 1 autoscale_mode on last_change 126 flags hashpspool,selfmanaged_snaps stripe_width 0 application rbd
pool 6 'ocs-storagecluster-cephblockpool-us-east-1c' replicated size 1 min_size 1 crush_rule 10 object_hash rjenkins pg_num 1 pgp_num 1 autoscale_mode on last_change 128 flags hashpspool,selfmanaged_snaps stripe_width 0 application rbd
pool 7 'ocs-storagecluster-cephblockpool-us-east-1a' replicated size 1 min_size 1 crush_rule 13 object_hash rjenkins pg_num 1 pgp_num 1 autoscale_mode on last_change 123 flags hashpspool,selfmanaged_snaps stripe_width 0 application rbd

Is there any workaround available to the best of your knowledge?
Yes, disable the reconciliation of cephblockpool and add 
spec:
  parameters:
    pg_num: '16'
    pgp_num: '16'

Comment 7 errata-xmlrpc 2024-07-17 13:19:33 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory (Important: Red Hat OpenShift Data Foundation 4.16.0 security, enhancement & bug fix update), and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHSA-2024:4591