Bug 2084541
| Summary: | CephClusterReadOnly alert is not getting triggered | ||
|---|---|---|---|
| Product: | [Red Hat Storage] Red Hat OpenShift Data Foundation | Reporter: | Filip Balák <fbalak> |
| Component: | odf-managed-service | Assignee: | Kaustav Majumder <kmajumde> |
| Status: | ASSIGNED --- | QA Contact: | Filip Balák <fbalak> |
| Severity: | high | Docs Contact: | |
| Priority: | high | ||
| Version: | 4.10 | CC: | aeyal, dbindra, kmajumde, odf-bz-bot, resoni |
| Target Milestone: | --- | Flags: | resoni:
needinfo?
(kmajumde) |
| Target Release: | --- | ||
| Hardware: | Unspecified | ||
| OS: | Unspecified | ||
| Whiteboard: | |||
| Fixed In Version: | Doc Type: | If docs needed, set a value | |
| Doc Text: | Story Points: | --- | |
| Clone Of: | Environment: | ||
| Last Closed: | Type: | Bug | |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
| Bug Depends On: | 2084014 | ||
| Bug Blocks: | |||
|
Description
Filip Balák
2022-05-12 12:01:48 UTC
Try it on the latest build Notifications for cluster utilization including CephClusterReadOnly are not working:
$ oc rsh -n openshift-storage $(oc get pods -n openshift-storage|grep tool|awk '{print$1}') ceph df
--- RAW STORAGE ---
CLASS SIZE AVAIL USED RAW USED %RAW USED
ssd 12 TiB 1.8 TiB 10 TiB 10 TiB 85.02
TOTAL 12 TiB 1.8 TiB 10 TiB 10 TiB 85.02
--- POOLS ---
POOL ID PGS STORED OBJECTS USED %USED MAX AVAIL
device_health_metrics 1 1 0 B 0 0 B 0 0 B
ocs-storagecluster-cephfilesystem-metadata 2 32 16 KiB 22 131 KiB 100.00 0 B
ocs-storagecluster-cephfilesystem-data0 3 256 0 B 0 0 B 0 0 B
cephblockpool-storageconsumer-fddd8f1a-09e4-42fc-be0d-7d70e5f02f79 4 64 3.4 TiB 893.22k 10 TiB 100.00 0 B
$ oc rsh -n openshift-storage $(oc get pods -n openshift-storage|grep tool|awk '{print$1}') ceph -s
cluster:
id: 9e2ee3a5-53ef-45f3-bbd7-2dc83b07993f
health: HEALTH_ERR
3 full osd(s)
4 pool(s) full
services:
mon: 3 daemons, quorum a,b,c (age 5h)
mgr: a(active, since 5h)
mds: 1/1 daemons up, 1 hot standby
osd: 3 osds: 3 up (since 5h), 3 in (since 5h)
data:
volumes: 1/1 healthy
pools: 4 pools, 353 pgs
objects: 893.24k objects, 3.4 TiB
usage: 10 TiB used, 1.8 TiB / 12 TiB avail
pgs: 353 active+clean
io:
client: 1.2 KiB/s rd, 2 op/s rd, 0 op/s wr
$ rosa describe addon-installation --cluster fbalak03-1-pr --addon ocs-provider-qe
Id: ocs-provider-qe
Href: /api/clusters_mgmt/v1/clusters/226dcb9q8ric7euo2o73oo9k3jg73rjq/addons/ocs-provider-qe
Addon state: ready
Parameters:
"size" : "4"
"onboarding-validation-key" : (...)
"notification-email-1" : "fbalak"
"notification-email-2" : "odf-ms-qe"
Tested with:
ocs-osd-deployer.v2.0.11
must-gather: http://magna002.ceph.redhat.com/ocsci-jenkins/openshift-clusters/fbalak03-1-pr/fbalak03-1-pr_20230301T100351/logs/testcases_1677687913/
@kmajumde can you please provide the latest update? |