Description of problem (please be detailed as possible and provide log snippests): The Rook-Ceph crashcollector can run on an unlabeled node, it shouldn't. Version of all relevant components (if applicable): Does this issue impact your ability to continue to work with the product (please explain in detail what is the user impact)? Is there any workaround available to the best of your knowledge? Rate from 1 - 5 the complexity of the scenario you performed that caused this bug (1 - very simple, 5 - very complex)? 1 Can this issue reproducible? yes Can this issue reproduce from the UI? If this is a regression, please provide more details to justify this: Steps to Reproduce: 1. 2. 3. Actual results: Expected results: Additional info:
Logs: http://rhsqe-repo.lab.eng.blr.redhat.com/OCS/ocs-qe-bugs/bz-1898501/
Mudit, the crash collector will run anywhere Ceph daemons run, is this really an issue with the crash collector here?
@Seb In the original BZ (https://bugzilla.redhat.com/show_bug.cgi?id=1883828), the crash collector pod was found to be running on a non-OCS node. I suspect there was a ceph pod running on that node at some point, though at the time of analysis the crash collector was the only one on that node. Does Rook remove the crash collector from a node if all the ceph daemons are removed as well? The bug seems to be that the crash collector remained even after ceph pods were no longer running there.
@Travis indeed, this is not handled, I've managed to repro, working on a fix.
Seb, is this already fixed?
Mudit, no it's not I'm working on it.
Doc text is added.
Doc text lgtm.
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Moderate: Red Hat OpenShift Container Storage 4.7.0 security, bug fix, and enhancement update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2021:2041
*** Bug 1965749 has been marked as a duplicate of this bug. ***