Bug 1509007
Summary: | [RFE] Add bad bond event into audit_log | ||
---|---|---|---|
Product: | Red Hat Enterprise Virtualization Manager | Reporter: | Javier Coscia <jcoscia> |
Component: | ovirt-engine | Assignee: | Nobody <nobody> |
Status: | CLOSED WONTFIX | QA Contact: | Michael Burman <mburman> |
Severity: | medium | Docs Contact: | |
Priority: | medium | ||
Version: | 4.1.6 | CC: | danken, dholler, jcoscia, lsurette, mburman, mgoldboi, michal.skrivanek, mkalinin, mmirecki, mperina, mtessun, nsoffer, sradco, srevivo |
Target Milestone: | --- | Keywords: | FutureFeature |
Target Release: | --- | ||
Hardware: | Unspecified | ||
OS: | Unspecified | ||
Whiteboard: | sync-to-jira | ||
Fixed In Version: | Doc Type: | If docs needed, set a value | |
Doc Text: | Story Points: | --- | |
Clone Of: | Environment: | ||
Last Closed: | 2021-03-24 15:24:35 UTC | Type: | Bug |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | Network | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: | |||
Bug Depends On: | 1240719, 1671876 | ||
Bug Blocks: |
Description
Javier Coscia
2017-11-02 16:43:54 UTC
The specific case in comment 0 is just one example. Once we improve the lvm filter configuration, this issue will be gone, but the general issue of reporting secondary failures (e.g. during flow cleanup, or during periodic checks) will always be a problem. I think this is best solved by a central logging system, collecting errors and warnings from all hosts and providing statistics on most common errors. Engine can integrated with such system to show alerts about hosts or storage domains that had such failures. (In reply to Nir Soffer from comment #1) > The specific case in comment 0 is just one example. Once we improve the lvm > filter configuration, this issue will be gone, but the general issue of > reporting > secondary failures (e.g. during flow cleanup, or during periodic checks) will > always be a problem. > > I think this is best solved by a central logging system, collecting errors > and > warnings from all hosts and providing statistics on most common errors. > > Engine can integrated with such system to show alerts about hosts or storage > domains that had such failures. The above is being done in the common logging effort, already for 4.2. We can have additional collection of logs (from journal, etc.) and set an alert for it. (In reply to Yaniv Kaul from comment #2) > The above is being done in the common logging effort, already for 4.2. > We can have additional collection of logs (from journal, etc.) and set an > alert for it. This is great - but we cannot depend on arbitrary text in log file for reporting events. If we treat this event as important event, it must be written in machine readable format (e.g json) or reported to the daemon collecting events on a host. (In reply to Yaniv Kaul from comment #2) > (In reply to Nir Soffer from comment #1) > > The specific case in comment 0 is just one example. Once we improve the lvm > > filter configuration, this issue will be gone, but the general issue of > > reporting > > secondary failures (e.g. during flow cleanup, or during periodic checks) will > > always be a problem. > > > > I think this is best solved by a central logging system, collecting errors > > and > > warnings from all hosts and providing statistics on most common errors. > > > > Engine can integrated with such system to show alerts about hosts or storage > > domains that had such failures. > > The above is being done in the common logging effort, already for 4.2. > We can have additional collection of logs (from journal, etc.) and set an > alert for it. Hi Yaniv, is there a BZ for this effort you mentioned in 4.2 ? If so, shouldn't this BZ be blocked by it ? Will add more cases or scenarios we might need to include besides the one in description as soon as I can come up with some. Thanks! Not that I'm aware of. Please include specific items - as we use collectd for monitoring, maybe it's something we already monitor or maybe it's something we can easily add. (In reply to Yaniv Kaul from comment #5) > Not that I'm aware of. Please include specific items - as we use collectd > for monitoring, maybe it's something we already monitor or maybe it's > something we can easily add. Forgot to set NEEDINFO on reporter for exact details of what's needed. This bug has not been marked as blocker for oVirt 4.3.0. Since we are releasing it tomorrow, January 29th, this bug has been re-targeted to 4.3.1. Dominik, I went over all comments and unless I'm missing something this is about adding bad bond event into audit_log, so moving to network team > Dominik, I went over all comments and unless I'm missing something this is about adding bad bond event into audit_log, so moving to network team
Ack, we could check for bad bonds the same way as UI already does on every getCaps and add a rate limited message to audit log.
I removed the target milestone to re-schedule this in the network team.
This bug didn't get any attention for a while, we didn't have the capacity to make any progress. If you deeply care about it or want to work on it please assign/target accordingly This bug didn't get any attention for a while, we didn't have the capacity to make any progress. If you deeply care about it or want to work on it please assign/target accordingly |