Description of problem: [RHVH3.6] Failed to add host to rhv-m 4.1 - Package collectd-disk cannot be found Unable to add host rhv-h 3.6 to 3.6 cluster in rhv-m 4.1 Version-Release number of selected component (if applicable): 4.1.1.8-0.1.el7 rhvh-3.6-0.20170413.0+1 vdsm-4.17.39-1.el7ev.noarch How reproducible: 100 Steps to Reproduce: 1. Try to add latest rhv-h 3.6 to cluster 3.6 in rhv-m 4.1 Actual results: Package collectd-disk cannot be found Expected results: Should work
There's no reason to use old 3.6 RHVH when new 4.1 is available, so closing as WONTFIX
In my opinion, as expressed also in idependently-opened bug 1438347 and bug 1433434, we do have a reason to add 3.6 hosts to 4.1 manager. Conservative customers that have big 3.6 clusters, would like to upgrade their manager to 4.1, test it with a small 4.1 cluster, while keeping their 3.6 cluster in full operation. Telling them that they would not be able to extend their 3.6 cluster once engine-4.1 is install may deter them from testing 4.1.
https://gerrit.ovirt.org/#/c/71539/7/src/plugins/ovirt-host-deploy/vdsm/packages.py@182 has added version-related logic to ovirt-host-deploy. Something similar can be done to pull collectd-disk only when needed by the cluster.
(In reply to Dan Kenigsberg from comment #7) > https://gerrit.ovirt.org/#/c/71539/7/src/plugins/ovirt-host-deploy/vdsm/ > packages.py@182 has added version-related logic to ovirt-host-deploy. > Something similar can be done to pull collectd-disk only when needed by the > cluster. I don't see how this is relevant - it checks for version of vdsm, not of cluster.
Two simple solutions I can think of: 1. Add an env key to allow preventing installing these packages. 2. Make the failure non-fatal, so that host-deploy will emit an error (or warning) but continue. (1.) will require adding a conf file in /etc to prevent attempting installation of these packages. (2.) will not require anything but has the downside that unrelated reasons preventing installation might go unnoticed.
(In reply to Yedidyah Bar David from comment #10) > Two simple solutions I can think of: > > 1. Add an env key to allow preventing installing these packages. > > 2. Make the failure non-fatal, so that host-deploy will emit an error (or > warning) but continue. > > (1.) will require adding a conf file in /etc to prevent attempting > installation of these packages. > > (2.) will not require anything but has the downside that unrelated reasons > preventing installation might go unnoticed. #2 sounds like a reasonable short-term solution to me.
*** Bug 1438005 has been marked as a duplicate of this bug. ***
(In reply to Yedidyah Bar David from comment #9) > > I don't see how this is relevant - it checks for version of vdsm, not of > cluster. It is enough for our use case: collectd-disk is available whereever a fresh vdsm-4.19 is.
*** Bug 1444450 has been marked as a duplicate of this bug. ***
(In reply to Dan Kenigsberg from comment #13) > (In reply to Yedidyah Bar David from comment #9) > > > > I don't see how this is relevant - it checks for version of vdsm, not of > > cluster. > > It is enough for our use case: collectd-disk is available whereever a fresh > vdsm-4.19 is. Does it handle also bug 1438005?
*** Bug 1440111 has been marked as a duplicate of this bug. ***
Please note that although the failure now is not fatal, engine still complains about it with misleading error message in the event log and this should be improved. Installing Host orchid-vds2.qa.lab.tlv.redhat.com. Stage: Package installation. Failed to install Host orchid-vds2.qa.lab.tlv.redhat.com. Yum Cannot queue package collectd: Package collectd cannot be found. Host orchid-vds2.qa.lab.tlv.redhat.com installation in progress . Failed to install collectd packages: Package collectd cannot be found. Failed to install Host orchid-vds2.qa.lab.tlv.redhat.com. Yum Cannot queue package fluentd: Package fluentd cannot be found. Host orchid-vds2.qa.lab.tlv.redhat.com installation in progress . Failed to install fluentd packages: Package fluentd cannot be found. Installing Host orchid-vds2.qa.lab.tlv.redhat.com. Yum Status: Downloading Packages. Those error messages are confusing as they report that host installation has failed, but it's succeeded eventually.
Please attach complete host-deploy log and relevant part of engine.log. Thanks.
2017-05-08 09:32:00,811+03 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (VdsDeploy) [327b5669] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509), Correlation ID: 327b5669, Call Stack: null, Cu stom Event ID: -1, Message: Installing Host orchid-vds2.qa.lab.tlv.redhat.com. Stage: Package installation. 2017-05-08 09:32:01,456+03 ERROR [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (VdsDeploy) [327b5669] EVENT_ID: VDS_INSTALL_IN_PROGRESS_ERROR(511), Correlation ID: 327b5669, Call Stack: nu ll, Custom Event ID: -1, Message: Failed to install Host orchid-vds2.qa.lab.tlv.redhat.com. Yum Cannot queue package collectd: Package collectd cannot be found. 2017-05-08 09:32:01,499+03 WARN [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (VdsDeploy) [327b5669] EVENT_ID: VDS_INSTALL_IN_PROGRESS_WARNING(510), Correlation ID: 327b5669, Call Stack: null, Custom Event ID: -1, Message: Host orchid-vds2.qa.lab.tlv.redhat.com installation in progress . Failed to install collectd packages: Package collectd cannot be found. 2017-05-08 09:32:01,527+03 ERROR [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (VdsDeploy) [327b5669] EVENT_ID: VDS_INSTALL_IN_PROGRESS_ERROR(511), Correlation ID: 327b5669, Call Stack: null, Custom Event ID: -1, Message: Failed to install Host orchid-vds2.qa.lab.tlv.redhat.com. Yum Cannot queue package fluentd: Package fluentd cannot be found. 2017-05-08 09:32:01,545+03 WARN [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (VdsDeploy) [327b5669] EVENT_ID: VDS_INSTALL_IN_PROGRESS_WARNING(510), Correlation ID: 327b5669, Call Stack: null, Custom Event ID: -1, Message: Host orchid-vds2.qa.lab.tlv.redhat.com installation in progress . Failed to install fluentd packages: Package fluentd cannot be found. 2017-05-08 09:32:03,308+03 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (VdsDeploy) [327b5669] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509), Correlation ID: 327b5669, Call Stack: null, Custom Event ID: -1, Message: Installing Host orchid-vds2.qa.lab.tlv.redhat.com. Yum Status: Downloading Packages.
Created attachment 1276978 [details] engine logs
1. I agree the message is confusing, but that's actually an engine bug - it emits it on every ERROR sent to it from host-deploy [1]. I suggest to open a bug to get this message changed, but it should be discussed with engine people. IMO it should be possible for host-deploy to emit an error and still do not fail, without a wrong message to the user. [1] https://gerrit.ovirt.org/gitweb?p=ovirt-engine.git;a=blob;f=backend/manager/modules/dal/src/main/resources/bundles/AuditLogMessages.properties;h=861a3ee3afcb26445b072ba54f64bedb58494246;hb=HEAD#l402 2. Pushed now a patch to cause host-deploy not emit these errors. Can you please try again with the jenkins build [2]? If you think it's important enough for current bug, you can move to ASSIGNED (or POST) (but then not sure it will enter 4.1.2), or open a new bug otherwise. [2] http://jenkins.ovirt.org/job/ovirt-host-deploy_master_check-patch-el7-x86_64/155/artifact/exported-artifacts/ 3. Please note that this does not prevent all the errors in the attached engine.log. In particular, the missing -nest package during "check-for-updates" is unrelated - should be handled by bug 1436655 for downstream, I do not think we have one for upstream - so we should probably open one and have bug 1426901 depend on it (or both).
(In reply to Yedidyah Bar David from comment #22) > 1. I agree the message is confusing, but that's actually an engine bug - it > emits it on every ERROR sent to it from host-deploy [1]. I suggest to open a > bug to get this message changed, but it should be discussed with engine > people. IMO it should be possible for host-deploy to emit an error and still > do not fail, without a wrong message to the user. Reported bug to improve the error message - BZ 1448798 > > [1] > https://gerrit.ovirt.org/gitweb?p=ovirt-engine.git;a=blob;f=backend/manager/ > modules/dal/src/main/resources/bundles/AuditLogMessages.properties; > h=861a3ee3afcb26445b072ba54f64bedb58494246;hb=HEAD#l402 > > 2. Pushed now a patch to cause host-deploy not emit these errors. Can you > please try again with the jenkins build [2]? If you think it's important > enough for current bug, you can move to ASSIGNED (or POST) (but then not > sure it will enter 4.1.2), or open a new bug otherwise. - If it's on latest master, then i will try it. - I don't think this is so important for the current bug. New bug reported^^ > > [2] > http://jenkins.ovirt.org/job/ovirt-host-deploy_master_check-patch-el7-x86_64/ > 155/artifact/exported-artifacts/ > > 3. Please note that this does not prevent all the errors in the attached > engine.log. In particular, the missing -nest package during > "check-for-updates" is unrelated - should be handled by bug 1436655 for > downstream, I do not think we have one for upstream - so we should probably > open one and have bug 1426901 depend on it (or both).
(In reply to Michael Burman from comment #23) > (In reply to Yedidyah Bar David from comment #22) > > 1. I agree the message is confusing, but that's actually an engine bug - it > > emits it on every ERROR sent to it from host-deploy [1]. I suggest to open a > > bug to get this message changed, but it should be discussed with engine > > people. IMO it should be possible for host-deploy to emit an error and still > > do not fail, without a wrong message to the user. > Reported bug to improve the error message - BZ 1448798 Thanks. Commented there and change the summary line. > > > > > [1] > > https://gerrit.ovirt.org/gitweb?p=ovirt-engine.git;a=blob;f=backend/manager/ > > modules/dal/src/main/resources/bundles/AuditLogMessages.properties; > > h=861a3ee3afcb26445b072ba54f64bedb58494246;hb=HEAD#l402 > > > > 2. Pushed now a patch to cause host-deploy not emit these errors. Can you > > please try again with the jenkins build [2]? If you think it's important > > enough for current bug, you can move to ASSIGNED (or POST) (but then not > > sure it will enter 4.1.2), or open a new bug otherwise. > - If it's on latest master, then i will try it. > - I don't think this is so important for the current bug. New bug reported^^ No, that's a different one. The patch of below patch is for host-deploy, not in the engine. It applies directly to current bug. Sandro - do we have time to include this in 4.1.2? Or postpone to later (and thus have another bug to track it, if at all)? > > [2] > > http://jenkins.ovirt.org/job/ovirt-host-deploy_master_check-patch-el7-x86_64/ > > 155/artifact/exported-artifacts/ This is the patch: https://gerrit.ovirt.org/76570
Target release should be placed once a package build is known to fix a issue. Since this bug is not modified, the target version has been reset. Please use target milestone to plan a fix for a oVirt release.
verified ovirt-host-deploy-1.6.5-1.el7ev.noarch host is installed even that package collectd is not avialable, warning is shown Host lunar installation in progress . Failed to install collectd packages.Please check the log for details.
3.6 NGN could be successfully added into 4.1 engine.