Bug 1244810
| Summary: | Scale out from 1 compute to 3 in a BM setup with Ceph, fails due to a virsh secret that wasn't created | ||||||
|---|---|---|---|---|---|---|---|
| Product: | Red Hat OpenStack | Reporter: | Udi Kalifon <ukalifon> | ||||
| Component: | python-rdomanager-oscplugin | Assignee: | Brad P. Crochet <brad> | ||||
| Status: | CLOSED ERRATA | QA Contact: | Udi Kalifon <ukalifon> | ||||
| Severity: | urgent | Docs Contact: | |||||
| Priority: | high | ||||||
| Version: | Director | CC: | brad, calfonso, gfidente, jslagle, mburns, mcornea, rhel-osp-director-maint, rrosa, sasha | ||||
| Target Milestone: | ga | Keywords: | Triaged | ||||
| Target Release: | Director | ||||||
| Hardware: | Unspecified | ||||||
| OS: | Unspecified | ||||||
| Whiteboard: | |||||||
| Fixed In Version: | python-rdomanager-oscplugin-0.0.8-42.el7ost | Doc Type: | Bug Fix | ||||
| Doc Text: | Story Points: | --- | |||||
| Clone Of: | Environment: | ||||||
| Last Closed: | 2015-08-05 13:59:55 UTC | Type: | Bug | ||||
| Regression: | --- | Mount Type: | --- | ||||
| Documentation: | --- | CRM: | |||||
| Verified Versions: | Category: | --- | |||||
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |||||
| Cloudforms Team: | --- | Target Upstream Version: | |||||
| Embargoed: | |||||||
| Bug Depends On: | 1243274 | ||||||
| Bug Blocks: | |||||||
| Attachments: |
|
||||||
Believed to be fixed by: https://review.gerrithub.io/#/c/239994/ Brad, this is a different BZ; we need to make sure the params at [1] are not re-created when updating an existing deployment. 1. https://github.com/rdo-management/python-rdomanager-oscplugin/blob/master/rdomanager_oscplugin/v1/overcloud_deploy.py#L314-L316 The previous fix would partially fix it. Here is the remainder: https://review.gerrithub.io/240650 *** Bug 1246023 has been marked as a duplicate of this bug. *** Verified in: python-rdomanager-oscplugin-0.0.8-43.el7ost.noarch Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHEA-2015:1549 |
Created attachment 1053921 [details] Failed resource after scale up Description of problem: I tried to scale up from 1 compute to 3 (on BMs with puddle 2015-07-13) and the stack failed on resource "ComputePuppetDeployment" of the 1st compute node (the one that already existed couldn't be updated). The failure reason is not very informative: "Error: Deployment to server failed: deploy_status_code : Deployment exited with non-zero status code: 6". Further debugging shows that on the compute node that failed, there was an attempt to run "virsh secret-set-value" while the secret table is really empty and the uuid of the secret didn't exist. It seems the fsid was updated during the scale-up when it should not have been. Additional info is attached to the bug. It shows the error from "heat deployment-show". Version-Release number of selected component (if applicable): python-rdomanager-oscplugin-0.0.8-32.el7ost.noarch How reproducible: 100% Steps to Reproduce: 1. Deploy with 3 controllers, 1 compute and 1 ceph. I deployed on bare metals, without network isolation. I deployed with tuskar. 2. Run the deployment command again and scale up to 3 computes Actual results: Scale up fails.