Description of problem: As per https://access.redhat.com/documentation/en-us/red_hat_openstack_platform/13/html/director_installation_and_usage/sect-scaling_the_overcloud#sect-Removing_Compute_Nodes When compute nodes are removed it necessary to manually clean up the service entries. Our client is requesting this be automated. Version-Release number of selected component (if applicable): OSP 10 How reproducible: Every time Steps to Reproduce: 1. Remove computes 2. Forced to clean up cruft service records Actual results: - Service entries require manual clean up. Expected results: - No manual clean up.
We now have scale down tasks and have landed the code to run scale down actions on nova-compute https://review.opendev.org/#/c/653893/
Moving back to ON_DEV: - overcloud with 10 computes, tried to drop compute-1, there's an issue because it seems to take compute-10 in addition - same overcloud, dropped compute-9: services are still up there, "enabled" but "down" Run log looks suspicious (see coming attachement) and, in addition, the /var/lib/mistral/<stack-name> is removed at some point, making debug complicated....
Created attachment 1643897 [details] Scaledown log This log was generated with the following command: openstack overcloud node delete -y --stack overcloud-0 464817dd-e129-402a-93bf-dc2ed216471a 2>&1 | tee ~/scaledown-1.log That's the only trace I can get for now since /var/lib/mistral/overcloud-0 is dropped :(.
Just created a new BZ for the ansible directory removal: https://bugzilla.redhat.com/show_bug.cgi?id=1782379
I just tried the new RPM, and it works fine! Please note, only nova services are cleaned right now, because Network DFG didn't provide relevant code for the auto-cleanup. Verification method: - deploy Director and Overcloud using the provided t-h-t package (and some dependencies). Overcloud has 1ctrl + 2 computes - remove one of the computes, and check the output of `openstack compute service list' using the overcloudrc env
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHEA-2020:0283