Description of problem: Deleted a failed overcloud, however, one node seems to be totally out of sync: (undercloud) [stack@f13-h17-b01-5039ms ~]$ nova list +----+------+--------+------------+-------------+----------+ | ID | Name | Status | Task State | Power State | Networks | +----+------+--------+------------+-------------+----------+ +----+------+--------+------------+-------------+----------+ (undercloud) [stack@f13-h17-b01-5039ms ~]$ openstack baremetal node list +--------------------------------------+------+--------------------------------------+-------------+--------------------+-------------+ | UUID | Name | Instance UUID | Power State | Provisioning State | Maintenance | +--------------------------------------+------+--------------------------------------+-------------+--------------------+-------------+ | 0470f116-78bc-45a4-bde2-735635570d89 | None | None | power off | available | False | | 55a398d0-c96d-42ad-b524-c66dc1f5268f | None | None | power off | available | False | | 2f2acd02-13fb-479b-9e22-6933fe4b3477 | None | None | power off | available | False | | e5836ee0-6ae3-4d56-bf03-893e6d69fb09 | None | None | power off | available | False | | 4956639d-ed42-49f0-80ea-ffeb87e260c9 | None | 3ae22c0f-d1cc-46a8-9d8b-7d5b0083ff82 | power off | deploying | False | | 858124ff-7802-489e-a58e-10dbd940ec22 | None | None | power off | available | False | | 995b580b-16a6-419a-bd11-4ace5a4af4b1 | None | None | power off | available | False | | ad39e4fb-b271-4104-9842-05f49aed24f4 | None | None | power off | available | False | | 8dd5afc4-05b7-4d0e-a051-ab0a89a51523 | None | None | power off | available | False | | bedd53ad-3ca6-4827-8eb2-4f44fa93f710 | None | None | power off | available | False | | 24e36e13-10d6-4dd6-ab9d-6225d97e4939 | None | None | power off | available | False | | 1fdf2237-d8ba-463e-b4a1-a98239e9b4a2 | None | None | power off | available | False | | 0ed86b28-7f66-4147-9333-147cf83e75ba | None | None | power off | available | False | | 550222e2-caca-479b-9feb-889a10aa02d3 | None | None | power off | available | False | | ad59d511-2494-45f8-b800-c745a7d92255 | None | None | power off | available | False | | f59572ff-2b54-45c6-a701-6bdc2beeed11 | None | None | power off | available | False | | 0e9df278-1d5f-448b-8582-0df11a757a7e | None | None | power off | available | False | | 7c255cd6-e944-479c-9a0f-13af04e98e94 | None | None | power off | available | False | | 913e60e4-0af5-48b8-91f1-4cfef4b728fb | None | None | power off | available | False | | 3f23ce06-403e-46d0-be7b-b614f2ca426c | None | None | power off | available | False | | 7a540ffa-fb70-48f6-8f9a-518a945c31cd | None | None | power off | available | False | | b2d7af36-5bd6-4589-bbf0-8944c6cfe9f6 | None | None | power off | available | False | | 094c22f0-4383-4461-a319-c8540a912da8 | None | None | power off | available | False | | 51ca3752-be7a-4da0-b385-e549eee85b04 | None | None | power off | available | False | | 03e4d1af-514b-4bf1-8d68-60572920e8f0 | None | None | power off | available | False | | 9798c5ad-5730-42b5-9b7e-5cd741fd42ea | None | None | power off | available | False | | a3fb4688-c7f9-4ad1-a4d2-ce63070578ce | None | None | power off | available | False | | f8dd23fe-9144-4f7d-80ae-cc9f9c028aac | None | None | power off | available | False | | dbdb580d-c30e-44fa-8c35-82d25afecbbb | None | None | power off | available | False | +--------------------------------------+------+--------------------------------------+-------------+--------------------+-------------+ (undercloud) [stack@f13-h17-b01-5039ms ~]$ openstack stack list Version-Release number of selected component (if applicable): OSP14 GA How reproducible: n/a
I was able to set the node to maintenance and remove it from the inventory.
Hi! I agree it's unfortunate, but it essentially works as intended. The "deploying" step is not interruptible, so when you delete your stack with nodes in this state, they cannot be immediately released. If you paste the conductor logs I can try to understand why the node got stuck, but it's probably just iscsi doing something. I wonder if we need a KB item on "deploying" and "cleaning" states..
ack - yeah having it documented how to get out of this hairy state would of been nice. I had to put the nodes into maintenance and clean up the deployment before re-kicking an overcloud deploy. Didn't know that the instance uuid shouldn't of been taken care of for the user -- hence the bug.
Triage: Include a definition list of provisioning states, describing each one
Provisioning states available on the Customer Portal: 17.0: https://access.redhat.com/documentation/en-us/red_hat_openstack_platform/17.0/html/bare_metal_provisioning/assembly_configuring-the-bare-metal-provisioning-service-after-deployment#ref_bare-metal-node-provisioning-states_bare-metal-post-deployment 16.2: https://access.redhat.com/documentation/en-us/red_hat_openstack_platform/16.2/html/bare_metal_provisioning/assembly_configuring-the-bare-metal-provisioning-service-after-deployment#ref_bare-metal-node-provisioning-states_bare-metal-post-deployment 16.1: https://access.redhat.com/documentation/en-us/red_hat_openstack_platform/16.1/html/bare_metal_provisioning/assembly_configuring-the-bare-metal-provisioning-service-after-deployment#ref_bare-metal-node-provisioning-states_bare-metal-post-deployment