Bug 1325375

Summary: rhel-osp-director: 7.3->8.0 upgrade with cinder/glance NFS backend fails during major-upgrade-pacemaker.yaml step. openstack-cinder-scheduler pcs resource fails to start on 2 out of 3 controllers.
Product: Red Hat OpenStack Reporter: Alexander Chuzhoy <sasha>
Component: rhosp-directorAssignee: Jiri Stransky <jstransk>
Status: CLOSED NOTABUG QA Contact: Arik Chernetsky <achernet>
Severity: high Docs Contact:
Priority: high    
Version: 8.0 (Liberty)CC: dbecker, mburns, morazi, rhel-osp-director-maint
Target Milestone: async   
Target Release: 8.0 (Liberty)   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: Doc Type: Bug Fix
Doc Text:
Story Points: ---
Clone Of: Environment:
Last Closed: 2016-04-13 18:13:09 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:

Description Alexander Chuzhoy 2016-04-08 15:32:46 UTC
rhel-osp-director: 7.3->8.0 upgrade with cinder/glance NFS backend fails during major-upgrade-pacemaker.yaml step. openstack-cinder-scheduler pcs resource fails to start on 2 out of 3 controllers.


Environment:
openstack-tripleo-heat-templates-0.8.14-7.el7ost.noarch
instack-undercloud-2.2.7-4.el7ost.noarch
openstack-puppet-modules-7.0.17-1.el7ost.noarch
openstack-tripleo-heat-templates-kilo-0.8.14-7.el7ost.noarch


Steps to reproduce:
1. deploy 7.3 with glance/cinder using NFS
2. Attempt to upgrade the setup to 8.0


Result:
Upgrade fails:
"ERROR: cluster remained unstable for more than 1800 seconds, exiting.\n


Checking the cluster  unable to revive the openstack-cinder-scheduler resources.


Expected result:
the upgrade should continue.

Comment 2 Jiri Stransky 2016-04-08 17:01:51 UTC
The upgrade failure on the investigated environment was induced by DNS instability.

Comment 3 Mike Burns 2016-04-13 18:13:09 UTC
based on comment 2 and based on bug 1325476 which completed an upgrade with NFS backends, closing this bug.  Please reopen if it reproduces.