Bug 1174389 - rubygem-staypuft: Failed actions in pcs status report after a successful deployment.
Summary: rubygem-staypuft: Failed actions in pcs status report after a successful dep...
Keywords:
Status: CLOSED CURRENTRELEASE
Alias: None
Product: Red Hat OpenStack
Classification: Red Hat
Component: openstack-foreman-installer
Version: unspecified
Hardware: x86_64
OS: Linux
high
high
Target Milestone: ---
: Installer
Assignee: Jason Guiditta
QA Contact: yeylon@redhat.com
URL:
Whiteboard:
: 1174784 1174800 1179493 (view as bug list)
Depends On:
Blocks: 1177026
TreeView+ depends on / blocked
 
Reported: 2014-12-15 19:09 UTC by Alexander Chuzhoy
Modified: 2016-04-18 06:54 UTC (History)
9 users (show)

Fixed In Version:
Doc Type: Bug Fix
Doc Text:
Clone Of:
Environment:
Last Closed: 2016-04-04 15:45:33 UTC
Target Upstream Version:
Embargoed:


Attachments (Terms of Use)
attaching logs from a host where pcs reports failure (461.16 KB, application/x-gzip)
2014-12-15 19:16 UTC, Alexander Chuzhoy
no flags Details
logs/conf from a controller reported with issues (10.20 MB, application/x-gzip)
2014-12-15 19:46 UTC, Alexander Chuzhoy
no flags Details
Logs from the second controller (8.81 MB, application/x-gzip)
2014-12-15 19:49 UTC, Alexander Chuzhoy
no flags Details
Logs from the third controller (8.04 MB, application/x-gzip)
2014-12-15 19:52 UTC, Alexander Chuzhoy
no flags Details

Description Alexander Chuzhoy 2014-12-15 19:09:56 UTC
rubygem-staypuft:  Failed actions in pscs status report after a successful deployment.
Environment:
openstack-foreman-installer-3.0.5-1.el7ost.noarch
ruby193-rubygem-staypuft-0.5.5-1.el7ost.noarch
ruby193-rubygem-foreman_openstack_simplify-0.0.6-8.el7ost.noarch
rhel-osp-installer-client-0.5.2-2.el7ost.noarch
openstack-puppet-modules-2014.2.7-2.el7ost.noarch
rhel-osp-installer-0.5.2-2.el7ost.noarch

Steps to reproduce:
1. Install rhel-osp-installer
2. Complete a successful deployment of HAneutron (3 controllers + 1 compute)
3. Check if there are failed actions in the output from 'pcs status':

Result:
Failed actions:
    openstack-ceilometer-central_monitor_30000 on pcmk-maca25400702875 'not running' (7): call=571, status=complete, last-rc-change='Mon Dec 15 13:50:33 2014', queued=0ms, exec=0ms
    neutron-server_monitor_30000 on pcmk-maca25400702875 'OCF_PENDING' (196): call=527, status=complete, last-rc-change='Mon Dec 15 13:49:29 2014', queued=0ms, exec=0ms

Expected result:
The 'Failed actions' section shouldn't report these.

Comment 1 Alexander Chuzhoy 2014-12-15 19:16:05 UTC
Created attachment 969236 [details]
attaching logs from a host where pcs reports failure

Comment 2 Alexander Chuzhoy 2014-12-15 19:17:08 UTC
Full output from pcs status:
[root@maca25400702875 ~]# pcs status                                                         
Cluster name: openstack                                                                      
Last updated: Mon Dec 15 14:16:47 2014                                                       
Last change: Mon Dec 15 11:34:13 2014 via crmd on pcmk-maca25400702875                       
Stack: corosync                                                                              
Current DC: pcmk-maca25400702875 (2) - partition with quorum                                 
Version: 1.1.10-32.el7_0.1-368c726                                                           
3 Nodes configured                                                                           
124 Resources configured                                                                     


Online: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]

Full list of resources:

 ip-192.168.0.4 (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702875 
 ip-192.168.0.3 (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702876 
 ip-192.168.0.29        (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702877 
 ip-192.168.0.31        (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702875 
 ip-192.168.0.2 (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702876         
 ip-192.168.0.32        (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702877 
 ip-192.168.0.23        (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702875 
 ip-192.168.0.24        (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702876 
 ip-192.168.0.25        (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702877 
 ip-192.168.0.36        (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702875 
 Clone Set: memcached-clone [memcached]                                              
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]     
 Clone Set: rabbitmq-server-clone [rabbitmq-server]                                  
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]     
 ip-192.168.0.30        (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702876 
 Clone Set: haproxy-clone [haproxy]                                                  
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]     
 ip-192.168.0.13        (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702877 
 Master/Slave Set: galera-master [galera]                                            
     Masters: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]     
 ip-192.168.0.26        (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702875 
 ip-192.168.0.28        (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702876 
 ip-192.168.0.27        (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702877 
 Clone Set: openstack-keystone-clone [openstack-keystone]                            
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]     
 ip-192.168.0.14        (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702875 
 Clone Set: fs-varlibglanceimages-clone [fs-varlibglanceimages]                      
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]     
 ip-192.168.0.16        (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702876 
 ip-192.168.0.15        (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702877 
 Clone Set: openstack-glance-registry-clone [openstack-glance-registry]              
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]     
 Clone Set: openstack-glance-api-clone [openstack-glance-api]                        
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]     
 ip-192.168.0.35        (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702875 
 ip-192.168.0.33        (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702876 
 ip-192.168.0.34        (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702877 
 Clone Set: openstack-nova-consoleauth-clone [openstack-nova-consoleauth]            
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]     
 Clone Set: openstack-nova-api-clone [openstack-nova-api]                            
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]     
 Clone Set: openstack-nova-novncproxy-clone [openstack-nova-novncproxy]              
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]     
 Clone Set: openstack-nova-conductor-clone [openstack-nova-conductor]                
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]     
 Clone Set: openstack-nova-scheduler-clone [openstack-nova-scheduler]                
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]     
 ip-192.168.0.6 (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702875         
 ip-192.168.0.5 (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702876         
 ip-192.168.0.12        (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702877 
 Clone Set: openstack-cinder-api-clone [openstack-cinder-api]                        
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]     
 Clone Set: openstack-cinder-scheduler-clone [openstack-cinder-scheduler]            
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]     
 openstack-cinder-volume        (systemd:openstack-cinder-volume):      Started pcmk-maca25400702875 
 Clone Set: neutron-server-clone [neutron-server]                                                    
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]                     
 ip-192.168.0.17        (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702876                 
 ip-192.168.0.18        (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702877                 
 ip-192.168.0.19        (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702875                 
 ip-192.168.0.20        (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702876                 
 ip-192.168.0.22        (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702877                 
 ip-192.168.0.21        (ocf::heartbeat:IPaddr2):       Started pcmk-maca25400702875                 
 Clone Set: neutron-ovs-cleanup-clone [neutron-ovs-cleanup]                                          
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]                     
 Clone Set: neutron-netns-cleanup-clone [neutron-netns-cleanup]                                      
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]                     
 Resource Group: neutron-agents                                                                      
     neutron-openvswitch-agent  (systemd:neutron-openvswitch-agent):    Started pcmk-maca25400702876 
     neutron-dhcp-agent (systemd:neutron-dhcp-agent):   Started pcmk-maca25400702876                 
     neutron-l3-agent   (systemd:neutron-l3-agent):     Started pcmk-maca25400702876                 
     neutron-metadata-agent     (systemd:neutron-metadata-agent):       Started pcmk-maca25400702876 
 Clone Set: openstack-heat-api-clone [openstack-heat-api]                                            
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]                     
 Resource Group: heat                                                                                
     openstack-heat-engine      (systemd:openstack-heat-engine):        Started pcmk-maca25400702877 
 Clone Set: openstack-heat-api-cfn-clone [openstack-heat-api-cfn]                                    
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]                     
 Clone Set: openstack-heat-api-cloudwatch-clone [openstack-heat-api-cloudwatch]                      
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]                     
 Clone Set: httpd-clone [httpd]                                                                      
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]                     
 Clone Set: mongod-clone [mongod]                                                                    
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]                     
 Clone Set: ceilometer-delay-clone [ceilometer-delay]                                                
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]                     
 openstack-ceilometer-central   (systemd:openstack-ceilometer-central): Started pcmk-maca25400702875
 Clone Set: openstack-ceilometer-api-clone [openstack-ceilometer-api]
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]
 Clone Set: openstack-ceilometer-alarm-notifier-clone [openstack-ceilometer-alarm-notifier]
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]
 Clone Set: openstack-ceilometer-collector-clone [openstack-ceilometer-collector]
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]
 Clone Set: openstack-ceilometer-notification-clone [openstack-ceilometer-notification]
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]
 Clone Set: openstack-ceilometer-alarm-evaluator-clone [openstack-ceilometer-alarm-evaluator]
     Started: [ pcmk-maca25400702875 pcmk-maca25400702876 pcmk-maca25400702877 ]

Failed actions:
    openstack-ceilometer-central_monitor_30000 on pcmk-maca25400702875 'not running' (7): call=571, status=complete, last-rc-change='Mon Dec 15 13:50:33 2014', queued=0ms, exec=0ms
    neutron-server_monitor_30000 on pcmk-maca25400702875 'OCF_PENDING' (196): call=527, status=complete, last-rc-change='Mon Dec 15 13:49:29 2014', queued=0ms, exec=0ms


PCSD Status:
  pcmk-maca25400702876: Online
  pcmk-maca25400702875: Online
  pcmk-maca25400702877: Online

Daemon Status:
  corosync: active/enabled
  pacemaker: active/enabled
  pcsd: active/enabled

Comment 4 Alexander Chuzhoy 2014-12-15 19:46:15 UTC
Created attachment 969276 [details]
logs/conf from a controller reported with issues

Comment 5 Alexander Chuzhoy 2014-12-15 19:49:03 UTC
Created attachment 969278 [details]
Logs from the second controller

Comment 6 Alexander Chuzhoy 2014-12-15 19:52:19 UTC
Created attachment 969280 [details]
Logs from the third controller

Comment 7 Crag Wolfe 2014-12-16 03:53:26 UTC
/var/log/ceilometer/central.log on c1 includes the following output (multiple times):

File Edit Options Buffers Tools Help
2014-12-15 11:20:12.409 23864 ERROR ceilometer.nova_client [-] The request you have made requires authentication. (HTTP 401)
2014-12-15 11:20:12.409 23864 TRACE ceilometer.nova_client Traceback (most recent call last):
2014-12-15 11:20:12.409 23864 TRACE ceilometer.nova_client   File "/usr/lib/python2.7/site-packages/ceilometer/nova_client.py", line 51, in with_logging
2014-12-15 11:20:12.409 23864 TRACE ceilometer.nova_client     return func(*args, **kwargs)
2014-12-15 11:20:12.409 23864 TRACE ceilometer.nova_client   File "/usr/lib/python2.7/site-packages/ceilometer/nova_client.py", line 155, in instance_get_all
2014-12-15 11:20:12.409 23864 TRACE ceilometer.nova_client     search_opts=search_opts)
2014-12-15 11:20:12.409 23864 TRACE ceilometer.nova_client   File "/usr/lib/python2.7/site-packages/novaclient/v1_1/servers.py", line 603, in list
2014-12-15 11:20:12.409 23864 TRACE ceilometer.nova_client     return self._list("/servers%s%s" % (detail, query_string), "servers")
2014-12-15 11:20:12.409 23864 TRACE ceilometer.nova_client   File "/usr/lib/python2.7/site-packages/novaclient/base.py", line 67, in _list
2014-12-15 11:20:12.409 23864 TRACE ceilometer.nova_client     _resp, body = self.api.client.get(url)
2014-12-15 11:20:12.409 23864 TRACE ceilometer.nova_client   File "/usr/lib/python2.7/site-packages/novaclient/client.py", line 487, in get
2014-12-15 11:20:12.409 23864 TRACE ceilometer.nova_client     return self._cs_request(url, 'GET', **kwargs)
2014-12-15 11:20:12.409 23864 TRACE ceilometer.nova_client   File "/usr/lib/python2.7/site-packages/novaclient/client.py", line 446, in _cs_request
2014-12-15 11:20:12.409 23864 TRACE ceilometer.nova_client     self.authenticate()
2014-12-15 11:20:12.409 23864 TRACE ceilometer.nova_client   File "/usr/lib/python2.7/site-packages/novaclient/client.py", line 586, in authenticate
2014-12-15 11:20:12.409 23864 TRACE ceilometer.nova_client     auth_url = self._v2_auth(auth_url)
2014-12-15 11:20:12.409 23864 TRACE ceilometer.nova_client   File "/usr/lib/python2.7/site-packages/novaclient/client.py", line 677, in _v2_auth
2014-12-15 11:20:12.409 23864 TRACE ceilometer.nova_client     return self._authenticate(url, body)
2014-12-15 11:20:12.409 23864 TRACE ceilometer.nova_client   File "/usr/lib/python2.7/site-packages/novaclient/client.py", line 690, in _authenticate
2014-12-15 11:20:12.409 23864 TRACE ceilometer.nova_client     **kwargs)
2014-12-15 11:20:12.409 23864 TRACE ceilometer.nova_client   File "/usr/lib/python2.7/site-packages/novaclient/client.py", line 439, in _time_request
2014-12-15 11:20:12.409 23864 TRACE ceilometer.nova_client     resp, body = self.request(url, method, **kwargs)
2014-12-15 11:20:12.409 23864 TRACE ceilometer.nova_client   File "/usr/lib/python2.7/site-packages/novaclient/client.py", line 433, in request
2014-12-15 11:20:12.409 23864 TRACE ceilometer.nova_client     raise exceptions.from_response(resp, body, url, method)
2014-12-15 11:20:12.409 23864 TRACE ceilometer.nova_client Unauthorized: The request you have made requires authentication. (HTTP 401)
2014-12-15 11:20:12.409 23864 TRACE ceilometer.nova_client

(and similar log lines for ceilometer.agent instead of ceilometer.nova_client).

to be continued...

Comment 8 Mike Burns 2014-12-16 14:03:19 UTC
*** Bug 1174784 has been marked as a duplicate of this bug. ***

Comment 9 Mike Burns 2014-12-16 14:44:38 UTC
*** Bug 1174800 has been marked as a duplicate of this bug. ***

Comment 10 Eoghan Glynn 2014-12-16 17:47:34 UTC
Looking at the cN-logs tarballs attached, I noticed that the configured auth_url differs between the services, specifically both ceilometer & neutron use 192.168.0.26:35357, whereas nova uses 192.168.0.28:35357 and glance uses localhost:5000, e.g.

$ find . -name "*.conf" | xargs grep auth_url | grep -v '#'
./etc/neutron/neutron.conf:nova_admin_auth_url =http://192.168.0.26:35357/v2.0
./etc/ceilometer/ceilometer.conf:os_auth_url=http://192.168.0.26:35357/v2.0
./etc/nova/nova.conf:admin_auth_url=http://192.168.0.28:35357/v2.0
./etc/glance/glance-cache.conf:auth_url = http://localhost:5000/v2.0

Do we expect that spread of keystone addresses for this kind of deployment?

Is the keystone instance addressable at 192.168.0.26 actually running cleanly?

Comment 11 Crag Wolfe 2014-12-16 18:14:02 UTC
An older version is probably the issue with ceilometer.  Looking at ceilometer.conf from an uploaded log:

$ grep rabbit_host c1/etc/ceilometer/ceilometer.conf 
#rabbit_host=localhost
rabbit_host=127.0.0.1
#rabbit_hosts=$rabbit_host:$rabbit_port
rabbit_hosts=127.0.0.1:5672

Instead of 127.0.0.1, it should be a rabbit vip which was fixed in:
https://bugzilla.redhat.com/show_bug.cgi?id=1173217

Comment 12 Crag Wolfe 2014-12-16 19:23:01 UTC
Regarding auth_url question, given haproxy.cfg that looks like:

listen keystone-admin
  bind 192.168.0.28:35357
  bind 192.168.0.27:35357
  bind 192.168.0.26:35357
  mode  tcp
  option  tcplog
  server pcmk-maca25400702876 192.168.0.8:35357  check inter 1s
  server pcmk-maca25400702875 192.168.0.7:35357  check inter 1s
  server pcmk-maca25400702877 192.168.0.10:35357  check inter 1s

either the .26 or .28 url should work.  The glance-cache.conf auth_url looks broken, however.

Comment 13 Mike Burns 2014-12-17 20:54:30 UTC
(In reply to Crag Wolfe from comment #12)

> 
> either the .26 or .28 url should work.  The glance-cache.conf auth_url looks
> broken, however.

This part was fixed as part of the openstack-foreman-installer-3.0.8-1.el7ost build

Comment 14 Crag Wolfe 2014-12-18 04:38:34 UTC
Regarding the neutron-server_monitor failed action -- pacemaker makes a number of attempts at starting neutron-server:

$ grep 'Recover neutron-server' messages
Dec 15 11:26:19 maca25400702875 pengine[15268]: notice: LogActions: Recover neutron-server:1    (Started pcmk-maca25400702876)
Dec 15 11:32:00 maca25400702875 pengine[15268]: notice: LogActions: Recover neutron-server:0    (Started pcmk-maca25400702875)
Dec 15 11:48:07 maca25400702875 pengine[15268]: notice: LogActions: Recover neutron-server:0    (Started pcmk-maca25400702875)
Dec 15 11:48:10 maca25400702875 pengine[15268]: notice: LogActions: Recover neutron-server:0    (Started pcmk-maca25400702875)
Dec 15 12:49:34 maca25400702875 pengine[15268]: notice: LogActions: Recover neutron-server:0    (Started pcmk-maca25400702875)
Dec 15 13:49:32 maca25400702875 pengine[15268]: notice: LogActions: Recover neutron-server:0    (Started pcmk-maca25400702875)
Dec 15 13:49:35 maca25400702875 pengine[15268]: notice: LogActions: Recover neutron-server:0    (Started pcmk-maca25400702875)

However, I'm not finding the smoking gun that caused pacemaker to log that.  And there is evidence of neutron-server being restarted periodically:

$ grep 'Config paste file' neutron/server.log

2014-12-15 11:08:04.147 3858 INFO neutron.common.config [-] Config paste file: /usr/share/neutron/api-paste.ini
2014-12-15 11:08:58.880 5028 INFO neutron.common.config [-] Config paste file: /usr/share/neutron/api-paste.ini
2014-12-15 11:25:21.637 1708 INFO neutron.common.config [-] Config paste file: /usr/share/neutron/api-paste.ini
2014-12-15 11:31:58.455 14468 INFO neutron.common.config [-] Config paste file: /usr/share/neutron/api-paste.ini
2014-12-15 11:32:07.814 14662 INFO neutron.common.config [-] Config paste file: /usr/share/neutron/api-paste.ini
2014-12-15 11:48:07.838 8659 INFO neutron.common.config [-] Config paste file: /usr/share/neutron/api-paste.ini
2014-12-15 11:48:15.858 8810 INFO neutron.common.config [-] Config paste file: /usr/share/neutron/api-paste.ini
2014-12-15 12:19:42.713 25450 INFO neutron.common.config [-] Config paste file: /usr/share/neutron/api-paste.ini
2014-12-15 12:49:25.953 6900 INFO neutron.common.config [-] Config paste file: /usr/share/neutron/api-paste.ini
2014-12-15 12:49:41.749 7483 INFO neutron.common.config [-] Config paste file: /usr/share/neutron/api-paste.ini
2014-12-15 13:19:33.439 21549 INFO neutron.common.config [-] Config paste file: /usr/share/neutron/api-paste.ini
2014-12-15 13:49:30.586 3604 INFO neutron.common.config [-] Config paste file: /usr/share/neutron/api-paste.ini
2014-12-15 13:49:38.705 3752 INFO neutron.common.config [-] Config paste file: /usr/share/neutron/api-paste.ini
2014-12-15 14:19:35.128 18144 INFO neutron.common.config [-] Config paste file: /usr/share/neutron/api-paste.ini

But it does not appear that neutron-server process was actually dead before these restarts took place.

I am not seeing any of these restarts or failed actions in my development environment, making this difficult to debug.

Going back to the well and seeing if David has any additional insights (thanks David!).

Comment 15 David Vossel 2014-12-18 17:28:23 UTC
Did you all increase the start timeout for neutron-server yet? I saw something similar to this last night and recommended setting 'op start timeout=90s' for neutron-server.

pcs resource create neutron-server systemd:neutron-server op start timeout=90s --clone

or to update a running deployment

pcs resource op add neutron-server start timeout=90s

-- David

Comment 16 Jason Guiditta 2014-12-18 18:39:55 UTC
(In reply to David Vossel from comment #15)
> Did you all increase the start timeout for neutron-server yet? I saw
> something similar to this last night and recommended setting 'op start
> timeout=90s' for neutron-server.
> 
> pcs resource create neutron-server systemd:neutron-server op start
> timeout=90s --clone
> 
> or to update a running deployment
> 
> pcs resource op add neutron-server start timeout=90s
> 
> -- David

We will be adding this as part of our effort to align with the newly released HA ref arch for OSP 6.  Fabio, the arch has a timeout of 60 seconds, and both this BZ and https://bugzilla.redhat.com/show_bug.cgi?id=1175525 suggest 90, which should we be using?

Comment 17 Fabio Massimo Di Nitto 2014-12-18 21:55:06 UTC
(In reply to Jason Guiditta from comment #16)
> (In reply to David Vossel from comment #15)
> > Did you all increase the start timeout for neutron-server yet? I saw
> > something similar to this last night and recommended setting 'op start
> > timeout=90s' for neutron-server.
> > 
> > pcs resource create neutron-server systemd:neutron-server op start
> > timeout=90s --clone
> > 
> > or to update a running deployment
> > 
> > pcs resource op add neutron-server start timeout=90s
> > 
> > -- David
> 
> We will be adding this as part of our effort to align with the newly
> released HA ref arch for OSP 6.  Fabio, the arch has a timeout of 60
> seconds, and both this BZ and
> https://bugzilla.redhat.com/show_bug.cgi?id=1175525 suggest 90, which should
> we be using?

90 seconds is fine. I'll fix the doc.

Comment 18 Mike Burns 2015-01-07 14:32:54 UTC
*** Bug 1179493 has been marked as a duplicate of this bug. ***


Note You need to log in before you can comment on or make changes to this bug.