Bug 1669078

Summary: Add support for configuring Octavia LB timeouts in OSP 13
Product: Red Hat OpenStack Reporter: Carlos Goncalves <cgoncalves>
Component: openstack-octaviaAssignee: Kamil Sambor <ksambor>
Status: CLOSED ERRATA QA Contact: Jon Uriarte <juriarte>
Severity: high Docs Contact:
Priority: high    
Version: 13.0 (Queens)CC: ihrachys, lpeer, majopela
Target Milestone: z5Keywords: Triaged, ZStream
Target Release: 13.0 (Queens)   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: openstack-octavia-2.0.2-5.el7ost Doc Type: Enhancement
Doc Text:
Story Points: ---
Clone Of: 1636496 Environment:
Last Closed: 2019-03-14 13:33:12 UTC Type: ---
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:
Bug Depends On: 1636496    
Bug Blocks: 1623989    

Comment 11 Jon Uriarte 2019-02-28 15:09:57 UTC
Waiting to [1] to be fixed, as it blocks Openshift installation.


[1] https://bugzilla.redhat.com/show_bug.cgi?id=1684077

Comment 12 Jon Uriarte 2019-03-01 14:51:13 UTC
Verified in OSP13 2019-02-25.2 puddle.

[stack@undercloud-0 ~]$ cat /etc/yum.repos.d/latest-installed 
13  -p 2019-02-25.2

[stack@undercloud-0 ~]$ rpm -qa | grep openstack-tripleo-heat-templates
openstack-tripleo-heat-templates-8.2.0-4.el7ost.noarch

[stack@undercloud-0 ~]$ rpm -qa | grep puppet-octavia
puppet-octavia-12.4.0-8.el7ost.noarch

- octavia_health_manager container:
openstack-octavia-common-2.0.3-2.el7ost.noarch
openstack-octavia-health-manager-2.0.3-2.el7ost.noarch

- octavia_api container:
openstack-octavia-common-2.0.3-2.el7ost.noarch
openstack-octavia-api-2.0.3-2.el7ost.noarch

- octavia_housekeeping container:
openstack-octavia-common-2.0.3-2.el7ost.noarch
openstack-octavia-housekeeping-2.0.3-2.el7ost.noarch

- octavia_worker container:
openstack-octavia-common-2.0.3-2.el7ost.noarch
openstack-octavia-worker-2.0.3-2.el7ost.noarch



Verification steps:

1. Deploy OSP 13 with octavia in a hybrid environment
  ยท Set in customized Jenkins job OVERCLOUD_DEPLOY_OVERRIDE_OPTIONS:

     --config-heat OctaviaTimeoutClientData=1200000 \
     --config-heat OctaviaTimeoutMemberData=1200000

2. Check the new timeouts values are reflected in configuration (in the controller):

    /var/lib/config-data/puppet-generated/octavia/etc/octavia/octavia.conf:timeout_client_data=1200000
    /var/lib/config-data/puppet-generated/octavia/etc/octavia/octavia.conf:timeout_member_connect=5000
    /var/lib/config-data/puppet-generated/octavia/etc/octavia/octavia.conf:timeout_member_data=1200000
    /var/lib/config-data/puppet-generated/octavia/etc/octavia/octavia.conf:timeout_tcp_inspect=0
    /var/lib/config-data/clustercheck/etc/puppet/hieradata/service_configs.json:    "octavia::controller::timeout_client_data": 1200000,
    /var/lib/config-data/clustercheck/etc/puppet/hieradata/service_configs.json:    "octavia::controller::timeout_member_connect": 5000,
    /var/lib/config-data/clustercheck/etc/puppet/hieradata/service_configs.json:    "octavia::controller::timeout_member_data": 1200000,
    /var/lib/config-data/clustercheck/etc/puppet/hieradata/service_configs.json:    "octavia::controller::timeout_tcp_inspect": 0,

3. Deploy OSP 3.11 with Kuryr and check the cluster is ready and all the pods are running:
Note: A workaround for [1] has been applied - "yum install openshift-ansible --enablerepo=rhelosp-rhel-7.6-server-opt"

[openshift@master-0 ~]$ oc version
oc v3.11.90
kubernetes v1.11.0+d4cacc0
features: Basic-Auth GSSAPI Kerberos SPNEGO

Server https://console.openshift.example.com:8443
openshift v3.11.90
kubernetes v1.11.0+d4cacc0


[openshift@master-0 ~]$ oc get nodes
NAME                                 STATUS    ROLES     AGE       VERSION
app-node-0.openshift.example.com     Ready     compute   2h        v1.11.0+d4cacc0
app-node-1.openshift.example.com     Ready     compute   2h        v1.11.0+d4cacc0
infra-node-0.openshift.example.com   Ready     infra     2h        v1.11.0+d4cacc0
master-0.openshift.example.com       Ready     master    2h        v1.11.0+d4cacc0


[openshift@master-0 ~]$ oc get pods --all-namespaces
NAMESPACE                   NAME                                                READY     STATUS    RESTARTS   AGE
default                     docker-registry-1-p6b2d                             1/1       Running   0          2h
default                     kuryr-pod-1958724533                                1/1       Running   0          1h
default                     registry-console-1-c7f79                            1/1       Running   0          2h
default                     router-1-vs7ld                                      1/1       Running   0          2h
kube-system                 master-api-master-0.openshift.example.com           1/1       Running   0          2h
kube-system                 master-controllers-master-0.openshift.example.com   1/1       Running   0          2h
kube-system                 master-etcd-master-0.openshift.example.com          1/1       Running   0          2h
kuryr-namespace-699852674   kuryr-pod-436421213                                 1/1       Running   0          1h
kuryr-namespace-920831344   kuryr-pod-826298564                                 1/1       Running   0          1h
openshift-console           console-5d5b6bd95d-m487h                            1/1       Running   0          2h
openshift-infra             kuryr-cni-ds-6w4hx                                  2/2       Running   0          1h
openshift-infra             kuryr-cni-ds-pdfx5                                  2/2       Running   0          1h
openshift-infra             kuryr-cni-ds-pkzxj                                  2/2       Running   0          1h
openshift-infra             kuryr-cni-ds-xpg25                                  2/2       Running   0          1h
openshift-infra             kuryr-controller-6c6d965c54-k68fn                   1/1       Running   0          1h
openshift-monitoring        alertmanager-main-0                                 3/3       Running   0          2h
openshift-monitoring        alertmanager-main-1                                 3/3       Running   0          2h
openshift-monitoring        alertmanager-main-2                                 3/3       Running   0          2h
openshift-monitoring        cluster-monitoring-operator-75c6b544dd-4swg4        1/1       Running   0          2h
openshift-monitoring        grafana-c7d5bc87c-qw92g                             2/2       Running   0          2h
openshift-monitoring        kube-state-metrics-6c64799586-gmjv5                 3/3       Running   0          2h
openshift-monitoring        node-exporter-57hdh                                 2/2       Running   0          2h
openshift-monitoring        node-exporter-mptl8                                 2/2       Running   0          2h
openshift-monitoring        node-exporter-qdqq5                                 2/2       Running   0          2h
openshift-monitoring        node-exporter-rqm49                                 2/2       Running   0          2h
openshift-monitoring        prometheus-k8s-0                                    4/4       Running   1          2h
openshift-monitoring        prometheus-k8s-1                                    4/4       Running   1          2h
openshift-monitoring        prometheus-operator-5b47ff445b-w5v4c                1/1       Running   0          2h
openshift-node              sync-fv2tv                                          1/1       Running   0          2h
openshift-node              sync-gklcl                                          1/1       Running   0          2h
openshift-node              sync-gr62d                                          1/1       Running   0          2h
openshift-node              sync-k6nn2                                          1/1       Running   0          2h



[1] https://bugzilla.redhat.com/show_bug.cgi?id=1684077

Comment 14 errata-xmlrpc 2019-03-14 13:33:12 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory, and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHSA-2019:0567