Bug 1685970
Summary: | Fluentd daemonset doesn't include a tolerate everything toleration(3.11) | ||
---|---|---|---|
Product: | OpenShift Container Platform | Reporter: | Juan Luis de Sousa-Valadas <jdesousa> |
Component: | Logging | Assignee: | Jeff Cantrill <jcantril> |
Status: | CLOSED ERRATA | QA Contact: | Siva Reddy <schituku> |
Severity: | medium | Docs Contact: | |
Priority: | medium | ||
Version: | 3.10.0 | CC: | aos-bugs, jdesousa, jokerman, mmccomas, rmeggins, schituku |
Target Milestone: | --- | Flags: | jdesousa:
needinfo-
|
Target Release: | 3.11.z | ||
Hardware: | Unspecified | ||
OS: | Unspecified | ||
Whiteboard: | |||
Fixed In Version: | Doc Type: | Bug Fix | |
Doc Text: |
Cause:
The fluentd daemonset doesn't include a tolerate everything toleration
Consequence:
If a customer taints a node, fluentd pod gets evicted
Fix:
Add a toleration
Result:
Fluentd pods don't get evicted
|
Story Points: | --- |
Clone Of: | Environment: | ||
Last Closed: | 2019-04-11 05:38:40 UTC | Type: | Bug |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: | |||
Bug Depends On: | |||
Bug Blocks: | 1690200 |
Description
Juan Luis de Sousa-Valadas
2019-03-06 12:15:14 UTC
Created a pull request to fix it. https://github.com/openshift/openshift-ansible/pull/11310 Please update the doctype and text. Version: openshift-ansible-3.11.95-1.git.0.d080cce.el7.noarch.rpm oc v3.11.95 kubernetes v1.11.0+d4cacc0 features: Basic-Auth GSSAPI Kerberos SPNEGO Server openshift v3.11.95 kubernetes v1.11.0+d4cacc0 Steps to reproduce: 1. taint the node # oc adm taint node $node NodeWithImpairedVolumes=true:NoExecute #oc describe node $node | grep -i taint 2. delete the fluentd pods # oc get ds logging-fluentd -o yaml > log-ds.yaml # oc delete ds logging-fluentd # oc create -f logging-fluentd 3. Note the logging pods The logging pods get created with out any issue in spite of taint being present on the node. oc get pods -n openshift-node -o wide | grep $node ; oc get pods -n openshift-sdn -o wide | grep $node ; oc get pods -n openshift-logging -o wide | grep $node sync-6gbl8 1/1 Running 0 2h 172.31.15.203 ip-172-31-15-203.us-west-2.compute.internal <none> ovs-xnxtq 1/1 Running 0 2h 172.31.15.203 ip-172-31-15-203.us-west-2.compute.internal <none> sdn-7dflr 1/1 Running 0 2h 172.31.15.203 ip-172-31-15-203.us-west-2.compute.internal <none> logging-fluentd-r8lc4 1/1 Running 0 1h 172.22.0.2 ip-172-31-15-203.us-west-2.compute.internal <none> Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2019:0636 |