Login
[x]
Log in using an account from:
Fedora Account System
Red Hat Associate
Red Hat Customer
Or login using a Red Hat Bugzilla account
Forgot Password
Login:
Hide Forgot
Create an Account
Red Hat Bugzilla – Attachment 1474781 Details for
Bug 1614479
Stopping master host causes another master controller to restart
[?]
New
Simple Search
Advanced Search
My Links
Browse
Requests
Reports
Current State
Search
Tabular reports
Graphical reports
Duplicates
Other Reports
User Changes
Plotly Reports
Bug Status
Bug Severity
Non-Defaults
|
Product Dashboard
Help
Page Help!
Bug Writing Guidelines
What's new
Browser Support Policy
5.0.4.rh83 Release notes
FAQ
Guides index
User guide
Web Services
Contact
Legal
This site requires JavaScript to be enabled to function correctly, please enable it.
master controller logs
master-controller.log (text/plain), 59.60 KB, created by
Vikas Laad
on 2018-08-09 17:23:01 UTC
(
hide
)
Description:
master controller logs
Filename:
MIME Type:
Creator:
Vikas Laad
Created:
2018-08-09 17:23:01 UTC
Size:
59.60 KB
patch
obsolete
>E0809 14:57:09.803544 1 helpers.go:134] Encountered config error json: unknown field "masterCount" in object *config.MasterConfig, raw JSON: >{"admissionConfig":{"pluginConfig":{"BuildDefaults":{"configuration":{"apiVersion":"v1","env":[],"kind":"BuildDefaultsConfig","resources":{"limits":{},"requests":{}}}},"BuildOverrides":{"configuration":{"apiVersion":"v1","kind":"BuildOverridesConfig"}},"openshift.io/ImagePolicy":{"configuration":{"apiVersion":"v1","executionRules":[{"matchImageAnnotations":[{"key":"images.openshift.io/deny-execution","value":"true"}],"name":"execution-denied","onResources":[{"resource":"pods"},{"resource":"builds"}],"reject":true,"skipOnResolutionFailure":true}],"kind":"ImagePolicyConfig"}}}},"aggregatorConfig":{"proxyClientInfo":{"certFile":"aggregator-front-proxy.crt","keyFile":"aggregator-front-proxy.key"}},"apiLevels":["v1"],"apiVersion":"v1","authConfig":{"requestHeader":{"clientCA":"front-proxy-ca.crt","clientCommonNames":["aggregator-front-proxy"],"extraHeaderPrefixes":["X-Remote-Extra-"],"groupHeaders":["X-Remote-Group"],"usernameHeaders":["X-Remote-User"]}},"controllerConfig":{"election":{"lockName":"openshift-master-controllers"},"serviceServingCert":{"signer":{"certFile":"service-signer.crt","keyFile":"service-signer.key"}}},"controllers":"*","corsAllowedOrigins":["(?i)//127\\.0\\.0\\.1(:|\\z)","(?i)//localhost(:|\\z)","(?i)//172\\.31\\.5\\.129(:|\\z)","(?i)//18\\.236\\.162\\.181(:|\\z)","(?i)//ip\\-172\\-31\\-5\\-129\\.us\\-west\\-2\\.compute\\.internal(:|\\z)","(?i)//kubernetes\\.default(:|\\z)","(?i)//kubernetes\\.default\\.svc\\.cluster\\.local(:|\\z)","(?i)//kubernetes(:|\\z)","(?i)//ec2\\-18\\-236\\-162\\-181\\.us\\-west\\-2\\.compute\\.amazonaws\\.com(:|\\z)","(?i)//openshift\\.default(:|\\z)","(?i)//openshift\\.default\\.svc(:|\\z)","(?i)//ec2\\-18\\-237\\-82\\-147\\.us\\-west\\-2\\.compute\\.amazonaws\\.com(:|\\z)","(?i)//openshift\\.default\\.svc\\.cluster\\.local(:|\\z)","(?i)//kubernetes\\.default\\.svc(:|\\z)","(?i)//openshift(:|\\z)","(?i)//172\\.24\\.0\\.1(:|\\z)"],"dnsConfig":{"bindAddress":"0.0.0.0:8053","bindNetwork":"tcp4"},"etcdClientInfo":{"ca":"master.etcd-ca.crt","certFile":"master.etcd-client.crt","keyFile":"master.etcd-client.key","urls":["https://ip-172-31-52-241.us-west-2.compute.internal:2379","https://ip-172-31-5-129.us-west-2.compute.internal:2379","https://ip-172-31-5-14.us-west-2.compute.internal:2379"]},"etcdStorageConfig":{"kubernetesStoragePrefix":"kubernetes.io","kubernetesStorageVersion":"v1","openShiftStoragePrefix":"openshift.io","openShiftStorageVersion":"v1"},"imageConfig":{"format":"registry.reg-aws.openshift.com:443/openshift3/ose-${component}:${version}","latest":false},"imagePolicyConfig":{"internalRegistryHostname":"docker-registry.default.svc:5000"},"kind":"MasterConfig","kubeletClientInfo":{"ca":"ca-bundle.crt","certFile":"master.kubelet-client.crt","keyFile":"master.kubelet-client.key","port":10250},"kubernetesMasterConfig":{"apiServerArguments":{"cloud-config":["/etc/origin/cloudprovider/aws.conf"],"cloud-provider":["aws"],"storage-backend":["etcd3"],"storage-media-type":["application/vnd.kubernetes.protobuf"]},"controllerArguments":{"cloud-config":["/etc/origin/cloudprovider/aws.conf"],"cloud-provider":["aws"],"cluster-signing-cert-file":["/etc/origin/master/ca.crt"],"cluster-signing-key-file":["/etc/origin/master/ca.key"],"disable-attach-detach-reconcile-sync":["true"]},"masterCount":3,"masterIP":"172.31.5.129","podEvictionTimeout":null,"proxyClientInfo":{"certFile":"master.proxy-client.crt","keyFile":"master.proxy-client.key"},"schedulerArguments":null,"schedulerConfigFile":"/etc/origin/master/scheduler.json","servicesNodePortRange":"","servicesSubnet":"172.24.0.0/14","staticNodeNames":[]},"masterClients":{"externalKubernetesClientConnectionOverrides":{"acceptContentTypes":"application/vnd.kubernetes.protobuf,application/json","burst":400,"contentType":"application/vnd.kubernetes.protobuf","qps":200},"externalKubernetesKubeConfig":"","openshiftLoopbackClientConnectionOverrides":{"acceptContentTypes":"application/vnd.kubernetes.protobuf,application/json","burst":600,"contentType":"application/vnd.kubernetes.protobuf","qps":300},"openshiftLoopbackKubeConfig":"openshift-master.kubeconfig"},"masterPublicURL":"https://ec2-18-237-82-147.us-west-2.compute.amazonaws.com:8443","networkConfig":{"clusterNetworks":[{"cidr":"172.20.0.0/14","hostSubnetLength":9}],"externalIPNetworkCIDRs":["0.0.0.0/0"],"networkPluginName":"redhat/openshift-ovs-networkpolicy","serviceNetworkCIDR":"172.24.0.0/14"},"oauthConfig":{"assetPublicURL":"https://ec2-18-237-82-147.us-west-2.compute.amazonaws.com:8443/console/","grantConfig":{"method":"auto"},"identityProviders":[{"challenge":true,"login":true,"mappingMethod":"claim","name":"allow_all","provider":{"apiVersion":"v1","kind":"AllowAllPasswordIdentityProvider"}}],"masterCA":"ca-bundle.crt","masterPublicURL":"https://ec2-18-237-82-147.us-west-2.compute.amazonaws.com:8443","masterURL":"https://ec2-18-237-82-147.us-west-2.compute.amazonaws.com:8443","sessionConfig":{"sessionMaxAgeSeconds":3600,"sessionName":"ssn","sessionSecretsFile":"/etc/origin/master/session-secrets.yaml"},"tokenConfig":{"accessTokenMaxAgeSeconds":86400,"authorizeTokenMaxAgeSeconds":500}},"pauseControllers":false,"policyConfig":{"bootstrapPolicyFile":"/etc/origin/master/policy.json","openshiftInfrastructureNamespace":"openshift-infra","openshiftSharedResourcesNamespace":"openshift"},"projectConfig":{"defaultNodeSelector":"node-role.kubernetes.io/compute=true","projectRequestMessage":"","projectRequestTemplate":"","securityAllocator":{"mcsAllocatorRange":"s0:/2","mcsLabelsPerProject":5,"uidAllocatorRange":"1000000000-1999999999/10000"}},"routingConfig":{"subdomain":"apps.0809-rvj.qe.rhcloud.com"},"serviceAccountConfig":{"limitSecretReferences":false,"managedNames":["default","builder","deployer"],"masterCA":"ca-bundle.crt","privateKeyFile":"serviceaccounts.private.key","publicKeyFiles":["serviceaccounts.public.key"]},"servingInfo":{"bindAddress":"0.0.0.0:8443","bindNetwork":"tcp4","certFile":"master.server.crt","clientCA":"ca.crt","keyFile":"master.server.key","maxRequestsInFlight":500,"requestTimeoutSeconds":3600},"volumeConfig":{"dynamicProvisioningEnabled":true}} >I0809 14:57:09.868768 1 plugins.go:84] Registered admission plugin "NamespaceLifecycle" >I0809 14:57:09.868789 1 plugins.go:84] Registered admission plugin "Initializers" >I0809 14:57:09.868796 1 plugins.go:84] Registered admission plugin "ValidatingAdmissionWebhook" >I0809 14:57:09.868801 1 plugins.go:84] Registered admission plugin "MutatingAdmissionWebhook" >I0809 14:57:09.868809 1 plugins.go:84] Registered admission plugin "AlwaysAdmit" >I0809 14:57:09.868814 1 plugins.go:84] Registered admission plugin "AlwaysPullImages" >I0809 14:57:09.868818 1 plugins.go:84] Registered admission plugin "LimitPodHardAntiAffinityTopology" >I0809 14:57:09.868824 1 plugins.go:84] Registered admission plugin "DefaultTolerationSeconds" >I0809 14:57:09.868831 1 plugins.go:84] Registered admission plugin "AlwaysDeny" >I0809 14:57:09.868838 1 plugins.go:84] Registered admission plugin "EventRateLimit" >I0809 14:57:09.868843 1 plugins.go:84] Registered admission plugin "DenyEscalatingExec" >I0809 14:57:09.868849 1 plugins.go:84] Registered admission plugin "DenyExecOnPrivileged" >I0809 14:57:09.868857 1 plugins.go:84] Registered admission plugin "ExtendedResourceToleration" >I0809 14:57:09.868870 1 plugins.go:84] Registered admission plugin "OwnerReferencesPermissionEnforcement" >I0809 14:57:09.868876 1 plugins.go:84] Registered admission plugin "ImagePolicyWebhook" >I0809 14:57:09.868881 1 plugins.go:84] Registered admission plugin "LimitRanger" >I0809 14:57:09.868886 1 plugins.go:84] Registered admission plugin "NamespaceAutoProvision" >I0809 14:57:09.868893 1 plugins.go:84] Registered admission plugin "NamespaceExists" >I0809 14:57:09.868898 1 plugins.go:84] Registered admission plugin "NodeRestriction" >I0809 14:57:09.868902 1 plugins.go:84] Registered admission plugin "PersistentVolumeLabel" >I0809 14:57:09.868907 1 plugins.go:84] Registered admission plugin "PodNodeSelector" >I0809 14:57:09.868917 1 plugins.go:84] Registered admission plugin "PodPreset" >I0809 14:57:09.868922 1 plugins.go:84] Registered admission plugin "PodTolerationRestriction" >I0809 14:57:09.868928 1 plugins.go:84] Registered admission plugin "ResourceQuota" >I0809 14:57:09.868938 1 plugins.go:84] Registered admission plugin "PodSecurityPolicy" >I0809 14:57:09.868949 1 plugins.go:84] Registered admission plugin "Priority" >I0809 14:57:09.868953 1 plugins.go:84] Registered admission plugin "SecurityContextDeny" >I0809 14:57:09.868968 1 plugins.go:84] Registered admission plugin "ServiceAccount" >I0809 14:57:09.868973 1 plugins.go:84] Registered admission plugin "DefaultStorageClass" >I0809 14:57:09.868980 1 plugins.go:84] Registered admission plugin "PersistentVolumeClaimResize" >I0809 14:57:09.868986 1 plugins.go:84] Registered admission plugin "StorageObjectInUseProtection" >I0809 14:57:09.872188 1 start_kube_scheduler.go:62] `kube-scheduler [--leader-elect=true --leader-elect-resource-lock=configmaps --port=0 --kubeconfig=/etc/origin/master/openshift-master.kubeconfig --policy-config-file=/etc/origin/master/scheduler.json --kube-api-content-type=application/vnd.kubernetes.protobuf --kube-api-qps=300 --kube-api-burst=600]` >I0809 14:57:09.873616 1 flags.go:27] FLAG: --address="0.0.0.0" >I0809 14:57:09.873872 1 flags.go:27] FLAG: --algorithm-provider="" >I0809 14:57:09.873879 1 flags.go:27] FLAG: --azure-container-registry-config="" >I0809 14:57:09.873885 1 flags.go:27] FLAG: --config="" >I0809 14:57:09.873891 1 flags.go:27] FLAG: --contention-profiling="false" >I0809 14:57:09.873902 1 flags.go:27] FLAG: --failure-domains="kubernetes.io/hostname,failure-domain.beta.kubernetes.io/zone,failure-domain.beta.kubernetes.io/region" >I0809 14:57:09.873912 1 flags.go:27] FLAG: --feature-gates="" >I0809 14:57:09.873921 1 flags.go:27] FLAG: --google-json-key="" >I0809 14:57:09.873926 1 flags.go:27] FLAG: --hard-pod-affinity-symmetric-weight="1" >I0809 14:57:09.873936 1 flags.go:27] FLAG: --kube-api-burst="600" >I0809 14:57:09.873941 1 flags.go:27] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" >I0809 14:57:09.873947 1 flags.go:27] FLAG: --kube-api-qps="300" >I0809 14:57:09.873954 1 flags.go:27] FLAG: --kubeconfig="/etc/origin/master/openshift-master.kubeconfig" >I0809 14:57:09.873960 1 flags.go:27] FLAG: --leader-elect="true" >I0809 14:57:09.873965 1 flags.go:27] FLAG: --leader-elect-lease-duration="15s" >I0809 14:57:09.873974 1 flags.go:27] FLAG: --leader-elect-renew-deadline="10s" >I0809 14:57:09.873979 1 flags.go:27] FLAG: --leader-elect-resource-lock="configmaps" >I0809 14:57:09.873985 1 flags.go:27] FLAG: --leader-elect-retry-period="2s" >I0809 14:57:09.873990 1 flags.go:27] FLAG: --lock-object-name="kube-scheduler" >I0809 14:57:09.873995 1 flags.go:27] FLAG: --lock-object-namespace="kube-system" >I0809 14:57:09.874002 1 flags.go:27] FLAG: --log-flush-frequency="5s" >I0809 14:57:09.874007 1 flags.go:27] FLAG: --master="" >I0809 14:57:09.874012 1 flags.go:27] FLAG: --policy-config-file="/etc/origin/master/scheduler.json" >I0809 14:57:09.874018 1 flags.go:27] FLAG: --policy-configmap="" >I0809 14:57:09.874022 1 flags.go:27] FLAG: --policy-configmap-namespace="kube-system" >I0809 14:57:09.874030 1 flags.go:27] FLAG: --port="0" >I0809 14:57:09.874037 1 flags.go:27] FLAG: --profiling="false" >I0809 14:57:09.874041 1 flags.go:27] FLAG: --scheduler-name="default-scheduler" >I0809 14:57:09.874047 1 flags.go:27] FLAG: --use-legacy-policy-config="false" >I0809 14:57:09.874052 1 flags.go:27] FLAG: --version="false" >I0809 14:57:09.874067 1 flags.go:27] FLAG: --write-config-to="" >I0809 14:57:09.874776 1 controller_manager.go:45] Starting controllers on 0.0.0.0:8444 (v3.11.0-0.11.0) >I0809 14:57:09.879133 1 server.go:126] Version: v1.11.0+d4cacc0 >I0809 14:57:09.883284 1 factory.go:960] Creating scheduler from configuration: {{ } [{NoVolumeZoneConflict <nil>} {MaxEBSVolumeCount <nil>} {MaxGCEPDVolumeCount <nil>} {MaxAzureDiskVolumeCount <nil>} {MatchInterPodAffinity <nil>} {NoDiskConflict <nil>} {GeneralPredicates <nil>} {PodToleratesNodeTaints <nil>} {CheckNodeMemoryPressure <nil>} {CheckNodeDiskPressure <nil>} {CheckVolumeBinding <nil>} {Region 0xc421d855c0}] [{SelectorSpreadPriority 1 <nil>} {InterPodAffinityPriority 1 <nil>} {LeastRequestedPriority 1 <nil>} {BalancedResourceAllocation 1 <nil>} {NodePreferAvoidPodsPriority 10000 <nil>} {NodeAffinityPriority 1 <nil>} {TaintTolerationPriority 1 <nil>} {Zone 2 0xc421d8b380}] [] 0 false} >I0809 14:57:09.883357 1 factory.go:977] Registering predicate: NoVolumeZoneConflict >I0809 14:57:09.883367 1 plugins.go:224] Predicate type NoVolumeZoneConflict already registered, reusing. >I0809 14:57:09.883378 1 factory.go:977] Registering predicate: MaxEBSVolumeCount >I0809 14:57:09.883382 1 plugins.go:224] Predicate type MaxEBSVolumeCount already registered, reusing. >I0809 14:57:09.883392 1 factory.go:977] Registering predicate: MaxGCEPDVolumeCount >I0809 14:57:09.883396 1 plugins.go:224] Predicate type MaxGCEPDVolumeCount already registered, reusing. >I0809 14:57:09.883406 1 factory.go:977] Registering predicate: MaxAzureDiskVolumeCount >I0809 14:57:09.883411 1 plugins.go:224] Predicate type MaxAzureDiskVolumeCount already registered, reusing. >I0809 14:57:09.883416 1 factory.go:977] Registering predicate: MatchInterPodAffinity >I0809 14:57:09.883420 1 plugins.go:224] Predicate type MatchInterPodAffinity already registered, reusing. >I0809 14:57:09.883425 1 factory.go:977] Registering predicate: NoDiskConflict >I0809 14:57:09.883438 1 plugins.go:224] Predicate type NoDiskConflict already registered, reusing. >I0809 14:57:09.883444 1 factory.go:977] Registering predicate: GeneralPredicates >I0809 14:57:09.883451 1 plugins.go:224] Predicate type GeneralPredicates already registered, reusing. >I0809 14:57:09.883457 1 factory.go:977] Registering predicate: PodToleratesNodeTaints >I0809 14:57:09.883466 1 plugins.go:224] Predicate type PodToleratesNodeTaints already registered, reusing. >I0809 14:57:09.883473 1 factory.go:977] Registering predicate: CheckNodeMemoryPressure >I0809 14:57:09.883483 1 plugins.go:224] Predicate type CheckNodeMemoryPressure already registered, reusing. >I0809 14:57:09.883494 1 factory.go:977] Registering predicate: CheckNodeDiskPressure >I0809 14:57:09.883502 1 plugins.go:224] Predicate type CheckNodeDiskPressure already registered, reusing. >I0809 14:57:09.883510 1 factory.go:977] Registering predicate: CheckVolumeBinding >I0809 14:57:09.883519 1 plugins.go:224] Predicate type CheckVolumeBinding already registered, reusing. >I0809 14:57:09.883528 1 factory.go:977] Registering predicate: Region >I0809 14:57:09.883539 1 factory.go:992] Registering priority: SelectorSpreadPriority >I0809 14:57:09.883560 1 plugins.go:336] Priority type SelectorSpreadPriority already registered, reusing. >I0809 14:57:09.883573 1 factory.go:992] Registering priority: InterPodAffinityPriority >I0809 14:57:09.883584 1 plugins.go:336] Priority type InterPodAffinityPriority already registered, reusing. >I0809 14:57:09.883597 1 factory.go:992] Registering priority: LeastRequestedPriority >I0809 14:57:09.883609 1 plugins.go:336] Priority type LeastRequestedPriority already registered, reusing. >I0809 14:57:09.883621 1 factory.go:992] Registering priority: BalancedResourceAllocation >I0809 14:57:09.883631 1 plugins.go:336] Priority type BalancedResourceAllocation already registered, reusing. >I0809 14:57:09.883644 1 factory.go:992] Registering priority: NodePreferAvoidPodsPriority >I0809 14:57:09.883655 1 plugins.go:336] Priority type NodePreferAvoidPodsPriority already registered, reusing. >I0809 14:57:09.883668 1 factory.go:992] Registering priority: NodeAffinityPriority >I0809 14:57:09.883679 1 plugins.go:336] Priority type NodeAffinityPriority already registered, reusing. >I0809 14:57:09.883690 1 factory.go:992] Registering priority: TaintTolerationPriority >I0809 14:57:09.883702 1 plugins.go:336] Priority type TaintTolerationPriority already registered, reusing. >I0809 14:57:09.883716 1 factory.go:992] Registering priority: Zone >I0809 14:57:09.883735 1 factory.go:1049] Creating scheduler with fit predicates 'map[MaxEBSVolumeCount:{} MaxAzureDiskVolumeCount:{} NoDiskConflict:{} PodToleratesNodeTaints:{} CheckNodeMemoryPressure:{} CheckNodeDiskPressure:{} CheckVolumeBinding:{} NoVolumeZoneConflict:{} MaxGCEPDVolumeCount:{} MatchInterPodAffinity:{} GeneralPredicates:{} Region:{}]' and priority functions 'map[InterPodAffinityPriority:{} LeastRequestedPriority:{} BalancedResourceAllocation:{} NodePreferAvoidPodsPriority:{} NodeAffinityPriority:{} TaintTolerationPriority:{} Zone:{} SelectorSpreadPriority:{}]' >I0809 14:57:09.936160 1 start_kube_controller_manager.go:87] `kube-controller-manager [--cluster-signing-cert-file= --use-service-account-credentials=true --cluster-signing-key-file= --leader-elect-retry-period=3s --service-account-private-key-file=/etc/origin/master/serviceaccounts.private.key --kubeconfig=/etc/origin/master/openshift-master.kubeconfig --port=0 --kube-api-qps=300 --kube-api-burst=600 --openshift-config=/tmp/openshift-kube-controller-manager-config-100803247/master-config.yaml --controllers=* --controllers=-ttl --controllers=-bootstrapsigner --controllers=-tokencleaner --root-ca-file=/etc/origin/master/ca-bundle.crt --kube-api-content-type=application/vnd.kubernetes.protobuf --pod-eviction-timeout=5m --enable-dynamic-provisioning=true --leader-elect-resource-lock=configmaps]` >I0809 14:57:09.936227 1 flags.go:27] FLAG: --address="0.0.0.0" >I0809 14:57:09.936243 1 flags.go:27] FLAG: --allocate-node-cidrs="false" >I0809 14:57:09.936250 1 flags.go:27] FLAG: --allow-untagged-cloud="false" >I0809 14:57:09.936255 1 flags.go:27] FLAG: --attach-detach-reconcile-sync-period="1m0s" >I0809 14:57:09.936265 1 flags.go:27] FLAG: --azure-container-registry-config="" >I0809 14:57:09.936272 1 flags.go:27] FLAG: --bind-address="0.0.0.0" >I0809 14:57:09.936278 1 flags.go:27] FLAG: --cert-dir="/var/run/kubernetes" >I0809 14:57:09.936287 1 flags.go:27] FLAG: --cidr-allocator-type="RangeAllocator" >I0809 14:57:09.936294 1 flags.go:27] FLAG: --cloud-config="" >I0809 14:57:09.936299 1 flags.go:27] FLAG: --cloud-provider="" >I0809 14:57:09.936307 1 flags.go:27] FLAG: --cluster-cidr="" >I0809 14:57:09.936313 1 flags.go:27] FLAG: --cluster-name="kubernetes" >I0809 14:57:09.936321 1 flags.go:27] FLAG: --cluster-signing-cert-file="" >I0809 14:57:09.936326 1 flags.go:27] FLAG: --cluster-signing-key-file="" >I0809 14:57:09.936335 1 flags.go:27] FLAG: --concurrent-deployment-syncs="5" >I0809 14:57:09.936343 1 flags.go:27] FLAG: --concurrent-endpoint-syncs="5" >I0809 14:57:09.936352 1 flags.go:27] FLAG: --concurrent-gc-syncs="20" >I0809 14:57:09.936359 1 flags.go:27] FLAG: --concurrent-namespace-syncs="10" >I0809 14:57:09.936366 1 flags.go:27] FLAG: --concurrent-replicaset-syncs="5" >I0809 14:57:09.936377 1 flags.go:27] FLAG: --concurrent-resource-quota-syncs="5" >I0809 14:57:09.936383 1 flags.go:27] FLAG: --concurrent-service-syncs="1" >I0809 14:57:09.936391 1 flags.go:27] FLAG: --concurrent-serviceaccount-token-syncs="5" >I0809 14:57:09.936402 1 flags.go:27] FLAG: --concurrent_rc_syncs="5" >I0809 14:57:09.936410 1 flags.go:27] FLAG: --configure-cloud-routes="true" >I0809 14:57:09.936419 1 flags.go:27] FLAG: --contention-profiling="false" >I0809 14:57:09.936430 1 flags.go:27] FLAG: --controller-start-interval="0s" >I0809 14:57:09.936439 1 flags.go:27] FLAG: --controllers="[*,-ttl,-bootstrapsigner,-tokencleaner]" >I0809 14:57:09.936454 1 flags.go:27] FLAG: --deleting-pods-burst="0" >I0809 14:57:09.936461 1 flags.go:27] FLAG: --deleting-pods-qps="0.1" >I0809 14:57:09.936467 1 flags.go:27] FLAG: --deployment-controller-sync-period="30s" >I0809 14:57:09.936475 1 flags.go:27] FLAG: --disable-attach-detach-reconcile-sync="false" >I0809 14:57:09.936480 1 flags.go:27] FLAG: --enable-dynamic-provisioning="true" >I0809 14:57:09.936485 1 flags.go:27] FLAG: --enable-garbage-collector="true" >I0809 14:57:09.936490 1 flags.go:27] FLAG: --enable-hostpath-provisioner="false" >I0809 14:57:09.936497 1 flags.go:27] FLAG: --enable-taint-manager="true" >I0809 14:57:09.936505 1 flags.go:27] FLAG: --experimental-cluster-signing-duration="8760h0m0s" >I0809 14:57:09.936510 1 flags.go:27] FLAG: --external-cloud-volume-plugin="" >I0809 14:57:09.936516 1 flags.go:27] FLAG: --feature-gates="" >I0809 14:57:09.936530 1 flags.go:27] FLAG: --flex-volume-plugin-dir="/usr/libexec/kubernetes/kubelet-plugins/volume/exec/" >I0809 14:57:09.936543 1 flags.go:27] FLAG: --google-json-key="" >I0809 14:57:09.936566 1 flags.go:27] FLAG: --horizontal-pod-autoscaler-downscale-delay="5m0s" >I0809 14:57:09.936573 1 flags.go:27] FLAG: --horizontal-pod-autoscaler-sync-period="30s" >I0809 14:57:09.936579 1 flags.go:27] FLAG: --horizontal-pod-autoscaler-tolerance="0.1" >I0809 14:57:09.936587 1 flags.go:27] FLAG: --horizontal-pod-autoscaler-upscale-delay="3m0s" >I0809 14:57:09.936592 1 flags.go:27] FLAG: --horizontal-pod-autoscaler-use-rest-clients="true" >I0809 14:57:09.936598 1 flags.go:27] FLAG: --http2-max-streams-per-connection="0" >I0809 14:57:09.936605 1 flags.go:27] FLAG: --insecure-experimental-approve-all-kubelet-csrs-for-group="" >I0809 14:57:09.936610 1 flags.go:27] FLAG: --kube-api-burst="600" >I0809 14:57:09.936620 1 flags.go:27] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" >I0809 14:57:09.936631 1 flags.go:27] FLAG: --kube-api-qps="300" >I0809 14:57:09.936639 1 flags.go:27] FLAG: --kubeconfig="/etc/origin/master/openshift-master.kubeconfig" >I0809 14:57:09.936649 1 flags.go:27] FLAG: --large-cluster-size-threshold="50" >I0809 14:57:09.936655 1 flags.go:27] FLAG: --leader-elect="true" >I0809 14:57:09.936660 1 flags.go:27] FLAG: --leader-elect-lease-duration="15s" >I0809 14:57:09.936666 1 flags.go:27] FLAG: --leader-elect-renew-deadline="10s" >I0809 14:57:09.936671 1 flags.go:27] FLAG: --leader-elect-resource-lock="configmaps" >I0809 14:57:09.936676 1 flags.go:27] FLAG: --leader-elect-retry-period="3s" >I0809 14:57:09.936685 1 flags.go:27] FLAG: --log-flush-frequency="5s" >I0809 14:57:09.936690 1 flags.go:27] FLAG: --master="" >I0809 14:57:09.936696 1 flags.go:27] FLAG: --min-resync-period="12h0m0s" >I0809 14:57:09.936705 1 flags.go:27] FLAG: --namespace-sync-period="5m0s" >I0809 14:57:09.936710 1 flags.go:27] FLAG: --node-cidr-mask-size="24" >I0809 14:57:09.936717 1 flags.go:27] FLAG: --node-eviction-rate="0.1" >I0809 14:57:09.936727 1 flags.go:27] FLAG: --node-monitor-grace-period="40s" >I0809 14:57:09.936732 1 flags.go:27] FLAG: --node-monitor-period="5s" >I0809 14:57:09.936738 1 flags.go:27] FLAG: --node-startup-grace-period="1m0s" >I0809 14:57:09.936743 1 flags.go:27] FLAG: --node-sync-period="0s" >I0809 14:57:09.936749 1 flags.go:27] FLAG: --openshift-config="/tmp/openshift-kube-controller-manager-config-100803247/master-config.yaml" >I0809 14:57:09.936756 1 flags.go:27] FLAG: --pod-eviction-timeout="5m0s" >I0809 14:57:09.936762 1 flags.go:27] FLAG: --port="0" >I0809 14:57:09.936767 1 flags.go:27] FLAG: --profiling="false" >I0809 14:57:09.936777 1 flags.go:27] FLAG: --pv-recycler-increment-timeout-nfs="30" >I0809 14:57:09.936785 1 flags.go:27] FLAG: --pv-recycler-minimum-timeout-hostpath="60" >I0809 14:57:09.936791 1 flags.go:27] FLAG: --pv-recycler-minimum-timeout-nfs="300" >I0809 14:57:09.936798 1 flags.go:27] FLAG: --pv-recycler-pod-template-filepath-hostpath="" >I0809 14:57:09.936803 1 flags.go:27] FLAG: --pv-recycler-pod-template-filepath-nfs="" >I0809 14:57:09.936813 1 flags.go:27] FLAG: --pv-recycler-timeout-increment-hostpath="30" >I0809 14:57:09.936818 1 flags.go:27] FLAG: --pvclaimbinder-sync-period="15s" >I0809 14:57:09.936823 1 flags.go:27] FLAG: --register-retry-count="10" >I0809 14:57:09.936829 1 flags.go:27] FLAG: --resource-quota-sync-period="5m0s" >I0809 14:57:09.936834 1 flags.go:27] FLAG: --root-ca-file="/etc/origin/master/ca-bundle.crt" >I0809 14:57:09.936845 1 flags.go:27] FLAG: --route-reconciliation-period="10s" >I0809 14:57:09.936850 1 flags.go:27] FLAG: --secondary-node-eviction-rate="0.01" >I0809 14:57:09.936856 1 flags.go:27] FLAG: --secure-port="0" >I0809 14:57:09.936865 1 flags.go:27] FLAG: --service-account-private-key-file="/etc/origin/master/serviceaccounts.private.key" >I0809 14:57:09.936876 1 flags.go:27] FLAG: --service-cluster-ip-range="" >I0809 14:57:09.936882 1 flags.go:27] FLAG: --terminated-pod-gc-threshold="12500" >I0809 14:57:09.936887 1 flags.go:27] FLAG: --tls-cert-file="" >I0809 14:57:09.936893 1 flags.go:27] FLAG: --tls-cipher-suites="[]" >I0809 14:57:09.936904 1 flags.go:27] FLAG: --tls-min-version="" >I0809 14:57:09.936909 1 flags.go:27] FLAG: --tls-private-key-file="" >I0809 14:57:09.936914 1 flags.go:27] FLAG: --tls-sni-cert-key="[]" >I0809 14:57:09.936921 1 flags.go:27] FLAG: --unhealthy-zone-threshold="0.55" >I0809 14:57:09.936927 1 flags.go:27] FLAG: --use-service-account-credentials="true" >I0809 14:57:09.936933 1 flags.go:27] FLAG: --version="false" >I0809 14:57:09.937329 1 standalone_apiserver.go:101] Started health checks at 0.0.0.0:8444 >I0809 14:57:09.937867 1 leaderelection.go:185] attempting to acquire leader lease kube-system/openshift-master-controllers... >I0809 14:57:09.944121 1 controllermanager.go:129] Version: v1.11.0+d4cacc0 >E0809 14:57:09.944143 1 controllermanager.go:134] unable to register configz: &{%!c(string=register config "componentconfig" twice)} >I0809 14:57:09.944179 1 leaderelection.go:185] attempting to acquire leader lease kube-system/kube-controller-manager... >E0809 14:57:09.952241 1 leaderelection.go:234] error retrieving resource lock kube-system/kube-controller-manager: Get https://ip-172-31-5-129.us-west-2.compute.internal:8443/api/v1/namespaces/kube-system/configmaps/kube-controller-manager: dial tcp 172.31.5.129:8443: connect: connection refused >E0809 14:57:09.952256 1 reflector.go:136] k8s.io/client-go/informers/factory.go:130: Failed to list *v1beta1.ReplicaSet: Get https://ip-172-31-5-129.us-west-2.compute.internal:8443/apis/extensions/v1beta1/replicasets?limit=500&resourceVersion=0: dial tcp 172.31.5.129:8443: connect: connection refused >E0809 14:57:09.952284 1 reflector.go:136] k8s.io/client-go/informers/factory.go:130: Failed to list *v1.Service: Get https://ip-172-31-5-129.us-west-2.compute.internal:8443/api/v1/services?limit=500&resourceVersion=0: dial tcp 172.31.5.129:8443: connect: connection refused >E0809 14:57:09.952259 1 reflector.go:136] k8s.io/client-go/informers/factory.go:130: Failed to list *v1beta1.PodDisruptionBudget: Get https://ip-172-31-5-129.us-west-2.compute.internal:8443/apis/policy/v1beta1/poddisruptionbudgets?limit=500&resourceVersion=0: dial tcp 172.31.5.129:8443: connect: connection refused >E0809 14:57:09.952337 1 reflector.go:136] k8s.io/client-go/informers/factory.go:130: Failed to list *v1.Node: Get https://ip-172-31-5-129.us-west-2.compute.internal:8443/api/v1/nodes?limit=500&resourceVersion=0: dial tcp 172.31.5.129:8443: connect: connection refused >E0809 14:57:09.952361 1 reflector.go:136] k8s.io/client-go/informers/factory.go:130: Failed to list *v1.ReplicationController: Get https://ip-172-31-5-129.us-west-2.compute.internal:8443/api/v1/replicationcontrollers?limit=500&resourceVersion=0: dial tcp 172.31.5.129:8443: connect: connection refused >E0809 14:57:09.952463 1 reflector.go:136] k8s.io/client-go/informers/factory.go:130: Failed to list *v1.PersistentVolume: Get https://ip-172-31-5-129.us-west-2.compute.internal:8443/api/v1/persistentvolumes?limit=500&resourceVersion=0: dial tcp 172.31.5.129:8443: connect: connection refused >E0809 14:57:09.952378 1 leaderelection.go:234] error retrieving resource lock kube-system/openshift-master-controllers: Get https://ip-172-31-5-129.us-west-2.compute.internal:8443/api/v1/namespaces/kube-system/configmaps/openshift-master-controllers: dial tcp 172.31.5.129:8443: connect: connection refused >E0809 14:57:09.952413 1 reflector.go:136] k8s.io/client-go/informers/factory.go:130: Failed to list *v1.StorageClass: Get https://ip-172-31-5-129.us-west-2.compute.internal:8443/apis/storage.k8s.io/v1/storageclasses?limit=500&resourceVersion=0: dial tcp 172.31.5.129:8443: connect: connection refused >E0809 14:57:09.952418 1 reflector.go:136] k8s.io/client-go/informers/factory.go:130: Failed to list *v1.PersistentVolumeClaim: Get https://ip-172-31-5-129.us-west-2.compute.internal:8443/api/v1/persistentvolumeclaims?limit=500&resourceVersion=0: dial tcp 172.31.5.129:8443: connect: connection refused >E0809 14:57:09.952453 1 reflector.go:136] k8s.io/client-go/informers/factory.go:130: Failed to list *v1beta1.StatefulSet: Get https://ip-172-31-5-129.us-west-2.compute.internal:8443/apis/apps/v1beta1/statefulsets?limit=500&resourceVersion=0: dial tcp 172.31.5.129:8443: connect: connection refused >E0809 14:57:09.952460 1 reflector.go:136] k8s.io/kubernetes/cmd/kube-scheduler/app/server.go:176: Failed to list *v1.Pod: Get https://ip-172-31-5-129.us-west-2.compute.internal:8443/api/v1/pods?fieldSelector=status.phase%21%3DFailed%2Cstatus.phase%21%3DSucceeded&limit=500&resourceVersion=0: dial tcp 172.31.5.129:8443: connect: connection refused >I0809 14:57:14.133202 1 healthz.go:72] /healthz/log check >E0809 14:57:20.963444 1 reflector.go:136] k8s.io/client-go/informers/factory.go:130: Failed to list *v1.StorageClass: Get https://ip-172-31-5-129.us-west-2.compute.internal:8443/apis/storage.k8s.io/v1/storageclasses?limit=500&resourceVersion=0: net/http: TLS handshake timeout >E0809 14:57:20.963444 1 reflector.go:136] k8s.io/client-go/informers/factory.go:130: Failed to list *v1beta1.PodDisruptionBudget: Get https://ip-172-31-5-129.us-west-2.compute.internal:8443/apis/policy/v1beta1/poddisruptionbudgets?limit=500&resourceVersion=0: net/http: TLS handshake timeout >E0809 14:57:20.963509 1 reflector.go:136] k8s.io/client-go/informers/factory.go:130: Failed to list *v1.PersistentVolume: Get https://ip-172-31-5-129.us-west-2.compute.internal:8443/api/v1/persistentvolumes?limit=500&resourceVersion=0: net/http: TLS handshake timeout >E0809 14:57:20.963573 1 reflector.go:136] k8s.io/client-go/informers/factory.go:130: Failed to list *v1.Node: Get https://ip-172-31-5-129.us-west-2.compute.internal:8443/api/v1/nodes?limit=500&resourceVersion=0: net/http: TLS handshake timeout >E0809 14:57:20.963623 1 reflector.go:136] k8s.io/client-go/informers/factory.go:130: Failed to list *v1.PersistentVolumeClaim: Get https://ip-172-31-5-129.us-west-2.compute.internal:8443/api/v1/persistentvolumeclaims?limit=500&resourceVersion=0: net/http: TLS handshake timeout >E0809 14:57:20.963716 1 reflector.go:136] k8s.io/kubernetes/cmd/kube-scheduler/app/server.go:176: Failed to list *v1.Pod: Get https://ip-172-31-5-129.us-west-2.compute.internal:8443/api/v1/pods?fieldSelector=status.phase%21%3DFailed%2Cstatus.phase%21%3DSucceeded&limit=500&resourceVersion=0: net/http: TLS handshake timeout >E0809 14:57:20.963718 1 reflector.go:136] k8s.io/client-go/informers/factory.go:130: Failed to list *v1.Service: Get https://ip-172-31-5-129.us-west-2.compute.internal:8443/api/v1/services?limit=500&resourceVersion=0: net/http: TLS handshake timeout >E0809 14:57:20.963763 1 reflector.go:136] k8s.io/client-go/informers/factory.go:130: Failed to list *v1beta1.StatefulSet: Get https://ip-172-31-5-129.us-west-2.compute.internal:8443/apis/apps/v1beta1/statefulsets?limit=500&resourceVersion=0: net/http: TLS handshake timeout >E0809 14:57:20.963765 1 reflector.go:136] k8s.io/client-go/informers/factory.go:130: Failed to list *v1beta1.ReplicaSet: Get https://ip-172-31-5-129.us-west-2.compute.internal:8443/apis/extensions/v1beta1/replicasets?limit=500&resourceVersion=0: net/http: TLS handshake timeout >E0809 14:57:20.963781 1 reflector.go:136] k8s.io/client-go/informers/factory.go:130: Failed to list *v1.ReplicationController: Get https://ip-172-31-5-129.us-west-2.compute.internal:8443/api/v1/replicationcontrollers?limit=500&resourceVersion=0: net/http: TLS handshake timeout >E0809 14:57:23.089785 1 leaderelection.go:234] error retrieving resource lock kube-system/openshift-master-controllers: Get https://ip-172-31-5-129.us-west-2.compute.internal:8443/api/v1/namespaces/kube-system/configmaps/openshift-master-controllers: net/http: TLS handshake timeout >E0809 14:57:23.903952 1 leaderelection.go:238] error initially creating leader election record: namespaces "kube-system" not found >I0809 14:57:24.128206 1 healthz.go:72] /healthz/log check >I0809 14:57:24.640832 1 controller_utils.go:1025] Waiting for caches to sync for scheduler controller >I0809 14:57:24.740971 1 controller_utils.go:1032] Caches are synced for scheduler controller >I0809 14:57:24.740995 1 leaderelection.go:185] attempting to acquire leader lease kube-system/kube-scheduler... >I0809 14:57:24.745086 1 leaderelection.go:194] successfully acquired lease kube-system/kube-scheduler >I0809 14:57:34.127655 1 healthz.go:72] /healthz/log check >I0809 14:57:44.127594 1 healthz.go:72] /healthz/log check >I0809 14:57:54.127458 1 healthz.go:72] /healthz/log check >I0809 14:58:04.127479 1 healthz.go:72] /healthz/log check >I0809 14:58:14.127577 1 healthz.go:72] /healthz/log check >I0809 14:58:24.127488 1 healthz.go:72] /healthz/log check >I0809 14:58:34.127453 1 healthz.go:72] /healthz/log check >I0809 14:58:44.127528 1 healthz.go:72] /healthz/log check >I0809 14:58:54.127762 1 healthz.go:72] /healthz/log check >I0809 14:59:04.128265 1 healthz.go:72] /healthz/log check >I0809 14:59:14.127432 1 healthz.go:72] /healthz/log check >I0809 14:59:24.127605 1 healthz.go:72] /healthz/log check >I0809 14:59:34.127629 1 healthz.go:72] /healthz/log check >I0809 14:59:44.127372 1 healthz.go:72] /healthz/log check >I0809 14:59:54.127493 1 healthz.go:72] /healthz/log check >I0809 15:00:04.127492 1 healthz.go:72] /healthz/log check >I0809 15:00:14.127731 1 healthz.go:72] /healthz/log check >I0809 15:00:24.127542 1 healthz.go:72] /healthz/log check >I0809 15:00:34.127459 1 healthz.go:72] /healthz/log check >I0809 15:01:16.138418 1 healthz.go:72] /healthz/log check >I0809 15:01:26.138405 1 healthz.go:72] /healthz/log check >I0809 15:01:36.138358 1 healthz.go:72] /healthz/log check >I0809 15:01:46.138357 1 healthz.go:72] /healthz/log check >I0809 15:02:15.596687 1 healthz.go:72] /healthz/log check >I0809 15:02:25.596419 1 healthz.go:72] /healthz/log check >I0809 15:02:35.596541 1 healthz.go:72] /healthz/log check >I0809 15:02:36.512165 1 scheduler.go:194] Failed to schedule pod: default/router-1-deploy >I0809 15:02:36.513114 1 factory.go:1416] Updating pod condition for default/router-1-deploy to (PodScheduled==False) >I0809 15:02:36.519214 1 scheduler.go:194] Failed to schedule pod: default/router-1-deploy >I0809 15:02:36.519277 1 factory.go:1416] Updating pod condition for default/router-1-deploy to (PodScheduled==False) >I0809 15:02:37.815743 1 scheduler.go:194] Failed to schedule pod: default/router-1-deploy >I0809 15:02:37.815809 1 factory.go:1416] Updating pod condition for default/router-1-deploy to (PodScheduled==False) >I0809 15:02:37.855093 1 scheduler.go:194] Failed to schedule pod: default/router-1-deploy >I0809 15:02:37.855172 1 factory.go:1416] Updating pod condition for default/router-1-deploy to (PodScheduled==False) >I0809 15:02:37.889464 1 scheduler.go:194] Failed to schedule pod: default/router-1-deploy >I0809 15:02:37.889521 1 factory.go:1416] Updating pod condition for default/router-1-deploy to (PodScheduled==False) >I0809 15:02:42.977688 1 scheduler.go:194] Failed to schedule pod: default/router-1-deploy >I0809 15:02:42.977750 1 factory.go:1416] Updating pod condition for default/router-1-deploy to (PodScheduled==False) >I0809 15:02:45.170450 1 scheduler.go:194] Failed to schedule pod: default/router-1-deploy >I0809 15:02:45.170517 1 factory.go:1416] Updating pod condition for default/router-1-deploy to (PodScheduled==False) >I0809 15:02:45.208238 1 scheduler.go:194] Failed to schedule pod: default/router-1-deploy >I0809 15:02:45.208311 1 factory.go:1416] Updating pod condition for default/router-1-deploy to (PodScheduled==False) >I0809 15:02:45.596398 1 healthz.go:72] /healthz/log check >I0809 15:02:55.596359 1 healthz.go:72] /healthz/log check >I0809 15:03:05.596479 1 healthz.go:72] /healthz/log check >I0809 15:03:15.596576 1 healthz.go:72] /healthz/log check >I0809 15:03:25.596442 1 healthz.go:72] /healthz/log check >I0809 15:03:35.596926 1 healthz.go:72] /healthz/log check >I0809 15:03:45.596514 1 healthz.go:72] /healthz/log check >I0809 15:03:55.596604 1 healthz.go:72] /healthz/log check >I0809 15:04:05.596377 1 healthz.go:72] /healthz/log check >I0809 15:04:15.596912 1 healthz.go:72] /healthz/log check >I0809 15:04:25.596484 1 healthz.go:72] /healthz/log check >I0809 15:04:35.596602 1 healthz.go:72] /healthz/log check >I0809 15:04:45.596383 1 healthz.go:72] /healthz/log check >I0809 15:04:55.596460 1 healthz.go:72] /healthz/log check >I0809 15:05:05.596466 1 healthz.go:72] /healthz/log check >I0809 15:05:15.596464 1 healthz.go:72] /healthz/log check >I0809 15:05:25.596612 1 healthz.go:72] /healthz/log check >I0809 15:05:35.596419 1 healthz.go:72] /healthz/log check >I0809 15:05:45.596356 1 healthz.go:72] /healthz/log check >I0809 15:05:55.596928 1 healthz.go:72] /healthz/log check >I0809 15:06:05.596423 1 healthz.go:72] /healthz/log check >I0809 15:06:15.596377 1 healthz.go:72] /healthz/log check >I0809 15:06:25.596356 1 healthz.go:72] /healthz/log check >I0809 15:06:35.596464 1 healthz.go:72] /healthz/log check >I0809 15:06:45.596513 1 healthz.go:72] /healthz/log check >I0809 15:06:55.597344 1 healthz.go:72] /healthz/log check >I0809 15:07:05.596411 1 healthz.go:72] /healthz/log check >I0809 15:07:15.596522 1 healthz.go:72] /healthz/log check >I0809 15:07:25.596857 1 healthz.go:72] /healthz/log check >I0809 15:07:35.596436 1 healthz.go:72] /healthz/log check >I0809 15:07:45.596464 1 healthz.go:72] /healthz/log check >I0809 15:07:55.596389 1 healthz.go:72] /healthz/log check >I0809 15:08:05.596437 1 healthz.go:72] /healthz/log check >I0809 15:08:15.596421 1 healthz.go:72] /healthz/log check >I0809 15:08:25.596490 1 healthz.go:72] /healthz/log check >I0809 15:08:35.596474 1 healthz.go:72] /healthz/log check >I0809 15:08:45.596659 1 healthz.go:72] /healthz/log check >I0809 15:08:55.596514 1 healthz.go:72] /healthz/log check >I0809 15:09:05.596400 1 healthz.go:72] /healthz/log check >I0809 15:09:15.596499 1 healthz.go:72] /healthz/log check >W0809 15:09:16.906899 1 reflector.go:272] k8s.io/client-go/informers/factory.go:130: watch of *v1.PersistentVolume ended with: The resourceVersion for the provided watch is too old. >I0809 15:09:25.596435 1 healthz.go:72] /healthz/log check >I0809 15:09:35.596906 1 healthz.go:72] /healthz/log check >I0809 15:09:45.596406 1 healthz.go:72] /healthz/log check >I0809 15:09:55.596469 1 healthz.go:72] /healthz/log check >I0809 15:10:05.596414 1 healthz.go:72] /healthz/log check >I0809 15:10:15.596525 1 healthz.go:72] /healthz/log check >W0809 15:10:19.889426 1 reflector.go:272] k8s.io/client-go/informers/factory.go:130: watch of *v1.PersistentVolumeClaim ended with: The resourceVersion for the provided watch is too old. >I0809 15:10:25.596616 1 healthz.go:72] /healthz/log check >I0809 15:10:35.596506 1 healthz.go:72] /healthz/log check >I0809 15:10:45.596432 1 healthz.go:72] /healthz/log check >I0809 15:10:55.596453 1 healthz.go:72] /healthz/log check >I0809 15:11:05.596394 1 healthz.go:72] /healthz/log check >I0809 15:11:15.596502 1 healthz.go:72] /healthz/log check >I0809 15:11:25.596349 1 healthz.go:72] /healthz/log check >I0809 15:11:35.596462 1 healthz.go:72] /healthz/log check >I0809 15:11:45.596416 1 healthz.go:72] /healthz/log check >I0809 15:11:55.596400 1 healthz.go:72] /healthz/log check >I0809 15:12:05.596424 1 healthz.go:72] /healthz/log check >I0809 15:12:15.596415 1 healthz.go:72] /healthz/log check >I0809 15:12:25.596514 1 healthz.go:72] /healthz/log check >I0809 15:12:35.596463 1 healthz.go:72] /healthz/log check >W0809 15:12:40.874394 1 reflector.go:272] k8s.io/client-go/informers/factory.go:130: watch of *v1beta1.ReplicaSet ended with: The resourceVersion for the provided watch is too old. >I0809 15:12:45.596398 1 healthz.go:72] /healthz/log check >I0809 15:12:55.596642 1 healthz.go:72] /healthz/log check >I0809 15:13:05.596485 1 healthz.go:72] /healthz/log check >W0809 15:13:05.814473 1 reflector.go:272] k8s.io/client-go/informers/factory.go:130: watch of *v1.StorageClass ended with: The resourceVersion for the provided watch is too old. >I0809 15:13:15.596360 1 healthz.go:72] /healthz/log check >I0809 15:13:25.596379 1 healthz.go:72] /healthz/log check >W0809 15:13:29.826357 1 reflector.go:272] k8s.io/client-go/informers/factory.go:130: watch of *v1beta1.PodDisruptionBudget ended with: The resourceVersion for the provided watch is too old. >I0809 15:13:35.596393 1 healthz.go:72] /healthz/log check >I0809 15:13:45.596411 1 healthz.go:72] /healthz/log check >I0809 15:13:55.596417 1 healthz.go:72] /healthz/log check >I0809 15:14:05.596458 1 healthz.go:72] /healthz/log check >I0809 15:14:15.596309 1 healthz.go:72] /healthz/log check >I0809 15:14:25.596442 1 healthz.go:72] /healthz/log check >I0809 15:14:35.596488 1 healthz.go:72] /healthz/log check >I0809 15:14:45.596905 1 healthz.go:72] /healthz/log check >I0809 15:14:55.596488 1 healthz.go:72] /healthz/log check >I0809 15:15:05.596399 1 healthz.go:72] /healthz/log check >I0809 15:15:15.596503 1 healthz.go:72] /healthz/log check >I0809 15:15:25.596396 1 healthz.go:72] /healthz/log check >I0809 15:15:35.596460 1 healthz.go:72] /healthz/log check >I0809 15:15:45.596820 1 healthz.go:72] /healthz/log check >I0809 15:15:55.596510 1 healthz.go:72] /healthz/log check >I0809 15:16:05.596433 1 healthz.go:72] /healthz/log check >I0809 15:16:15.596413 1 healthz.go:72] /healthz/log check >I0809 15:16:25.596370 1 healthz.go:72] /healthz/log check >W0809 15:16:25.808263 1 reflector.go:272] k8s.io/client-go/informers/factory.go:130: watch of *v1beta1.StatefulSet ended with: The resourceVersion for the provided watch is too old. >I0809 15:16:35.596420 1 healthz.go:72] /healthz/log check >I0809 15:16:45.596412 1 healthz.go:72] /healthz/log check >I0809 15:16:55.596385 1 healthz.go:72] /healthz/log check >I0809 15:17:05.596581 1 healthz.go:72] /healthz/log check >I0809 15:17:15.596433 1 healthz.go:72] /healthz/log check >I0809 15:17:25.604288 1 healthz.go:72] /healthz/log check >I0809 15:17:35.596449 1 healthz.go:72] /healthz/log check >I0809 15:17:45.596393 1 healthz.go:72] /healthz/log check >I0809 15:17:55.596749 1 healthz.go:72] /healthz/log check >I0809 15:18:05.596417 1 healthz.go:72] /healthz/log check >I0809 15:18:15.596387 1 healthz.go:72] /healthz/log check >I0809 15:18:25.596459 1 healthz.go:72] /healthz/log check >I0809 15:18:35.596389 1 healthz.go:72] /healthz/log check >I0809 15:18:45.596458 1 healthz.go:72] /healthz/log check >I0809 15:18:55.596409 1 healthz.go:72] /healthz/log check >I0809 15:19:05.597102 1 healthz.go:72] /healthz/log check >I0809 15:19:15.596491 1 healthz.go:72] /healthz/log check >I0809 15:19:25.597444 1 healthz.go:72] /healthz/log check >I0809 15:19:35.596464 1 healthz.go:72] /healthz/log check >I0809 15:19:45.596359 1 healthz.go:72] /healthz/log check >I0809 15:19:55.596429 1 healthz.go:72] /healthz/log check >I0809 15:20:05.596511 1 healthz.go:72] /healthz/log check >I0809 15:20:15.596452 1 healthz.go:72] /healthz/log check >I0809 15:20:25.596391 1 healthz.go:72] /healthz/log check >I0809 15:20:35.596398 1 healthz.go:72] /healthz/log check >I0809 15:20:45.596355 1 healthz.go:72] /healthz/log check >I0809 15:20:55.596373 1 healthz.go:72] /healthz/log check >I0809 15:21:05.596489 1 healthz.go:72] /healthz/log check >I0809 15:21:15.596377 1 healthz.go:72] /healthz/log check >I0809 15:21:25.596821 1 healthz.go:72] /healthz/log check >I0809 15:21:35.596428 1 healthz.go:72] /healthz/log check >I0809 15:21:45.596394 1 healthz.go:72] /healthz/log check >I0809 15:21:55.596531 1 healthz.go:72] /healthz/log check >I0809 15:22:05.596391 1 healthz.go:72] /healthz/log check >I0809 15:22:15.596397 1 healthz.go:72] /healthz/log check >I0809 15:22:25.597277 1 healthz.go:72] /healthz/log check >I0809 15:22:35.596356 1 healthz.go:72] /healthz/log check >I0809 15:22:45.596454 1 healthz.go:72] /healthz/log check >W0809 15:22:51.908493 1 reflector.go:272] k8s.io/client-go/informers/factory.go:130: watch of *v1beta1.StatefulSet ended with: The resourceVersion for the provided watch is too old. >I0809 15:22:55.596375 1 healthz.go:72] /healthz/log check >I0809 15:23:05.596446 1 healthz.go:72] /healthz/log check >I0809 15:23:15.596364 1 healthz.go:72] /healthz/log check >I0809 15:23:25.596311 1 healthz.go:72] /healthz/log check >I0809 15:23:35.596374 1 healthz.go:72] /healthz/log check >I0809 15:23:45.596459 1 healthz.go:72] /healthz/log check >I0809 15:23:55.596437 1 healthz.go:72] /healthz/log check >I0809 15:24:05.596405 1 healthz.go:72] /healthz/log check >I0809 15:24:15.596418 1 healthz.go:72] /healthz/log check >I0809 15:24:25.596336 1 healthz.go:72] /healthz/log check >I0809 15:24:35.596430 1 healthz.go:72] /healthz/log check >I0809 15:24:45.596341 1 healthz.go:72] /healthz/log check >I0809 15:24:55.596430 1 healthz.go:72] /healthz/log check >W0809 15:24:57.953746 1 reflector.go:272] k8s.io/client-go/informers/factory.go:130: watch of *v1.PersistentVolume ended with: The resourceVersion for the provided watch is too old. >I0809 15:25:05.597281 1 healthz.go:72] /healthz/log check >W0809 15:25:07.963333 1 reflector.go:272] k8s.io/client-go/informers/factory.go:130: watch of *v1.PersistentVolumeClaim ended with: The resourceVersion for the provided watch is too old. >I0809 15:25:15.596419 1 healthz.go:72] /healthz/log check >I0809 15:25:25.596343 1 healthz.go:72] /healthz/log check >I0809 15:25:35.596405 1 healthz.go:72] /healthz/log check >I0809 15:25:45.596461 1 healthz.go:72] /healthz/log check >I0809 15:25:55.596372 1 healthz.go:72] /healthz/log check >I0809 15:26:05.596379 1 healthz.go:72] /healthz/log check >I0809 15:26:15.596367 1 healthz.go:72] /healthz/log check >I0809 15:26:25.596311 1 healthz.go:72] /healthz/log check >I0809 15:26:35.596425 1 healthz.go:72] /healthz/log check >I0809 15:26:45.596499 1 healthz.go:72] /healthz/log check >I0809 15:26:55.596693 1 healthz.go:72] /healthz/log check >I0809 15:27:05.596433 1 healthz.go:72] /healthz/log check >I0809 15:27:15.596448 1 healthz.go:72] /healthz/log check >W0809 15:27:21.914655 1 reflector.go:272] k8s.io/client-go/informers/factory.go:130: watch of *v1.StorageClass ended with: The resourceVersion for the provided watch is too old. >I0809 15:27:25.600231 1 healthz.go:72] /healthz/log check >I0809 15:27:35.596438 1 healthz.go:72] /healthz/log check >I0809 15:27:45.596481 1 healthz.go:72] /healthz/log check >I0809 15:27:55.596377 1 healthz.go:72] /healthz/log check >I0809 15:28:05.596434 1 healthz.go:72] /healthz/log check >I0809 15:28:15.596998 1 healthz.go:72] /healthz/log check >I0809 15:28:25.596484 1 healthz.go:72] /healthz/log check >I0809 15:28:35.596422 1 healthz.go:72] /healthz/log check >I0809 15:28:45.596427 1 healthz.go:72] /healthz/log check >I0809 15:28:55.596452 1 healthz.go:72] /healthz/log check >I0809 15:29:05.596456 1 healthz.go:72] /healthz/log check >I0809 15:29:15.596487 1 healthz.go:72] /healthz/log check >I0809 15:29:25.596306 1 healthz.go:72] /healthz/log check >I0809 15:29:35.596438 1 healthz.go:72] /healthz/log check >I0809 15:29:45.596424 1 healthz.go:72] /healthz/log check >W0809 15:29:47.891940 1 reflector.go:272] k8s.io/client-go/informers/factory.go:130: watch of *v1beta1.ReplicaSet ended with: The resourceVersion for the provided watch is too old. >I0809 15:29:55.596465 1 healthz.go:72] /healthz/log check >I0809 15:30:05.596429 1 healthz.go:72] /healthz/log check >I0809 15:30:15.596462 1 healthz.go:72] /healthz/log check >I0809 15:30:25.596540 1 healthz.go:72] /healthz/log check >I0809 15:30:35.596470 1 healthz.go:72] /healthz/log check >I0809 15:30:45.596360 1 healthz.go:72] /healthz/log check >I0809 15:30:55.596441 1 healthz.go:72] /healthz/log check >I0809 15:31:05.596511 1 healthz.go:72] /healthz/log check >I0809 15:31:15.596461 1 healthz.go:72] /healthz/log check >I0809 15:31:25.596515 1 healthz.go:72] /healthz/log check >W0809 15:31:32.912003 1 reflector.go:272] k8s.io/client-go/informers/factory.go:130: watch of *v1beta1.PodDisruptionBudget ended with: The resourceVersion for the provided watch is too old. >I0809 15:31:35.596404 1 healthz.go:72] /healthz/log check >I0809 15:31:45.596466 1 healthz.go:72] /healthz/log check >I0809 15:31:55.596382 1 healthz.go:72] /healthz/log check >I0809 15:32:05.596473 1 healthz.go:72] /healthz/log check >I0809 15:32:15.596410 1 healthz.go:72] /healthz/log check >I0809 15:32:25.597110 1 healthz.go:72] /healthz/log check >I0809 15:32:35.596387 1 healthz.go:72] /healthz/log check >I0809 15:32:45.596389 1 healthz.go:72] /healthz/log check >I0809 15:32:55.596365 1 healthz.go:72] /healthz/log check >I0809 15:33:05.596481 1 healthz.go:72] /healthz/log check >W0809 15:33:14.012759 1 reflector.go:272] k8s.io/client-go/informers/factory.go:130: watch of *v1.PersistentVolume ended with: The resourceVersion for the provided watch is too old. >I0809 15:33:15.596384 1 healthz.go:72] /healthz/log check >I0809 15:33:25.596389 1 healthz.go:72] /healthz/log check >I0809 15:33:35.596399 1 healthz.go:72] /healthz/log check >I0809 15:33:45.596907 1 healthz.go:72] /healthz/log check >I0809 15:33:55.596464 1 healthz.go:72] /healthz/log check >I0809 15:34:05.596484 1 healthz.go:72] /healthz/log check >I0809 15:34:15.596306 1 healthz.go:72] /healthz/log check >I0809 15:34:25.596460 1 healthz.go:72] /healthz/log check >I0809 15:34:35.596419 1 healthz.go:72] /healthz/log check >I0809 15:34:45.596371 1 healthz.go:72] /healthz/log check >I0809 15:34:55.596372 1 healthz.go:72] /healthz/log check >I0809 15:35:05.596313 1 healthz.go:72] /healthz/log check >I0809 15:35:15.596324 1 healthz.go:72] /healthz/log check >I0809 15:35:25.596366 1 healthz.go:72] /healthz/log check >I0809 15:35:35.596405 1 healthz.go:72] /healthz/log check >I0809 15:35:45.596370 1 healthz.go:72] /healthz/log check >I0809 15:35:55.596488 1 healthz.go:72] /healthz/log check >I0809 15:36:05.596420 1 healthz.go:72] /healthz/log check >I0809 15:36:15.596409 1 healthz.go:72] /healthz/log check >I0809 15:36:25.596491 1 healthz.go:72] /healthz/log check >I0809 15:36:35.596913 1 healthz.go:72] /healthz/log check >I0809 15:36:45.596367 1 healthz.go:72] /healthz/log check >I0809 15:36:55.596447 1 healthz.go:72] /healthz/log check >I0809 15:37:05.596495 1 healthz.go:72] /healthz/log check >I0809 15:37:15.596503 1 healthz.go:72] /healthz/log check >I0809 15:37:25.596854 1 healthz.go:72] /healthz/log check >I0809 15:37:35.596310 1 healthz.go:72] /healthz/log check >I0809 15:37:45.596349 1 healthz.go:72] /healthz/log check >I0809 15:37:55.596425 1 healthz.go:72] /healthz/log check >I0809 15:38:05.596543 1 healthz.go:72] /healthz/log check >W0809 15:38:15.333251 1 reflector.go:272] k8s.io/client-go/informers/factory.go:130: watch of *v1.StorageClass ended with: The resourceVersion for the provided watch is too old. >I0809 15:38:15.596431 1 healthz.go:72] /healthz/log check >W0809 15:38:16.339188 1 reflector.go:272] k8s.io/client-go/informers/factory.go:130: watch of *v1.Service ended with: too old resource version: 6462 (10496) >I0809 15:38:22.620857 1 leaderelection.go:194] successfully acquired lease kube-system/kube-controller-manager >I0809 15:38:22.621082 1 event.go:221] Event(v1.ObjectReference{Kind:"ConfigMap", Namespace:"kube-system", Name:"kube-controller-manager", UID:"862a3d8d-9be4-11e8-ab01-02a810770ee6", APIVersion:"v1", ResourceVersion:"10512", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' ip-172-31-5-129.us-west-2.compute.internal_7dc0e228-9be4-11e8-9f15-02ecef0b760e became leader >E0809 15:38:22.640806 1 controllermanager.go:383] unable to get all supported resources from server: unable to retrieve the complete list of server APIs: servicecatalog.k8s.io/v1beta1: the server is currently unable to handle the request >I0809 15:38:22.641787 1 aws.go:1032] Building AWS cloudprovider >E0809 15:38:22.660973 1 memcache.go:147] couldn't get resource list for servicecatalog.k8s.io/v1beta1: the server is currently unable to handle the request >I0809 15:38:22.899950 1 tags.go:76] AWS cloud filtering on ClusterID: vlaad-311 >I0809 15:38:22.911189 1 controller_utils.go:1025] Waiting for caches to sync for tokens controller >I0809 15:38:22.944835 1 controllermanager.go:477] Starting "serviceaccount" >I0809 15:38:22.981665 1 controllermanager.go:487] Started "serviceaccount" >I0809 15:38:22.981905 1 controllermanager.go:477] Starting "job" >I0809 15:38:22.981875 1 serviceaccounts_controller.go:117] Starting service account controller >I0809 15:38:22.981984 1 controller_utils.go:1025] Waiting for caches to sync for service account controller >I0809 15:38:23.000807 1 controllermanager.go:487] Started "job" >I0809 15:38:23.000828 1 controllermanager.go:477] Starting "disruption" >I0809 15:38:23.000850 1 job_controller.go:143] Starting job controller >I0809 15:38:23.000858 1 controller_utils.go:1025] Waiting for caches to sync for job controller >I0809 15:38:23.011594 1 controller_utils.go:1032] Caches are synced for tokens controller >I0809 15:38:23.017463 1 controllermanager.go:487] Started "disruption" >I0809 15:38:23.017481 1 controllermanager.go:477] Starting "persistentvolume-binder" >I0809 15:38:23.017485 1 disruption.go:288] Starting disruption controller >I0809 15:38:23.017503 1 controller_utils.go:1025] Waiting for caches to sync for disruption controller >I0809 15:38:23.036254 1 plugins.go:500] Loaded volume plugin "kubernetes.io/host-path" >I0809 15:38:23.036276 1 plugins.go:500] Loaded volume plugin "kubernetes.io/nfs" >I0809 15:38:23.036287 1 plugins.go:500] Loaded volume plugin "kubernetes.io/glusterfs" >I0809 15:38:23.036298 1 plugins.go:500] Loaded volume plugin "kubernetes.io/rbd" >I0809 15:38:23.036309 1 plugins.go:500] Loaded volume plugin "kubernetes.io/quobyte" >I0809 15:38:23.036324 1 plugins.go:500] Loaded volume plugin "kubernetes.io/azure-file" >I0809 15:38:23.036333 1 plugins.go:500] Loaded volume plugin "kubernetes.io/flocker" >I0809 15:38:23.036346 1 plugins.go:500] Loaded volume plugin "kubernetes.io/portworx-volume" >I0809 15:38:23.036357 1 plugins.go:500] Loaded volume plugin "kubernetes.io/scaleio" >I0809 15:38:23.036369 1 plugins.go:500] Loaded volume plugin "kubernetes.io/local-volume" >I0809 15:38:23.036384 1 plugins.go:500] Loaded volume plugin "kubernetes.io/storageos" >I0809 15:38:23.036401 1 plugins.go:500] Loaded volume plugin "kubernetes.io/aws-ebs" >I0809 15:38:23.036415 1 plugins.go:500] Loaded volume plugin "kubernetes.io/gce-pd" >I0809 15:38:23.036433 1 plugins.go:500] Loaded volume plugin "kubernetes.io/cinder" >I0809 15:38:23.036450 1 plugins.go:500] Loaded volume plugin "kubernetes.io/vsphere-volume" >I0809 15:38:23.036465 1 plugins.go:500] Loaded volume plugin "kubernetes.io/azure-disk" >I0809 15:38:23.036477 1 plugins.go:500] Loaded volume plugin "kubernetes.io/photon-pd" >I0809 15:38:23.036596 1 controllermanager.go:487] Started "persistentvolume-binder" >I0809 15:38:23.036611 1 controllermanager.go:477] Starting "replicaset" >I0809 15:38:23.036668 1 pv_controller_base.go:271] Starting persistent volume controller >I0809 15:38:23.036680 1 controller_utils.go:1025] Waiting for caches to sync for persistent volume controller >I0809 15:38:23.050178 1 controllermanager.go:487] Started "replicaset" >I0809 15:38:23.050197 1 controllermanager.go:477] Starting "cronjob" >I0809 15:38:23.050283 1 replica_set.go:182] Starting replicaset controller >I0809 15:38:23.050293 1 controller_utils.go:1025] Waiting for caches to sync for ReplicaSet controller >I0809 15:38:23.063040 1 controllermanager.go:487] Started "cronjob" >I0809 15:38:23.063058 1 controllermanager.go:477] Starting "persistentvolume-expander" >I0809 15:38:23.063149 1 cronjob_controller.go:94] Starting CronJob Manager >I0809 15:38:23.079538 1 plugins.go:500] Loaded volume plugin "kubernetes.io/aws-ebs" >I0809 15:38:23.079571 1 plugins.go:500] Loaded volume plugin "kubernetes.io/gce-pd" >I0809 15:38:23.079581 1 plugins.go:500] Loaded volume plugin "kubernetes.io/cinder" >I0809 15:38:23.079595 1 plugins.go:500] Loaded volume plugin "kubernetes.io/portworx-volume" >I0809 15:38:23.079608 1 plugins.go:500] Loaded volume plugin "kubernetes.io/vsphere-volume" >I0809 15:38:23.079616 1 plugins.go:500] Loaded volume plugin "kubernetes.io/glusterfs" >I0809 15:38:23.079625 1 plugins.go:500] Loaded volume plugin "kubernetes.io/rbd" >I0809 15:38:23.079633 1 plugins.go:500] Loaded volume plugin "kubernetes.io/azure-disk" >I0809 15:38:23.079645 1 plugins.go:500] Loaded volume plugin "kubernetes.io/azure-file" >I0809 15:38:23.079661 1 plugins.go:500] Loaded volume plugin "kubernetes.io/photon-pd" >I0809 15:38:23.079670 1 plugins.go:500] Loaded volume plugin "kubernetes.io/scaleio" >I0809 15:38:23.079684 1 plugins.go:500] Loaded volume plugin "kubernetes.io/storageos" >I0809 15:38:23.079695 1 plugins.go:500] Loaded volume plugin "kubernetes.io/fc" >I0809 15:38:23.079793 1 controllermanager.go:487] Started "persistentvolume-expander" >I0809 15:38:23.079808 1 controllermanager.go:477] Starting "resourcequota" >I0809 15:38:23.079863 1 expand_controller.go:150] Starting expand controller >I0809 15:38:23.079875 1 controller_utils.go:1025] Waiting for caches to sync for expand controller >E0809 15:38:23.099658 1 controllermanager.go:480] Error starting "resourcequota" >F0809 15:38:23.099677 1 controllermanager.go:185] error starting controllers: failed to discover resources: unable to retrieve the complete list of server APIs: servicecatalog.k8s.io/v1beta1: the server is currently unable to handle the request
You cannot view the attachment while viewing its details because your browser does not support IFRAMEs.
View the attachment on a separate page
.
View Attachment As Raw
Actions:
View
Attachments on
bug 1614479
: 1474781