Bug 1733830 - many "Error on ingesting results from rule evaluation with different value but same timestamp" msg in prometheus-k8s pod logs
Summary: many "Error on ingesting results from rule evaluation with different value bu...
Keywords:
Status: CLOSED ERRATA
Alias: None
Product: OpenShift Container Platform
Classification: Red Hat
Component: Monitoring
Version: 4.2.0
Hardware: Unspecified
OS: Unspecified
low
low
Target Milestone: ---
: 4.2.0
Assignee: Lili Cosic
QA Contact: Viacheslav
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2019-07-29 02:49 UTC by Junqi Zhao
Modified: 2019-10-16 06:34 UTC (History)
7 users (show)

Fixed In Version:
Doc Type: If docs needed, set a value
Doc Text:
Clone Of:
Environment:
Last Closed: 2019-10-16 06:33:51 UTC
Target Upstream Version:
Embargoed:


Attachments (Terms of Use)
"Error on ingesting results from rule evaluation with different value but same timestamp" msg (110.78 KB, text/plain)
2019-07-29 02:49 UTC, Junqi Zhao
no flags Details


Links
System ID Private Priority Status Summary Last Updated
Github coreos kube-prometheus issues 134 0 None closed Duplicate records in prometheus-rules 2021-02-19 13:35:58 UTC
Github kubernetes-monitoring kubernetes-mixin issues 230 0 None closed Recording rule generating "duplicate sample for timestamp" errors 2021-02-19 13:35:59 UTC
Github openshift cluster-monitoring-operator pull 442 0 None closed Bug 1733830: Bump kubernetes-mixin 2021-02-19 13:35:59 UTC
Github openshift cluster-monitoring-operator pull 456 0 None closed Update node_exporter mixin and kubernetes mixin 2021-02-19 13:35:59 UTC
Red Hat Product Errata RHBA-2019:2922 0 None None None 2019-10-16 06:34:05 UTC

Description Junqi Zhao 2019-07-29 02:49:03 UTC
Created attachment 1594135 [details]
"Error on ingesting results from rule evaluation with different value but same timestamp" msg

Description of problem:
See the attached log file, there are a lot of
"Error on ingesting results from rule evaluation with different value but same timestamp" msg in prometheus-k8s pod logs

It does not affect the prometheus function by far

# oc -n openshift-monitoring logs -c prometheus prometheus-k8s-0 
level=warn ts=2019-07-29T00:58:25.748Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-07-29T00:58:55.744Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=9
level=warn ts=2019-07-29T00:58:55.748Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-07-29T00:59:25.744Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=9
level=warn ts=2019-07-29T00:59:25.748Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-07-29T00:59:55.744Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=10
level=warn ts=2019-07-29T00:59:55.748Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-07-29T01:00:25.744Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=11
level=warn ts=2019-07-29T01:00:25.748Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
Version-Release number of selected component (if applicable):
Always

How reproducible:
4.2.0-0.nightly-2019-07-28-222114

Steps to Reproduce:
1. oc -n openshift-monitoring logs -c prometheus prometheus-k8s-0 
2.
3.

Actual results:
"Error on ingesting results from rule evaluation with different value but same timestamp" msg

Expected results:
Should not have such error

Additional info:

Comment 1 Andrew Pickering 2019-07-29 13:10:36 UTC
Simon, could you please take a look?

Comment 2 Simon Pasquier 2019-07-29 14:13:45 UTC
Sure I'll look into it.

Comment 3 Simon Pasquier 2019-07-30 08:17:56 UTC
I've enabled debug logging for Prometheus and can see the details about the offending rules.

> level=debug ts=2019-07-30T07:59:55.749Z caller=manager.go:542 component="rule manager" group=k8s.rules msg="Rule evaluation result discarded" err="duplicate sample for timestamp" sample="{__name__=\"namespace:container_memory_usage_bytes:sum\", namespace=\"openshift-image-registry\"} => 24219648 @[1564473595729]"
> level=debug ts=2019-07-30T07:59:55.749Z caller=manager.go:542 component="rule manager" group=k8s.rules msg="Rule evaluation result discarded" err="duplicate sample for timestamp" sample="{__name__=\"namespace:container_memory_usage_bytes:sum\", namespace=\"openshift-monitoring\"} => 2301251584 @[1564473595729]"
> level=debug ts=2019-07-30T07:59:55.749Z caller=manager.go:542 component="rule manager" group=k8s.rules msg="Rule evaluation result discarded" err="duplicate sample for timestamp" sample="{__name__=\"namespace:container_memory_usage_bytes:sum\", namespace=\"openshift-marketplace\"} => 210898944 @[1564473595729]"
> level=debug ts=2019-07-30T07:59:55.749Z caller=manager.go:542 component="rule manager" group=k8s.rules msg="Rule evaluation result discarded" err="duplicate sample for timestamp" sample="{__name__=\"namespace:container_memory_usage_bytes:sum\", namespace=\"openshift-cluster-node-tuning-operator\"} => 246198272 @[1564473595729]"
> level=debug ts=2019-07-30T07:59:55.749Z caller=manager.go:542 component="rule manager" group=k8s.rules msg="Rule evaluation result discarded" err="duplicate sample for timestamp" sample="{__name__=\"namespace:container_memory_usage_bytes:sum\", namespace=\"openshift-machine-config-operator\"} => 204333056 @[1564473595729]"
> level=warn ts=2019-07-30T07:59:55.749Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5

There are indeed 2 recording rules that populate the same metric but with different label sets.

It has been flagged upstream already:
https://github.com/coreos/kube-prometheus/issues/134

I've created an issue in the kubernetes-mixin repository too:
https://github.com/kubernetes-monitoring/kubernetes-mixin/issues/230

Comment 6 Viacheslav 2019-08-19 12:58:36 UTC
Fixed.

Comment 10 Viacheslav 2019-08-20 09:10:13 UTC
Filed after retest on latest night build 4.2.0-0.nightly-2019-08-20-043744


level=warn ts=2019-08-20T08:45:55.739Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=6
level=warn ts=2019-08-20T08:46:25.740Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=6
level=warn ts=2019-08-20T08:46:55.739Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T08:47:25.739Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T08:47:55.740Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T08:48:25.740Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T08:48:55.740Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T08:49:25.741Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T08:49:55.741Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T08:50:25.741Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T08:50:55.741Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T08:51:25.741Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T08:51:55.741Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T08:52:25.742Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T08:52:55.741Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T08:53:25.742Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T08:53:55.742Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T08:54:25.742Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T08:54:55.742Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T08:55:25.742Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T08:55:55.743Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T08:56:25.743Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T08:56:55.743Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=6
level=warn ts=2019-08-20T08:57:25.743Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=6
level=warn ts=2019-08-20T08:57:55.743Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=6
level=warn ts=2019-08-20T08:58:25.743Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=6
level=warn ts=2019-08-20T08:58:55.743Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T08:59:25.743Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T08:59:55.744Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T09:00:25.743Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T09:00:55.743Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T09:01:25.744Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T09:01:55.744Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T09:02:25.744Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T09:02:55.744Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5
level=warn ts=2019-08-20T09:03:25.744Z caller=manager.go:554 component="rule manager" group=k8s.rules msg="Error on ingesting results from rule evaluation with different value but same timestamp" numDropped=5

Comment 12 Lili Cosic 2019-08-21 11:42:09 UTC
Simon: I brought over the changes you made in mixins that you mentioned above [1] to CMO[2] master so 4.2, so your fix should be in master - 4.2.

[1]: https://github.com/kubernetes-monitoring/kubernetes-mixin/pull/232
[2]: https://github.com/openshift/cluster-monitoring-operator/pull/442/files#diff-1dca87d186c04a487d72e52ab0b4dde5L24-L30

Comment 13 Sergiusz Urbaniak 2019-08-21 12:35:32 UTC
@viacheslav, @junqi: do you mind to retest? This should be fixed, setting to MODIFIED.

Comment 14 Junqi Zhao 2019-08-22 03:40:18 UTC
(In reply to Sergiusz Urbaniak from comment #13)
> @viacheslav, @junqi: do you mind to retest? This should be fixed, setting to
> MODIFIED.

tested with 4.2.0-0.nightly-2019-08-21-235427, still see the error

Comment 15 Viacheslav 2019-08-22 08:18:34 UTC
errors still present grouped near the bottom of log, was checked on 4.2.0-0.nightly-2019-08-22-04381

Comment 19 Junqi Zhao 2019-09-09 02:34:18 UTC
no such issue with 4.2.0-0.nightly-2019-09-08-180038

Comment 20 errata-xmlrpc 2019-10-16 06:33:51 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory, and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHBA-2019:2922


Note You need to log in before you can comment on or make changes to this bug.