Bug 1314729
Summary: | change severity of some messages related to "keep alive" entries | ||
---|---|---|---|
Product: | Red Hat Enterprise Linux 6 | Reporter: | Jaroslav Reznik <jreznik> |
Component: | 389-ds-base | Assignee: | Noriko Hosoi <nhosoi> |
Status: | CLOSED ERRATA | QA Contact: | Viktor Ashirov <vashirov> |
Severity: | urgent | Docs Contact: | |
Priority: | urgent | ||
Version: | 6.7 | CC: | amsharma, ekeck, gparente, nhosoi, nkinder, rmeggins, tbordaz |
Target Milestone: | rc | Keywords: | ZStream |
Target Release: | --- | ||
Hardware: | x86_64 | ||
OS: | Linux | ||
Whiteboard: | |||
Fixed In Version: | 389-ds-base-1.2.11.15-72.el6_7 | Doc Type: | Bug Fix |
Doc Text: |
Keep alive entries are used to prevent skipped updates from being evaluated several times in a fractional replication. If a large number of updates is skipped, these entries can be updated very frequently. Before the keep alive entries are updated, it is tested that they exist. Previously, the test was being logged at a "Fatal" log level, so it was always logged no matter which log level had been set. As a consequence, error logs were filled with unnecessary messages. With this update, the log level for keep alive entry creation has been changed from "Fatal" to "Replication debugging" (8192), and the error log file is no longer filled with "Fatal" messages.
|
Story Points: | --- |
Clone Of: | 1313258 | Environment: | |
Last Closed: | 2016-03-22 14:47:51 UTC | Type: | --- |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: | |||
Bug Depends On: | 1313258 | ||
Bug Blocks: |
Description
Jaroslav Reznik
2016-03-04 10:59:25 UTC
[root@ipaqa64vml ~]# rpm -qa | grep 389 389-ds-base-libs-1.2.11.15-72.el6_7.x86_64 389-ds-base-1.2.11.15-72.el6_7.x86_64 Errors logs without replication log level:: [root@ipaqa64vml ~]# tail -f /var/log/dirsrv/slapd-M2/errors [10/Mar/2016:02:31:20 -0500] - import userRoot: Workers finished; cleaning up... [10/Mar/2016:02:31:20 -0500] - import userRoot: Workers cleaned up. [10/Mar/2016:02:31:20 -0500] - import userRoot: Indexing complete. Post-processing... [10/Mar/2016:02:31:20 -0500] - import userRoot: Generating numSubordinates complete. [10/Mar/2016:02:31:20 -0500] - import userRoot: Flushing caches... [10/Mar/2016:02:31:20 -0500] - import userRoot: Closing files... [10/Mar/2016:02:31:20 -0500] - import userRoot: Import complete. Processed 11 entries in 3 seconds. (3.67 entries/sec) [10/Mar/2016:02:31:21 -0500] NSMMReplicationPlugin - multimaster_be_state_change: replica dc=example,dc=com is coming online; enabling replication [root@ipaqa64vml yum.repos.d]# tail -f /var/log/dirsrv/slapd-M1/errors [10/Mar/2016:02:20:08 -0500] - slapd started. Listening on All Interfaces port 389 for LDAP requests [10/Mar/2016:02:24:02 -0500] NSMMReplicationPlugin - Beginning total update of replica "agmt="cn=M1_to_M2" (localhost:1389)". [10/Mar/2016:02:24:02 -0500] NSMMReplicationPlugin - Need to create replication keep alive entry <cn=repl keep alive 1,dc=example,dc=com> [10/Mar/2016:02:24:02 -0500] NSMMReplicationPlugin - add dn: cn=repl keep alive 1,dc=example,dc=com objectclass: top objectclass: ldapsubentry objectclass: extensibleObject cn: repl keep alive 1 [10/Mar/2016:02:24:06 -0500] NSMMReplicationPlugin - Finished total update of replica "agmt="cn=M1_to_M2" (localhost:1389)". Sent 10 entries. With Replication logs:: [10/Mar/2016:02:27:20 -0500] NSMMReplicationPlugin - agmt="cn=M1_to_M2" (localhost:1389): Replica was successfully acquired. [10/Mar/2016:02:27:20 -0500] NSMMReplicationPlugin - Beginning total update of replica "agmt="cn=M1_to_M2" (localhost:1389)". [10/Mar/2016:02:27:20 -0500] NSMMReplicationPlugin - replication keep alive entry <cn=repl keep alive 1,dc=example,dc=com> already exists [10/Mar/2016:02:27:20 -0500] - repl5_tot_waitfor_async_results: 0 15 [10/Mar/2016:02:27:21 -0500] - repl5_tot_waitfor_async_results: 15 15 [10/Mar/2016:02:27:23 -0500] NSMMReplicationPlugin - agmt="cn=M1_to_M2" (localhost:1389): Successfully released consumer [10/Mar/2016:02:27:23 -0500] NSMMReplicationPlugin - agmt="cn=M1_to_M2" (localhost:1389): Beginning linger on the connection [10/Mar/2016:02:27:23 -0500] NSMMReplicationPlugin - Finished total update of replica "agmt="cn=M1_to_M2" (localhost:1389)". Sent 11 entries. [10/Mar/2016:02:27:23 -0500] NSMMReplicationPlugin - agmt="cn=M1_to_M2" (localhost:1389): Cancelling linger on the connection [10/Mar/2016:02:27:23 -0500] NSMMReplicationPlugin - agmt="cn=M1_to_M2" (localhost:1389): Disconnected from the consumer [10/Mar/2016:02:27:23 -0500] NSMMReplicationPlugin - agmt="cn=M1_to_M2" (localhost:1389): State: start -> ready_to_acquire_replica [10/Mar/2016:02:27:23 -0500] NSMMReplicationPlugin - agmt="cn=M1_to_M2" (localhost:1389): Trying non-secure slapi_ldap_init_ext [10/Mar/2016:02:27:23 -0500] NSMMReplicationPlugin - agmt="cn=M1_to_M2" (localhost:1389): binddn = cn=replication manager,cn=config, passwd = {DES}2mO8LgMr/w5ognn2BRK+gQ== [10/Mar/2016:02:27:23 -0500] NSMMReplicationPlugin - agmt="cn=M1_to_M2" (localhost:1389): No linger to cancel on the connection [10/Mar/2016:02:27:23 -0500] - _csngen_adjust_local_time: gen state before 56e121d70001:1457594839:0:0 [10/Mar/2016:02:27:23 -0500] - _csngen_adjust_local_time: gen state after 56e121db0000:1457594843:0:0 [10/Mar/2016:02:27:23 -0500] NSMMReplicationPlugin - agmt="cn=M1_to_M2" (localhost:1389): Replica was successfully acquired. [10/Mar/2016:02:27:23 -0500] NSMMReplicationPlugin - agmt="cn=M1_to_M2" (localhost:1389): State: ready_to_acquire_replica -> sending_updates [10/Mar/2016:02:27:23 -0500] - csngen_adjust_time: gen state before 56e121db0001:1457594843:0:0 [10/Mar/2016:02:27:23 -0500] NSMMReplicationPlugin - changelog program - _cl5GetDBFile: found DB object 7f425c015c60 for database /var/lib/dirsrv/slapd-M1/changelogdb/cb96b982-e69011e5-b456b189-0a306cbb_56e120a6000000010000.db4 [10/Mar/2016:02:27:23 -0500] - _cl5PositionCursorForReplay (agmt="cn=M1_to_M2" (localhost:1389)): Consumer RUV: [10/Mar/2016:02:27:23 -0500] NSMMReplicationPlugin - agmt="cn=M1_to_M2" (localhost:1389): {replicageneration} 56e120a6000000010000 [10/Mar/2016:02:27:23 -0500] NSMMReplicationPlugin - agmt="cn=M1_to_M2" (localhost:1389): {replica 2 ldap://ipaqa64vml.idmqe.lab.eng.bos.redhat.com:1389} [10/Mar/2016:02:27:23 -0500] NSMMReplicationPlugin - agmt="cn=M1_to_M2" (localhost:1389): {replica 1 ldap://ipaqa64vml.idmqe.lab.eng.bos.redhat.com:389} 56e12112000100010000 56e1216d000000010000 00000000 [10/Mar/2016:02:27:23 -0500] - _cl5PositionCursorForReplay (agmt="cn=M1_to_M2" (localhost:1389)): Supplier RUV: [10/Mar/2016:02:27:23 -0500] NSMMReplicationPlugin - agmt="cn=M1_to_M2" (localhost:1389): {replicageneration} 56e120a6000000010000 Hence marking VERIFIED. Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://rhn.redhat.com/errata/RHBA-2016-0468.html |