Bug 1641604
| Summary: | [Deployment] Logs are full of NULL (0 binary) | |||
|---|---|---|---|---|
| Product: | Red Hat OpenStack | Reporter: | Noam Manos <nmanos> | |
| Component: | openstack-tripleo-heat-templates | Assignee: | Victor Pickard <vpickard> | |
| Status: | CLOSED ERRATA | QA Contact: | Noam Manos <nmanos> | |
| Severity: | urgent | Docs Contact: | ||
| Priority: | high | |||
| Version: | 14.0 (Rocky) | CC: | jchhatba, jjoyce, jschluet, mburns, mkolesni, nyechiel, slinaber, tfreger, tvignaud, vpickard, wznoinsk | |
| Target Milestone: | rc | Keywords: | Triaged | |
| Target Release: | 14.0 (Rocky) | |||
| Hardware: | Unspecified | |||
| OS: | Unspecified | |||
| Whiteboard: | Deployment | |||
| Fixed In Version: | openstack-tripleo-heat-templates-9.0.1-0.20181013060884.el7ost.noarch.rpm | Doc Type: | If docs needed, set a value | |
| Doc Text: | Story Points: | --- | ||
| Clone Of: | ||||
| : | 1648348 (view as bug list) | Environment: | ||
| Last Closed: | 2019-01-11 11:54:07 UTC | Type: | Bug | |
| Regression: | --- | Mount Type: | --- | |
| Documentation: | --- | CRM: | ||
| Verified Versions: | Category: | --- | ||
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | ||
| Cloudforms Team: | --- | Target Upstream Version: | ||
| Embargoed: | ||||
| Bug Depends On: | ||||
| Bug Blocks: | 1648348 | |||
|
Description
Noam Manos
2018-10-22 10:14:01 UTC
After investigating the issue with Janki it seems that the cause of this problem is that the log file is also being rotated by a general log rotation policy from TripleO, which causes Log4J to freak out and output tons of NULL (0 binary) into the log. Vic is already working on a fix upstream so assigning this to him. In the meanwhile it's recommended to apply the fix manually before deploying. Using puddle 2018-10-17.2 + patch https://review.openstack.org/#/c/612468 By adding to to deployment OVERCLOUD_DEPLOY_OVERRIDE_OPTIONS : --fetchfiles-undercloud https://git.openstack.org/cgit/openstack/tripleo-heat-templates/tree/docker/services/logging/files/opendaylight-api.yaml?h=refs/changes/68/612468/1,/usr/share/openstack-tripleo-heat-templates/docker/services/logging/files/opendaylight-api.yaml I can see that the karaf.log have moved into new sub directories: /var/log/containers/opendaylight/karaf/logs Rather than directly in opendaylight dir. Is it by design ? (undercloud) [stack@undercloud-0 ~]$ ssh heat-admin@controller-1 "ll /var/log/containers/opendaylight/"; bash: ll: command not found (undercloud) [stack@undercloud-0 ~]$ ssh heat-admin@controller-1 "ls /var/log/containers/opendaylight/"; karaf (undercloud) [stack@undercloud-0 ~]$ ssh heat-admin@controller-1 "ls /var/log/containers/opendaylight/karaf"; logs (undercloud) [stack@undercloud-0 ~]$ ssh heat-admin@controller-1 "ls /var/log/containers/opendaylight/karaf/logs"; karaf.log (undercloud) [stack@undercloud-0 ~]$ ssh heat-admin@controller-1 "ls -lah /var/log/containers/opendaylight/karaf/logs"; total 5.6G drwxr-xr-x. 2 42462 42462 23 Oct 24 13:27 . drwxr-xr-x. 3 root root 18 Oct 24 13:11 .. -rw-r--r--. 1 42462 42462 5.1G Oct 28 14:56 karaf.log However, the NULL binary bytes is not there, and a normal log output is created: (undercloud) [stack@undercloud-0 ~]$ ssh heat-admin@controller-1 "tail /var/log/containers/opendaylight/karaf/logs/*"; 2018-10-28T14:53:32,508 | INFO | nioEventLoopGroup-11-8 | AbstractConnectionAdapter | 387 - org.opendaylight.openflowplugin.openflowjava.openflow-protocol-impl - 0.6.4.redhat-4 | The channel outbound queue size:1024 2018-10-28T14:53:32,508 | INFO | nioEventLoopGroup-9-13 | SystemNotificationsListenerImpl | 378 - org.opendaylight.openflowplugin.impl - 0.6.4.redhat-4 | ConnectionEvent: Connection closed by device, Device:/127.0.0.1:37564, NodeId:null 2018-10-28T14:53:32,509 | INFO | nioEventLoopGroup-11-8 | SystemNotificationsListenerImpl | 378 - org.opendaylight.openflowplugin.impl - 0.6.4.redhat-4 | ConnectionEvent: Connection closed by device, Device:/127.0.0.1:58424, NodeId:null 2018-10-28T14:53:32,509 | INFO | nioEventLoopGroup-9-14 | AbstractConnectionAdapter | 387 - org.opendaylight.openflowplugin.openflowjava.openflow-protocol-impl - 0.6.4.redhat-4 | The channel outbound queue size:1024 2018-10-28T14:53:32,511 | INFO | nioEventLoopGroup-11-9 | AbstractConnectionAdapter | 387 - org.opendaylight.openflowplugin.openflowjava.openflow-protocol-impl - 0.6.4.redhat-4 | The channel outbound queue size:1024 2018-10-28T14:53:32,511 | INFO | nioEventLoopGroup-9-14 | SystemNotificationsListenerImpl | 378 - org.opendaylight.openflowplugin.impl - 0.6.4.redhat-4 | ConnectionEvent: Connection closed by device, Device:/127.0.0.1:37568, NodeId:null 2018-10-28T14:53:32,512 | INFO | nioEventLoopGroup-11-9 | SystemNotificationsListenerImpl | 378 - org.opendaylight.openflowplugin.impl - 0.6.4.redhat-4 | ConnectionEvent: Connection closed by device, Device:/127.0.0.1:58430, NodeId:null 2018-10-28T14:53:32,533 | DEBUG | org.opendaylight.yang.gen.v1.urn.tbd.params.xml.ns.yang.network.topology.rev131021.network.topology.topology.Node_AsyncDataTreeChangeListenerBase-DataTreeChangeHandler-0 | ElanOvsdbNodeListener | 344 - org.opendaylight.netvirt.elanmanager-impl - 0.6.4.redhat-4 | ElanOvsdbNodeListener.update, updated node detected. original: Node{getNodeId=Uri [_value=ovsdb://uuid/ef0201ff-886a-4427-a667-e3e8b9abe8ce], augmentations={interface org.opendaylight.yang.gen.v1.urn.opendaylight.params.xml.ns.yang.ovsdb.rev150105.OvsdbNodeAugmentation=OvsdbNodeAugmentation{getConnectionInfo=ConnectionInfo{getLocalIp=IpAddress [_ipv4Address=Ipv4Address [_value=172.17.1.17]] Hi, Yes, the karaf logs are now located here, as you stated: /var/log/containers/karaf/logs/karaf.log Please refer to the bug below for more details https://bugs.launchpad.net/tripleo/+bug/1793566 RPM with the fix can be found here: http://brew-task-repos.usersys.redhat.com/repos/official/openstack-tripleo-heat-templates/9.0.1/0.20181013060884.el7ost/noarch/openstack-tripleo-heat-templates-9.0.1-0.20181013060884.el7ost.noarch.rpm Running OSP14 deployment with openvswitch2.10-2.10.0-21.el7fdn.x86_64.rpm: http://download.eng.bos.redhat.com/brewroot/packages/openvswitch2.10/2.10.0/21.el7fdn/x86_64/openvswitch2.10-2.10.0-21.el7fdn.x86_64.rpm But there's no new file structure for odl logs, and disk is 100% full after deployment: ###### controller-0: ssh heat-admin.24.11 ###### b7c98e16d013 192.168.24.1:8787/rhosp14/openstack-gnocchi-metricd:2018-11-09.3 "kolla_start" 4 hours ago Up 4 hours (unhealthy) gnocchi_metricd a15023406b91 192.168.24.1:8787/rhosp14/openstack-ceilometer-central:2018-11-09.3 "kolla_start" 4 hours ago Up 4 hours (unhealthy) ceilometer_agent_central fc00f047c54c 192.168.24.1:8787/rhosp14/openstack-opendaylight:2018-11-09.3 "kolla_start" 4 hours ago Up 4 hours (unhealthy) opendaylight_api Filesystem Size Used Avail Use% Mounted on /dev/vda2 20G 20G 20K 100% / devtmpfs 16G 0 16G 0% /dev tmpfs 16G 39M 16G 1% /dev/shm tmpfs 16G 151M 16G 1% /run tmpfs 16G 0 16G 0% /sys/fs/cgroup tmpfs 3.2G 0 3.2G 0% /run/user/0 tmpfs 3.2G 0 3.2G 0% /run/user/1000 [heat-admin@controller-0 ~]$ ll /var/log/containers/opendaylight/ total 106476 -rw-r--r--. 1 42462 42462 106352640 Nov 18 15:13 karaf.log -rw-r--r--. 1 42462 42462 88886182 Nov 18 13:00 karaf.log.1 -rw-r--r--. 1 42462 42462 2667866 Nov 18 12:00 karaf.log.2.gz [heat-admin@controller-0 /]$ [heat-admin@controller-0 /]$ tail -1000 /var/log/containers/opendaylight/karaf.log 2018-11-18T15:18:35,261 | INFO | nioEventLoopGroup-11-3 | AbstractConnectionAdapter | 387 - org.opendaylight.openflowplugin.openflowjava.openflow-protocol-impl - 0.6.4.redhat-6 | The channel outbound queue size:1024 2018-11-18T15:18:35,261 | INFO | nioEventLoopGroup-11-2 | AbstractConnectionAdapter | 387 - org.opendaylight.openflowplugin.openflowjava.openflow-protocol-impl - 0.6.4.redhat-6 | The channel outbound queue size:1024 2018-11-18T15:18:35,261 | INFO | nioEventLoopGroup-11-3 | SystemNotificationsListenerImpl | 378 - org.opendaylight.openflowplugin.impl - 0.6.4.redhat-6 | ConnectionEvent: Connection closed by device, Device:/127.0.0.1:46372, NodeId:null 2018-11-18T15:18:35,262 | INFO | nioEventLoopGroup-11-2 | SystemNotificationsListenerImpl | 378 - org.opendaylight.openflowplugin.impl - 0.6.4.redhat-6 | ConnectionEvent: Connection closed by device, Device:/127.0.0.1:46368, NodeId:null 2018-11-18T15:18:36,699 | ERROR | qtp626219007-21545 | DatastoreServiceStatusProvider | 262 - org.opendaylight.genius.mdsalutil-impl - 0.4.4.redhat-6 | Unable to obtain the datastore status java.lang.reflect.UndeclaredThrowableException: null at com.sun.proxy.$Proxy110.getRaftState(Unknown Source) [231:org.opendaylight.controller.sal-distributed-datastore:1.7.4.redhat-6] at org.opendaylight.genius.mdsalutil.diagstatus.internal.DatastoreServiceStatusProvider.getServiceDescriptor(DatastoreServiceStatusProvider.java:89) [262:org.opendaylight.genius.mdsalutil-impl:0.4.4.redhat-6] at Proxy241f9f36_ab17_4d9d_9a59_3fad77bf0588.getServiceDescriptor(Unknown Source) [?:?] at Proxyd6f0d70e_6629_4345_b144_55579e64838d.getServiceDescriptor(Unknown Source) [?:?] at org.opendaylight.infrautils.diagstatus.internal.DiagStatusServiceImpl.updateServiceStatusMap(DiagStatusServiceImpl.java:137) [272:org.opendaylight.infrautils.diagstatus-impl:1.3.4.redhat-6] at org.opendaylight.infrautils.diagstatus.internal.DiagStatusServiceImpl.getAllServiceDescriptors(DiagStatusServiceImpl.java:86) [272:org.opendaylight.infrautils.diagstatus-impl:1.3.4.redhat-6] at org.opendaylight.infrautils.diagstatus.internal.DiagStatusServiceImpl.isOperational(DiagStatusServiceImpl.java:95) [272:org.opendaylight.infrautils.diagstatus-impl:1.3.4.redhat-6] at Proxyea07b505_74b4_459e_95c3_86daec3e8e0b.isOperational(Unknown Source) [?:?] at Proxy108b1804_c30c_4378_a025_b6f5674a1215.isOperational(Unknown Source) [?:?] at org.opendaylight.infrautils.diagstatus.web.DiagStatusServlet.doGet(DiagStatusServlet.java:42) [274:org.opendaylight.infrautils.diagstatus-web:1.3.4.redhat-6] at javax.servlet.http.HttpServlet.service(HttpServlet.java:687) [72:javax.servlet-api:3.1.0] at javax.servlet.http.HttpServlet.service(HttpServlet.java:790) [72:javax.servlet-api:3.1.0] at org.eclipse.jetty.servlet.ServletHolder.handle(ServletHolder.java:848) [163:org.eclipse.jetty.servlet:9.3.21.v20170918] at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1772) [163:org.eclipse.jetty.servlet:9.3.21.v20170918] at org.eclipse.jetty.websocket.server.WebSocketUpgradeFilter.doFilter(WebSocketUpgradeFilter.java:205) [173:org.eclipse.jetty.websocket.server:9.3.21.v20170918] at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1759) [163:org.eclipse.jetty.servlet:9.3.21.v20170918] at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:582) [163:org.eclipse.jetty.servlet:9.3.21.v20170918] at org.ops4j.pax.web.service.jetty.internal.HttpServiceServletHandler.doHandle(HttpServiceServletHandler.java:71) [441:org.ops4j.pax.web.pax-web-jetty:6.0.9] at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:143) [162:org.eclipse.jetty.server:9.3.21.v20170918] at org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:548) [160:org.eclipse.jetty.security:9.3.21.v20170918] at org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:226) [162:org.eclipse.jetty.server:9.3.21.v20170918] at org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1180) [162:org.eclipse.jetty.server:9.3.21.v20170918] at org.ops4j.pax.web.service.jetty.internal.HttpServiceContext.doHandle(HttpServiceContext.java:284) [441:org.ops4j.pax.web.pax-web-jetty:6.0.9] at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:512) [163:org.eclipse.jetty.servlet:9.3.21.v20170918] at org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:185) [162:org.eclipse.jetty.server:9.3.21.v20170918] at org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1112) [162:org.eclipse.jetty.server:9.3.21.v20170918] at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141) [162:org.eclipse.jetty.server:9.3.21.v20170918] at org.ops4j.pax.web.service.jetty.internal.JettyServerHandlerCollection.handle(JettyServerHandlerCollection.java:80) [441:org.ops4j.pax.web.pax-web-jetty:6.0.9] at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:134) [162:org.eclipse.jetty.server:9.3.21.v20170918] at org.eclipse.jetty.server.Server.handle(Server.java:534) [162:org.eclipse.jetty.server:9.3.21.v20170918] at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:333) [162:org.eclipse.jetty.server:9.3.21.v20170918] at org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:251) [162:org.eclipse.jetty.server:9.3.21.v20170918] at org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:283) [154:org.eclipse.jetty.io:9.3.21.v20170918] at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:108) [154:org.eclipse.jetty.io:9.3.21.v20170918] at org.eclipse.jetty.io.SelectChannelEndPoint$2.run(SelectChannelEndPoint.java:93) [154:org.eclipse.jetty.io:9.3.21.v20170918] at org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.executeProduceConsume(ExecuteProduceConsume.java:303) [165:org.eclipse.jetty.util:9.3.21.v20170918] at org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.produceConsume(ExecuteProduceConsume.java:148) [165:org.eclipse.jetty.util:9.3.21.v20170918] at org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.run(ExecuteProduceConsume.java:136) [165:org.eclipse.jetty.util:9.3.21.v20170918] at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:671) [165:org.eclipse.jetty.util:9.3.21.v20170918] at org.eclipse.jetty.util.thread.QueuedThreadPool$2.run(QueuedThreadPool.java:589) [165:org.eclipse.jetty.util:9.3.21.v20170918] at java.lang.Thread.run(Thread.java:748) [?:?] Caused by: javax.management.InstanceNotFoundException: org.opendaylight.controller:type=DistributedConfigDatastore,Category=Shards,name=member-0-shard-default-config at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.getMBean(DefaultMBeanServerInterceptor.java:1095) ~[?:?] at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.getAttribute(DefaultMBeanServerInterceptor.java:643) ~[?:?] at com.sun.jmx.mbeanserver.JmxMBeanServer.getAttribute(JmxMBeanServer.java:678) ~[?:?] at com.sun.jmx.mbeanserver.MXBeanProxy$GetHandler.invoke(MXBeanProxy.java:122) ~[?:?] at com.sun.jmx.mbeanserver.MXBeanProxy.invoke(MXBeanProxy.java:167) ~[?:?] at javax.management.MBeanServerInvocationHandler.invoke(MBeanServerInvocationHandler.java:258) ~[?:?] ... 41 more 2018-11-18T15:18:36,700 | INFO | nioEventLoopGroup-9-9 | AbstractConnectionAdapter | 387 - org.opendaylight.openflowplugin.openflowjava.openflow-protocol-impl - 0.6.4.redhat-6 | The channel outbound queue size:1024 2018-11-18T15:18:36,700 | INFO | nioEventLoopGroup-11-4 | AbstractConnectionAdapter | 387 - org.opendaylight.openflowplugin.openflowjava.openflow-protocol-impl - 0.6.4.redhat-6 | The channel outbound queue size:1024 2018-11-18T15:18:36,701 | ERROR | qtp626219007-21545 | DatastoreServiceStatusProvider | 262 - org.opendaylight.genius.mdsalutil-impl - 0.4.4.redhat-6 | Unable to obtain the datastore status java.lang.reflect.UndeclaredThrowableException: null at com.sun.proxy.$Proxy110.getRaftState(Unknown Source) [231:org.opendaylight.controller.sal-distributed-datastore:1.7.4.redhat-6] at org.opendaylight.genius.mdsalutil.diagstatus.internal.DatastoreServiceStatusProvider.getServiceDescriptor(DatastoreServiceStatusProvider.java:89) [262:org.opendaylight.genius.mdsalutil-impl:0.4.4.redhat-6] at Proxy241f9f36_ab17_4d9d_9a59_3fad77bf0588.getServiceDescriptor(Unknown Source) [?:?] at Proxyd6f0d70e_6629_4345_b144_55579e64838d.getServiceDescriptor(Unknown Source) [?:?] at org.opendaylight.infrautils.diagstatus.internal.DiagStatusServiceImpl.updateServiceStatusMap(DiagStatusServiceImpl.java:137) [272:org.opendaylight.infrautils.diagstatus-impl:1.3.4.redhat-6] at org.opendaylight.infrautils.diagstatus.internal.DiagStatusServiceImpl.getAllServiceDescriptors(DiagStatusServiceImpl.java:86) [272:org.opendaylight.infrautils.diagstatus-impl:1.3.4.redhat-6] at org.opendaylight.infrautils.diagstatus.internal.DiagStatusServiceImpl.isOperational(DiagStatusServiceImpl.java:95) [272:org.opendaylight.infrautils.diagstatus-impl:1.3.4.redhat-6] at org.opendaylight.infrautils.diagstatus.internal.DiagStatusServiceImpl.getAllServiceDescriptorsAsJSON(DiagStatusServiceImpl.java:110) [272:org.opendaylight.infrautils.diagstatus-impl:1.3.4.redhat-6] at Proxyea07b505_74b4_459e_95c3_86daec3e8e0b.getAllServiceDescriptorsAsJSON(Unknown Source) [?:?] at Proxy108b1804_c30c_4378_a025_b6f5674a1215.getAllServiceDescriptorsAsJSON(Unknown Source) [?:?] at org.opendaylight.infrautils.diagstatus.web.DiagStatusServlet.doGet(DiagStatusServlet.java:52) [274:org.opendaylight.infrautils.diagstatus-web:1.3.4.redhat-6] at javax.servlet.http.HttpServlet.service(HttpServlet.java:687) [72:javax.servlet-api:3.1.0] at javax.servlet.http.HttpServlet.service(HttpServlet.java:790) [72:javax.servlet-api:3.1.0] at org.eclipse.jetty.servlet.ServletHolder.handle(ServletHolder.java:848) [163:org.eclipse.jetty.servlet:9.3.21.v20170918] at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1772) [163:org.eclipse.jetty.servlet:9.3.21.v20170918] at org.eclipse.jetty.websocket.server.WebSocketUpgradeFilter.doFilter(WebSocketUpgradeFilter.java:205) [173:org.eclipse.jetty.websocket.server:9.3.21.v20170918] at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1759) [163:org.eclipse.jetty.servlet:9.3.21.v20170918] at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:582) [163:org.eclipse.jetty.servlet:9.3.21.v20170918] at org.ops4j.pax.web.service.jetty.internal.HttpServiceServletHandler.doHandle(HttpServiceServletHandler.java:71) [441:org.ops4j.pax.web.pax-web-jetty:6.0.9] at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:143) [162:org.eclipse.jetty.server:9.3.21.v20170918] at org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:548) [160:org.eclipse.jetty.security:9.3.21.v20170918] at org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:226) [162:org.eclipse.jetty.server:9.3.21.v20170918] at org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1180) [162:org.eclipse.jetty.server:9.3.21.v20170918] at org.ops4j.pax.web.service.jetty.internal.HttpServiceContext.doHandle(HttpServiceContext.java:284) [441:org.ops4j.pax.web.pax-web-jetty:6.0.9] at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:512) [163:org.eclipse.jetty.servlet:9.3.21.v20170918] at org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:185) [162:org.eclipse.jetty.server:9.3.21.v20170918] at org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1112) [162:org.eclipse.jetty.server:9.3.21.v20170918] at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141) [162:org.eclipse.jetty.server:9.3.21.v20170918] at org.ops4j.pax.web.service.jetty.internal.JettyServerHandlerCollection.handle(JettyServerHandlerCollection.java:80) [441:org.ops4j.pax.web.pax-web-jetty:6.0.9] at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:134) [162:org.eclipse.jetty.server:9.3.21.v20170918] at org.eclipse.jetty.server.Server.handle(Server.java:534) [162:org.eclipse.jetty.server:9.3.21.v20170918] at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:333) [162:org.eclipse.jetty.server:9.3.21.v20170918] at org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:251) [162:org.eclipse.jetty.server:9.3.21.v20170918] at org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:283) [154:org.eclipse.jetty.io:9.3.21.v20170918] at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:108) [154:org.eclipse.jetty.io:9.3.21.v20170918] at org.eclipse.jetty.io.SelectChannelEndPoint$2.run(SelectChannelEndPoint.java:93) [154:org.eclipse.jetty.io:9.3.21.v20170918] at org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.executeProduceConsume(ExecuteProduceConsume.java:303) [165:org.eclipse.jetty.util:9.3.21.v20170918] at org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.produceConsume(ExecuteProduceConsume.java:148) [165:org.eclipse.jetty.util:9.3.21.v20170918] at org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.run(ExecuteProduceConsume.java:136) [165:org.eclipse.jetty.util:9.3.21.v20170918] at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:671) [165:org.eclipse.jetty.util:9.3.21.v20170918] at org.eclipse.jetty.util.thread.QueuedThreadPool$2.run(QueuedThreadPool.java:589) [165:org.eclipse.jetty.util:9.3.21.v20170918] at java.lang.Thread.run(Thread.java:748) [?:?] Caused by: javax.management.InstanceNotFoundException: org.opendaylight.controller:type=DistributedConfigDatastore,Category=Shards,name=member-0-shard-default-config at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.getMBean(DefaultMBeanServerInterceptor.java:1095) ~[?:?] at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.getAttribute(DefaultMBeanServerInterceptor.java:643) ~[?:?] at com.sun.jmx.mbeanserver.JmxMBeanServer.getAttribute(JmxMBeanServer.java:678) ~[?:?] at com.sun.jmx.mbeanserver.MXBeanProxy$GetHandler.invoke(MXBeanProxy.java:122) ~[?:?] at com.sun.jmx.mbeanserver.MXBeanProxy.invoke(MXBeanProxy.java:167) ~[?:?] at javax.management.MBeanServerInvocationHandler.invoke(MBeanServerInvocationHandler.java:258) ~[?:?] ... 42 more The bug was still occurring even when overcloud nodes got the new rpm: [heat-admin@controller-0 ~]$ rpm -qa | grep openstack-tripleo-heat-templates openstack-tripleo-heat-templates-9.0.1-0.20181013060884.el7ost.noarch Apparently, we need to apply this rpm into undercloud, and not into overcloud. Undercloud was still using old rpm: (undercloud) [stack@undercloud-0 ~]$ rpm -qa | grep trip openstack-tripleo-heat-templates openstack-tripleo-heat-templates-9.0.1-0.20181013060879.el7ost.noarch I'm redeploying OSP with undercloud-packages option added to OVERCLOUD_DEPLOY_OVERRIDE_OPTIONS: --undercloud-packages http://brew.redhat.com/official/openstack-tripleo-heat-templates/9.0.1/0.20181013060884.el7ost/noarch/openstack-tripleo-heat-templates-9.0.1-0.20181013060884.el7ost.noarch.rpm Deploying on OSP14 2018-11-07.3 + openstack-tripleo-heat-templates-9.0.1-0.20181013060879.el7ost.noarch
Passed with following results. Notice:
1) The log dir of Karaf is directly under /var/log/containers/opendaylight.
2) There is no issue of NULLs anymore, but there is perhaps other issue in neutron log: Failed during periodic task operation journal_recovery.
---
(overcloud) [stack@undercloud-0 ~]$ rpm -qa | grep openstack-tripleo-heat-templates
openstack-tripleo-heat-templates-9.0.1-0.20181013060879.el7ost.noarch
(undercloud) [stack@undercloud-0 ~]$ openstack server list --all
+--------------------------------------+-------------------+--------+--------------------------------------------------------------+----------+---------------+
| ID | Name | Status | Networks | Image | Flavor |
+--------------------------------------+-------------------+--------+--------------------------------------------------------------+----------+---------------+
| 5ecae502-3861-4aec-8ab8-0a3959ab3f50 | cirros35_vm2_net2 | ACTIVE | net_ipv64_2=2002::f816:3eff:fe81:a9f2, 10.0.2.22, 10.0.0.211 | cirros35 | cirros_flavor |
| e53bc071-ffb9-4aae-97cf-1efe0c386362 | cirros35_vm1_net2 | ACTIVE | net_ipv64_2=2002::f816:3eff:fe47:f78c, 10.0.2.24, 10.0.0.220 | cirros35 | cirros_flavor |
| 2d6f2598-9266-4ab4-ae8b-5fabcc69e63f | cirros35_vm2_net1 | ACTIVE | net_ipv64_1=10.0.1.10, 2001::f816:3eff:fec4:2570, 10.0.0.226 | cirros35 | cirros_flavor |
| f57bf00e-ded7-483b-8e89-ba725d717d22 | cirros35_vm1_net1 | ACTIVE | net_ipv64_1=10.0.1.5, 2001::f816:3eff:fee1:ee13, 10.0.0.213 | cirros35 | cirros_flavor |
+--------------------------------------+-------------------+--------+--------------------------------------------------------------+----------+---------------+
###### controller-1: ssh heat-admin.24.14 ######
Filesystem Size Used Avail Use% Mounted on
/dev/vda2 40G 18G 23G 43% /
devtmpfs 16G 0 16G 0% /dev
tmpfs 16G 54M 16G 1% /dev/shm
tmpfs 16G 6.9M 16G 1% /run
tmpfs 16G 0 16G 0% /sys/fs/cgroup
tmpfs 3.2G 0 3.2G 0% /run/user/1000
(undercloud) [stack@undercloud-0 ~]$ ssh heat-admin.24.14 "ls -lh /var/log/containers/opendaylight/"
total 4.7G
-rw-r--r--. 1 42462 42462 4.7G Nov 21 09:18 karaf.log
-rw-r--r--. 1 42462 42462 4.6G Nov 21 09:01 karaf.log.1
-rw-r--r--. 1 42462 42462 5.9M Nov 21 00:01 karaf.log.10.gz
-rw-r--r--. 1 42462 42462 5.8M Nov 20 23:01 karaf.log.11.gz
-rw-r--r--. 1 42462 42462 5.8M Nov 20 22:01 karaf.log.12.gz
-rw-r--r--. 1 42462 42462 5.6M Nov 20 21:00 karaf.log.13.gz
-rw-r--r--. 1 42462 42462 5.5M Nov 20 20:00 karaf.log.14.gz
-rw-r--r--. 1 42462 42462 6.8M Nov 21 08:02 karaf.log.2.gz
-rw-r--r--. 1 42462 42462 6.6M Nov 21 07:01 karaf.log.3.gz
-rw-r--r--. 1 42462 42462 6.5M Nov 21 06:01 karaf.log.4.gz
-rw-r--r--. 1 42462 42462 6.4M Nov 21 05:02 karaf.log.5.gz
-rw-r--r--. 1 42462 42462 6.4M Nov 21 04:02 karaf.log.6.gz
-rw-r--r--. 1 42462 42462 6.2M Nov 21 03:01 karaf.log.7.gz
-rw-r--r--. 1 42462 42462 6.1M Nov 21 02:02 karaf.log.8.gz
-rw-r--r--. 1 42462 42462 6.0M Nov 21 01:01 karaf.log.9.gz
*** /var/log/containers/neutron/server.log ***
2018-11-21 09:11:09.718 37 INFO networking_odl.journal.periodic_task [req-32bbc692-6b7d-4c29-8419-ef73aea0c343 - - - - -] Finished full_sync phase of maintenance task.
2018-11-21 09:11:09.718 37 INFO networking_odl.journal.periodic_task [req-32bbc692-6b7d-4c29-8419-ef73aea0c343 - - - - -] Starting journal_recovery phase of periodic task maintenance.
2018-11-21 09:11:09.746 37 DEBUG networking_odl.journal.recovery [req-32bbc692-6b7d-4c29-8419-ef73aea0c343 - - - - -] Attempting recovery of journal entry <networking_odl.db.models.OpenDaylightJournal[object at 7fd8308f9050] {seqnum=602, object_type=u'network', object_uuid=u'cd5cb086-22e4-434d-8eef-de38adf9e183', operation=u'update', data={'provider:physical_network': None, 'ipv6_address_scope': None, 'revision_number': 4, 'port_security_enabled': True, 'mtu': 1450, 'id': u'cd5cb086-22e4-434d-8eef-de38adf9e183', 'router:external': False, 'availability_zone_hints': [], 'availability_zones': [u'nova'], 'ipv4_address_scope': None, 'shared': False, 'project_id': u'3873ba1c11a2484bbe6b5cf2c94a742f', 'status': u'ACTIVE', 'subnets': [], 'description': u'', 'tags': [], 'updated_at': '2018-11-19T12:50:35Z', 'provider:segmentation_id': None, 'name': u'tempest-network-smoke--241992513', 'admin_state_up': True, 'tenant_id': u'3873ba1c11a2484bbe6b5cf2c94a742f', 'created_at': '2018-11-19T12:47:35Z', 'provider:network_type': None, 'vlan_transparent': None}, state='failed', retry_count=5, last_retried=datetime.datetime(2018, 11, 19, 12, 50, 36), version_id=13}>. journal_recovery /usr/lib/python2.7/site-packages/networking_odl/journal/recovery.py:39
2018-11-21 09:11:09.747 37 DEBUG networking_odl.common.client [req-32bbc692-6b7d-4c29-8419-ef73aea0c343 - - - - -] Sending METHOD (get) URL (http://172.17.1.20:8081/controller/nb/v2/neutron/networks/cd5cb086-22e4-434d-8eef-de38adf9e183) JSON (None) request /usr/lib/python2.7/site-packages/networking_odl/common/client.py:89
2018-11-21 09:11:09.786 37 ERROR networking_odl.journal.periodic_task [req-32bbc692-6b7d-4c29-8419-ef73aea0c343 - - - - -] Failed during periodic task operation journal_recovery.: InvalidRequestError: Object '<OpenDaylightJournal at 0x7fd8308f9050>' is already attached to session '35892' (this is '35901')
2018-11-21 09:11:09.786 37 ERROR networking_odl.journal.periodic_task Traceback (most recent call last):
2018-11-21 09:11:09.786 37 ERROR networking_odl.journal.periodic_task File "/usr/lib/python2.7/site-packages/networking_odl/journal/periodic_task.py", line 65, in _execute_op
2018-11-21 09:11:09.786 37 ERROR networking_odl.journal.periodic_task operation(context)
2018-11-21 09:11:09.786 37 ERROR networking_odl.journal.periodic_task File "/usr/lib/python2.7/site-packages/neutron_lib/db/api.py", line 233, in wrapped
2018-11-21 09:11:09.786 37 ERROR networking_odl.journal.periodic_task return method(*args, **kwargs)
2018-11-21 09:11:09.786 37 ERROR networking_odl.journal.periodic_task File "/usr/lib/python2.7/site-packages/neutron_lib/db/api.py", line 140, in wrapped
2018-11-21 09:11:09.786 37 ERROR networking_odl.journal.periodic_task setattr(e, '_RETRY_EXCEEDED', True)
2018-11-21 09:11:09.786 37 ERROR networking_odl.journal.periodic_task File "/usr/lib/python2.7/site-packages/oslo_utils/excutils.py", line 220, in __exit__
2018-11-21 09:11:09.786 37 ERROR networking_odl.journal.periodic_task self.force_reraise()
2018-11-21 09:11:09.786 37 ERROR networking_odl.journal.periodic_task File "/usr/lib/python2.7/site-packages/oslo_utils/excutils.py", line 196, in force_reraise
2018-11-21 09:11:09.786 37 ERROR networking_odl.journal.periodic_task six.reraise(self.type_, self.value, self.tb)
2018-11-21 09:11:09.786 37 ERROR networking_odl.journal.periodic_task File "/usr/lib/python2.7/site-packages/neutron_lib/db/api.py", line 136, in wrapped
2018-11-21 09:11:09.786 37 ERROR networking_odl.journal.periodic_task return f(*args, **kwargs)
2018-11-21 09:11:09.786 37 ERROR networking_odl.journal.periodic_task File "/usr/lib/python2.7/site-packages/oslo_db/api.py", line 154, in wrapper
2018-11-21 09:11:09.786 37 ERROR networking_odl.journal.periodic_task ectxt.value = e.inner_exc
2018-11-21 09:11:09.786 37 ERROR networking_odl.journal.periodic_task File "/usr/lib/python2.7/site-packages/oslo_utils/excutils.py", line 220, in __exit__
2018-11-21 09:11:09.786 37 ERROR networking_odl.journal.periodic_task self.force_reraise()
2018-11-21 09:11:09.786 37 ERROR networking_odl.journal.periodic_task File "/usr/lib/python2.7/site-packages/oslo_utils/excutils.py", line 196, in force_reraise
2018-11-21 09:11:09.786 37 ERROR networking_odl.journal.periodic_task six.reraise(self.type_, self.value, self.tb)
2018-11-21 09:11:09.786 37 ERROR networking_odl.journal.periodic_task File "/usr/lib/python2.7/site-packages/oslo_db/api.py", line 142, in wrapper
2018-11-21 09:11:09.786 37 ERROR networking_odl.journal.periodic_task return f(*args, **kwargs)
Log rotation is fixed on OSP14 Puddle 2018-11-21.2: ###### controller-0: ssh heat-admin.24.14 ###### /var/log/containers/opendaylight/karaf/logs/karaf.log ###### controller-1: ssh heat-admin.24.7 ###### /var/log/containers/opendaylight/karaf/logs/karaf.log ###### controller-2: ssh heat-admin.24.8 ###### /var/log/containers/opendaylight/karaf/logs/karaf.log Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHEA-2019:0045 |