Bug 2131836
Summary: | PANIC podman API service endpoint handler panic | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|
Product: | Red Hat Enterprise Linux 8 | Reporter: | navabharathi.gorantl | ||||||||
Component: | podman | Assignee: | Jindrich Novy <jnovy> | ||||||||
Status: | CLOSED ERRATA | QA Contact: | Edward Shen <weshen> | ||||||||
Severity: | high | Docs Contact: | |||||||||
Priority: | unspecified | ||||||||||
Version: | 8.6 | CC: | bbaude, dornelas, dwalsh, jligon, jnovy, lsm5, mamccoma, mangesh.panche, mheon, pthomas, toneata, tsweeney, umohnani, vikas.goel, ypu | ||||||||
Target Milestone: | rc | Keywords: | Triaged, ZStream | ||||||||
Target Release: | --- | Flags: | pm-rhel:
mirror+
|
||||||||
Hardware: | Unspecified | ||||||||||
OS: | Unspecified | ||||||||||
Whiteboard: | |||||||||||
Fixed In Version: | podman-4.3.1-1.el8 | Doc Type: | If docs needed, set a value | ||||||||
Doc Text: | Story Points: | --- | |||||||||
Clone Of: | |||||||||||
: | 2132412 2132413 (view as bug list) | Environment: | |||||||||
Last Closed: | 2023-05-16 08:22:22 UTC | Type: | Bug | ||||||||
Regression: | --- | Mount Type: | --- | ||||||||
Documentation: | --- | CRM: | |||||||||
Verified Versions: | Category: | --- | |||||||||
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |||||||||
Cloudforms Team: | --- | Target Upstream Version: | |||||||||
Embargoed: | |||||||||||
Bug Depends On: | |||||||||||
Bug Blocks: | 2132412, 2132413, 2136287 | ||||||||||
Attachments: |
|
Can I ask what is bringing up the containers - is it the app talking to the Podman API? Podman itself does not restart containers after a reboot, so I'm curious. Regardless, the panic is definitely an issue, and ought to be an easy/fast fix. please attach your compose file in question. This bugzilla is actually two problems. The first problem is a problem with an image layer after reboots, is being actively worked on elsewhere. This needs to be confirmed. The second problem is that when a remote compat inspect is being done with an image that has a problem, the error reporting attempts to access nil to get the image id. Fix for the panic upstream -> https://github.com/containers/podman/pull/16050 The problem exists in both 4.1 and 4.2 Lastly, as Matt says, there is no mechanism for having containers come up from reboot unless systemd is being used. Created attachment 1916060 [details]
compose yml
We use VCS (Veritas cluster server) which brings up the container after system start. Attached compose file Can you suggest a workaround for now to get the system to working state? It's hard to say, since the panic is swallowing what I suspect to be the actual error. Is this happening on more than one system? Does it reproduce consistently? If it is only one system: We recommend saving any data on the system, then doing a `podman system reset` to remove all images. It seems like the storage library was interrupted midway through writing an image by power loss, resulting in the image being corrupted. This is a system test machine and have not heard this happening on any other system yet. Cannot use "podman system reset" as it resets all configurations system test is able to reproduce the issue again. Attached the latest logs Created attachment 1918107 [details]
/var/log/messages
Brent, please see previous comment. Also, could not run "podman system reset". Any other way to get podman running again? Error: error opening database /var/lib/containers/storage/libpod/bolt_state.db: open /var/lib/containers/storage/libpod/bolt_state.db: input/output error The panic here is different than the previous one. I spent a little time looking at it this morning,. I think this is wholly related to the DB being corrupted; and the calls to Info are part of the API so to speak. @mheon do you agree? And if so, should it be deferred to a new bug? It's definitely different, but this does not strike me as the underlying cause. It looks like something is causing Podman to try and lock a null c/storage lock; none of those locks should ever be nil, so there's something going wrong higher up the chain than the panic. For a manual removal / reset of the system, you can remove `/var/lib/containers/storage` and `/var/run/libpod`. This should get things stabilized sufficiently to run `podman system reset`. Tried manually deleting '/var/lib/containers/storage' but could not due to device/resource busy [root@eagappflx040 containers]# rm -rf storage/ rm: cannot remove 'storage/overlay/dbeba618543dbe64859a7bddf4c14757a9d80bb0268493be760133765235bb6d/merged': Device or resource busy rm: cannot remove 'storage/overlay/96f544efc96a41b8e13f4ea868ee42ef3b57c1c2ef681892aece4349119d52e6/merged': Device or resource busy rm: cannot remove 'storage/overlay/8ec99d6acf1d1cac52a3800f32dc1e4eea7f4963694507b99f11c53f435aef07/merged': Device or resource busy rm: cannot remove 'storage/overlay/fb15fc9745b6b8d801cf55756039b10b3419137eda983bd8d3373dfc97d8b096/merged': Device or resource busy rm: cannot remove 'storage/overlay/81fc77b8480a3b4082bd6741a9541382f753b51c0d1b7d9235a19ac77b2ea45a/merged': Device or resource busy rm: cannot remove 'storage/overlay/1d8807fcb9bdcdc9fce76996ddca9e7e23d7f7a63752c85f3fc2f0c7a20f0f67/merged': Device or resource busy rm: cannot remove 'storage/overlay/2771138a08b8667213edb044b588704e481fa5457dfae6068264837d76bf4ab3/merged': Device or resource busy rm: cannot remove 'storage/overlay/ae13097ea3e3058821985e23e0e6945edcbe25d58ccd541d0b49f38280e3e9cd/merged': Device or resource busy rm: cannot remove 'storage/overlay-containers/bcbb498d5de15870e5d00097058008b549670c490a107fe2213a638a9558d506/userdata/shm': Device or resource busy rm: cannot remove 'storage/overlay-containers/3ff041442067b22138bb93b7ea4130484b520bc5eae65c9fc3e3801165756679/userdata/shm': Device or resource busy rm: cannot remove 'storage/overlay-containers/0c7c7faa696e69159ceed1b2649f5a00536b10e6aac1a0bdbe90f71398ffa0fa/userdata/shm': Device or resource busy rm: cannot remove 'storage/overlay-containers/bbae8c62c21e4317d1a0d1af4db9ce5fe0d39269282195f4fb18df67b0c8fae3/userdata/shm': Device or resource busy rm: cannot remove 'storage/overlay-containers/852db27b6583b1987d68f232c28969d065a73e7527172bf1a122038228337ea4/userdata/shm': Device or resource busy rm: cannot remove 'storage/overlay-containers/80c7e2b97bdbd9a22ef118d0c29aafa4a1a23b2a2d942e8834dd5969cdee6d95/userdata/shm': Device or resource busy rm: cannot remove 'storage/overlay-containers/5849a9d2f66f96f3e120c549091ad246c91befd4951144a06539c722c355bbe8/userdata/shm': Device or resource busy rm: cannot remove 'storage/overlay-containers/2e3dcd37487aef2f20caa3c296c67bb4f4d404a6624c01ec380bfa2ad1ebb3e3/userdata/shm': Device or resource busy Looks like you have a lot of running containers or at least leaked overlay mount points. mount | grep overlay-container And then unmount those mount points. Today we ran into another set of podman panic issues during power cycle on multiple setups. In all cases, it ran into issues while running docker network ls | grep macvlan | awk '{print $2}' On two setups we see the similar error: <30>1 2022-11-22T04:04:40.309+00:00 nbapp837 infra-network-control.sh 11201 - - panic: invalid freelist page: 5918356779941454958, page type is branch <30>1 2022-11-22T04:04:40.309+00:00 nbapp837 infra-network-control.sh 11201 - - goroutine 1 [running]: <30>1 2022-11-22T04:04:40.338+00:00 nbapp837 infra-network-control.sh 11201 - - panic({0x55a4df2c6460, 0xc0003c8110}) <30>1 2022-11-22T04:04:40.338+00:00 nbapp837 infra-network-control.sh 11201 - - #011/usr/lib/golang/src/runtime/panic.go:1147 +0x3a8 fp=0xc0004cfb28 sp=0xc0004cfa68 pc=0x55a4dde990e8 <30>1 2022-11-22T04:04:40.347+00:00 nbapp837 infra-network-control.sh 11201 - - github.com/containers/podman/vendor/go.etcd.io/bbolt.(*freelist).read(0x55a4ded74981, 0x7f01e453c000) <30>1 2022-11-22T04:04:40.347+00:00 nbapp837 infra-network-control.sh 11201 - - #011/builddir/build/BUILD/containers-podman-7f5e2fd/_build/src/github.com/containers/podman/vendor/go.etcd.io/bbolt/freelist.go:266 +0x234 fp=0xc0004cfbc8 sp=0xc0004cfb28 pc=0x55a4de634e94 <30>1 2022-11-22T04:04:40.347+00:00 nbapp837 infra-network-control.sh 11201 - - github.com/containers/podman/vendor/go.etcd.io/bbolt.(*DB).loadFreelist.func1() <30>1 2022-11-22T04:04:40.348+00:00 nbapp837 infra-network-control.sh 11201 - - #011/builddir/build/BUILD/containers-podman-7f5e2fd/_build/src/github.com/containers/podman/vendor/go.etcd.io/bbolt/db.go:323 +0xae fp=0xc0004cfbf8 sp=0xc0004cfbc8 pc=0x55a4de62fe0e <30>1 2022-11-22T04:04:40.348+00:00 nbapp837 infra-network-control.sh 11201 - - sync.(*Once).doSlow(0xc000d261c8, 0x10) <30>1 2022-11-22T04:04:40.348+00:00 nbapp837 infra-network-control.sh 11201 - - #011/usr/lib/golang/src/sync/once.go:68 +0xd2 fp=0xc0004cfc60 sp=0xc0004cfbf8 pc=0x55a4dded76f2 <30>1 2022-11-22T04:04:40.348+00:00 nbapp837 infra-network-control.sh 11201 - - sync.(*Once).Do(...) <30>1 2022-11-22T04:04:40.348+00:00 nbapp837 infra-network-control.sh 11201 - - #011/usr/lib/golang/src/sync/once.go:59 <30>1 2022-11-22T04:04:40.348+00:00 nbapp837 infra-network-control.sh 11201 - - github.com/containers/podman/vendor/go.etcd.io/bbolt.(*DB).loadFreelist(0xc000d26000) <30>1 2022-11-22T04:04:40.348+00:00 nbapp837 infra-network-control.sh 11201 - - #011/builddir/build/BUILD/containers-podman-7f5e2fd/_build/src/github.com/containers/podman/vendor/go.etcd.io/bbolt/db.go:316 +0x47 fp=0xc0004cfc90 sp=0xc0004cfc60 pc=0x55a4de62fd27 <30>1 2022-11-22T04:04:40.348+00:00 nbapp837 infra-network-control.sh 11201 - - github.com/containers/podman/vendor/go.etcd.io/bbolt.Open({0xc0002000f0, 0x30}, 0xdeda7550, 0x0) <30>1 2022-11-22T04:04:40.348+00:00 nbapp837 infra-network-control.sh 11201 - - #011/builddir/build/BUILD/containers-podman-7f5e2fd/_build/src/github.com/containers/podman/vendor/go.etcd.io/bbolt/db.go:293 +0x46b fp=0xc0004d0d68 sp=0xc0004cfc90 pc=0x55a4de62fa8b <30>1 2022-11-22T04:04:40.365+00:00 nbapp837 infra-network-control.sh 11201 - - github.com/containers/podman/libpod.NewBoltState({0xc0002000f0, 0x30}, 0xc000c99380) <30>1 2022-11-22T04:04:40.365+00:00 nbapp837 infra-network-control.sh 11201 - - #011/builddir/build/BUILD/containers-podman-7f5e2fd/_build/src/github.com/containers/podman/libpod/boltdb_state.go:77 +0x152 fp=0xc0004d0f80 sp=0xc0004d0d68 pc=0x55a4dea61b92 <30>1 2022-11-22T04:04:40.365+00:00 nbapp837 infra-network-control.sh 11201 - - github.com/containers/podman/libpod.makeRuntime(0xc000c99380) <30>1 2022-11-22T04:04:40.365+00:00 nbapp837 infra-network-control.sh 11201 - - #011/builddir/build/BUILD/containers-podman-7f5e2fd/_build/src/github.com/containers/podman/libpod/runtime.go:325 +0x189 fp=0xc0004d13b0 sp=0xc0004d0f80 pc=0x55a4deb28989 <30>1 2022-11-22T04:04:40.365+00:00 nbapp837 infra-network-control.sh 11201 - - github.com/containers/podman/libpod.newRuntimeFromConfig(0xc000cc7c00, {0xc0004d17b0, 0x0, 0x55a4df5115b0}) <30>1 2022-11-22T04:04:40.365+00:00 nbapp837 infra-network-control.sh 11201 - - #011/builddir/build/BUILD/containers-podman-7f5e2fd/_build/src/github.com/containers/podman/libpod/runtime.go:227 +0x3d7 fp=0xc0004d16d0 sp=0xc0004d13b0 pc=0x55a4deb281d7 <30>1 2022-11-22T04:04:40.365+00:00 nbapp837 infra-network-control.sh 11201 - - github.com/containers/podman/libpod.NewRuntime({0xc000c7ca00, 0x55a4ded770cb}, {0xc0004d17b0, 0x0, 0x0}) <30>1 2022-11-22T04:04:40.365+00:00 nbapp837 infra-network-control.sh 11201 - - #011/builddir/build/BUILD/containers-podman-7f5e2fd/_build/src/github.com/containers/podman/libpod/runtime.go:170 +0x51 fp=0xc0004d1700 sp=0xc0004d16d0 pc=0x55a4deb27db1 <30>1 2022-11-22T04:04:40.366+00:00 nbapp837 infra-network-control.sh 11201 - - github.com/containers/podman/pkg/domain/infra.getRuntime({0x55a4df560498, 0xc000040048}, 0xc000c7ca00, 0xc0004d1a80) <30>1 2022-11-22T04:04:40.366+00:00 nbapp837 infra-network-control.sh 11201 - - #011/builddir/build/BUILD/containers-podman-7f5e2fd/_build/src/github.com/containers/podman/pkg/domain/infra/runtime_libpod.go:254 +0x19b7 fp=0xc0004d1a60 sp=0xc0004d1700 pc=0x55a4dec28437 <30>1 2022-11-22T04:04:40.366+00:00 nbapp837 infra-network-control.sh 11201 - - github.com/containers/podman/pkg/domain/infra.GetRuntime.func1() <30>1 2022-11-22T04:04:40.366+00:00 nbapp837 infra-network-control.sh 11201 - - #011/builddir/build/BUILD/containers-podman-7f5e2fd/_build/src/github.com/containers/podman/pkg/domain/infra/runtime_libpod.go:80 +0x49 fp=0xc0004d1ab0 sp=0xc0004d1a60 pc=0x55a4dec26a09 <30>1 2022-11-22T20:10:23.219+00:00 nbapp817 infra-network-control.sh 192536 - - panic: invalid freelist page: 66, page type is leaf <30>1 2022-11-22T20:10:23.219+00:00 nbapp817 infra-network-control.sh 192536 - - goroutine 1 [running]: <30>1 2022-11-22T20:10:23.220+00:00 nbapp817 infra-network-control.sh 192536 - - panic({0x55d5c5777460, 0xc00026e920}) <30>1 2022-11-22T20:10:23.220+00:00 nbapp817 infra-network-control.sh 192536 - - #011/usr/lib/golang/src/runtime/panic.go:1147 +0x3a8 fp=0xc000a05b28 sp=0xc000a05a68 pc=0x55d5c434a0e8 <30>1 2022-11-22T20:10:23.220+00:00 nbapp817 infra-network-control.sh 192536 - - github.com/containers/podman/vendor/go.etcd.io/bbolt.(*freelist).read(0x55d5c5225981, 0x7f7f34263000) <30>1 2022-11-22T20:10:23.220+00:00 nbapp817 infra-network-control.sh 192536 - - #011/builddir/build/BUILD/containers-podman-7f5e2fd/_build/src/github.com/containers/podman/vendor/go.etcd.io/bbolt/freelist.go:266 +0x234 fp=0xc000a05bc8 sp=0xc000a05b28 pc=0x55d5c4ae5e94 <30>1 2022-11-22T20:10:23.220+00:00 nbapp817 infra-network-control.sh 192536 - - github.com/containers/podman/vendor/go.etcd.io/bbolt.(*DB).loadFreelist.func1() <30>1 2022-11-22T20:10:23.220+00:00 nbapp817 infra-network-control.sh 192536 - - #011/builddir/build/BUILD/containers-podman-7f5e2fd/_build/src/github.com/containers/podman/vendor/go.etcd.io/bbolt/db.go:323 +0xae fp=0xc000a05bf8 sp=0xc000a05bc8 pc=0x55d5c4ae0e0e <30>1 2022-11-22T20:10:23.220+00:00 nbapp817 infra-network-control.sh 192536 - - sync.(*Once).doSlow(0xc0001d7608, 0x10) <30>1 2022-11-22T20:10:23.220+00:00 nbapp817 infra-network-control.sh 192536 - - #011/usr/lib/golang/src/sync/once.go:68 +0xd2 fp=0xc000a05c60 sp=0xc000a05bf8 pc=0x55d5c43886f2 <30>1 2022-11-22T20:10:23.220+00:00 nbapp817 infra-network-control.sh 192536 - - sync.(*Once).Do(...) <30>1 2022-11-22T20:10:23.220+00:00 nbapp817 infra-network-control.sh 192536 - - #011/usr/lib/golang/src/sync/once.go:59 <30>1 2022-11-22T20:10:23.220+00:00 nbapp817 infra-network-control.sh 192536 - - github.com/containers/podman/vendor/go.etcd.io/bbolt.(*DB).loadFreelist(0xc0001d7440) <30>1 2022-11-22T20:10:23.220+00:00 nbapp817 infra-network-control.sh 192536 - - #011/builddir/build/BUILD/containers-podman-7f5e2fd/_build/src/github.com/containers/podman/vendor/go.etcd.io/bbolt/db.go:316 +0x47 fp=0xc000a05c90 sp=0xc000a05c60 pc=0x55d5c4ae0d27 <30>1 2022-11-22T20:10:23.220+00:00 nbapp817 infra-network-control.sh 192536 - - github.com/containers/podman/vendor/go.etcd.io/bbolt.Open({0xc0001c7650, 0x30}, 0xc5258550, 0x0) <30>1 2022-11-22T20:10:23.220+00:00 nbapp817 infra-network-control.sh 192536 - - #011/builddir/build/BUILD/containers-podman-7f5e2fd/_build/src/github.com/containers/podman/vendor/go.etcd.io/bbolt/db.go:293 +0x46b fp=0xc000a06d68 sp=0xc000a05c90 pc=0x55d5c4ae0a8b <30>1 2022-11-22T20:10:23.220+00:00 nbapp817 infra-network-control.sh 192536 - - github.com/containers/podman/libpod.NewBoltState({0xc0001c7650, 0x30}, 0xc000f696c0) <30>1 2022-11-22T20:10:23.220+00:00 nbapp817 infra-network-control.sh 192536 - - #011/builddir/build/BUILD/containers-podman-7f5e2fd/_build/src/github.com/containers/podman/libpod/boltdb_state.go:77 +0x152 fp=0xc000a06f80 sp=0xc000a06d68 pc=0x55d5c4f12b92 <30>1 2022-11-22T20:10:23.221+00:00 nbapp817 infra-network-control.sh 192536 - - github.com/containers/podman/libpod.makeRuntime(0xc000f696c0) <30>1 2022-11-22T20:10:23.221+00:00 nbapp817 infra-network-control.sh 192536 - - #011/builddir/build/BUILD/containers-podman-7f5e2fd/_build/src/github.com/containers/podman/libpod/runtime.go:325 +0x189 fp=0xc000a073b0 sp=0xc000a06f80 pc=0x55d5c4fd9989 <30>1 2022-11-22T20:10:23.221+00:00 nbapp817 infra-network-control.sh 192536 - - github.com/containers/podman/libpod.newRuntimeFromConfig(0xc000fa7c00, {0xc000a077b0, 0x0, 0x55d5c59c25b0}) <30>1 2022-11-22T20:10:23.221+00:00 nbapp817 infra-network-control.sh 192536 - - #011/builddir/build/BUILD/containers-podman-7f5e2fd/_build/src/github.com/containers/podman/libpod/runtime.go:227 +0x3d7 fp=0xc000a076d0 sp=0xc000a073b0 pc=0x55d5c4fd91d7 <30>1 2022-11-22T20:10:23.221+00:00 nbapp817 infra-network-control.sh 192536 - - github.com/containers/podman/libpod.NewRuntime({0xc000f5ca00, 0x55d5c52280cb}, {0xc000a077b0, 0x0, 0x0}) <30>1 2022-11-22T20:10:23.221+00:00 nbapp817 infra-network-control.sh 192536 - - #011/builddir/build/BUILD/containers-podman-7f5e2fd/_build/src/github.com/containers/podman/libpod/runtime.go:170 +0x51 fp=0xc000a07700 sp=0xc000a076d0 pc=0x55d5c4fd8db1 <30>1 2022-11-22T20:10:23.221+00:00 nbapp817 infra-network-control.sh 192536 - - github.com/containers/podman/pkg/domain/infra.getRuntime({0x55d5c5a11498, 0xc000040048}, 0xc000f5ca00, 0xc000a07a80) <30>1 2022-11-22T20:10:23.221+00:00 nbapp817 infra-network-control.sh 192536 - - #011/builddir/build/BUILD/containers-podman-7f5e2fd/_build/src/github.com/containers/podman/pkg/domain/infra/runtime_libpod.go:254 +0x19b7 fp=0xc000a07a60 sp=0xc000a07700 pc=0x55d5c50d9437 <30>1 2022-11-22T20:10:23.221+00:00 nbapp817 infra-network-control.sh 192536 - - github.com/containers/podman/pkg/domain/infra.GetRuntime.func1() <30>1 2022-11-22T20:10:23.221+00:00 nbapp817 infra-network-control.sh 192536 - - #011/builddir/build/BUILD/containers-podman-7f5e2fd/_build/src/github.com/containers/podman/pkg/domain/infra/runtime_libpod.go:80 +0x49 fp=0xc000a07ab0 sp=0xc000a07a60 pc=0x55d5c50d7a09 On two setups we see the following error: <30>1 2022-11-22T03:58:14.521+00:00 nbapp836 infra-network-control.sh 10637 - - Error: container 3ef225efde16ba8c3395134ca80ae8908b9d57de38a42e450b1363e7e20c2438 missing state in DB: internal libpod error <30>1 2022-11-22T03:58:14.522+00:00 nbapp836 podman 12681 - - Error: container 3ef225efde16ba8c3395134ca80ae8908b9d57de38a42e450b1363e7e20c2438 missing state in DB: internal libpod error <29>1 2022-11-22T03:58:14.527+00:00 nbapp836 systemd 1 - - podman.service: Main process exited, code=exited, status=125/n/a <28>1 2022-11-22T03:58:14.527+00:00 nbapp836 systemd 1 - - podman.service: Failed with result 'exit-code'. https://access.redhat.com/support/cases/#/case/03330920/discussion?commentId=a0a6R00000U4HJFQA3 @bbaude please see comment: https://bugzilla.redhat.com/show_bug.cgi?id=2131836#c24 Suggest filing another bug if the orginal issue was resolved and another one found. The original issue also referred to boltdb corruption which caused the panic. It looks like that is not addressed, hence we are seeing this issue. Please let us know, if this needs to opened as separate issue. Also, we would like to know, if there is any workaround available for this issue. I have created a separate bugzilla 2149112 for this issue. Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Moderate: container-tools:rhel8 security, bug fix, and enhancement update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2023:2758 The needinfo request[s] on this closed bug have been removed as they have been unresolved for 120 days |
Created attachment 1915831 [details] journalctl logs for podman API server Description of problem: After node power cycle, podman APIs panic resulting is error bringing up containers. Attached journalctl logs for podman API server Version-Release number of selected component (if applicable): podman-4.1.1 How reproducible: Steps to Reproduce: 1. Create containers 2. Power cycle node. Power recycle happened at Fri Sep 30 01:32 3. Unable to start containers Actual results: Containers fail to start Expected results: Containers should be ONLINE after node restart Additional info: [root@n58-h109 /]# last hostadmi pts/2 10.98.181.193 Mon Oct 3 11:25 still logged in hostadmi pts/1 10.98.181.193 Mon Oct 3 09:18 still logged in hostadmi pts/1 10.98.180.194 Fri Sep 30 15:20 - 18:22 (03:01) hostadmi pts/0 10.80.98.154 Fri Sep 30 02:40 still logged in reboot system boot 4.18.0-372.26.1. Fri Sep 30 01:32 still running hostadmi pts/0 10.80.98.154 Fri Sep 30 00:21 - 01:28 (01:07) hostadmi pts/0 172.28.9.253 Thu Sep 29 21:13 - 00:14 (03:00) reboot system boot 4.18.0-372.26.1. Thu Sep 29 20:50 - 01:28 (04:38) reboot system boot 4.18.0-372.26.1. Thu Sep 29 20:44 - 01:28 (04:44) hostadmi pts/4 rsvlmvc01vm445.r Thu Sep 29 20:00 - 20:40 (00:39) root pts/5 Wed Sep 28 20:21 - crash (1+00:23) hostadmi pts/4 haadpf36f80a.com Wed Sep 28 20:21 - 20:27 (00:06) root pts/4 Tue Sep 27 00:53 - 20:21 (1+19:27) hostadmi pts/3 rsvlmvc01vm445.r Tue Sep 27 00:52 - 01:13 (00:20) root pts/5 Mon Sep 26 23:15 - 20:21 (1+21:05) hostadmi pts/4 rsvlmvc01vm445.r Mon Sep 26 23:15 - 23:28 (00:12) hostadmi pts/4 172.28.9.250 Mon Sep 26 19:43 - 19:53 (00:10) hostadmi pts/3 rsvlmvc01vm445.r Mon Sep 26 05:16 - 05:42 (00:25) root pts/0 Mon Sep 26 04:44 - 04:44 (00:00) root pts/0 Mon Sep 26 04:44 - 04:44 (00:00) hostadmi tty1 Mon Sep 26 04:38 - 20:40 (3+16:01) reboot system boot 3.10.0-1160.45.1 Mon Sep 26 03:23 - 01:28 (3+22:04) reboot system boot 3.10.0-1160.45.1 Mon Sep 26 03:20 - 01:28 (3+22:08) wtmp begins Mon Sep 26 03:20:30 2022