● npc99a84aae6d34 State: degraded Units: 434 loaded (incl. loaded aliases) Jobs: 0 queued Failed: 1 units Since: Tue 2026-04-14 22:07:32 UTC; 44min ago systemd: 257.9-1~deb13u1 Tainted: unmerged-bin CGroup: / ├─init.scope │ └─1 /sbin/init nofb ├─system.slice │ ├─apache-htcacheclean.service │ │ └─15613 /usr/bin/htcacheclean -d 120 -p /var/cache/apache2/mod_cache_disk -l 300M -n │ ├─apache2.service │ │ ├─91335 /usr/sbin/apache2 -k start │ │ ├─91340 /usr/sbin/apache2 -k start │ │ └─91341 /usr/sbin/apache2 -k start │ ├─atd.service │ │ └─814 /usr/sbin/atd -f │ ├─cron.service │ │ └─805 /usr/sbin/cron -f │ ├─dbus.service │ │ └─589 /usr/bin/dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─epmd.service │ │ └─21889 /usr/bin/epmd -systemd │ ├─haproxy.service │ │ ├─15053 /usr/sbin/haproxy -Ws -f /etc/haproxy/haproxy.cfg -p /run/haproxy.pid -S /run/haproxy-master.sock │ │ └─15055 /usr/sbin/haproxy -Ws -f /etc/haproxy/haproxy.cfg -p /run/haproxy.pid -S /run/haproxy-master.sock │ ├─haveged.service │ │ └─432 /usr/sbin/haveged --Foreground --verbose=1 │ ├─iscsid.service │ │ ├─43268 /usr/sbin/iscsid │ │ └─43269 /usr/sbin/iscsid │ ├─ksmtuned.service │ │ ├─ 6269 /bin/bash /usr/sbin/ksmtuned │ │ └─121050 sleep 60 │ ├─libvirtd.service │ │ └─87645 /usr/sbin/libvirtd --timeout 120 │ ├─mariadb.service │ │ └─62213 /usr/sbin/mariadbd │ ├─memcached.service │ │ └─88828 /usr/bin/memcached -m 64 -p 11211 -u memcache -l 127.0.0.1 -l ::1 -P /var/run/memcached/memcached.pid │ ├─netavark-firewalld-reload.service │ │ └─13083 /usr/lib/podman/netavark firewalld-reload │ ├─ovn-controller-vtep.service │ │ └─93768 ovn-controller-vtep -vconsole:emer -vsyslog:err -vfile:info --vtep-db=/var/run/openvswitch/db.sock --ovnsb-db=/var/run/ovn/ovnsb_db.sock --no-chdir --log-file=/var/log/ovn/ovn-controller-vtep.log --pidfile=/var/run/ovn/ovn-controller-vtep.pid --detach │ ├─ovn-controller.service │ │ └─94497 ovn-controller unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --no-chdir --log-file=/var/log/ovn/ovn-controller.log --pidfile=/var/run/ovn/ovn-controller.pid --detach │ ├─ovn-northd.service │ │ └─94198 ovn-northd -vconsole:emer -vsyslog:err -vfile:info --ovnnb-db=unix:/var/run/ovn/ovnnb_db.sock --ovnsb-db=unix:/var/run/ovn/ovnsb_db.sock --no-chdir --log-file=/var/log/ovn/ovn-northd.log --pidfile=/var/run/ovn/ovn-northd.pid --detach │ ├─ovn-ovsdb-server-nb.service │ │ └─94125 ovsdb-server -vconsole:off -vfile:info --log-file=/var/log/ovn/ovsdb-server-nb.log --pidfile=/var/run/ovn/ovnnb_db.pid --remote=punix:/var/run/ovn/ovnnb_db.sock --unixctl=/var/run/ovn/ovnnb_db.ctl --remote=db:OVN_Northbound,NB_Global,connections --private-key=db:OVN_Northbound,SSL,private_key --certificate=db:OVN_Northbound,SSL,certificate --ca-cert=db:OVN_Northbound,SSL,ca_cert --ssl-protocols=db:OVN_Northbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Northbound,SSL,ssl_ciphers --ssl-ciphersuites=db:OVN_Northbound,SSL,ssl_ciphersuites /var/lib/ovn/ovnnb_db.db │ ├─ovn-ovsdb-server-sb.service │ │ └─94127 ovsdb-server -vconsole:off -vfile:info --log-file=/var/log/ovn/ovsdb-server-sb.log --pidfile=/var/run/ovn/ovnsb_db.pid --remote=punix:/var/run/ovn/ovnsb_db.sock --unixctl=/var/run/ovn/ovnsb_db.ctl --remote=db:OVN_Southbound,SB_Global,connections --private-key=db:OVN_Southbound,SSL,private_key --certificate=db:OVN_Southbound,SSL,certificate --ca-cert=db:OVN_Southbound,SSL,ca_cert --ssl-protocols=db:OVN_Southbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Southbound,SSL,ssl_ciphers --ssl-ciphersuites=db:OVN_Southbound,SSL,ssl_ciphersuites /var/lib/ovn/ovnsb_db.db │ ├─ovs-vswitchd.service │ │ └─93684 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach │ ├─ovsdb-server.service │ │ └─93449 ovsdb-server /etc/openvswitch/conf.db -vconsole:emer -vsyslog:err -vfile:info --remote=punix:/var/run/openvswitch/db.sock --private-key=db:Open_vSwitch,SSL,private_key --certificate=db:Open_vSwitch,SSL,certificate --bootstrap-ca-cert=db:Open_vSwitch,SSL,ca_cert --no-chdir --log-file=/var/log/openvswitch/ovsdb-server.log --pidfile=/var/run/openvswitch/ovsdb-server.pid --detach │ ├─polkit.service │ │ └─105627 /usr/lib/polkit-1/polkitd --no-debug --log-level=notice │ ├─rabbitmq-server.service │ │ ├─22008 /usr/lib/erlang/erts-15.2.7/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -- -root /usr/lib/erlang -bindir /usr/lib/erlang/erts-15.2.7/bin -progname erl -- -home /var/lib/rabbitmq -- -pa "" -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger "[]" -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -- │ │ ├─22018 erl_child_setup 65536 │ │ ├─22049 sh -s disksup │ │ ├─22051 /usr/lib/erlang/lib/os_mon-2.10.1/priv/bin/memsup │ │ ├─22052 /usr/lib/erlang/lib/os_mon-2.10.1/priv/bin/cpu_sup │ │ ├─22054 /usr/lib/erlang/erts-15.2.7/bin/inet_gethost 4 │ │ ├─22055 /usr/lib/erlang/erts-15.2.7/bin/inet_gethost 4 │ │ └─22058 /bin/sh -s rabbit_disk_monitor │ ├─rsync.service │ │ └─88855 /usr/bin/rsync --daemon --no-detach │ ├─rsyslog.service │ │ └─19101 /usr/sbin/rsyslogd -n -iNONE │ ├─ssh.service │ │ └─13390 "sshd: /usr/sbin/sshd -D [listener] 0 of 30-100 startups" │ ├─system-ceph\x2d929e05b1\x2d4e8c\x2d4c8f\x2db13f\x2d4521b44f10e1.slice │ │ ├─ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@crash.npc99a84aae6d34.service │ │ │ ├─libpod-payload-91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc │ │ │ │ ├─54984 /run/podman-init -- /usr/bin/ceph-crash -n client.crash.npc99a84aae6d34 │ │ │ │ └─54986 /usr/bin/python3 -s /usr/bin/ceph-crash -n client.crash.npc99a84aae6d34 │ │ │ └─runtime │ │ │ └─54981 /usr/bin/conmon --api-version 1 -c 91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc -u 91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc/userdata -p /run/containers/storage/overlay-containers/91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc/userdata/pidfile -n ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1-crash-npc99a84aae6d34 --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc --full-attach -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc/userdata/oci-log --conmon-pidfile /run/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@crash.npc99a84aae6d34.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg "" --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg --rm --exit-command-arg 91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc │ │ ├─ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mgr.npc99a84aae6d34.krobmi.service │ │ │ ├─libpod-payload-3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274 │ │ │ │ ├─55422 /run/podman-init -- /usr/bin/ceph-mgr -n mgr.npc99a84aae6d34.krobmi -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false │ │ │ │ └─55424 /usr/bin/ceph-mgr -n mgr.npc99a84aae6d34.krobmi -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false │ │ │ └─runtime │ │ │ └─55417 /usr/bin/conmon --api-version 1 -c 3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274 -u 3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274 -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274/userdata -p /run/containers/storage/overlay-containers/3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274/userdata/pidfile -n ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1-mgr-npc99a84aae6d34-krobmi --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274 --full-attach -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274/userdata/oci-log --conmon-pidfile /run/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mgr.npc99a84aae6d34.krobmi.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg "" --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg --rm --exit-command-arg 3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274 │ │ ├─ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mgr.npc99a84aae6d34.mxlymt.service │ │ │ ├─libpod-payload-c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73 │ │ │ │ ├─51960 /run/podman-init -- /usr/bin/ceph-mgr -n mgr.npc99a84aae6d34.mxlymt -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false │ │ │ │ └─51962 /usr/bin/ceph-mgr -n mgr.npc99a84aae6d34.mxlymt -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false │ │ │ └─runtime │ │ │ └─51958 /usr/bin/conmon --api-version 1 -c c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73 -u c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73 -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73/userdata -p /run/containers/storage/overlay-containers/c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73/userdata/pidfile -n ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1-mgr-npc99a84aae6d34-mxlymt --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73 --full-attach -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73/userdata/oci-log --conmon-pidfile /run/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mgr.npc99a84aae6d34.mxlymt.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg "" --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg --rm --exit-command-arg c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73 │ │ ├─ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mon.npc99a84aae6d34.service │ │ │ ├─libpod-payload-b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942 │ │ │ │ ├─51752 /run/podman-init -- /usr/bin/ceph-mon -n mon.npc99a84aae6d34 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false │ │ │ │ └─51754 /usr/bin/ceph-mon -n mon.npc99a84aae6d34 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false │ │ │ └─runtime │ │ │ └─51750 /usr/bin/conmon --api-version 1 -c b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942 -u b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942 -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942/userdata -p /run/containers/storage/overlay-containers/b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942/userdata/pidfile -n ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1-mon-npc99a84aae6d34 --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942 --full-attach -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942/userdata/oci-log --conmon-pidfile /run/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mon.npc99a84aae6d34.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg "" --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg --rm --exit-command-arg b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942 │ │ └─ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@osd.0.service │ │ ├─libpod-payload-45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee │ │ │ ├─57930 /run/podman-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --osd-objectstore=bluestore │ │ │ └─57932 /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --osd-objectstore=bluestore │ │ └─runtime │ │ └─57928 /usr/bin/conmon --api-version 1 -c 45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee -u 45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee/userdata -p /run/containers/storage/overlay-containers/45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee/userdata/pidfile -n ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1-osd-0 --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee --full-attach -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee/userdata/oci-log --conmon-pidfile /run/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@osd.0.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg "" --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg --rm --exit-command-arg 45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee │ ├─system-devstack.slice │ │ ├─devstack@br-ex-tcpdump.service │ │ │ └─110919 /bin/tcpdump -i br-ex arp or rarp or icmp or icmp6 -enlX │ │ ├─devstack@br-int-flows.service │ │ │ ├─111525 /bin/sh -c "set +e; while true; do echo ovs-ofctl dump-flows br-int; ovs-ofctl dump-flows br-int ; sleep 30; done; " │ │ │ └─121349 sleep 30 │ │ ├─devstack@c-api.service │ │ │ ├─105804 "cinder-apiuWSGI master" │ │ │ ├─105808 "cinder-apiuWSGI worker 1" │ │ │ └─105809 "cinder-apiuWSGI worker 2" │ │ ├─devstack@c-bak.service │ │ │ └─107200 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/cinder-backup --config-file /etc/cinder/cinder.conf │ │ ├─devstack@c-sch.service │ │ │ └─106585 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/cinder-scheduler --config-file /etc/cinder/cinder.conf │ │ ├─devstack@c-vol.service │ │ │ ├─107810 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/cinder-volume --config-file /etc/cinder/cinder.conf │ │ │ └─108126 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/cinder-volume --config-file /etc/cinder/cinder.conf │ │ ├─devstack@etcd.service │ │ │ └─64223 /opt/stack/bin/etcd --name npc99a84aae6d34 --data-dir /opt/stack/data/etcd --initial-cluster-state new --initial-cluster-token etcd-cluster-01 --initial-cluster "npc99a84aae6d34=http://[2001:41d0:302:1000::cbc]:2380" --initial-advertise-peer-urls "http://[2001:41d0:302:1000::cbc]:2380" --advertise-client-urls "http://[2001:41d0:302:1000::cbc]:2379" --listen-peer-urls http://0.0.0.0:2380 --listen-client-urls "http://[2001:41d0:302:1000::cbc]:2379" --log-level=debug │ │ ├─devstack@file_tracker.service │ │ │ ├─ 63498 /bin/bash /opt/stack/devstack/tools/file_tracker.sh │ │ │ └─122592 sleep 20 │ │ ├─devstack@g-api.service │ │ │ ├─108887 "glance-apiuWSGI master" │ │ │ ├─108888 "glance-apiuWSGI worker 1" │ │ │ └─108889 "glance-apiuWSGI worker 2" │ │ ├─devstack@keystone.service │ │ │ ├─65424 "keystoneuWSGI master" │ │ │ ├─65425 "keystoneuWSGI worker 1" │ │ │ └─65426 "keystoneuWSGI worker 2" │ │ ├─devstack@memory_tracker.service │ │ │ ├─ 62947 /bin/bash /opt/stack/devstack/tools/memory_tracker.sh │ │ │ └─122596 sleep 20 │ │ ├─devstack@n-api-meta.service │ │ │ ├─101619 "nova-api-metauWSGI master" │ │ │ ├─101621 "nova-api-metauWSGI worker 1" │ │ │ ├─101622 "nova-api-metauWSGI worker 2" │ │ │ └─101623 "nova-api-metauWSGI http 1" │ │ ├─devstack@n-api.service │ │ │ ├─92903 "nova-apiuWSGI master" │ │ │ ├─92904 "nova-apiuWSGI worker 1" │ │ │ └─92905 "nova-apiuWSGI worker 2" │ │ ├─devstack@n-cond-cell1.service │ │ │ ├─103735 "nova-conductor: master process [/opt/stack/data/venv/bin/nova-conductor --config-file /etc/nova/nova_cell1.conf]" │ │ │ ├─104902 "nova-conductor: ServiceWrapper worker(0)" │ │ │ └─104904 "nova-conductor: ServiceWrapper worker(1)" │ │ ├─devstack@n-cpu.service │ │ │ └─104923 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/nova-compute --config-file /etc/nova/nova-cpu.conf │ │ ├─devstack@n-novnc-cell1.service │ │ │ └─102318 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/nova-novncproxy --config-file /etc/nova/nova_cell1.conf --web /opt/stack/novnc │ │ ├─devstack@n-sch.service │ │ │ ├─100952 "nova-scheduler: master process [/opt/stack/data/venv/bin/nova-scheduler --config-file /etc/nova/nova.conf]" │ │ │ ├─102220 "nova-scheduler: ServiceWrapper worker(0)" │ │ │ └─102223 "nova-scheduler: ServiceWrapper worker(1)" │ │ ├─devstack@n-super-cond.service │ │ │ ├─103157 "nova-conductor: master process [/opt/stack/data/venv/bin/nova-conductor --config-file /etc/nova/nova.conf]" │ │ │ ├─104721 "nova-conductor: ServiceWrapper worker(0)" │ │ │ └─104725 "nova-conductor: ServiceWrapper worker(1)" │ │ ├─devstack@neutron-api.service │ │ │ ├─96269 "neutron-apiuWSGI master" │ │ │ ├─96270 "neutron-apiuWSGI worker 1" │ │ │ └─96271 "neutron-apiuWSGI worker 2" │ │ ├─devstack@neutron-ovn-maintenance-worker.service │ │ │ ├─97897 "neutron-ovn-maintenance-worker: master process [/opt/stack/data/venv/bin/neutron-ovn-maintenance-worker --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" │ │ │ └─98765 "neutron-server: maintenance worker (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-ovn-maintenance-worker --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ ├─devstack@neutron-periodic-workers.service │ │ │ ├─97349 "neutron-periodic-workers: master process [/opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" │ │ │ ├─98512 "neutron-server: Periodic worker for \"AgentSchedulerDbMixin\" (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ │ ├─98520 "neutron-server: Periodic worker for \"AgentSchedulerDbMixin\" (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ │ ├─98524 "neutron-server: Periodic worker for \"DbQuotaNoLockDriver\" (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ │ └─98532 "neutron-server: Periodic worker for \"L3_NAT_dbonly_mixin\" (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ ├─devstack@neutron-rpc-server.service │ │ │ ├─96802 "neutron-rpc-server: master process [/opt/stack/data/venv/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" │ │ │ └─98148 "neutron-server: rpc worker (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ ├─devstack@openstack-cli-server.service │ │ │ └─61256 /opt/stack/data/venv/bin/python3 /opt/stack/devstack/files/openstack-cli-server/openstack-cli-server │ │ ├─devstack@placement-api.service │ │ │ ├─98699 "placementuWSGI master" │ │ │ ├─98700 "placementuWSGI worker 1" │ │ │ └─98701 "placementuWSGI worker 2" │ │ ├─devstack@q-ovn-agent.service │ │ │ ├─95086 "neutron-ovn-agent: master process [/opt/stack/data/venv/bin/neutron-ovn-agent --config-file /etc/neutron/plugins/ml2/ovn_agent.ini]" │ │ │ ├─95991 "neutron-ovn-agent: ServiceWrapper worker(0)" │ │ │ ├─96262 /opt/stack/data/venv/bin/python3.13 /usr/local/bin/privsep-helper --config-file /etc/neutron/plugins/ml2/ovn_agent.ini --privsep_context neutron.privileged.namespace_cmd --privsep_sock_path /tmp/tmp38ne4qkg/privsep.sock │ │ │ └─99558 /opt/stack/data/venv/bin/python3.13 /usr/local/bin/privsep-helper --config-file /etc/neutron/plugins/ml2/ovn_agent.ini --privsep_context neutron.privileged.default --privsep_sock_path /tmp/tmpqd4w1pg4/privsep.sock │ │ ├─devstack@s-account.service │ │ │ ├─90424 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-account-server /etc/swift/account-server/1.conf -v │ │ │ ├─90719 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-account-server /etc/swift/account-server/1.conf -v │ │ │ └─90721 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-account-server /etc/swift/account-server/1.conf -v │ │ ├─devstack@s-container-sync.service │ │ │ └─91065 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-container-sync /etc/swift/container-server/1.conf │ │ ├─devstack@s-container.service │ │ │ ├─89898 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-container-server /etc/swift/container-server/1.conf -v │ │ │ ├─90188 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-container-server /etc/swift/container-server/1.conf -v │ │ │ └─90190 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-container-server /etc/swift/container-server/1.conf -v │ │ ├─devstack@s-object.service │ │ │ ├─89369 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-object-server /etc/swift/object-server/1.conf -v │ │ │ ├─89690 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-object-server /etc/swift/object-server/1.conf -v │ │ │ └─89691 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-object-server /etc/swift/object-server/1.conf -v │ │ └─devstack@s-proxy.service │ │ ├─91903 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-proxy-server /etc/swift/proxy-server.conf -v │ │ └─91982 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-proxy-server /etc/swift/proxy-server.conf -v │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─821 /sbin/agetty -o "-- \\u" --noreset --noclear - linux │ ├─system-glean.slice │ │ └─glean@ens3.service │ │ └─743 dhclient -4 -v -i -pf /run/dhclient.ens3.pid -lf /var/lib/dhcp/dhclient.ens3.leases -I -df /var/lib/dhcp/dhclient6.ens3.leases ens3 │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─823 /sbin/agetty -o "-- \\u" --noreset --noclear --keep-baud 115200,57600,38400,9600 - vt220 │ ├─systemd-journald.service │ │ └─20737 /usr/lib/systemd/systemd-journald │ ├─systemd-logind.service │ │ └─605 /usr/lib/systemd/systemd-logind │ ├─systemd-machined.service │ │ └─41980 /usr/lib/systemd/systemd-machined │ ├─systemd-timesyncd.service │ │ └─411 /usr/lib/systemd/systemd-timesyncd │ ├─systemd-udevd.service │ │ └─udev │ │ └─422 /usr/lib/systemd/systemd-udevd │ ├─unbound.service │ │ └─1203 /usr/sbin/unbound -d -p │ ├─uuidd.service │ │ └─96046 /usr/sbin/uuidd --socket-activation --cont-clock │ └─virtlogd.service │ └─47685 /usr/sbin/virtlogd └─user.slice └─user-1000.slice ├─session-1.scope │ └─1087 /usr/bin/python3.13 ├─session-34.scope │ ├─120787 "sshd-session: zuul [priv]" │ ├─120794 "sshd-session: zuul@notty" │ ├─122625 /bin/sh -c "sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-mjejnubreskapnsupqhyuesybxywnovx ; /usr/bin/python3.13' && sleep 0" │ ├─122626 sudo -H -S -n -u root /bin/sh -c "echo BECOME-SUCCESS-mjejnubreskapnsupqhyuesybxywnovx ; /usr/bin/python3.13" │ ├─122628 /bin/sh -c "echo BECOME-SUCCESS-mjejnubreskapnsupqhyuesybxywnovx ; /usr/bin/python3.13" │ ├─122629 /usr/bin/python3.13 │ ├─122630 /bin/bash -c "sudo iptables-save > /home/zuul/iptables.txt\n\n# NOTE(sfernand): Run 'df' with a 60s timeout to prevent hangs from\n# stale NFS mounts.\ntimeout -s 9 60s df -h > /home/zuul/df.txt || true\n# If 'df' times out, the mount output helps debug which NFS share\n# is unresponsive.\nmount > /home/zuul/mount.txt\n\nfor py_ver in 2 3; do\n if [[ \`which python\${py_ver}\` ]]; then\n python\${py_ver} -m pip freeze > /home/zuul/pip\${py_ver}-freeze.txt\n fi\ndone\n\nif [ \`command -v dpkg\` ]; then\n dpkg -l> /home/zuul/dpkg-l.txt\nfi\nif [ \`command -v rpm\` ]; then\n rpm -qa | sort > /home/zuul/rpm-qa.txt\nfi\n\n# Services status\nsudo systemctl status --all > services.txt 2>/dev/null\n\n# NOTE(kchamart) The 'audit.log' can be useful in cases when QEMU\n# failed to start due to denials from SELinux — useful for CentOS\n# and Fedora machines. For Ubuntu (which runs AppArmor), DevStack\n# already captures the contents of /var/log/kern.log (via\n# \`journalctl -t kernel\` redirected into syslog.txt.gz), which\n# contains AppArmor-related messages.\nif [ -f /var/log/audit/audit.log ] ; then\n sudo cp /var/log/audit/audit.log /home/zuul/audit.log &&\n chmod +r /home/zuul/audit.log;\nfi\n\n# gzip and save any coredumps in /var/core\nif [ -d /var/core ]; then\n sudo gzip -r /var/core\n sudo cp -r /var/core /home/zuul/\nfi\n\nsudo ss -lntup | grep ':53' > /home/zuul/listen53.txt\n\n# NOTE(andreaf) Service logs are already in logs/ thanks for the\n# export-devstack-journal log. Apache logs are under apache/ thans to the\n# apache-logs-conf role.\ngrep -i deprecat /home/zuul/logs/*.txt /home/zuul/apache/*.log | \\\n sed -r 's/[0-9]{1,2}\\:[0-9]{1,2}\\:[0-9]{1,2}\\.[0-9]{1,3}/ /g' | \\\n sed -r 's/[0-9]{1,2}\\:[0-9]{1,2}\\:[0-9]{1,2}/ /g' | \\\n sed -r 's/[0-9]{1,4}-[0-9]{1,2}-[0-9]{1,4}/ /g' |\n sed -r 's/\\[.*\\]/ /g' | \\\n sed -r 's/\\s[0-9]+\\s/ /g' | \\\n awk '{if (\$0 in seen) {seen[\$0]++} else {out[++n]=\$0;seen[\$0]=1}} END { for (i=1; i<=n; i++) print seen[out[i]]\" :: \" out[i] }' > /home/zuul/deprecations.log\n" │ ├─122644 sudo systemctl status --all │ └─122646 systemctl status --all └─user@1000.service ├─init.scope │ ├─859 /usr/lib/systemd/systemd --user │ └─862 "(sd-pam)" └─session.slice └─dbus.service └─12675 /usr/bin/dbus-daemon --session --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only ● proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point Loaded: loaded (/usr/lib/systemd/system/proc-sys-fs-binfmt_misc.automount; static) Active: active (running) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 3d97285309dc460da29926a2098e1105 Triggers: ● proc-sys-fs-binfmt_misc.mount Where: /proc/sys/fs/binfmt_misc Docs: https://docs.kernel.org/admin-guide/binfmt-misc.html https://systemd.io/API_FILE_SYSTEMS Apr 14 22:07:33 debian systemd[1]: proc-sys-fs-binfmt_misc.automount: Got automount request for /proc/sys/fs/binfmt_misc, triggered by 420 (systemd-binfmt) Notice: journal has been rotated since unit was started, output may be incomplete. ● dev-cdrom.device - QEMU_DVD-ROM config-2 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:01.1-ata1-host0-target0:0:0-0:0:0:0-block-sr0.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: b76f8db1e04c4a0eb4dec5354419ce17 Device: /sys/devices/pci0000:00/0000:00:01.1/ata1/host0/target0:0:0/0:0:0:0/block/sr0 ● dev-ceph_vg-ceph_lv_data.device - /dev/ceph_vg/ceph_lv_data Follows: unit currently follows state of sys-devices-virtual-block-dm\x2d0.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:26:36 UTC; 25min ago Invocation: 8f91e3e17fc64379a84d12a049eb0f1c Device: /sys/devices/virtual/block/dm-0 ● dev-disk-by\x2ddiskseq-11.device - QEMU_DVD-ROM config-2 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:01.1-ata1-host0-target0:0:0-0:0:0:0-block-sr0.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 42aabb96b950468492c46c66b1bfaad9 Device: /sys/devices/pci0000:00/0000:00:01.1/ata1/host0/target0:0:0/0:0:0:0/block/sr0 ● dev-disk-by\x2ddiskseq-20.device - /dev/disk/by-diskseq/20 Follows: unit currently follows state of sys-devices-virtual-block-loop0.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:26:36 UTC; 25min ago Invocation: 538480ee1c454e39a18489407c108099 Device: /sys/devices/virtual/block/loop0 ● dev-disk-by\x2ddiskseq-21.device - /dev/disk/by-diskseq/21 Follows: unit currently follows state of sys-devices-virtual-block-dm\x2d0.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:26:36 UTC; 25min ago Invocation: f234fb5854f448f8b1dee06b5fa04860 Device: /sys/devices/virtual/block/dm-0 ● dev-disk-by\x2ddiskseq-22.device - /dev/disk/by-diskseq/22 Follows: unit currently follows state of sys-devices-virtual-block-loop1.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:31:46 UTC; 20min ago Invocation: 5c03c616f375496096897dfa0d45e6d7 Device: /sys/devices/virtual/block/loop1 ● dev-disk-by\x2ddiskseq-9.device - /dev/disk/by-diskseq/9 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio1-block-vda.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 6a2b1582f34e4b7d9522bc7239719594 Device: /sys/devices/pci0000:00/0000:00:04.0/virtio1/block/vda ● dev-disk-by\x2ddiskseq-9\x2dpart1.device - /dev/disk/by-diskseq/9-part1 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio1-block-vda-vda1.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 3280f8528eed46209e4138dd0413cc59 Device: /sys/devices/pci0000:00/0000:00:04.0/virtio1/block/vda/vda1 ● dev-disk-by\x2did-ata\x2dQEMU_DVD\x2dROM_QM00001.device - QEMU_DVD-ROM config-2 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:01.1-ata1-host0-target0:0:0-0:0:0:0-block-sr0.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: f93dddefe3d648d48c870aa036671a72 Device: /sys/devices/pci0000:00/0000:00:01.1/ata1/host0/target0:0:0/0:0:0:0/block/sr0 ● dev-disk-by\x2did-dm\x2dname\x2dceph_vg\x2dceph_lv_data.device - /dev/disk/by-id/dm-name-ceph_vg-ceph_lv_data Follows: unit currently follows state of sys-devices-virtual-block-dm\x2d0.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:26:36 UTC; 25min ago Invocation: c0a8c9c42e0240f9a1357f80aa1997f5 Device: /sys/devices/virtual/block/dm-0 ● dev-disk-by\x2did-dm\x2duuid\x2dLVM\x2dREaBFzJ1pVpbKTsMl8oqD8ZH20nu3Isyv3G5PvfVxL5FnQaj7C0y3BEt3r3MSGae.device - /dev/disk/by-id/dm-uuid-LVM-REaBFzJ1pVpbKTsMl8oqD8ZH20nu3Isyv3G5PvfVxL5FnQaj7C0y3BEt3r3MSGae Follows: unit currently follows state of sys-devices-virtual-block-dm\x2d0.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:26:36 UTC; 25min ago Invocation: 6ad55e5b399c4383be17b1a9e6d363c9 Device: /sys/devices/virtual/block/dm-0 ● dev-disk-by\x2did-lvm\x2dpv\x2duuid\x2dtsmaQE\x2dW9eU\x2ds35L\x2demXc\x2d4hGv\x2d0vnf\x2deFO7do.device - /dev/disk/by-id/lvm-pv-uuid-tsmaQE-W9eU-s35L-emXc-4hGv-0vnf-eFO7do Follows: unit currently follows state of sys-devices-virtual-block-loop0.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:26:36 UTC; 25min ago Invocation: ac4e1ae375c8446eba420ba53d9e3ff0 Device: /sys/devices/virtual/block/loop0 ● dev-disk-by\x2dlabel-cloudimg\x2drootfs.device - /dev/disk/by-label/cloudimg-rootfs Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio1-block-vda-vda1.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 3608524dcd4d476eb487108910ec4de7 Device: /sys/devices/pci0000:00/0000:00:04.0/virtio1/block/vda/vda1 ● dev-disk-by\x2dlabel-config\x2d2.device - QEMU_DVD-ROM config-2 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:01.1-ata1-host0-target0:0:0-0:0:0:0-block-sr0.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 99bba5b4d5a1481ba80d749bc2646c45 Device: /sys/devices/pci0000:00/0000:00:01.1/ata1/host0/target0:0:0/0:0:0:0/block/sr0 ● dev-disk-by\x2dloop\x2dinode-254:1\x2d2751675.device - /dev/disk/by-loop-inode/254:1-2751675 Follows: unit currently follows state of sys-devices-virtual-block-loop0.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:26:36 UTC; 25min ago Invocation: 40e42e01a8b5460bb8314a3d563bcf6f Device: /sys/devices/virtual/block/loop0 ● dev-disk-by\x2dloop\x2dinode-254:1\x2d8386259.device - /dev/disk/by-loop-inode/254:1-8386259 Follows: unit currently follows state of sys-devices-virtual-block-loop1.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:31:46 UTC; 20min ago Invocation: cb9f797c01ff4e8d89458f301eb66a93 Device: /sys/devices/virtual/block/loop1 ● dev-disk-by\x2dloop\x2dref-\x5cx2fopt\x5cx2fstack\x5cx2fceph\x2dosd.img.device - /dev/disk/by-loop-ref/\x2fopt\x2fstack\x2fceph-osd.img Follows: unit currently follows state of sys-devices-virtual-block-loop0.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:26:36 UTC; 25min ago Invocation: 1289b1823c7843a2914811af5c704083 Device: /sys/devices/virtual/block/loop0 ● dev-disk-by\x2dloop\x2dref-\x5cx2fopt\x5cx2fstack\x5cx2fdata\x5cx2fswift\x5cx2fdrives\x5cx2fimages\x5cx2fswift.img.device - /dev/disk/by-loop-ref/\x2fopt\x2fstack\x2fdata\x2fswift\x2fdrives\x2fimages\x2fswift.img Follows: unit currently follows state of sys-devices-virtual-block-loop1.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:31:46 UTC; 20min ago Invocation: 08c70a2a6fe1439db4d03c56b9ab312d Device: /sys/devices/virtual/block/loop1 ● dev-disk-by\x2dpartuuid-7e01cb06\x2d01.device - /dev/disk/by-partuuid/7e01cb06-01 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio1-block-vda-vda1.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: a912f3dc141c4ed89b6a2efbdd0bb713 Device: /sys/devices/pci0000:00/0000:00:04.0/virtio1/block/vda/vda1 ● dev-disk-by\x2dpath-pci\x2d0000:00:01.1\x2data\x2d1.0.device - QEMU_DVD-ROM config-2 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:01.1-ata1-host0-target0:0:0-0:0:0:0-block-sr0.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 75b94846fd764f8791df586a93f2c45b Device: /sys/devices/pci0000:00/0000:00:01.1/ata1/host0/target0:0:0/0:0:0:0/block/sr0 ● dev-disk-by\x2dpath-pci\x2d0000:00:01.1\x2data\x2d1.device - QEMU_DVD-ROM config-2 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:01.1-ata1-host0-target0:0:0-0:0:0:0-block-sr0.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: b162d93c9dc14946bc9d08a14f0c770c Device: /sys/devices/pci0000:00/0000:00:01.1/ata1/host0/target0:0:0/0:0:0:0/block/sr0 ● dev-disk-by\x2dpath-pci\x2d0000:00:04.0.device - /dev/disk/by-path/pci-0000:00:04.0 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio1-block-vda.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 2d363274832a4c748362001c7f49e499 Device: /sys/devices/pci0000:00/0000:00:04.0/virtio1/block/vda ● dev-disk-by\x2dpath-pci\x2d0000:00:04.0\x2dpart-by\x2dlabel-cloudimg\x2drootfs.device - /dev/disk/by-path/pci-0000:00:04.0-part/by-label/cloudimg-rootfs Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio1-block-vda-vda1.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 32c76d08c7fb4d42b5f1c7ef9611fb35 Device: /sys/devices/pci0000:00/0000:00:04.0/virtio1/block/vda/vda1 ● dev-disk-by\x2dpath-pci\x2d0000:00:04.0\x2dpart-by\x2dpartnum-1.device - /dev/disk/by-path/pci-0000:00:04.0-part/by-partnum/1 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio1-block-vda-vda1.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: ad1bfdd4c66c4515966e88a737fa96db Device: /sys/devices/pci0000:00/0000:00:04.0/virtio1/block/vda/vda1 ● dev-disk-by\x2dpath-pci\x2d0000:00:04.0\x2dpart-by\x2dpartuuid-7e01cb06\x2d01.device - /dev/disk/by-path/pci-0000:00:04.0-part/by-partuuid/7e01cb06-01 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio1-block-vda-vda1.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 8b2d36f64a754ed2ba91d7cf4baafb96 Device: /sys/devices/pci0000:00/0000:00:04.0/virtio1/block/vda/vda1 ● dev-disk-by\x2dpath-pci\x2d0000:00:04.0\x2dpart-by\x2duuid-355e9d64\x2d3765\x2d4012\x2db042\x2d14affed2b0bf.device - /dev/disk/by-path/pci-0000:00:04.0-part/by-uuid/355e9d64-3765-4012-b042-14affed2b0bf Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio1-block-vda-vda1.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 6a3f9545d8174646b0f5dd5095dbc7c6 Device: /sys/devices/pci0000:00/0000:00:04.0/virtio1/block/vda/vda1 ● dev-disk-by\x2dpath-pci\x2d0000:00:04.0\x2dpart1.device - /dev/disk/by-path/pci-0000:00:04.0-part1 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio1-block-vda-vda1.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: bfe3d82d8c004c08a1920b2665f81e43 Device: /sys/devices/pci0000:00/0000:00:04.0/virtio1/block/vda/vda1 ● dev-disk-by\x2dpath-virtio\x2dpci\x2d0000:00:04.0.device - /dev/disk/by-path/virtio-pci-0000:00:04.0 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio1-block-vda.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 4d227723017541f880069f87a5f4eb3b Device: /sys/devices/pci0000:00/0000:00:04.0/virtio1/block/vda ● dev-disk-by\x2dpath-virtio\x2dpci\x2d0000:00:04.0\x2dpart1.device - /dev/disk/by-path/virtio-pci-0000:00:04.0-part1 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio1-block-vda-vda1.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 00f7a2990f374765b927d35176537780 Device: /sys/devices/pci0000:00/0000:00:04.0/virtio1/block/vda/vda1 ● dev-disk-by\x2duuid-2026\x2d04\x2d14\x2d22\x2d07\x2d23\x2d00.device - QEMU_DVD-ROM config-2 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:01.1-ata1-host0-target0:0:0-0:0:0:0-block-sr0.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 1f25eb340a864084b3476668af19fae8 Device: /sys/devices/pci0000:00/0000:00:01.1/ata1/host0/target0:0:0/0:0:0:0/block/sr0 ● dev-disk-by\x2duuid-355e9d64\x2d3765\x2d4012\x2db042\x2d14affed2b0bf.device - /dev/disk/by-uuid/355e9d64-3765-4012-b042-14affed2b0bf Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio1-block-vda-vda1.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 922df879dba64d2d9a8ca9d8df00c5ac Device: /sys/devices/pci0000:00/0000:00:04.0/virtio1/block/vda/vda1 ● dev-disk-by\x2duuid-51f7c90b\x2d5782\x2d4c30\x2daa45\x2dc5615703a18c.device - /dev/disk/by-uuid/51f7c90b-5782-4c30-aa45-c5615703a18c Follows: unit currently follows state of sys-devices-virtual-block-loop1.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:31:46 UTC; 20min ago Invocation: 755b05ffacb34dadafc78c9f97cb4258 Device: /sys/devices/virtual/block/loop1 ● dev-dm\x2d0.device - /dev/dm-0 Follows: unit currently follows state of sys-devices-virtual-block-dm\x2d0.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:26:36 UTC; 25min ago Invocation: 01fa1d5732a44f3286efad1e598e95f2 Device: /sys/devices/virtual/block/dm-0 ● dev-loop0.device - /dev/loop0 Follows: unit currently follows state of sys-devices-virtual-block-loop0.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:26:36 UTC; 25min ago Invocation: a5aae8db536d410a9de8984bcc44bd47 Device: /sys/devices/virtual/block/loop0 ● dev-loop1.device - /dev/loop1 Follows: unit currently follows state of sys-devices-virtual-block-loop1.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:31:46 UTC; 20min ago Invocation: b76adfee36cd41d689ffcefcbd132c49 Device: /sys/devices/virtual/block/loop1 ● dev-mapper-ceph_vg\x2dceph_lv_data.device - /dev/mapper/ceph_vg-ceph_lv_data Follows: unit currently follows state of sys-devices-virtual-block-dm\x2d0.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:26:36 UTC; 25min ago Invocation: 3728047e9e7c49b1b280d8f2047142e7 Device: /sys/devices/virtual/block/dm-0 ● dev-sr0.device - QEMU_DVD-ROM config-2 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:01.1-ata1-host0-target0:0:0-0:0:0:0-block-sr0.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 34f0b4bb29f44e489cbeb0af8eabc599 Device: /sys/devices/pci0000:00/0000:00:01.1/ata1/host0/target0:0:0/0:0:0:0/block/sr0 ○ dev-tpmrm0.device - /dev/tpmrm0 Loaded: loaded Active: inactive (dead) ● dev-ttyS0.device - /dev/ttyS0 Follows: unit currently follows state of sys-devices-pnp0-00:00-00:00:0-00:00:0.0-tty-ttyS0.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 406740f955f74b8296ea17f6c35b0fb5 Device: /sys/devices/pnp0/00:00/00:00:0/00:00:0.0/tty/ttyS0 Apr 14 22:07:34 debian systemd[1]: Found device dev-ttyS0.device - /dev/ttyS0. ● dev-ttyS1.device - /dev/ttyS1 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.1-tty-ttyS1.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: e06437230bbc441ca4df33d93fca4261 Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.1/tty/ttyS1 ● dev-ttyS2.device - /dev/ttyS2 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.2-tty-ttyS2.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 7117c81d63a2457aa8a654cee62756ec Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.2/tty/ttyS2 ● dev-ttyS3.device - /dev/ttyS3 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.3-tty-ttyS3.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: c6e76d48ff3b4f3e8a9b3dc81a1eebc1 Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.3/tty/ttyS3 ● dev-vda.device - /dev/vda Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio1-block-vda.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 4c143c5e434d41a182581ee938e0914a Device: /sys/devices/pci0000:00/0000:00:04.0/virtio1/block/vda ● dev-vda1.device - /dev/vda1 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio1-block-vda-vda1.device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 0e311f4869604da4b4ad7d1d15d612ec Device: /sys/devices/pci0000:00/0000:00:04.0/virtio1/block/vda/vda1 Notice: journal has been rotated since unit was started, output may be incomplete. ● sys-devices-pci0000:00-0000:00:01.1-ata1-host0-target0:0:0-0:0:0:0-block-sr0.device - QEMU_DVD-ROM config-2 Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: b0a8bdb868a44a6e9cd126bbd796b8f8 Device: /sys/devices/pci0000:00/0000:00:01.1/ata1/host0/target0:0:0/0:0:0:0/block/sr0 ● sys-devices-pci0000:00-0000:00:03.0-virtio0-net-ens3.device - Virtio network device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 56d4d4e28c8c4239965b88b11a1e518e Device: /sys/devices/pci0000:00/0000:00:03.0/virtio0/net/ens3 ● sys-devices-pci0000:00-0000:00:04.0-virtio1-block-vda-vda1.device - /sys/devices/pci0000:00/0000:00:04.0/virtio1/block/vda/vda1 Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 43f53f9084f34beface4753f30f1d2f8 Device: /sys/devices/pci0000:00/0000:00:04.0/virtio1/block/vda/vda1 ● sys-devices-pci0000:00-0000:00:04.0-virtio1-block-vda.device - /sys/devices/pci0000:00/0000:00:04.0/virtio1/block/vda Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 06133fa0110c4e2bacde47c5888bb86a Device: /sys/devices/pci0000:00/0000:00:04.0/virtio1/block/vda ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.1-tty-ttyS1.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.1/tty/ttyS1 Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 27b38ba223f0486c8d1d9065cd8453d2 Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.1/tty/ttyS1 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.2-tty-ttyS2.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.2/tty/ttyS2 Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 37b80d995d55485f895934994f2a9388 Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.2/tty/ttyS2 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.3-tty-ttyS3.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.3/tty/ttyS3 Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 09da71d3d6254fd8997e94e054ec3ada Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.3/tty/ttyS3 ● sys-devices-pnp0-00:00-00:00:0-00:00:0.0-tty-ttyS0.device - /sys/devices/pnp0/00:00/00:00:0/00:00:0.0/tty/ttyS0 Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: e196a329625d4c8db121ced01f9cc524 Device: /sys/devices/pnp0/00:00/00:00:0/00:00:0.0/tty/ttyS0 ● sys-devices-virtual-block-dm\x2d0.device - /sys/devices/virtual/block/dm-0 Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:26:36 UTC; 25min ago Invocation: 4d426315078b46d38647c32cf16c973f Device: /sys/devices/virtual/block/dm-0 ● sys-devices-virtual-block-loop0.device - /sys/devices/virtual/block/loop0 Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:26:36 UTC; 25min ago Invocation: 12295e76d3c646c9b514035b708390b9 Device: /sys/devices/virtual/block/loop0 ● sys-devices-virtual-block-loop1.device - /sys/devices/virtual/block/loop1 Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:31:46 UTC; 20min ago Invocation: c3a4a8389f20400e9e63ddbf9ac14161 Device: /sys/devices/virtual/block/loop1 ● sys-devices-virtual-net-br\x2dex.device - /sys/devices/virtual/net/br-ex Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:33:35 UTC; 18min ago Invocation: 9d7bc94b495840d9ab7eb98d8a6a8997 Device: /sys/devices/virtual/net/br-ex ● sys-devices-virtual-net-br\x2dinfra.device - /sys/devices/virtual/net/br-infra Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:12:40 UTC; 39min ago Invocation: 28d88a7d1fac440ea8c88fb4cd98c070 Device: /sys/devices/virtual/net/br-infra ● sys-devices-virtual-net-br\x2dint.device - /sys/devices/virtual/net/br-int Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:23:25 UTC; 28min ago Invocation: 33576d86862944bd929b8c8ef9f0d2b7 Device: /sys/devices/virtual/net/br-int ● sys-devices-virtual-net-lo.device - /sys/devices/virtual/net/lo Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 3807892d188546ddbf87ca53d368bc5e Device: /sys/devices/virtual/net/lo ● sys-devices-virtual-net-ovs\x2dsystem.device - /sys/devices/virtual/net/ovs-system Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:12:40 UTC; 39min ago Invocation: d2b938b4b84b4dfca8c4aa31a3ed3be3 Device: /sys/devices/virtual/net/ovs-system ● sys-devices-virtual-net-vxlan_sys_4789.device - /sys/devices/virtual/net/vxlan_sys_4789 Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:12:48 UTC; 39min ago Invocation: 775de7587db54f11bca7494e1d8d9332 Device: /sys/devices/virtual/net/vxlan_sys_4789 ● sys-module-configfs.device - /sys/module/configfs Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 3e07966c3411434587b98e2ccfb19c4d Device: /sys/module/configfs ● sys-module-fuse.device - /sys/module/fuse Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: af0497ee0aa84af7a25677817918eeec Device: /sys/module/fuse ● sys-subsystem-net-devices-br\x2dex.device - /sys/subsystem/net/devices/br-ex Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:33:35 UTC; 18min ago Invocation: 4ad2170097e8455484a060d6df90b7e9 Device: /sys/devices/virtual/net/br-ex ● sys-subsystem-net-devices-br\x2dinfra.device - /sys/subsystem/net/devices/br-infra Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:12:40 UTC; 39min ago Invocation: bf10e3786420461090b6b48c32d047e5 Device: /sys/devices/virtual/net/br-infra ● sys-subsystem-net-devices-br\x2dint.device - /sys/subsystem/net/devices/br-int Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:23:25 UTC; 28min ago Invocation: 0e4d299bc4114145b08fe86f7c2dbb99 Device: /sys/devices/virtual/net/br-int ● sys-subsystem-net-devices-ens3.device - Virtio network device Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 770645d43b1146ee81361c01251ea9ed Device: /sys/devices/pci0000:00/0000:00:03.0/virtio0/net/ens3 ● sys-subsystem-net-devices-ovs\x2dsystem.device - /sys/subsystem/net/devices/ovs-system Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:12:40 UTC; 39min ago Invocation: 553ce310db414b1c8655e3fc10e85fd7 Device: /sys/devices/virtual/net/ovs-system ● sys-subsystem-net-devices-vxlan_sys_4789.device - /sys/subsystem/net/devices/vxlan_sys_4789 Loaded: loaded Active: active (plugged) since Tue 2026-04-14 22:12:48 UTC; 39min ago Invocation: 46d53a27f253436e977ea4f4acfd2851 Device: /sys/devices/virtual/net/vxlan_sys_4789 ● -.mount - Root Mount Loaded: loaded (/etc/fstab; generated) Active: active (mounted) since Tue 2026-04-14 22:07:33 UTC; 44min ago Where: / What: /dev/vda1 Docs: man:fstab(5) man:systemd-fstab-generator(8) Notice: journal has been rotated since unit was started, output may be incomplete. ● dev-hugepages.mount - Huge Pages File System Loaded: loaded (/usr/lib/systemd/system/dev-hugepages.mount; static) Active: active (mounted) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: f8ffc814ed2040c4af8a45993504538d Where: /dev/hugepages What: hugetlbfs Docs: https://docs.kernel.org/admin-guide/mm/hugetlbpage.html https://systemd.io/API_FILE_SYSTEMS Tasks: 0 (limit: 9256) Memory: 164K (peak: 1.7M) CPU: 28ms CGroup: /dev-hugepages.mount Notice: journal has been rotated since unit was started, output may be incomplete. ● dev-mqueue.mount - POSIX Message Queue File System Loaded: loaded (/proc/self/mountinfo; static) Active: active (mounted) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 91f5714b2d114b458a24e6296cee4d3f Where: /dev/mqueue What: mqueue Docs: man:mq_overview(7) https://systemd.io/API_FILE_SYSTEMS Tasks: 0 (limit: 9256) Memory: 116K (peak: 1.7M) CPU: 26ms CGroup: /dev-mqueue.mount Notice: journal has been rotated since unit was started, output may be incomplete. ● mnt-config.mount - /mnt/config Loaded: loaded (/proc/self/mountinfo) Active: active (mounted) since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: 1fbcab90263c41a9a970d6120e8fc22f Where: /mnt/config What: /dev/sr0 ● opt-stack-data-etcd.mount - /opt/stack/data/etcd Loaded: loaded (/proc/self/mountinfo) Active: active (mounted) since Tue 2026-04-14 22:19:52 UTC; 32min ago Invocation: 8d7b2e9ccf0f4517a06f772fccf56e90 Where: /opt/stack/data/etcd What: tmpfs ● opt-stack-data-swift-drives-sdb1.mount - /opt/stack/data/swift/drives/sdb1 Loaded: loaded (/etc/fstab; generated) Active: active (mounted) since Tue 2026-04-14 22:31:47 UTC; 20min ago Invocation: 54ff4feb6e5d4aa5b5843659942307c7 Where: /opt/stack/data/swift/drives/sdb1 What: /dev/loop1 Docs: man:fstab(5) man:systemd-fstab-generator(8) ● proc-sys-fs-binfmt_misc.mount - Arbitrary Executable File Formats File System Loaded: loaded (/usr/lib/systemd/system/proc-sys-fs-binfmt_misc.mount; disabled; preset: disabled) Active: active (mounted) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: df54eec1a0dd4e0ea7ae52a550dcf494 TriggeredBy: ● proc-sys-fs-binfmt_misc.automount Where: /proc/sys/fs/binfmt_misc What: binfmt_misc Docs: https://docs.kernel.org/admin-guide/binfmt-misc.html https://systemd.io/API_FILE_SYSTEMS Tasks: 0 (limit: 9256) Memory: 8K (peak: 1.9M) CPU: 21ms CGroup: /proc-sys-fs-binfmt_misc.mount Apr 14 22:07:34 debian systemd[1]: Mounting proc-sys-fs-binfmt_misc.mount - Arbitrary Executable File Formats File System... Apr 14 22:07:34 debian systemd[1]: Mounted proc-sys-fs-binfmt_misc.mount - Arbitrary Executable File Formats File System. ● run-lock.mount - Legacy Locks Directory /run/lock Loaded: loaded (/proc/self/mountinfo; disabled; preset: enabled) Active: active (mounted) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 3833cd7ac14c42eaa858a0d5448f5563 Where: /run/lock What: tmpfs Tasks: 0 (limit: 9256) Memory: 104K (peak: 1.7M) CPU: 23ms CGroup: /system.slice/run-lock.mount Notice: journal has been rotated since unit was started, output may be incomplete. ● run-netns.mount - /run/netns Loaded: loaded (/proc/self/mountinfo) Active: active (mounted) since Tue 2026-04-14 22:25:42 UTC; 26min ago Invocation: 3b4e8512ea564e639db5075834f63c94 Where: /run/netns What: tmpfs ● run-user-1000.mount - /run/user/1000 Loaded: loaded (/proc/self/mountinfo) Active: active (mounted) since Tue 2026-04-14 22:09:41 UTC; 42min ago Invocation: 8627f7ff79954eefa0d3e7c49b213df0 Where: /run/user/1000 What: tmpfs ● sys-fs-fuse-connections.mount - FUSE Control File System Loaded: loaded (/usr/lib/systemd/system/sys-fs-fuse-connections.mount; static) Active: active (mounted) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: ecf589c9c6ef4754b389f86df024fab3 Where: /sys/fs/fuse/connections What: fusectl Docs: https://docs.kernel.org/filesystems/fuse.html https://systemd.io/API_FILE_SYSTEMS Tasks: 0 (limit: 9256) Memory: 4K (peak: 1.7M) CPU: 19ms CGroup: /sys-fs-fuse-connections.mount Apr 14 22:07:33 debian systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. ● sys-kernel-config.mount - Kernel Configuration File System Loaded: loaded (/usr/lib/systemd/system/sys-kernel-config.mount; static) Active: active (mounted) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: dfb743f89666408188d010b3d7e7e9fd Where: /sys/kernel/config What: configfs Docs: https://docs.kernel.org/filesystems/configfs.html https://systemd.io/API_FILE_SYSTEMS Tasks: 0 (limit: 9256) Memory: 4K (peak: 1.7M) CPU: 10ms CGroup: /sys-kernel-config.mount Apr 14 22:07:33 debian systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. ● sys-kernel-debug.mount - Kernel Debug File System Loaded: loaded (/proc/self/mountinfo; static) Active: active (mounted) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 5f244d1edda04454adfdbc967f4da1f8 Where: /sys/kernel/debug What: debugfs Docs: https://docs.kernel.org/filesystems/debugfs.html https://systemd.io/API_FILE_SYSTEMS Tasks: 0 (limit: 9256) Memory: 216K (peak: 1.7M) CPU: 22ms CGroup: /sys-kernel-debug.mount Notice: journal has been rotated since unit was started, output may be incomplete. ● sys-kernel-tracing.mount - Kernel Trace File System Loaded: loaded (/proc/self/mountinfo; static) Active: active (mounted) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 4b5558ff826d44feb053d004431016d1 Where: /sys/kernel/tracing What: tracefs Docs: https://docs.kernel.org/trace/ftrace.html https://systemd.io/API_FILE_SYSTEMS Tasks: 0 (limit: 9256) Memory: 4K (peak: 1.7M) CPU: 21ms CGroup: /sys-kernel-tracing.mount Notice: journal has been rotated since unit was started, output may be incomplete. ● tmp.mount - Temporary Directory /tmp Loaded: loaded (/usr/lib/systemd/system/tmp.mount; static) Active: active (mounted) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: be6222762ec84bec964a0d1fc164946c Where: /tmp What: tmpfs Docs: https://systemd.io/TEMPORARY_DIRECTORIES man:file-hierarchy(7) https://systemd.io/API_FILE_SYSTEMS Tasks: 0 (limit: 9256) Memory: 4K (peak: 1.9M) CPU: 21ms CGroup: /system.slice/tmp.mount Notice: journal has been rotated since unit was started, output may be incomplete. ● var-lib-containers-storage-overlay-35cd129b4b90ed18029cbfc03c9f153c30cd9e46db179ac8df66cd10e6f114e7-merged.mount - /var/lib/containers/storage/overlay/35cd129b4b90ed18029cbfc03c9f153c30cd9e46db179ac8df66cd10e6f114e7/merged Loaded: loaded (/proc/self/mountinfo) Active: active (mounted) since Tue 2026-04-14 22:25:55 UTC; 26min ago Invocation: 7feac33140d64fad8702611c560aed3b Where: /var/lib/containers/storage/overlay/35cd129b4b90ed18029cbfc03c9f153c30cd9e46db179ac8df66cd10e6f114e7/merged What: overlay ● var-lib-containers-storage-overlay-3e180a3d54009312700addca927fbf624203846c13fbff5b4aa63cab8c6c7f2c-merged.mount - /var/lib/containers/storage/overlay/3e180a3d54009312700addca927fbf624203846c13fbff5b4aa63cab8c6c7f2c/merged Loaded: loaded (/proc/self/mountinfo) Active: active (mounted) since Tue 2026-04-14 22:27:02 UTC; 24min ago Invocation: 84b8cabc5f5c42b9852889ba5dd697fe Where: /var/lib/containers/storage/overlay/3e180a3d54009312700addca927fbf624203846c13fbff5b4aa63cab8c6c7f2c/merged What: overlay ● var-lib-containers-storage-overlay-4cb29e49758a348d8ea4c41a003a791b18342b2175e965161ea7877b2c1624c6-merged.mount - /var/lib/containers/storage/overlay/4cb29e49758a348d8ea4c41a003a791b18342b2175e965161ea7877b2c1624c6/merged Loaded: loaded (/proc/self/mountinfo) Active: active (mounted) since Tue 2026-04-14 22:25:56 UTC; 25min ago Invocation: 1afa81b7d7924f8bba54378e5d6623ad Where: /var/lib/containers/storage/overlay/4cb29e49758a348d8ea4c41a003a791b18342b2175e965161ea7877b2c1624c6/merged What: overlay ● var-lib-containers-storage-overlay-8b85dd6aeeccfc36246abff3298332ce9a111fd871434707b4970ee7af9e377e-merged.mount - /var/lib/containers/storage/overlay/8b85dd6aeeccfc36246abff3298332ce9a111fd871434707b4970ee7af9e377e/merged Loaded: loaded (/proc/self/mountinfo) Active: active (mounted) since Tue 2026-04-14 22:26:43 UTC; 25min ago Invocation: 281ea995350b45afbd14899ff9fc8034 Where: /var/lib/containers/storage/overlay/8b85dd6aeeccfc36246abff3298332ce9a111fd871434707b4970ee7af9e377e/merged What: overlay ● var-lib-containers-storage-overlay-b027375590453567f4715c261b8082aad4c4ff363d9bbef7a737d0872c8a87b0-merged.mount - /var/lib/containers/storage/overlay/b027375590453567f4715c261b8082aad4c4ff363d9bbef7a737d0872c8a87b0/merged Loaded: loaded (/proc/self/mountinfo) Active: active (mounted) since Tue 2026-04-14 22:26:46 UTC; 25min ago Invocation: ee73d053b6164180833fae434d263707 Where: /var/lib/containers/storage/overlay/b027375590453567f4715c261b8082aad4c4ff363d9bbef7a737d0872c8a87b0/merged What: overlay ● var-lib-containers-storage-overlay.mount - /var/lib/containers/storage/overlay Loaded: loaded (/proc/self/mountinfo) Active: active (mounted) since Tue 2026-04-14 22:25:55 UTC; 26min ago Invocation: 42f16d4cf93d4ae190e5c6bd0fd98540 Where: /var/lib/containers/storage/overlay What: /dev/vda1 ○ var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) Loaded: loaded (/usr/lib/systemd/system/var-lib-machines.mount; static) Active: inactive (dead) Condition: start condition unmet at Tue 2026-04-14 22:24:05 UTC; 27min ago Where: /var/lib/machines What: /var/lib/machines.raw Apr 14 22:23:56 npc99a84aae6d34 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Apr 14 22:24:05 npc99a84aae6d34 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). ● acpid.path - ACPI Events Check Loaded: loaded (/usr/lib/systemd/system/acpid.path; enabled; preset: enabled) Active: active (waiting) since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: a46de782439348a3ac63a26d42baad25 Triggers: ● acpid.service Apr 14 22:07:35 debian systemd[1]: Started acpid.path - ACPI Events Check. ● systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch Loaded: loaded (/usr/lib/systemd/system/systemd-ask-password-console.path; static) Active: active (waiting) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: d7df0796a75d44bcb3eff52dda555e70 Triggers: ● systemd-ask-password-console.service Docs: man:systemd-ask-password-console.path(8) Notice: journal has been rotated since unit was started, output may be incomplete. ● systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch Loaded: loaded (/usr/lib/systemd/system/systemd-ask-password-wall.path; static) Active: active (waiting) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 4f9df1a36eb14b768dcddd24dc06e69d Triggers: ● systemd-ask-password-wall.service Docs: man:systemd-ask-password-wall.path(8) Notice: journal has been rotated since unit was started, output may be incomplete. ● init.scope - System and Service Manager Loaded: loaded Transient: yes Active: active (running) since Tue 2026-04-14 22:07:33 UTC; 44min ago Docs: man:systemd(1) Tasks: 1 (limit: 9256) Memory: 13.4M (peak: 28.6M, swap: 0B, swap peak: 4K) CPU: 1min 33.224s CGroup: /init.scope └─1 /sbin/init nofb Apr 14 22:39:10 npc99a84aae6d34 systemd[1]: Started session-30.scope - Session 30 of User zuul. Apr 14 22:39:11 npc99a84aae6d34 systemd[1]: session-30.scope: Deactivated successfully. Apr 14 22:39:13 npc99a84aae6d34 systemd[1]: Started session-31.scope - Session 31 of User zuul. Apr 14 22:39:14 npc99a84aae6d34 systemd[1]: session-31.scope: Deactivated successfully. Apr 14 22:50:12 npc99a84aae6d34 systemd[1]: Started session-32.scope - Session 32 of User zuul. Apr 14 22:50:34 npc99a84aae6d34 systemd[1]: session-32.scope: Deactivated successfully. Apr 14 22:50:34 npc99a84aae6d34 systemd[1]: session-32.scope: Consumed 11.856s CPU time, 123.4M memory peak. Apr 14 22:50:36 npc99a84aae6d34 systemd[1]: Started session-33.scope - Session 33 of User zuul. Apr 14 22:50:43 npc99a84aae6d34 systemd[1]: session-33.scope: Deactivated successfully. Apr 14 22:50:46 npc99a84aae6d34 systemd[1]: Started session-34.scope - Session 34 of User zuul. ● session-1.scope - Session 1 of User zuul Loaded: loaded (/run/systemd/transient/session-1.scope; transient) Transient: yes Active: active (abandoned) since Tue 2026-04-14 22:09:41 UTC; 42min ago Invocation: eeb4155d5ff44c1787ee919be9933fd5 Tasks: 2 Memory: 1G (peak: 7G, swap: 64K, swap peak: 108K, zswap: 1.3K) CPU: 16min 50.580s CGroup: /user.slice/user-1000.slice/session-1.scope └─1087 /usr/bin/python3.13 Apr 14 22:39:22 npc99a84aae6d34 sudo[118174]: pam_unix(sudo:session): session opened for user root(uid=0) by zuul(uid=1000) Apr 14 22:39:22 npc99a84aae6d34 sudo[118179]: zuul : PWD=/home/zuul ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-fveuspxemvcebhrsagexdojmwynlmziy ; /usr/bin/python3.13' Apr 14 22:39:22 npc99a84aae6d34 sudo[118179]: pam_unix(sudo:session): session opened for user root(uid=0) by zuul(uid=1000) Apr 14 22:39:23 npc99a84aae6d34 python3.13[118177]: ansible-ansible.posix.synchronize Invoked with owner=True group=True perms=True src=/etc/ceph/ dest=zuul@213.32.73.159:/etc/ceph/ _local_rsync_path=rsync dest_port=22 private_key=None _local_rsync_password=NOT_LOGGING_PARAMETER rsync_path=sudo -u root rsync delete=False _substitute_controller=False archive=True checksum=False compress=True existing_only=False dirs=False copy_links=False set_remote_user=True rsync_timeout=0 rsync_opts=[] use_ssh_args=False ssh_connection_multiplexing=False partial=False verify_host=False delay_updates=True mode=push recursive=None links=None times=None _ssh_args=None link_dest=None Apr 14 22:39:23 npc99a84aae6d34 python3.13[118182]: ansible-ansible.posix.synchronize Invoked with owner=True group=True perms=True src=/etc/ceph/ dest=/etc/ceph/ _local_rsync_path=rsync dest_port=22 _local_rsync_password=NOT_LOGGING_PARAMETER rsync_path=None delete=False _substitute_controller=False archive=True checksum=False compress=True existing_only=False dirs=False copy_links=False set_remote_user=True rsync_timeout=0 rsync_opts=[] use_ssh_args=False ssh_connection_multiplexing=False partial=False verify_host=False delay_updates=True mode=push private_key=None recursive=None links=None times=None _ssh_args=None link_dest=None Apr 14 22:39:23 npc99a84aae6d34 sudo[118179]: pam_unix(sudo:session): session closed for user root Apr 14 22:39:23 npc99a84aae6d34 sudo[118174]: pam_unix(sudo:session): session closed for user root Apr 14 22:40:23 npc99a84aae6d34 sshd-session[884]: Received disconnect from 104.130.127.165 port 47140:11: disconnected by user Apr 14 22:40:23 npc99a84aae6d34 sshd-session[884]: Disconnected from user zuul 104.130.127.165 port 47140 Apr 14 22:40:23 npc99a84aae6d34 sshd-session[854]: pam_unix(sshd:session): session closed for user zuul ● session-34.scope - Session 34 of User zuul Loaded: loaded (/run/systemd/transient/session-34.scope; transient) Transient: yes Active: active (running) since Tue 2026-04-14 22:50:46 UTC; 1min 9s ago Invocation: bcd7832ab2054a978b2ffa5d629ecd7f Tasks: 10 Memory: 59.8M (peak: 79.7M) CPU: 31.073s CGroup: /user.slice/user-1000.slice/session-34.scope ├─120787 "sshd-session: zuul [priv]" ├─120794 "sshd-session: zuul@notty" ├─122625 /bin/sh -c "sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-mjejnubreskapnsupqhyuesybxywnovx ; /usr/bin/python3.13' && sleep 0" ├─122626 sudo -H -S -n -u root /bin/sh -c "echo BECOME-SUCCESS-mjejnubreskapnsupqhyuesybxywnovx ; /usr/bin/python3.13" ├─122628 /bin/sh -c "echo BECOME-SUCCESS-mjejnubreskapnsupqhyuesybxywnovx ; /usr/bin/python3.13" ├─122629 /usr/bin/python3.13 ├─122630 /bin/bash -c "sudo iptables-save > /home/zuul/iptables.txt\n\n# NOTE(sfernand): Run 'df' with a 60s timeout to prevent hangs from\n# stale NFS mounts.\ntimeout -s 9 60s df -h > /home/zuul/df.txt || true\n# If 'df' times out, the mount output helps debug which NFS share\n# is unresponsive.\nmount > /home/zuul/mount.txt\n\nfor py_ver in 2 3; do\n if [[ \`which python\${py_ver}\` ]]; then\n python\${py_ver} -m pip freeze > /home/zuul/pip\${py_ver}-freeze.txt\n fi\ndone\n\nif [ \`command -v dpkg\` ]; then\n dpkg -l> /home/zuul/dpkg-l.txt\nfi\nif [ \`command -v rpm\` ]; then\n rpm -qa | sort > /home/zuul/rpm-qa.txt\nfi\n\n# Services status\nsudo systemctl status --all > services.txt 2>/dev/null\n\n# NOTE(kchamart) The 'audit.log' can be useful in cases when QEMU\n# failed to start due to denials from SELinux — useful for CentOS\n# and Fedora machines. For Ubuntu (which runs AppArmor), DevStack\n# already captures the contents of /var/log/kern.log (via\n# \`journalctl -t kernel\` redirected into syslog.txt.gz), which\n# contains AppArmor-related messages.\nif [ -f /var/log/audit/audit.log ] ; then\n sudo cp /var/log/audit/audit.log /home/zuul/audit.log &&\n chmod +r /home/zuul/audit.log;\nfi\n\n# gzip and save any coredumps in /var/core\nif [ -d /var/core ]; then\n sudo gzip -r /var/core\n sudo cp -r /var/core /home/zuul/\nfi\n\nsudo ss -lntup | grep ':53' > /home/zuul/listen53.txt\n\n# NOTE(andreaf) Service logs are already in logs/ thanks for the\n# export-devstack-journal log. Apache logs are under apache/ thans to the\n# apache-logs-conf role.\ngrep -i deprecat /home/zuul/logs/*.txt /home/zuul/apache/*.log | \\\n sed -r 's/[0-9]{1,2}\\:[0-9]{1,2}\\:[0-9]{1,2}\\.[0-9]{1,3}/ /g' | \\\n sed -r 's/[0-9]{1,2}\\:[0-9]{1,2}\\:[0-9]{1,2}/ /g' | \\\n sed -r 's/[0-9]{1,4}-[0-9]{1,2}-[0-9]{1,4}/ /g' |\n sed -r 's/\\[.*\\]/ /g' | \\\n sed -r 's/\\s[0-9]+\\s/ /g' | \\\n awk '{if (\$0 in seen) {seen[\$0]++} else {out[++n]=\$0;seen[\$0]=1}} END { for (i=1; i<=n; i++) print seen[out[i]]\" :: \" out[i] }' > /home/zuul/deprecations.log\n" ├─122644 sudo systemctl status --all └─122646 systemctl status --all Apr 14 22:51:53 npc99a84aae6d34 python3.13[122619]: ansible-ansible.legacy.command Invoked with _raw_params=cp -pRL /etc/openstack /home/zuul/etc/ zuul_no_log=False zuul_log_id=bc764e01-af26-4e37-8285-000000000030-1-controller zuul_output_max_bytes=1073741824 zuul_ansible_split_streams=False _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Apr 14 22:51:53 npc99a84aae6d34 sudo[122616]: pam_unix(sudo:session): session closed for user root Apr 14 22:51:54 npc99a84aae6d34 sudo[122626]: zuul : PWD=/home/zuul ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-mjejnubreskapnsupqhyuesybxywnovx ; /usr/bin/python3.13' Apr 14 22:51:54 npc99a84aae6d34 sudo[122626]: pam_unix(sudo:session): session opened for user root(uid=0) by zuul(uid=1000) Apr 14 22:51:55 npc99a84aae6d34 python3.13[122629]: ansible-ansible.legacy.command Invoked with executable=/bin/bash _raw_params=sudo iptables-save > /home/zuul/iptables.txt # NOTE(sfernand): Run 'df' with a 60s timeout to prevent hangs from # stale NFS mounts. timeout -s 9 60s df -h > /home/zuul/df.txt || true # If 'df' times out, the mount output helps debug which NFS share # is unresponsive. mount > /home/zuul/mount.txt for py_ver in 2 3; do if [[ `which python${py_ver}` ]]; then python${py_ver} -m pip freeze > /home/zuul/pip${py_ver}-freeze.txt fi done if [ `command -v dpkg` ]; then dpkg -l> /home/zuul/dpkg-l.txt fi if [ `command -v rpm` ]; then rpm -qa | sort > /home/zuul/rpm-qa.txt fi # Services status sudo systemctl status --all > services.txt 2>/dev/null # NOTE(kchamart) The 'audit.log' can be useful in cases when QEMU # failed to start due to denials from SELinux — useful for CentOS # and Fedora machines. For Ubuntu (which runs AppArmor), DevStack # already captures the contents of /var/log/kern.log (via # `journalctl -t kernel` redirected into syslog.txt.gz), which # contains AppArmor-related messages. if [ -f /var/log/audit/audit.log ] ; then sudo cp /var/log/audit/audit.log /home/zuul/audit.log && chmod +r /home/zuul/audit.log; fi # gzip and save any coredumps in /var/core if [ -d /var/core ]; then sudo gzip -r /var/core sudo cp -r /var/core /home/zuul/ fi sudo ss -lntup | grep ':53' > /home/zuul/listen53.txt # NOTE(andreaf) Service logs are already in logs/ thanks for the # export-devstack-journal log. Apache logs are under apache/ thans to the # apache-logs-conf role. grep -i deprecat /home/zuul/logs/*.txt /home/zuul/apache/*.log | \ sed -r 's/[0-9]{1,2}\:[0-9]{1,2}\:[0-9]{1,2}\.[0-9]{1,3}/ /g' | \ sed -r 's/[0-9]{1,2}\:[0-9]{1,2}\:[0-9]{1,2}/ /g' | \ sed -r 's/[0-9]{1,4}-[0-9]{1,2}-[0-9]{1,4}/ /g' | sed -r 's/\[.*\]/ /g' | \ sed -r 's/\s[0-9]+\s/ /g' | \ awk '{if ($0 in seen) {seen[$0]++} else {out[++n]=$0;seen[$0]=1}} END { for (i=1; i<=n; i++) print seen[out[i]]" :: " out[i] }' > /home/zuul/deprecations.log _uses_shell=True zuul_no_log=False zuul_log_id=bc764e01-af26-4e37-8285-000000000034-1-controller zuul_output_max_bytes=1073741824 zuul_ansible_split_streams=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None creates=None removes=None stdin=None Apr 14 22:51:55 npc99a84aae6d34 sudo[122632]: root : PWD=/home/zuul ; USER=root ; COMMAND=/usr/sbin/iptables-save Apr 14 22:51:55 npc99a84aae6d34 sudo[122632]: pam_unix(sudo:session): session opened for user root(uid=0) by zuul(uid=0) Apr 14 22:51:55 npc99a84aae6d34 sudo[122632]: pam_unix(sudo:session): session closed for user root Apr 14 22:51:55 npc99a84aae6d34 sudo[122644]: root : PWD=/home/zuul ; USER=root ; COMMAND=/usr/bin/systemctl status --all Apr 14 22:51:55 npc99a84aae6d34 sudo[122644]: pam_unix(sudo:session): session opened for user root(uid=0) by zuul(uid=0) ○ acpid.service - ACPI event daemon Loaded: loaded (/usr/lib/systemd/system/acpid.service; disabled; preset: enabled) Active: inactive (dead) TriggeredBy: ● acpid.path ● acpid.socket Docs: man:acpid(8) ● apache-htcacheclean.service - Disk Cache Cleaning Daemon for Apache HTTP Server Loaded: loaded (/usr/lib/systemd/system/apache-htcacheclean.service; disabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:17:54 UTC; 34min ago Invocation: 504c9f34ed784c94ae6f284303343daa Docs: https://httpd.apache.org/docs/2.4/programs/htcacheclean.html Main PID: 15613 (htcacheclean) Tasks: 1 (limit: 9256) Memory: 288K (peak: 1.8M) CPU: 131ms CGroup: /system.slice/apache-htcacheclean.service └─15613 /usr/bin/htcacheclean -d 120 -p /var/cache/apache2/mod_cache_disk -l 300M -n Apr 14 22:17:54 npc99a84aae6d34 systemd[1]: Starting apache-htcacheclean.service - Disk Cache Cleaning Daemon for Apache HTTP Server... Apr 14 22:17:54 npc99a84aae6d34 systemd[1]: Started apache-htcacheclean.service - Disk Cache Cleaning Daemon for Apache HTTP Server. ● apache2.service - The Apache HTTP Server Loaded: loaded (/usr/lib/systemd/system/apache2.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:33:14 UTC; 18min ago Invocation: 6cbddb20b55b48d9a0eec44031eb0069 Docs: https://httpd.apache.org/docs/2.4/ Main PID: 91335 (apache2) Tasks: 69 (limit: 9256) Memory: 18M (peak: 34.1M, swap: 3.1M, swap peak: 3.1M, zswap: 734.8K) CPU: 2.620s CGroup: /system.slice/apache2.service ├─91335 /usr/sbin/apache2 -k start ├─91340 /usr/sbin/apache2 -k start └─91341 /usr/sbin/apache2 -k start Apr 14 22:33:14 npc99a84aae6d34 systemd[1]: Starting apache2.service - The Apache HTTP Server... Apr 14 22:33:14 npc99a84aae6d34 apachectl[91334]: AH00558: apache2: Could not reliably determine the server's fully qualified domain name, using 2001:41d0:302:1000::cbc. Set the 'ServerName' directive globally to suppress this message Apr 14 22:33:14 npc99a84aae6d34 systemd[1]: Started apache2.service - The Apache HTTP Server. ○ apt-daily-upgrade.service - Daily apt upgrade and clean activities Loaded: loaded (/usr/lib/systemd/system/apt-daily-upgrade.service; static) Active: inactive (dead) TriggeredBy: ● apt-daily-upgrade.timer Docs: man:apt(8) ○ apt-daily.service - Daily apt download activities Loaded: loaded (/usr/lib/systemd/system/apt-daily.service; static) Active: inactive (dead) TriggeredBy: ● apt-daily.timer Docs: man:apt(8) ● atd.service - Deferred execution scheduler Loaded: loaded (/usr/lib/systemd/system/atd.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:07:36 UTC; 44min ago Invocation: 564048987d274900accbd54364e3fcb9 Docs: man:atd(8) Main PID: 814 (atd) Tasks: 1 (limit: 9256) Memory: 312K (peak: 2M) CPU: 38ms CGroup: /system.slice/atd.service └─814 /usr/sbin/atd -f Apr 14 22:07:36 npc99a84aae6d34 systemd[1]: Starting atd.service - Deferred execution scheduler... Apr 14 22:07:36 npc99a84aae6d34 systemd[1]: Started atd.service - Deferred execution scheduler. ● blk-availability.service - Availability of block devices Loaded: loaded (/usr/lib/systemd/system/blk-availability.service; enabled; preset: enabled) Active: active (exited) since Tue 2026-04-14 22:18:00 UTC; 33min ago Invocation: f75bc04a493b461b9d5525664574428a Main PID: 16088 (code=exited, status=0/SUCCESS) Mem peak: 1.7M CPU: 20ms Apr 14 22:17:59 npc99a84aae6d34 systemd[1]: Starting blk-availability.service - Availability of block devices... Apr 14 22:18:00 npc99a84aae6d34 systemd[1]: Finished blk-availability.service - Availability of block devices. ● ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@crash.npc99a84aae6d34.service - Ceph crash.npc99a84aae6d34 for 929e05b1-4e8c-4c8f-b13f-4521b44f10e1 Loaded: loaded (/etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:26:43 UTC; 25min ago Invocation: 737d53e2b88c4b2e8842f3ccc6d3924a Main PID: 54981 (conmon) Tasks: 3 (limit: 9256) Memory: 4M (peak: 24M, swap: 6M, swap peak: 6.1M, zswap: 2M) CPU: 705ms CGroup: /system.slice/system-ceph\x2d929e05b1\x2d4e8c\x2d4c8f\x2db13f\x2d4521b44f10e1.slice/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@crash.npc99a84aae6d34.service ├─libpod-payload-91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc │ ├─54984 /run/podman-init -- /usr/bin/ceph-crash -n client.crash.npc99a84aae6d34 │ └─54986 /usr/bin/python3 -s /usr/bin/ceph-crash -n client.crash.npc99a84aae6d34 └─runtime └─54981 /usr/bin/conmon --api-version 1 -c 91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc -u 91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc/userdata -p /run/containers/storage/overlay-containers/91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc/userdata/pidfile -n ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1-crash-npc99a84aae6d34 --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc --full-attach -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc/userdata/oci-log --conmon-pidfile /run/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@crash.npc99a84aae6d34.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg "" --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg --rm --exit-command-arg 91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc Apr 14 22:34:54 npc99a84aae6d34 systemd[1]: /etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@.service:22: Unit uses KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update the service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. Apr 14 22:34:55 npc99a84aae6d34 systemd[1]: /etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@.service:22: Unit uses KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update the service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. Apr 14 22:34:57 npc99a84aae6d34 systemd[1]: /etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@.service:22: Unit uses KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update the service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. Apr 14 22:34:57 npc99a84aae6d34 systemd[1]: /etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@.service:22: Unit uses KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update the service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. Apr 14 22:35:02 npc99a84aae6d34 systemd[1]: /etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@.service:22: Unit uses KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update the service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. Apr 14 22:35:02 npc99a84aae6d34 systemd[1]: /etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@.service:22: Unit uses KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update the service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. Apr 14 22:35:23 npc99a84aae6d34 systemd[1]: /etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@.service:22: Unit uses KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update the service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. Apr 14 22:35:23 npc99a84aae6d34 systemd[1]: /etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@.service:22: Unit uses KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update the service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. Apr 14 22:35:25 npc99a84aae6d34 systemd[1]: /etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@.service:22: Unit uses KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update the service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. Apr 14 22:35:26 npc99a84aae6d34 systemd[1]: /etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@.service:22: Unit uses KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update the service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. ● ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mgr.npc99a84aae6d34.krobmi.service - Ceph mgr.npc99a84aae6d34.krobmi for 929e05b1-4e8c-4c8f-b13f-4521b44f10e1 Loaded: loaded (/etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:26:46 UTC; 25min ago Invocation: 8633ff64f5414ff48ae71ca03cbf2463 Main PID: 55417 (conmon) Tasks: 19 (limit: 9256) Memory: 422.9M (peak: 473.6M, swap: 64.1M, swap peak: 64.1M, zswap: 23.2M) CPU: 22.001s CGroup: /system.slice/system-ceph\x2d929e05b1\x2d4e8c\x2d4c8f\x2db13f\x2d4521b44f10e1.slice/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mgr.npc99a84aae6d34.krobmi.service ├─libpod-payload-3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274 │ ├─55422 /run/podman-init -- /usr/bin/ceph-mgr -n mgr.npc99a84aae6d34.krobmi -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false │ └─55424 /usr/bin/ceph-mgr -n mgr.npc99a84aae6d34.krobmi -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false └─runtime └─55417 /usr/bin/conmon --api-version 1 -c 3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274 -u 3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274 -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274/userdata -p /run/containers/storage/overlay-containers/3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274/userdata/pidfile -n ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1-mgr-npc99a84aae6d34-krobmi --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274 --full-attach -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274/userdata/oci-log --conmon-pidfile /run/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mgr.npc99a84aae6d34.krobmi.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg "" --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg --rm --exit-command-arg 3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274 Apr 14 22:34:54 npc99a84aae6d34 systemd[1]: /etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@.service:22: Unit uses KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update the service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. Apr 14 22:34:55 npc99a84aae6d34 systemd[1]: /etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@.service:22: Unit uses KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update the service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. Apr 14 22:34:57 npc99a84aae6d34 systemd[1]: /etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@.service:22: Unit uses KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update the service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. Apr 14 22:34:57 npc99a84aae6d34 systemd[1]: /etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@.service:22: Unit uses KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update the service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. Apr 14 22:35:02 npc99a84aae6d34 systemd[1]: /etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@.service:22: Unit uses KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update the service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. Apr 14 22:35:02 npc99a84aae6d34 systemd[1]: /etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@.service:22: Unit uses KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update the service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. Apr 14 22:35:23 npc99a84aae6d34 systemd[1]: /etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@.service:22: Unit uses KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update the service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. Apr 14 22:35:23 npc99a84aae6d34 systemd[1]: /etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@.service:22: Unit uses KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update the service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. Apr 14 22:35:25 npc99a84aae6d34 systemd[1]: /etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@.service:22: Unit uses KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update the service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. Apr 14 22:35:26 npc99a84aae6d34 systemd[1]: /etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@.service:22: Unit uses KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update the service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. ● ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mgr.npc99a84aae6d34.mxlymt.service - Ceph mgr.npc99a84aae6d34.mxlymt for 929e05b1-4e8c-4c8f-b13f-4521b44f10e1 Loaded: loaded (/etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:25:56 UTC; 25min ago Invocation: 75a3044ec08a4de29987e63065e5db39 Main PID: 51958 (conmon) Tasks: 90 (limit: 9256) Memory: 448.2M (peak: 503.4M, swap: 61.7M, swap peak: 61.8M, zswap: 22.1M) CPU: 47.629s CGroup: /system.slice/system-ceph\x2d929e05b1\x2d4e8c\x2d4c8f\x2db13f\x2d4521b44f10e1.slice/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mgr.npc99a84aae6d34.mxlymt.service ├─libpod-payload-c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73 │ ├─51960 /run/podman-init -- /usr/bin/ceph-mgr -n mgr.npc99a84aae6d34.mxlymt -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false │ └─51962 /usr/bin/ceph-mgr -n mgr.npc99a84aae6d34.mxlymt -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false └─runtime └─51958 /usr/bin/conmon --api-version 1 -c c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73 -u c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73 -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73/userdata -p /run/containers/storage/overlay-containers/c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73/userdata/pidfile -n ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1-mgr-npc99a84aae6d34-mxlymt --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73 --full-attach -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73/userdata/oci-log --conmon-pidfile /run/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mgr.npc99a84aae6d34.mxlymt.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg "" --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg --rm --exit-command-arg c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73 Apr 14 22:51:42 npc99a84aae6d34 ceph-mgr[51962]: [volumes INFO mgr_util] cleaning up connections: [] Apr 14 22:51:42 npc99a84aae6d34 ceph-mgr[51962]: [volumes INFO mgr_util] scanning for idle connections.. Apr 14 22:51:42 npc99a84aae6d34 ceph-mgr[51962]: [volumes INFO mgr_util] cleaning up connections: [] Apr 14 22:51:42 npc99a84aae6d34 ceph-mgr[51962]: log_channel(cluster) log [DBG] : pgmap v571: 129 pgs: 129 active+clean; 62 MiB data, 88 MiB used, 30 GiB / 30 GiB avail Apr 14 22:51:44 npc99a84aae6d34 ceph-mgr[51962]: log_channel(cluster) log [DBG] : pgmap v572: 129 pgs: 129 active+clean; 62 MiB data, 88 MiB used, 30 GiB / 30 GiB avail Apr 14 22:51:46 npc99a84aae6d34 ceph-mgr[51962]: log_channel(cluster) log [DBG] : pgmap v573: 129 pgs: 129 active+clean; 62 MiB data, 88 MiB used, 30 GiB / 30 GiB avail Apr 14 22:51:48 npc99a84aae6d34 ceph-mgr[51962]: log_channel(cluster) log [DBG] : pgmap v574: 129 pgs: 129 active+clean; 62 MiB data, 88 MiB used, 30 GiB / 30 GiB avail Apr 14 22:51:50 npc99a84aae6d34 ceph-mgr[51962]: log_channel(cluster) log [DBG] : pgmap v575: 129 pgs: 129 active+clean; 62 MiB data, 88 MiB used, 30 GiB / 30 GiB avail Apr 14 22:51:52 npc99a84aae6d34 ceph-mgr[51962]: log_channel(cluster) log [DBG] : pgmap v576: 129 pgs: 129 active+clean; 62 MiB data, 88 MiB used, 30 GiB / 30 GiB avail Apr 14 22:51:54 npc99a84aae6d34 ceph-mgr[51962]: log_channel(cluster) log [DBG] : pgmap v577: 129 pgs: 129 active+clean; 62 MiB data, 88 MiB used, 30 GiB / 30 GiB avail ● ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mon.npc99a84aae6d34.service - Ceph mon.npc99a84aae6d34 for 929e05b1-4e8c-4c8f-b13f-4521b44f10e1 Loaded: loaded (/etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:25:55 UTC; 26min ago Invocation: 049c4e15fb354c25b7db407cffa47c80 Main PID: 51750 (conmon) Tasks: 27 (limit: 9256) Memory: 52.6M (peak: 64.7M, swap: 5.3M, swap peak: 5.3M, zswap: 803.5K) CPU: 19.242s CGroup: /system.slice/system-ceph\x2d929e05b1\x2d4e8c\x2d4c8f\x2db13f\x2d4521b44f10e1.slice/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mon.npc99a84aae6d34.service ├─libpod-payload-b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942 │ ├─51752 /run/podman-init -- /usr/bin/ceph-mon -n mon.npc99a84aae6d34 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false │ └─51754 /usr/bin/ceph-mon -n mon.npc99a84aae6d34 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false └─runtime └─51750 /usr/bin/conmon --api-version 1 -c b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942 -u b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942 -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942/userdata -p /run/containers/storage/overlay-containers/b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942/userdata/pidfile -n ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1-mon-npc99a84aae6d34 --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942 --full-attach -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942/userdata/oci-log --conmon-pidfile /run/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mon.npc99a84aae6d34.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg "" --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg --rm --exit-command-arg b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942 Apr 14 22:51:46 npc99a84aae6d34 ceph-mon[51754]: rocksdb: (Original Log Time 2026/04/14-22:51:46.469885) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Apr 14 22:51:46 npc99a84aae6d34 ceph-mon[51754]: rocksdb: (Original Log Time 2026/04/14-22:51:46.469888) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Apr 14 22:51:46 npc99a84aae6d34 ceph-mon[51754]: rocksdb: (Original Log Time 2026/04/14-22:51:46.469892) [db/db_impl/db_impl_compaction_flush.cc:1903] [default] Manual compaction starting Apr 14 22:51:47 npc99a84aae6d34 ceph-mon[51754]: pgmap v573: 129 pgs: 129 active+clean; 62 MiB data, 88 MiB used, 30 GiB / 30 GiB avail Apr 14 22:51:47 npc99a84aae6d34 ceph-mon[51754]: mon.npc99a84aae6d34@0(leader).osd e40 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Apr 14 22:51:50 npc99a84aae6d34 ceph-mon[51754]: pgmap v574: 129 pgs: 129 active+clean; 62 MiB data, 88 MiB used, 30 GiB / 30 GiB avail Apr 14 22:51:52 npc99a84aae6d34 ceph-mon[51754]: pgmap v575: 129 pgs: 129 active+clean; 62 MiB data, 88 MiB used, 30 GiB / 30 GiB avail Apr 14 22:51:52 npc99a84aae6d34 ceph-mon[51754]: mon.npc99a84aae6d34@0(leader).osd e40 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Apr 14 22:51:54 npc99a84aae6d34 ceph-mon[51754]: pgmap v576: 129 pgs: 129 active+clean; 62 MiB data, 88 MiB used, 30 GiB / 30 GiB avail Apr 14 22:51:56 npc99a84aae6d34 ceph-mon[51754]: pgmap v577: 129 pgs: 129 active+clean; 62 MiB data, 88 MiB used, 30 GiB / 30 GiB avail ● ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@osd.0.service - Ceph osd.0 for 929e05b1-4e8c-4c8f-b13f-4521b44f10e1 Loaded: loaded (/etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:27:02 UTC; 24min ago Invocation: 595b9355e53743edbe5711f4e2283796 Main PID: 57928 (conmon) Tasks: 61 (limit: 9256) Memory: 130.6M (peak: 518M, swap: 10.7M, swap peak: 10.9M, zswap: 1015.7K) CPU: 12.400s CGroup: /system.slice/system-ceph\x2d929e05b1\x2d4e8c\x2d4c8f\x2db13f\x2d4521b44f10e1.slice/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@osd.0.service ├─libpod-payload-45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee │ ├─57930 /run/podman-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --osd-objectstore=bluestore │ └─57932 /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --osd-objectstore=bluestore └─runtime └─57928 /usr/bin/conmon --api-version 1 -c 45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee -u 45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee/userdata -p /run/containers/storage/overlay-containers/45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee/userdata/pidfile -n ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1-osd-0 --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee --full-attach -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee/userdata/oci-log --conmon-pidfile /run/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@osd.0.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg "" --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg --rm --exit-command-arg 45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee Apr 14 22:40:19 npc99a84aae6d34 ceph-osd[57932]: log_channel(cluster) log [DBG] : 2.4 scrub starts Apr 14 22:40:19 npc99a84aae6d34 ceph-osd[57932]: log_channel(cluster) log [DBG] : 2.4 scrub ok Apr 14 22:40:20 npc99a84aae6d34 ceph-osd[57932]: log_channel(cluster) log [DBG] : 2.2 scrub starts Apr 14 22:40:20 npc99a84aae6d34 ceph-osd[57932]: log_channel(cluster) log [DBG] : 2.2 scrub ok Apr 14 22:40:25 npc99a84aae6d34 ceph-osd[57932]: log_channel(cluster) log [DBG] : 2.1 scrub starts Apr 14 22:40:25 npc99a84aae6d34 ceph-osd[57932]: log_channel(cluster) log [DBG] : 2.1 scrub ok Apr 14 22:40:31 npc99a84aae6d34 ceph-osd[57932]: log_channel(cluster) log [DBG] : 2.0 scrub starts Apr 14 22:40:31 npc99a84aae6d34 ceph-osd[57932]: log_channel(cluster) log [DBG] : 2.0 scrub ok Apr 14 22:47:03 npc99a84aae6d34 ceph-osd[57932]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- Apr 14 22:47:03 npc99a84aae6d34 ceph-osd[57932]: rocksdb: [db/db_impl/db_impl.cc:1111] ** DB Stats ** Uptime(secs): 1200.1 total, 600.0 interval Cumulative writes: 4753 writes, 21K keys, 4753 commit groups, 1.0 writes per commit group, ingest: 0.02 GB, 0.01 MB/s Cumulative WAL: 4753 writes, 643 syncs, 7.39 writes per sync, written: 0.02 GB, 0.01 MB/s Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent Interval writes: 300 writes, 373 keys, 300 commit groups, 1.0 writes per commit group, ingest: 0.09 MB, 0.00 MB/s Interval WAL: 300 writes, 150 syncs, 2.00 writes per sync, written: 0.00 GB, 0.00 MB/s Interval stall: 00:00:0.000 H:M:S, 0.0 percent ** Compaction Stats [default] ** Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ L0 2/0 2.65 KB 0.2 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.3 0.00 0.00 1 0.004 0 0 0.0 0.0 Sum 2/0 2.65 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.3 0.00 0.00 1 0.004 0 0 0.0 0.0 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 ** Compaction Stats [default] ** Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.3 0.00 0.00 1 0.004 0 0 0.0 0.0 Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 Uptime(secs): 1200.1 total, 600.0 interval Flush(GB): cumulative 0.000, interval 0.000 AddFile(GB): cumulative 0.000, interval 0.000 AddFile(Total Files): cumulative 0, interval 0 AddFile(L0 Files): cumulative 0, interval 0 AddFile(Keys): cumulative 0, interval 0 Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count Block cache BinnedLRUCache@0x55a7f20e31f0#2 capacity: 388.00 MB usage: 2.09 KB table_size: 0 occupancy: 18446744073709551615 collections: 3 last_copies: 8 last_secs: 0.000104 secs_since: 0 Block cache entry stats(count,size,portion): DataBlock(3,1.42 KB,0.000357874%) FilterBlock(3,0.33 KB,8.25862e-05%) IndexBlock(3,0.34 KB,8.65189e-05%) Misc(1,0.00 KB,0%) ** File Read Latency Histogram By Level [default] ** ** Compaction Stats [m-0] ** Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 ** Compaction Stats [m-0] ** Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 Uptime(secs): 1200.1 total, 600.0 interval Flush(GB): cumulative 0.000, interval 0.000 AddFile(GB): cumulative 0.000, interval 0.000 AddFile(Total Files): cumulative 0, interval 0 AddFile(L0 Files): cumulative 0, interval 0 AddFile(Keys): cumulative 0, interval 0 Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count Block cache BinnedLRUCache@0x55a7f20e31f0#2 capacity: 388.00 MB usage: 2.09 KB table_size: 0 occupancy: 18446744073709551615 collections: 3 last_copies: 8 last_secs: 0.000104 secs_since: 0 Block cache entry stats(count,size,portion): DataBlock(3,1.42 KB,0.000357874%) FilterBlock(3,0.33 KB,8.25862e-05%) IndexBlock(3,0.34 KB,8.65189e-05%) Misc(1,0.00 KB,0%) ** File Read Latency Histogram By Level [m-0] ** ** Compaction Stats [m-1] ** Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 ** Compaction Stats [m-1] ** Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 Uptime(secs): 1200.1 total, 600.0 interval Flush(GB): cumulative 0.000, interval 0.000 AddFile(GB): cumulative 0.000, interval 0.000 AddFile(Total Files): cumulative 0, interval 0 AddFile(L0 Files): cumulative 0, interval 0 AddFile(Keys): cumulative 0, interval 0 Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count Block cache BinnedLRUCache@0x55a7f20e31f0#2 capacity: 388.00 MB usage: 2.09 KB table_size: 0 occupancy: 18446744073709551615 collections: 3 last_copies: 8 last_secs: 0.000104 secs_since: 0 Block cache entry stats(count,size,portion): DataBlock(3,1.42 KB,0.000357874%) FilterBlock(3,0.33 KB,8.25862e-05%) IndexBlock(3,0.34 KB,8.65189e-05%) Misc(1,0.00 KB,0%) ** File Read Latency Histogram By Level [m-1] ** ** Compaction Stats [m-2] ** Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 ** Compaction Stats [m-2] ** Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 Uptime(secs): 1200.1 total, 600.0 interval Flush(GB): cumulative 0.000, interval 0.000 AddFile(GB): cumulative 0.000, interval 0.000 AddFile(Total Files): cumulative 0, interval 0 AddFile(L0 Files): cumulative 0, interval 0 AddFile(Keys): cumulative 0, interval 0 Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count Block cache BinnedLRUCache@0x55a7f20e31f0#2 capacity: 388.00 MB usage: 2.09 KB table_size: 0 occupancy: 18446744073709551615 collections: 3 last_copies: 8 last_secs: 0.000104 secs_since: 0 Block cache entry stats(count,size,portion): DataBlock(3,1.42 KB,0.000357874%) FilterBlock(3,0.33 KB,8.25862e-05%) IndexBlock(3,0.34 KB,8.65189e-05%) Misc(1,0.00 KB,0%) ** File Read Latency Histogram By Level [m-2] ** ** Compaction Stats [p-0] ** Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ L0 1/0 1.56 KB 0.1 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.7 0.00 0.00 1 0.002 0 0 0.0 0.0 Sum 1/0 1.56 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.7 0.00 0.00 1 0.002 0 0 0.0 0.0 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 ** Compaction Stats [p-0] ** Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.7 0.00 0.00 1 0.002 0 0 0.0 0.0 Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 Uptime(secs): 1200.1 total, 600.0 interval Flush(GB): cumulative 0.000, interval 0.000 AddFile(GB): cumulative 0.000, interval 0.000 AddFile(Total Files): cumulative 0, interval 0 AddFile(L0 Files): cumulative 0, interval 0 AddFile(Keys): cumulative 0, interval 0 Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count Block cache BinnedLRUCache@0x55a7f20e31f0#2 capacity: 388.00 MB usage: 2.09 KB table_size: 0 occupancy: 18446744073709551615 collections: 3 last_copies: 8 last_secs: 0.000104 secs_since: 0 Block cache entry stats(count,size,portion): DataBlock(3,1.42 KB,0.000357874%) FilterBlock(3,0.33 KB,8.25862e-05%) IndexBlock(3,0.34 KB,8.65189e-05%) Misc(1,0.00 KB,0%) ** File Read Latency Histogram By Level [p-0] ** ** Compaction Stats [p-1] ** Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 ** Compaction Stats [p-1] ** Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 Uptime(secs): 1200.1 total, 600.0 interval Flush(GB): cumulative 0.000, interval 0.000 AddFile(GB): cumulative 0.000, interval 0.000 AddFile(Total Files): cumulative 0, interval 0 AddFile(L0 Files): cumulative 0, interval 0 AddFile(Keys): cumulative 0, interval 0 Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count Block cache BinnedLRUCache@0x55a7f20e31f0#2 capacity: 388.00 MB usage: 2.09 KB table_size: 0 occupancy: 18446744073709551615 collections: 3 last_copies: 8 last_secs: 0.000104 secs_since: 0 Block cache entry stats(count,size,portion): DataBlock(3,1.42 KB,0.000357874%) FilterBlock(3,0.33 KB,8.25862e-05%) IndexBlock(3,0.34 KB,8.65189e-05%) Misc(1,0.00 KB,0%) ** File Read Latency Histogram By Level [p-1] ** ** Compaction Stats [p-2] ** Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 ** Compaction Stats [p-2] ** Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 Uptime(secs): 1200.1 total, 600.0 interval Flush(GB): cumulative 0.000, interval 0.000 AddFile(GB): cumulative 0.000, interval 0.000 AddFile(Total Files): cumulative 0, interval 0 AddFile(L0 Files): cumulative 0, interval 0 AddFile(Keys): cumulative 0, interval 0 Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count Block cache BinnedLRUCache@0x55a7f20e31f0#2 capacity: 388.00 MB usage: 2.09 KB table_size: 0 occupancy: 18446744073709551615 collections: 3 last_copies: 8 last_secs: 0.000104 secs_since: 0 Block cache entry stats(count,size,portion): DataBlock(3,1.42 KB,0.000357874%) FilterBlock(3,0.33 KB,8.25862e-05%) IndexBlock(3,0.34 KB,8.65189e-05%) Misc(1,0.00 KB,0%) ** File Read Latency Histogram By Level [p-2] ** ** Compaction Stats [O-0] ** Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 ** Compaction Stats [O-0] ** Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 Uptime(secs): 1200.1 total, 600.0 interval Flush(GB): cumulative 0.000, interval 0.000 AddFile(GB): cumulative 0.000, interval 0.000 AddFile(Total Files): cumulative 0, interval 0 AddFile(L0 Files): cumulative 0, interval 0 AddFile(Keys): cumulative 0, interval 0 Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count Block cache BinnedLRUCache@0x55a7f20e3a30#2 capacity: 152.00 MB usage: 0.45 KB table_size: 0 occupancy: 18446744073709551615 collections: 3 last_copies: 2 last_secs: 1.2e-05 secs_since: 0 Block cache entry stats(count,size,portion): DataBlock(1,0.20 KB,0.000130503%) FilterBlock(1,0.11 KB,7.02707e-05%) IndexBlock(1,0.14 KB,9.03481e-05%) Misc(1,0.00 KB,0%) ** File Read Latency Histogram By Level [O-0] ** ** Compaction Stats [O-1] ** Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 ** Compaction Stats [O-1] ** Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 Uptime(secs): 1200.1 total, 600.0 interval Flush(GB): cumulative 0.000, interval 0.000 AddFile(GB): cumulative 0.000, interval 0.000 AddFile(Total Files): cumulative 0, interval 0 AddFile(L0 Files): cumulative 0, interval 0 AddFile(Keys): cumulative 0, interval 0 Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count Block cache BinnedLRUCache@0x55a7f20e3a30#2 capacity: 152.00 MB usage: 0.45 KB table_size: 0 occupancy: 18446744073709551615 collections: 3 last_copies: 2 last_secs: 1.2e-05 secs_since: 0 Block cache entry stats(count,size,portion): DataBlock(1,0.20 KB,0.000130503%) FilterBlock(1,0.11 KB,7.02707e-05%) IndexBlock(1,0.14 KB,9.03481e-05%) Misc(1,0.00 KB,0%) ** File Read Latency Histogram By Level [O-1] ** ** Compaction Stats [O-2] ** Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ L0 1/0 1.25 KB 0.1 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.6 0.00 0.00 1 0.002 0 0 0.0 0.0 Sum 1/0 1.25 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.6 0.00 0.00 1 0.002 0 0 0.0 0.0 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 ** Compaction Stats [O-2] ** Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.6 0.00 0.00 1 0.002 0 0 0.0 0.0 Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 Uptime(secs): 1200.1 total, 600.0 interval Flush(GB): cumulative 0.000, interval 0.000 AddFile(GB): cumulative 0.000, interval 0.000 AddFile(Total Files): cumulative 0, interval 0 AddFile(L0 Files): cumulative 0, interval 0 AddFile(Keys): cumulative 0, interval 0 Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count Block cache BinnedLRUCache@0x55a7f20e3a30#2 capacity: 152.00 MB usage: 0.45 KB table_size: 0 occupancy: 18446744073709551615 collections: 3 last_copies: 2 last_secs: 1.2e-05 secs_since: 0 Block cache entry stats(count,size,portion): DataBlock(1,0.20 KB,0.000130503%) FilterBlock(1,0.11 KB,7.02707e-05%) IndexBlock(1,0.14 KB,9.03481e-05%) Misc(1,0.00 KB,0%) ** File Read Latency Histogram By Level [O-2] ** ** Compaction Stats [L] ** Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ L0 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 1 0.001 0 0 0.0 0.0 Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 1 0.001 0 0 0.0 0.0 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 ** Compaction Stats [L] ** Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 1 0.001 0 0 0.0 0.0 Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 Uptime(secs): 1200.1 total, 600.0 interval Flush(GB): cumulative 0.000, interval 0.000 AddFile(GB): cumulative 0.000, interval 0.000 AddFile(Total Files): cumulative 0, interval 0 AddFile(L0 Files): cumulative 0, interval 0 AddFile(Keys): cumulative 0, interval 0 Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count Block cache BinnedLRUCache@0x55a7f20e31f0#2 capacity: 388.00 MB usage: 2.09 KB table_size: 0 occupancy: 18446744073709551615 collections: 3 last_copies: 8 last_secs: 0.000104 secs_since: 0 Block cache entry stats(count,size,portion): DataBlock(3,1.42 KB,0.000357874%) FilterBlock(3,0.33 KB,8.25862e-05%) IndexBlock(3,0.34 KB,8.65189e-05%) Misc(1,0.00 KB,0%) ** File Read Latency Histogram By Level [L] ** ** Compaction Stats [P] ** Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 ** Compaction Stats [P] ** Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 Uptime(secs): 1200.1 total, 600.0 interval Flush(GB): cumulative 0.000, interval 0.000 AddFile(GB): cumulative 0.000, interval 0.000 AddFile(Total Files): cumulative 0, interval 0 AddFile(L0 Files): cumulative 0, interval 0 AddFile(Keys): cumulative 0, interval 0 Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count Block cache BinnedLRUCache@0x55a7f20e31f0#2 capacity: 388.00 MB usage: 2.09 KB table_size: 0 occupancy: 18446744073709551615 collections: 3 last_copies: 8 last_secs: 0.000104 secs_since: 0 Block cache entry stats(count,size,portion): DataBlock(3,1.42 KB,0.000357874%) FilterBlock(3,0.33 KB,8.25862e-05%) IndexBlock(3,0.34 KB,8.65189e-05%) Misc(1,0.00 KB,0%) ** File Read Latency Histogram By Level [P] ** ● cron.service - Regular background program processing daemon Loaded: loaded (/usr/lib/systemd/system/cron.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:07:36 UTC; 44min ago Invocation: 7ba5623b3a5a4c928676169e3a3df8ca Docs: man:cron(8) Main PID: 805 (cron) Tasks: 1 (limit: 9256) Memory: 240K (peak: 1.7M, swap: 172K, swap peak: 172K, zswap: 31.8K) CPU: 45ms CGroup: /system.slice/cron.service └─805 /usr/sbin/cron -f Apr 14 22:07:36 npc99a84aae6d34 systemd[1]: Started cron.service - Regular background program processing daemon. Apr 14 22:07:36 npc99a84aae6d34 cron[805]: (CRON) INFO (pidfile fd = 3) Apr 14 22:07:36 npc99a84aae6d34 cron[805]: (CRON) INFO (Running @reboot jobs) Apr 14 22:17:01 npc99a84aae6d34 CRON[10922]: pam_unix(cron:session): session opened for user root(uid=0) by root(uid=0) Apr 14 22:17:01 npc99a84aae6d34 CRON[10924]: (root) CMD (cd / && run-parts --report /etc/cron.hourly) Apr 14 22:17:01 npc99a84aae6d34 CRON[10922]: pam_unix(cron:session): session closed for user root ● dbus.service - D-Bus System Message Bus Loaded: loaded (/usr/lib/systemd/system/dbus.service; static) Active: active (running) since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: 449f9012fe64485c941ef1d66239cf35 TriggeredBy: ● dbus.socket Docs: man:dbus-daemon(1) Main PID: 589 (dbus-daemon) Tasks: 1 (limit: 9256) Memory: 2.1M (peak: 4.1M) CPU: 12.761s CGroup: /system.slice/dbus.service └─589 /usr/bin/dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only Apr 14 22:23:54 npc99a84aae6d34 dbus-daemon[589]: [system] Reloaded configuration Apr 14 22:24:01 npc99a84aae6d34 dbus-daemon[589]: Unknown username "polkitd" in message bus configuration file Apr 14 22:24:01 npc99a84aae6d34 dbus-daemon[589]: Unknown username "polkitd" in message bus configuration file Apr 14 22:24:01 npc99a84aae6d34 dbus-daemon[589]: [system] Reloaded configuration Apr 14 22:24:02 npc99a84aae6d34 systemd[1]: Reloading dbus.service - D-Bus System Message Bus... Apr 14 22:24:02 npc99a84aae6d34 dbus-daemon[589]: [system] Reloaded configuration Apr 14 22:24:02 npc99a84aae6d34 dbus-send[41622]: method return time=1776205442.150030 sender=org.freedesktop.DBus -> destination=:1.58 serial=3 reply_serial=2 Apr 14 22:24:02 npc99a84aae6d34 systemd[1]: Reloaded dbus.service - D-Bus System Message Bus. Apr 14 22:34:46 npc99a84aae6d34 dbus-daemon[589]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.496' (uid=0 pid=87645 comm="/usr/sbin/libvirtd --timeout 120") Apr 14 22:34:46 npc99a84aae6d34 dbus-daemon[589]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' ● devstack@br-ex-tcpdump.service - Devstack devstack@br-ex-tcpdump.service Loaded: loaded (/etc/systemd/system/devstack@br-ex-tcpdump.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:35:24 UTC; 16min ago Invocation: 03d415e39f424a3d94e04452528c452a Main PID: 110919 (tcpdump) Tasks: 1 (limit: 9256) Memory: 2.5M (peak: 2.7M) CPU: 22ms CGroup: /system.slice/system-devstack.slice/devstack@br-ex-tcpdump.service └─110919 /bin/tcpdump -i br-ex arp or rarp or icmp or icmp6 -enlX Apr 14 22:35:24 npc99a84aae6d34 systemd[1]: Started devstack@br-ex-tcpdump.service - Devstack devstack@br-ex-tcpdump.service. Apr 14 22:35:24 npc99a84aae6d34 tcpdump[110919]: tcpdump: verbose output suppressed, use -v[v]... for full protocol decode Apr 14 22:35:24 npc99a84aae6d34 tcpdump[110919]: listening on br-ex, link-type EN10MB (Ethernet), snapshot length 262144 bytes ● devstack@br-int-flows.service - Devstack devstack@br-int-flows.service Loaded: loaded (/etc/systemd/system/devstack@br-int-flows.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:35:26 UTC; 16min ago Invocation: c07d2fea3e0b4267bdec44a648214573 Main PID: 111525 (sh) Tasks: 2 (limit: 9256) Memory: 2.8M (peak: 4.7M) CPU: 713ms CGroup: /system.slice/system-devstack.slice/devstack@br-int-flows.service ├─111525 /bin/sh -c "set +e; while true; do echo ovs-ofctl dump-flows br-int; ovs-ofctl dump-flows br-int ; sleep 30; done; " └─121349 sleep 30 Apr 14 22:51:27 npc99a84aae6d34 sh[121348]: cookie=0xa59e6423, duration=1030.256s, table=65, n_packets=0, n_bytes=0, idle_age=1030, priority=100,reg15=0x1,metadata=0x2 actions=clone(ct_clear,load:0->NXM_NX_REG11[],load:0->NXM_NX_REG12[],load:0->NXM_NX_REG13[0..15],load:0x5->NXM_NX_REG11[],load:0x6->NXM_NX_REG12[],load:0x1->OXM_OF_METADATA[],load:0x2->NXM_NX_REG14[],load:0->NXM_NX_REG10[],load:0->NXM_NX_REG15[],load:0->NXM_NX_REG0[],load:0->NXM_NX_REG1[],load:0->NXM_NX_REG2[],load:0->NXM_NX_REG3[],load:0->NXM_NX_REG4[],load:0->NXM_NX_REG5[],load:0->NXM_NX_REG6[],load:0->NXM_NX_REG7[],load:0->NXM_NX_REG8[],load:0->NXM_NX_REG9[],resubmit(,8)) Apr 14 22:51:27 npc99a84aae6d34 sh[121348]: cookie=0xfba4b1cc, duration=1030.256s, table=65, n_packets=0, n_bytes=0, idle_age=1030, priority=100,reg15=0x2,metadata=0x2 actions=clone(ct_clear,load:0->NXM_NX_REG11[],load:0->NXM_NX_REG12[],load:0->NXM_NX_REG13[0..15],load:0x1->NXM_NX_REG11[],load:0x2->NXM_NX_REG12[],load:0x3->OXM_OF_METADATA[],load:0x3->NXM_NX_REG14[],load:0->NXM_NX_REG10[],load:0->NXM_NX_REG15[],load:0->NXM_NX_REG0[],load:0->NXM_NX_REG1[],load:0->NXM_NX_REG2[],load:0->NXM_NX_REG3[],load:0->NXM_NX_REG4[],load:0->NXM_NX_REG5[],load:0->NXM_NX_REG6[],load:0->NXM_NX_REG7[],load:0->NXM_NX_REG8[],load:0->NXM_NX_REG9[],resubmit(,8)) Apr 14 22:51:27 npc99a84aae6d34 sh[121348]: cookie=0x94247a9, duration=1030.228s, table=65, n_packets=5, n_bytes=210, idle_age=1014, priority=100,reg15=0x1,metadata=0x3 actions=output:1 Apr 14 22:51:27 npc99a84aae6d34 sh[121348]: cookie=0x595271ba, duration=1026.377s, table=65, n_packets=0, n_bytes=0, idle_age=1026, priority=100,reg15=0x4,metadata=0x2 actions=clone(ct_clear,load:0->NXM_NX_REG11[],load:0->NXM_NX_REG12[],load:0->NXM_NX_REG13[0..15],load:0x5->NXM_NX_REG11[],load:0x6->NXM_NX_REG12[],load:0x1->OXM_OF_METADATA[],load:0x3->NXM_NX_REG14[],load:0->NXM_NX_REG10[],load:0->NXM_NX_REG15[],load:0->NXM_NX_REG0[],load:0->NXM_NX_REG1[],load:0->NXM_NX_REG2[],load:0->NXM_NX_REG3[],load:0->NXM_NX_REG4[],load:0->NXM_NX_REG5[],load:0->NXM_NX_REG6[],load:0->NXM_NX_REG7[],load:0->NXM_NX_REG8[],load:0->NXM_NX_REG9[],resubmit(,8)) Apr 14 22:51:27 npc99a84aae6d34 sh[121348]: cookie=0xcc9b5b27, duration=1026.378s, table=65, n_packets=0, n_bytes=0, idle_age=1026, priority=100,reg15=0x3,metadata=0x1 actions=clone(ct_clear,load:0->NXM_NX_REG11[],load:0->NXM_NX_REG12[],load:0->NXM_NX_REG13[0..15],load:0x4->NXM_NX_REG11[],load:0x3->NXM_NX_REG12[],load:0x2->OXM_OF_METADATA[],load:0x4->NXM_NX_REG14[],load:0->NXM_NX_REG10[],load:0->NXM_NX_REG15[],load:0->NXM_NX_REG0[],load:0->NXM_NX_REG1[],load:0->NXM_NX_REG2[],load:0->NXM_NX_REG3[],load:0->NXM_NX_REG4[],load:0->NXM_NX_REG5[],load:0->NXM_NX_REG6[],load:0->NXM_NX_REG7[],load:0->NXM_NX_REG8[],load:0->NXM_NX_REG9[],resubmit(,8)) Apr 14 22:51:27 npc99a84aae6d34 sh[121348]: cookie=0x0, duration=1067.969s, table=65, n_packets=17, n_bytes=1482, idle_age=1014, priority=0 actions=drop Apr 14 22:51:27 npc99a84aae6d34 sh[121348]: cookie=0x0, duration=1067.969s, table=80, n_packets=0, n_bytes=0, idle_age=1067, priority=0 actions=resubmit(,8) Apr 14 22:51:27 npc99a84aae6d34 sh[121348]: cookie=0x0, duration=1067.969s, table=81, n_packets=0, n_bytes=0, idle_age=1067, priority=100,ct_state=+est+trk,ip actions=move:NXM_NX_CT_NW_DST[]->NXM_NX_REG1[] Apr 14 22:51:27 npc99a84aae6d34 sh[121348]: cookie=0x0, duration=1067.969s, table=82, n_packets=0, n_bytes=0, idle_age=1067, priority=100,ct_state=+est+trk,ipv6 actions=move:NXM_NX_CT_IPV6_DST[]->NXM_NX_XXREG1[] Apr 14 22:51:27 npc99a84aae6d34 sh[121348]: cookie=0x0, duration=1067.969s, table=83, n_packets=0, n_bytes=0, idle_age=1067, priority=100,ct_state=+est+trk actions=move:NXM_NX_CT_TP_DST[]->NXM_NX_REG2[0..15] ● devstack@c-api.service - Devstack devstack@c-api.service Loaded: loaded (/etc/systemd/system/devstack@c-api.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:34:47 UTC; 17min ago Invocation: 4b8491ebc76640b4bfd9055cc5652ec5 Main PID: 105804 (uwsgi) Status: "uWSGI is ready" Tasks: 7 (limit: 9256) Memory: 253M (peak: 261.1M) CPU: 8.635s CGroup: /system.slice/system-devstack.slice/devstack@c-api.service ├─105804 "cinder-apiuWSGI master" ├─105808 "cinder-apiuWSGI worker 1" └─105809 "cinder-apiuWSGI worker 2" Apr 14 22:35:40 npc99a84aae6d34 devstack@c-api.service[105809]: DEBUG cinder.api.openstack.wsgi [None req-183d14b1-4ff6-4145-af80-a294195f4c44 None None] Empty body provided in request {{(pid=105809) get_body /opt/stack/cinder/cinder/api/openstack/wsgi.py:725}} Apr 14 22:35:40 npc99a84aae6d34 devstack@c-api.service[105809]: DEBUG cinder.api.openstack.wsgi [None req-183d14b1-4ff6-4145-af80-a294195f4c44 None None] Calling method 'all' {{(pid=105809) _process_stack /opt/stack/cinder/cinder/api/openstack/wsgi.py:878}} Apr 14 22:35:40 npc99a84aae6d34 devstack@c-api.service[105809]: INFO cinder.api.openstack.wsgi [None req-183d14b1-4ff6-4145-af80-a294195f4c44 None None] https://[2001:41d0:302:1000::cbc]/volume// returned with HTTP 300 Apr 14 22:35:40 npc99a84aae6d34 devstack@c-api.service[105809]: [pid: 105809|app: 0|req: 4/6] 2001:41d0:302:1000::cbc () {66 vars in 1526 bytes} [Tue Apr 14 22:35:40 2026] GET /volume/ => generated 401 bytes in 11 msecs (HTTP/1.1 300) 7 headers in 299 bytes (1 switches on core 0) Apr 14 22:37:54 npc99a84aae6d34 devstack@c-api.service[105808]: DEBUG cinder.api.middleware.request_id [None req-fe82ae1e-3dc7-4acd-b777-c9767148e2d0 None None] RequestId filter calling following filter/app {{(pid=105808) _context_setter /opt/stack/cinder/cinder/api/middleware/request_id.py:62}} Apr 14 22:37:54 npc99a84aae6d34 devstack@c-api.service[105808]: INFO cinder.api.openstack.wsgi [None req-fe82ae1e-3dc7-4acd-b777-c9767148e2d0 tempest-verify_tempest_config-1808879246 tempest-verify_tempest_config-1808879246-project-member] GET https://[2001:41d0:302:1000::cbc]/volume/v3/extensions Apr 14 22:37:54 npc99a84aae6d34 devstack@c-api.service[105808]: DEBUG cinder.api.openstack.wsgi [None req-fe82ae1e-3dc7-4acd-b777-c9767148e2d0 tempest-verify_tempest_config-1808879246 tempest-verify_tempest_config-1808879246-project-member] Empty body provided in request {{(pid=105808) get_body /opt/stack/cinder/cinder/api/openstack/wsgi.py:725}} Apr 14 22:37:54 npc99a84aae6d34 devstack@c-api.service[105808]: DEBUG cinder.api.openstack.wsgi [None req-fe82ae1e-3dc7-4acd-b777-c9767148e2d0 tempest-verify_tempest_config-1808879246 tempest-verify_tempest_config-1808879246-project-member] Calling method 'index' {{(pid=105808) _process_stack /opt/stack/cinder/cinder/api/openstack/wsgi.py:878}} Apr 14 22:37:54 npc99a84aae6d34 devstack@c-api.service[105808]: INFO cinder.api.openstack.wsgi [None req-fe82ae1e-3dc7-4acd-b777-c9767148e2d0 tempest-verify_tempest_config-1808879246 tempest-verify_tempest_config-1808879246-project-member] https://[2001:41d0:302:1000::cbc]/volume/v3/extensions returned with HTTP 200 Apr 14 22:37:54 npc99a84aae6d34 devstack@c-api.service[105808]: [pid: 105808|app: 0|req: 3/7] 2001:41d0:302:1000::cbc () {66 vars in 1347 bytes} [Tue Apr 14 22:37:54 2026] GET /volume/v3/extensions => generated 5245 bytes in 607 msecs (HTTP/1.1 200) 7 headers in 286 bytes (1 switches on core 0) ● devstack@c-bak.service - Devstack devstack@c-bak.service Loaded: loaded (/etc/systemd/system/devstack@c-bak.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:34:55 UTC; 17min ago Invocation: f10c496fbc494627a57b2ffc222d7c34 Main PID: 107200 (cinder-backup) Tasks: 61 (limit: 9256) Memory: 103.2M (peak: 104.8M) CPU: 3.616s CGroup: /system.slice/system-devstack.slice/devstack@c-bak.service └─107200 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/cinder-backup --config-file /etc/cinder/cinder.conf Apr 14 22:48:27 npc99a84aae6d34 cinder-backup[107200]: DEBUG oslo_service.periodic_task [None req-7d6cfd3f-621a-4e9e-a884-944058ce2f2a None None] Running periodic task BackupManager.publish_service_capabilities {{(pid=107200) run_periodic_tasks /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/periodic_task.py:210}} Apr 14 22:48:27 npc99a84aae6d34 cinder-backup[107200]: DEBUG cinder.manager [None req-7d6cfd3f-621a-4e9e-a884-944058ce2f2a None None] Notifying Schedulers of capabilities ... {{(pid=107200) _publish_service_capabilities /opt/stack/cinder/cinder/manager.py:202}} Apr 14 22:49:08 npc99a84aae6d34 cinder-backup[107200]: DEBUG dbcounter [-] [107200] Writing DB stats cinder:SELECT=1,cinder:UPDATE=1 {{(pid=107200) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:49:27 npc99a84aae6d34 cinder-backup[107200]: DEBUG oslo_service.periodic_task [None req-7d6cfd3f-621a-4e9e-a884-944058ce2f2a None None] Running periodic task BackupManager.publish_service_capabilities {{(pid=107200) run_periodic_tasks /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/periodic_task.py:210}} Apr 14 22:49:27 npc99a84aae6d34 cinder-backup[107200]: DEBUG cinder.manager [None req-7d6cfd3f-621a-4e9e-a884-944058ce2f2a None None] Notifying Schedulers of capabilities ... {{(pid=107200) _publish_service_capabilities /opt/stack/cinder/cinder/manager.py:202}} Apr 14 22:50:27 npc99a84aae6d34 cinder-backup[107200]: DEBUG oslo_service.periodic_task [None req-7d6cfd3f-621a-4e9e-a884-944058ce2f2a None None] Running periodic task BackupManager.publish_service_capabilities {{(pid=107200) run_periodic_tasks /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/periodic_task.py:210}} Apr 14 22:50:27 npc99a84aae6d34 cinder-backup[107200]: DEBUG cinder.manager [None req-7d6cfd3f-621a-4e9e-a884-944058ce2f2a None None] Notifying Schedulers of capabilities ... {{(pid=107200) _publish_service_capabilities /opt/stack/cinder/cinder/manager.py:202}} Apr 14 22:51:08 npc99a84aae6d34 cinder-backup[107200]: DEBUG dbcounter [-] [107200] Writing DB stats cinder:SELECT=1,cinder:UPDATE=1 {{(pid=107200) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:51:27 npc99a84aae6d34 cinder-backup[107200]: DEBUG oslo_service.periodic_task [None req-7d6cfd3f-621a-4e9e-a884-944058ce2f2a None None] Running periodic task BackupManager.publish_service_capabilities {{(pid=107200) run_periodic_tasks /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/periodic_task.py:210}} Apr 14 22:51:27 npc99a84aae6d34 cinder-backup[107200]: DEBUG cinder.manager [None req-7d6cfd3f-621a-4e9e-a884-944058ce2f2a None None] Notifying Schedulers of capabilities ... {{(pid=107200) _publish_service_capabilities /opt/stack/cinder/cinder/manager.py:202}} ● devstack@c-sch.service - Devstack devstack@c-sch.service Loaded: loaded (/etc/systemd/system/devstack@c-sch.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:34:53 UTC; 17min ago Invocation: 23cec2f89df44ee7af4fb0a7155f1722 Main PID: 106585 (cinder-schedule) Tasks: 1 (limit: 9256) Memory: 113.7M (peak: 114.9M) CPU: 3.955s CGroup: /system.slice/system-devstack.slice/devstack@c-sch.service └─106585 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/cinder-scheduler --config-file /etc/cinder/cinder.conf Apr 14 22:47:59 npc99a84aae6d34 cinder-scheduler[106585]: DEBUG cinder.scheduler.host_manager [None req-cb60450e-b80b-49be-8fc0-7c21f09ad8f4 None None] Received volume service update from npc99a84aae6d34@ceph: {'vendor_name': 'Open Source', 'driver_version': '1.3.0', 'storage_protocol': 'ceph', 'total_capacity_gb': 28.41, 'free_capacity_gb': 28.41, 'reserved_percentage': 0, 'multiattach': True, 'thin_provisioning_support': True, 'max_over_subscription_ratio': '20.0', 'location_info': 'ceph:/etc/ceph/ceph.conf:929e05b1-4e8c-4c8f-b13f-4521b44f10e1:cinder:volumes', 'backend_state': 'up', 'qos_support': True, 'volume_backend_name': 'ceph', 'replication_enabled': False, 'allocated_capacity_gb': 0, 'filter_function': None, 'goodness_function': None} {{(pid=106585) update_service_capabilities /opt/stack/cinder/cinder/scheduler/host_manager.py:629}} Apr 14 22:48:27 npc99a84aae6d34 cinder-scheduler[106585]: DEBUG cinder.scheduler.host_manager [None req-69bfc9a1-92ff-41ce-9fbb-3d254a86ea2a None None] Received backup service update from npc99a84aae6d34: {'backend_state': True, 'driver_name': 'cinder.backup.drivers.ceph.CephBackupDriver', 'availability_zone': 'nova'} {{(pid=106585) update_service_capabilities /opt/stack/cinder/cinder/scheduler/host_manager.py:598}} Apr 14 22:48:59 npc99a84aae6d34 cinder-scheduler[106585]: DEBUG cinder.scheduler.host_manager [None req-d63c3162-2be2-4819-b645-b9c5cc1fdf04 None None] Received volume service update from npc99a84aae6d34@ceph: {'vendor_name': 'Open Source', 'driver_version': '1.3.0', 'storage_protocol': 'ceph', 'total_capacity_gb': 28.41, 'free_capacity_gb': 28.41, 'reserved_percentage': 0, 'multiattach': True, 'thin_provisioning_support': True, 'max_over_subscription_ratio': '20.0', 'location_info': 'ceph:/etc/ceph/ceph.conf:929e05b1-4e8c-4c8f-b13f-4521b44f10e1:cinder:volumes', 'backend_state': 'up', 'qos_support': True, 'volume_backend_name': 'ceph', 'replication_enabled': False, 'allocated_capacity_gb': 0, 'filter_function': None, 'goodness_function': None} {{(pid=106585) update_service_capabilities /opt/stack/cinder/cinder/scheduler/host_manager.py:629}} Apr 14 22:49:11 npc99a84aae6d34 cinder-scheduler[106585]: DEBUG dbcounter [-] [106585] Writing DB stats cinder:SELECT=1,cinder:UPDATE=1 {{(pid=106585) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:49:27 npc99a84aae6d34 cinder-scheduler[106585]: DEBUG cinder.scheduler.host_manager [None req-76e41f95-5e9b-420c-a4a2-bca730713405 None None] Received backup service update from npc99a84aae6d34: {'backend_state': True, 'driver_name': 'cinder.backup.drivers.ceph.CephBackupDriver', 'availability_zone': 'nova'} {{(pid=106585) update_service_capabilities /opt/stack/cinder/cinder/scheduler/host_manager.py:598}} Apr 14 22:49:59 npc99a84aae6d34 cinder-scheduler[106585]: DEBUG cinder.scheduler.host_manager [None req-a6f9fe9e-e4bc-44e4-89df-4ff879598069 None None] Received volume service update from npc99a84aae6d34@ceph: {'vendor_name': 'Open Source', 'driver_version': '1.3.0', 'storage_protocol': 'ceph', 'total_capacity_gb': 28.41, 'free_capacity_gb': 28.41, 'reserved_percentage': 0, 'multiattach': True, 'thin_provisioning_support': True, 'max_over_subscription_ratio': '20.0', 'location_info': 'ceph:/etc/ceph/ceph.conf:929e05b1-4e8c-4c8f-b13f-4521b44f10e1:cinder:volumes', 'backend_state': 'up', 'qos_support': True, 'volume_backend_name': 'ceph', 'replication_enabled': False, 'allocated_capacity_gb': 0, 'filter_function': None, 'goodness_function': None} {{(pid=106585) update_service_capabilities /opt/stack/cinder/cinder/scheduler/host_manager.py:629}} Apr 14 22:50:27 npc99a84aae6d34 cinder-scheduler[106585]: DEBUG cinder.scheduler.host_manager [None req-95b8b995-590d-41bb-83ac-91f66653a46e None None] Received backup service update from npc99a84aae6d34: {'backend_state': True, 'driver_name': 'cinder.backup.drivers.ceph.CephBackupDriver', 'availability_zone': 'nova'} {{(pid=106585) update_service_capabilities /opt/stack/cinder/cinder/scheduler/host_manager.py:598}} Apr 14 22:50:59 npc99a84aae6d34 cinder-scheduler[106585]: DEBUG cinder.scheduler.host_manager [None req-c129fa44-ae5f-4444-9ba7-c6128086b305 None None] Received volume service update from npc99a84aae6d34@ceph: {'vendor_name': 'Open Source', 'driver_version': '1.3.0', 'storage_protocol': 'ceph', 'total_capacity_gb': 28.41, 'free_capacity_gb': 28.41, 'reserved_percentage': 0, 'multiattach': True, 'thin_provisioning_support': True, 'max_over_subscription_ratio': '20.0', 'location_info': 'ceph:/etc/ceph/ceph.conf:929e05b1-4e8c-4c8f-b13f-4521b44f10e1:cinder:volumes', 'backend_state': 'up', 'qos_support': True, 'volume_backend_name': 'ceph', 'replication_enabled': False, 'allocated_capacity_gb': 0, 'filter_function': None, 'goodness_function': None} {{(pid=106585) update_service_capabilities /opt/stack/cinder/cinder/scheduler/host_manager.py:629}} Apr 14 22:51:11 npc99a84aae6d34 cinder-scheduler[106585]: DEBUG dbcounter [-] [106585] Writing DB stats cinder:SELECT=1,cinder:UPDATE=1 {{(pid=106585) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:51:27 npc99a84aae6d34 cinder-scheduler[106585]: DEBUG cinder.scheduler.host_manager [None req-5b952546-c9f9-4373-b825-bd2516b4cd9b None None] Received backup service update from npc99a84aae6d34: {'backend_state': True, 'driver_name': 'cinder.backup.drivers.ceph.CephBackupDriver', 'availability_zone': 'nova'} {{(pid=106585) update_service_capabilities /opt/stack/cinder/cinder/scheduler/host_manager.py:598}} ● devstack@c-vol.service - Devstack devstack@c-vol.service Loaded: loaded (/etc/systemd/system/devstack@c-vol.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:34:57 UTC; 16min ago Invocation: ec9f4aa39eb6479db9dbb269037d1b07 Main PID: 107810 (cinder-volume) Tasks: 22 (limit: 9256) Memory: 151M (peak: 153.6M) CPU: 20.696s CGroup: /system.slice/system-devstack.slice/devstack@c-vol.service ├─107810 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/cinder-volume --config-file /etc/cinder/cinder.conf └─108126 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/cinder-volume --config-file /etc/cinder/cinder.conf Apr 14 22:49:10 npc99a84aae6d34 cinder-volume[108126]: DEBUG dbcounter [-] [108126] Writing DB stats cinder:SELECT=1,cinder:UPDATE=1 {{(pid=108126) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:49:59 npc99a84aae6d34 cinder-volume[108126]: DEBUG oslo_service.periodic_task [None req-50b4d842-c78a-40ef-bcac-e6d54218b21b None None] Running periodic task VolumeManager.publish_service_capabilities {{(pid=108126) run_periodic_tasks /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/periodic_task.py:210}} Apr 14 22:49:59 npc99a84aae6d34 cinder-volume[108126]: DEBUG cinder.volume.drivers.rbd [None req-50b4d842-c78a-40ef-bcac-e6d54218b21b None None] connecting to cinder@ceph (conf=/etc/ceph/ceph.conf, timeout=-1). {{(pid=108126) _do_conn /opt/stack/cinder/cinder/volume/drivers/rbd.py:603}} Apr 14 22:49:59 npc99a84aae6d34 cinder-volume[108126]: DEBUG cinder.volume.drivers.rbd [None req-50b4d842-c78a-40ef-bcac-e6d54218b21b None None] connecting to cinder@ceph (conf=/etc/ceph/ceph.conf, timeout=-1). {{(pid=108126) _do_conn /opt/stack/cinder/cinder/volume/drivers/rbd.py:603}} Apr 14 22:49:59 npc99a84aae6d34 cinder-volume[108126]: DEBUG cinder.manager [None req-50b4d842-c78a-40ef-bcac-e6d54218b21b None None] Notifying Schedulers of capabilities ... {{(pid=108126) _publish_service_capabilities /opt/stack/cinder/cinder/manager.py:202}} Apr 14 22:50:59 npc99a84aae6d34 cinder-volume[108126]: DEBUG oslo_service.periodic_task [None req-50b4d842-c78a-40ef-bcac-e6d54218b21b None None] Running periodic task VolumeManager.publish_service_capabilities {{(pid=108126) run_periodic_tasks /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/periodic_task.py:210}} Apr 14 22:50:59 npc99a84aae6d34 cinder-volume[108126]: DEBUG cinder.volume.drivers.rbd [None req-50b4d842-c78a-40ef-bcac-e6d54218b21b None None] connecting to cinder@ceph (conf=/etc/ceph/ceph.conf, timeout=-1). {{(pid=108126) _do_conn /opt/stack/cinder/cinder/volume/drivers/rbd.py:603}} Apr 14 22:50:59 npc99a84aae6d34 cinder-volume[108126]: DEBUG cinder.volume.drivers.rbd [None req-50b4d842-c78a-40ef-bcac-e6d54218b21b None None] connecting to cinder@ceph (conf=/etc/ceph/ceph.conf, timeout=-1). {{(pid=108126) _do_conn /opt/stack/cinder/cinder/volume/drivers/rbd.py:603}} Apr 14 22:50:59 npc99a84aae6d34 cinder-volume[108126]: DEBUG cinder.manager [None req-50b4d842-c78a-40ef-bcac-e6d54218b21b None None] Notifying Schedulers of capabilities ... {{(pid=108126) _publish_service_capabilities /opt/stack/cinder/cinder/manager.py:202}} Apr 14 22:51:10 npc99a84aae6d34 cinder-volume[108126]: DEBUG dbcounter [-] [108126] Writing DB stats cinder:SELECT=1,cinder:UPDATE=1 {{(pid=108126) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} ● devstack@etcd.service - Devstack devstack@etcd.service Loaded: loaded (/etc/systemd/system/devstack@etcd.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:27:54 UTC; 24min ago Invocation: 811138313e284c75a392b85d2988be08 Main PID: 64223 (etcd) Tasks: 13 (limit: 9256) Memory: 25.2M (peak: 155.9M, swap: 122.1M, swap peak: 122.1M, zswap: 9.4K) CPU: 13.577s CGroup: /system.slice/system-devstack.slice/devstack@etcd.service └─64223 /opt/stack/bin/etcd --name npc99a84aae6d34 --data-dir /opt/stack/data/etcd --initial-cluster-state new --initial-cluster-token etcd-cluster-01 --initial-cluster "npc99a84aae6d34=http://[2001:41d0:302:1000::cbc]:2380" --initial-advertise-peer-urls "http://[2001:41d0:302:1000::cbc]:2380" --advertise-client-urls "http://[2001:41d0:302:1000::cbc]:2379" --listen-peer-urls http://0.0.0.0:2380 --listen-client-urls "http://[2001:41d0:302:1000::cbc]:2379" --log-level=debug Apr 14 22:51:19 npc99a84aae6d34 etcd[64223]: {"level":"debug","ts":"2026-04-14T22:51:19.730412Z","caller":"etcdserver/server.go:1230","msg":"The member is active, skip checking leadership","latestTickTs":"2026-04-14T22:51:19.665946Z","now":"2026-04-14T22:51:19.730410Z"} Apr 14 22:51:19 npc99a84aae6d34 etcd[64223]: {"level":"debug","ts":"2026-04-14T22:51:19.772705Z","caller":"etcdserver/server.go:1230","msg":"The member is active, skip checking leadership","latestTickTs":"2026-04-14T22:51:19.765851Z","now":"2026-04-14T22:51:19.772703Z"} Apr 14 22:51:28 npc99a84aae6d34 etcd[64223]: {"level":"debug","ts":"2026-04-14T22:51:28.386511Z","caller":"etcdserver/server.go:1230","msg":"The member is active, skip checking leadership","latestTickTs":"2026-04-14T22:51:28.366346Z","now":"2026-04-14T22:51:28.386501Z"} Apr 14 22:51:30 npc99a84aae6d34 etcd[64223]: {"level":"debug","ts":"2026-04-14T22:51:30.849326Z","caller":"etcdserver/server.go:1230","msg":"The member is active, skip checking leadership","latestTickTs":"2026-04-14T22:51:30.766371Z","now":"2026-04-14T22:51:30.849320Z"} Apr 14 22:51:34 npc99a84aae6d34 etcd[64223]: {"level":"debug","ts":"2026-04-14T22:51:34.734560Z","caller":"etcdserver/server.go:1230","msg":"The member is active, skip checking leadership","latestTickTs":"2026-04-14T22:51:34.666760Z","now":"2026-04-14T22:51:34.734557Z"} Apr 14 22:51:34 npc99a84aae6d34 etcd[64223]: {"level":"debug","ts":"2026-04-14T22:51:34.775398Z","caller":"etcdserver/server.go:1230","msg":"The member is active, skip checking leadership","latestTickTs":"2026-04-14T22:51:34.765861Z","now":"2026-04-14T22:51:34.775395Z"} Apr 14 22:51:43 npc99a84aae6d34 etcd[64223]: {"level":"debug","ts":"2026-04-14T22:51:43.389251Z","caller":"etcdserver/server.go:1230","msg":"The member is active, skip checking leadership","latestTickTs":"2026-04-14T22:51:43.366576Z","now":"2026-04-14T22:51:43.389249Z"} Apr 14 22:51:45 npc99a84aae6d34 etcd[64223]: {"level":"debug","ts":"2026-04-14T22:51:45.850967Z","caller":"etcdserver/server.go:1230","msg":"The member is active, skip checking leadership","latestTickTs":"2026-04-14T22:51:45.766492Z","now":"2026-04-14T22:51:45.850965Z"} Apr 14 22:51:49 npc99a84aae6d34 etcd[64223]: {"level":"debug","ts":"2026-04-14T22:51:49.738103Z","caller":"etcdserver/server.go:1230","msg":"The member is active, skip checking leadership","latestTickTs":"2026-04-14T22:51:49.666682Z","now":"2026-04-14T22:51:49.738101Z"} Apr 14 22:51:49 npc99a84aae6d34 etcd[64223]: {"level":"debug","ts":"2026-04-14T22:51:49.778631Z","caller":"etcdserver/server.go:1230","msg":"The member is active, skip checking leadership","latestTickTs":"2026-04-14T22:51:49.766179Z","now":"2026-04-14T22:51:49.778628Z"} ● devstack@file_tracker.service - Devstack devstack@file_tracker.service Loaded: loaded (/etc/systemd/system/devstack@file_tracker.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:27:50 UTC; 24min ago Invocation: 3fc9efb2d8184937b1d056f0c61b3b6b Main PID: 63498 (file_tracker.sh) Tasks: 2 (limit: 9256) Memory: 488K (peak: 1.8M, swap: 132K, swap peak: 132K, zswap: 23.3K) CPU: 512ms CGroup: /system.slice/system-devstack.slice/devstack@file_tracker.service ├─ 63498 /bin/bash /opt/stack/devstack/tools/file_tracker.sh └─122592 sleep 20 Apr 14 22:48:50 npc99a84aae6d34 file_tracker.sh[119357]: 6304 0 9223372036854775807 Apr 14 22:49:10 npc99a84aae6d34 file_tracker.sh[119454]: 6336 0 9223372036854775807 Apr 14 22:49:30 npc99a84aae6d34 file_tracker.sh[119468]: 6304 0 9223372036854775807 Apr 14 22:49:50 npc99a84aae6d34 file_tracker.sh[119483]: 6336 0 9223372036854775807 Apr 14 22:50:10 npc99a84aae6d34 file_tracker.sh[119584]: 6336 0 9223372036854775807 Apr 14 22:50:30 npc99a84aae6d34 file_tracker.sh[120721]: 6496 0 9223372036854775807 Apr 14 22:50:50 npc99a84aae6d34 file_tracker.sh[120813]: 6528 0 9223372036854775807 Apr 14 22:51:10 npc99a84aae6d34 file_tracker.sh[121184]: 6368 0 9223372036854775807 Apr 14 22:51:30 npc99a84aae6d34 file_tracker.sh[121585]: 6528 0 9223372036854775807 Apr 14 22:51:50 npc99a84aae6d34 file_tracker.sh[122591]: 6368 0 9223372036854775807 ● devstack@g-api.service - Devstack devstack@g-api.service Loaded: loaded (/etc/systemd/system/devstack@g-api.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:35:02 UTC; 16min ago Invocation: fd7d62d319234e0598ed3177ab081727 Main PID: 108887 (uwsgi) Status: "uWSGI is ready" Tasks: 11 (limit: 9256) Memory: 289M (peak: 325.4M) CPU: 9.512s CGroup: /system.slice/system-devstack.slice/devstack@g-api.service ├─108887 "glance-apiuWSGI master" ├─108888 "glance-apiuWSGI worker 1" └─108889 "glance-apiuWSGI worker 2" Apr 14 22:50:32 npc99a84aae6d34 devstack@g-api.service[108889]: DEBUG glance.api.middleware.version_negotiation [None req-c54809bd-b3e9-4717-8caa-e7555a6987e6 demo admin] new path /v2/images/d3170f3d-6e44-4717-a8b5-74b0b9638323 {{(pid=108889) process_request /opt/stack/glance/glance/api/middleware/version_negotiation.py:70}} Apr 14 22:50:32 npc99a84aae6d34 devstack@g-api.service[108889]: WARNING keystonemiddleware.auth_token [None req-c54809bd-b3e9-4717-8caa-e7555a6987e6 demo admin] A valid token was submitted as a service token, but it was not a valid service token. This is incorrect but backwards compatible behaviour. This will be removed in future releases. Apr 14 22:50:32 npc99a84aae6d34 devstack@g-api.service[108889]: [pid: 108889|app: 0|req: 10/21] ::1 () {44 vars in 1192 bytes} [Tue Apr 14 22:50:32 2026] GET /v2/images/d3170f3d-6e44-4717-a8b5-74b0b9638323 => generated 1111 bytes in 79 msecs (HTTP/1.1 200) 4 headers in 158 bytes (1 switches on core 0) Apr 14 22:50:32 npc99a84aae6d34 devstack@g-api.service[108888]: DEBUG glance.api.middleware.version_negotiation [None req-b52006c1-fb30-4c4b-8ecd-80088210eaf3 admin admin] Determining version of request: GET /v2/schemas/image Accept: */* {{(pid=108888) process_request /opt/stack/glance/glance/api/middleware/version_negotiation.py:44}} Apr 14 22:50:32 npc99a84aae6d34 devstack@g-api.service[108888]: DEBUG glance.api.middleware.version_negotiation [None req-b52006c1-fb30-4c4b-8ecd-80088210eaf3 admin admin] Using url versioning {{(pid=108888) process_request /opt/stack/glance/glance/api/middleware/version_negotiation.py:57}} Apr 14 22:50:32 npc99a84aae6d34 devstack@g-api.service[108888]: DEBUG glance.api.middleware.version_negotiation [None req-b52006c1-fb30-4c4b-8ecd-80088210eaf3 admin admin] Matched version: v2 {{(pid=108888) process_request /opt/stack/glance/glance/api/middleware/version_negotiation.py:69}} Apr 14 22:50:32 npc99a84aae6d34 devstack@g-api.service[108888]: DEBUG glance.api.middleware.version_negotiation [None req-b52006c1-fb30-4c4b-8ecd-80088210eaf3 admin admin] new path /v2/schemas/image {{(pid=108888) process_request /opt/stack/glance/glance/api/middleware/version_negotiation.py:70}} Apr 14 22:50:32 npc99a84aae6d34 devstack@g-api.service[108888]: WARNING keystonemiddleware.auth_token [None req-b52006c1-fb30-4c4b-8ecd-80088210eaf3 admin admin] A valid token was submitted as a service token, but it was not a valid service token. This is incorrect but backwards compatible behaviour. This will be removed in future releases. Apr 14 22:50:32 npc99a84aae6d34 devstack@g-api.service[108888]: [pid: 108888|app: 0|req: 12/22] ::1 () {44 vars in 1132 bytes} [Tue Apr 14 22:50:32 2026] GET /v2/schemas/image => generated 6083 bytes in 8 msecs (HTTP/1.1 200) 4 headers in 158 bytes (1 switches on core 0) Apr 14 22:50:42 npc99a84aae6d34 devstack@g-api.service[108889]: DEBUG dbcounter [-] [108889] Writing DB stats glance:SELECT=5 {{(pid=108889) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} ● devstack@keystone.service - Devstack devstack@keystone.service Loaded: loaded (/etc/systemd/system/devstack@keystone.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:28:10 UTC; 23min ago Invocation: 5a3e8a2ed148412faee0f578dd05c77b Main PID: 65424 (uwsgi) Status: "uWSGI is ready" Tasks: 5 (limit: 9256) Memory: 247.3M (peak: 256.6M, swap: 6.7M, swap peak: 6.7M, zswap: 2.1M) CPU: 1min 54.302s CGroup: /system.slice/system-devstack.slice/devstack@keystone.service ├─65424 "keystoneuWSGI master" ├─65425 "keystoneuWSGI worker 1" └─65426 "keystoneuWSGI worker 2" Apr 14 22:51:01 npc99a84aae6d34 devstack@keystone.service[65426]: DEBUG keystone.server.flask.request_processing.middleware.auth_context [None req-bc4b3c64-27f2-46fd-bb9c-313608864c6e None None] Validating token access rules against request {{(pid=65426) validate_allowed_request /opt/stack/data/venv/lib/python3.13/site-packages/keystonemiddleware/auth_token/__init__.py:545}} Apr 14 22:51:01 npc99a84aae6d34 devstack@keystone.service[65426]: DEBUG keystone.server.flask.request_processing.middleware.auth_context [None req-1f4ba164-7669-486a-a8ab-166eaa373bdb service placement] RBAC: auth_context: {'token': , 'domain_id': None, 'trust_id': None, 'trustor_id': None, 'trustee_id': None, 'domain_name': None, 'group_ids': [], 'user_id': 'cdb1b8011e0740978f1f074fc6cab814', 'user_domain_id': 'default', 'system_scope': None, 'project_id': '938d5fb78ba44eefa201afd3873e898e', 'project_domain_id': 'default', 'roles': ['reader', 'service', 'member', 'admin', 'manager'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65426) fill_context /opt/stack/keystone/keystone/server/flask/request_processing/middleware/auth_context.py:509}} Apr 14 22:51:01 npc99a84aae6d34 devstack@keystone.service[65426]: DEBUG keystone.server.flask.request_processing.req_logging [None req-1f4ba164-7669-486a-a8ab-166eaa373bdb service placement] REQUEST_METHOD: `GET` {{(pid=65426) log_request_info /opt/stack/keystone/keystone/server/flask/request_processing/req_logging.py:26}} Apr 14 22:51:01 npc99a84aae6d34 devstack@keystone.service[65426]: DEBUG keystone.server.flask.request_processing.req_logging [None req-1f4ba164-7669-486a-a8ab-166eaa373bdb service placement] SCRIPT_NAME: `/identity` {{(pid=65426) log_request_info /opt/stack/keystone/keystone/server/flask/request_processing/req_logging.py:27}} Apr 14 22:51:01 npc99a84aae6d34 devstack@keystone.service[65426]: DEBUG keystone.server.flask.request_processing.req_logging [None req-1f4ba164-7669-486a-a8ab-166eaa373bdb service placement] PATH_INFO: `/v3/auth/tokens` {{(pid=65426) log_request_info /opt/stack/keystone/keystone/server/flask/request_processing/req_logging.py:28}} Apr 14 22:51:01 npc99a84aae6d34 devstack@keystone.service[65426]: DEBUG keystone.common.fernet_utils [None req-1f4ba164-7669-486a-a8ab-166eaa373bdb service placement] Loaded 2 Fernet keys from /etc/keystone/fernet-keys/, but `[fernet_tokens] max_active_keys = 3`; perhaps there have not been enough key rotations to reach `max_active_keys` yet? {{(pid=65426) load_keys /opt/stack/keystone/keystone/common/fernet_utils.py:297}} Apr 14 22:51:01 npc99a84aae6d34 devstack@keystone.service[65426]: DEBUG keystone.common.rbac_enforcer.enforcer [None req-1f4ba164-7669-486a-a8ab-166eaa373bdb service placement] RBAC: Authorizing `identity:validate_token()` {{(pid=65426) enforce_call /opt/stack/keystone/keystone/common/rbac_enforcer/enforcer.py:494}} Apr 14 22:51:01 npc99a84aae6d34 devstack@keystone.service[65426]: DEBUG keystone.common.rbac_enforcer.enforcer [None req-1f4ba164-7669-486a-a8ab-166eaa373bdb service placement] RBAC: Authorization granted {{(pid=65426) enforce_call /opt/stack/keystone/keystone/common/rbac_enforcer/enforcer.py:505}} Apr 14 22:51:01 npc99a84aae6d34 devstack@keystone.service[65426]: [pid: 65426|app: 0|req: 606/1214] 2001:41d0:302:1000::cbc () {68 vars in 1578 bytes} [Tue Apr 14 22:51:01 2026] GET /identity/v3/auth/tokens => generated 3245 bytes in 205 msecs (HTTP/1.1 200) 6 headers in 380 bytes (1 switches on core 0) Apr 14 22:51:11 npc99a84aae6d34 devstack@keystone.service[65426]: DEBUG dbcounter [-] [65426] Writing DB stats keystone:SELECT=70 {{(pid=65426) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} ● devstack@memory_tracker.service - Devstack devstack@memory_tracker.service Loaded: loaded (/etc/systemd/system/devstack@memory_tracker.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:27:47 UTC; 24min ago Invocation: befa6acd92c147d9a0a0151f8761144f Main PID: 62947 (memory_tracker.) Tasks: 2 (limit: 9256) Memory: 1.7M (peak: 9.3M, swap: 140K, swap peak: 140K, zswap: 14.9K) CPU: 3.259s CGroup: /system.slice/system-devstack.slice/devstack@memory_tracker.service ├─ 62947 /bin/bash /opt/stack/devstack/tools/memory_tracker.sh └─122596 sleep 20 Apr 14 22:47:30 npc99a84aae6d34 memory_tracker.sh[62947]: [[[ Apr 14 22:47:30 npc99a84aae6d34 memory_tracker.sh[119217]: Tue Apr 14 22:47:30 UTC 2026 Apr 14 22:47:30 npc99a84aae6d34 memory_tracker.sh[62947]: --- Apr 14 22:47:30 npc99a84aae6d34 memory_tracker.sh[119219]: [iscsid (pid:43269)]=15412KB; [ovs-vswitchd (pid:93684)]=904824KB Apr 14 22:47:30 npc99a84aae6d34 memory_tracker.sh[62947]: ]]] Apr 14 22:47:50 npc99a84aae6d34 memory_tracker.sh[62947]: [[[ Apr 14 22:47:50 npc99a84aae6d34 memory_tracker.sh[119230]: Tue Apr 14 22:47:50 UTC 2026 Apr 14 22:47:50 npc99a84aae6d34 memory_tracker.sh[62947]: --- Apr 14 22:47:50 npc99a84aae6d34 memory_tracker.sh[119232]: [iscsid (pid:43269)]=15412KB; [ovs-vswitchd (pid:93684)]=904824KB Apr 14 22:47:50 npc99a84aae6d34 memory_tracker.sh[62947]: ]]] ● devstack@n-api-meta.service - Devstack devstack@n-api-meta.service Loaded: loaded (/etc/systemd/system/devstack@n-api-meta.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:34:29 UTC; 17min ago Invocation: 01f5c69e78214748ba9c3e9790190e28 Main PID: 101619 (uwsgi) Status: "uWSGI is ready" Tasks: 16 (limit: 9256) Memory: 210.6M (peak: 227.3M, swap: 26.2M, swap peak: 26.2M, zswap: 10.3M) CPU: 9.236s CGroup: /system.slice/system-devstack.slice/devstack@n-api-meta.service ├─101619 "nova-api-metauWSGI master" ├─101621 "nova-api-metauWSGI worker 1" ├─101622 "nova-api-metauWSGI worker 2" └─101623 "nova-api-metauWSGI http 1" Apr 14 22:34:34 npc99a84aae6d34 devstack@n-api-meta.service[101622]: DEBUG nova.api.openstack.wsgi_app [None req-55d744c2-317c-4946-add1-e473a218552f None None] ******************************************************************************** {{(pid=101622) log_opt_values /opt/stack/data/venv/lib/python3.13/site-packages/oslo_config/cfg.py:3013}} Apr 14 22:34:34 npc99a84aae6d34 devstack@n-api-meta.service[101622]: WSGI app 0 (mountpoint='') ready in 4 seconds on interpreter 0x7f8bae129610 pid: 101622 (default app) Apr 14 22:34:43 npc99a84aae6d34 devstack@n-api-meta.service[101621]: DEBUG dbcounter [-] [101621] Writing DB stats nova_api:SELECT=1 {{(pid=101621) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:34:43 npc99a84aae6d34 devstack@n-api-meta.service[101621]: DEBUG dbcounter [-] [101621] Writing DB stats nova_cell1:SELECT=1 {{(pid=101621) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:34:43 npc99a84aae6d34 devstack@n-api-meta.service[101621]: DEBUG dbcounter [-] [101621] Writing DB stats nova_cell0:SELECT=1 {{(pid=101621) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:34:43 npc99a84aae6d34 devstack@n-api-meta.service[101621]: DEBUG dbcounter [-] [101621] Writing DB stats nova_cell0:SELECT=2,nova_cell0:INSERT=1 {{(pid=101621) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:34:43 npc99a84aae6d34 devstack@n-api-meta.service[101622]: DEBUG dbcounter [-] [101622] Writing DB stats nova_api:SELECT=1 {{(pid=101622) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:34:44 npc99a84aae6d34 devstack@n-api-meta.service[101622]: DEBUG dbcounter [-] [101622] Writing DB stats nova_cell1:SELECT=1 {{(pid=101622) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:34:44 npc99a84aae6d34 devstack@n-api-meta.service[101622]: DEBUG dbcounter [-] [101622] Writing DB stats nova_cell0:SELECT=1 {{(pid=101622) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:34:44 npc99a84aae6d34 devstack@n-api-meta.service[101622]: DEBUG dbcounter [-] [101622] Writing DB stats nova_cell0:SELECT=1 {{(pid=101622) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} ● devstack@n-api.service - Devstack devstack@n-api.service Loaded: loaded (/etc/systemd/system/devstack@n-api.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:33:22 UTC; 18min ago Invocation: 1421191c68db42fa9a85fe4dc334aba3 Main PID: 92903 (uwsgi) Status: "uWSGI is ready" Tasks: 18 (limit: 9256) Memory: 195.4M (peak: 278.9M, swap: 137.1M, swap peak: 143.5M, zswap: 45.5M) CPU: 11.086s CGroup: /system.slice/system-devstack.slice/devstack@n-api.service ├─92903 "nova-apiuWSGI master" ├─92904 "nova-apiuWSGI worker 1" └─92905 "nova-apiuWSGI worker 2" Apr 14 22:50:33 npc99a84aae6d34 devstack@n-api.service[92905]: DEBUG nova.virt.hardware [None req-59c4bb06-d4f3-4fd6-9f20-7bc224ddec87 demo admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=92905) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:500}} Apr 14 22:50:33 npc99a84aae6d34 devstack@n-api.service[92905]: WARNING nova.compute.api [None req-59c4bb06-d4f3-4fd6-9f20-7bc224ddec87 demo admin] No host-to-cell mapping found for host np7792c2587c804.: nova.exception.HostMappingNotFound: Host 'np7792c2587c804' is not mapped to any cell Apr 14 22:50:33 npc99a84aae6d34 devstack@n-api.service[92905]: INFO nova.api.openstack.wsgi [None req-59c4bb06-d4f3-4fd6-9f20-7bc224ddec87 demo admin] HTTP exception thrown: Compute host np7792c2587c804 could not be found. Apr 14 22:50:33 npc99a84aae6d34 devstack@n-api.service[92905]: DEBUG nova.api.openstack.wsgi [None req-59c4bb06-d4f3-4fd6-9f20-7bc224ddec87 demo admin] Returning 400 to user: Compute host np7792c2587c804 could not be found. {{(pid=92905) __call__ /opt/stack/nova/nova/api/openstack/wsgi.py:908}} Apr 14 22:50:33 npc99a84aae6d34 devstack@n-api.service[92905]: INFO nova.api.openstack.requestlog [None req-59c4bb06-d4f3-4fd6-9f20-7bc224ddec87 demo admin] 2001:41d0:302:1000::cbc "POST /compute/v2.1/servers" status: 400 len: 92 microversion: 2.74 time: 1.227231 Apr 14 22:50:33 npc99a84aae6d34 devstack@n-api.service[92905]: [pid: 92905|app: 0|req: 17/35] 2001:41d0:302:1000::cbc () {72 vars in 1499 bytes} [Tue Apr 14 22:50:31 2026] POST /compute/v2.1/servers => generated 92 bytes in 1228 msecs (HTTP/1.1 400) 9 headers in 382 bytes (1 switches on core 0) Apr 14 22:50:41 npc99a84aae6d34 devstack@n-api.service[92904]: DEBUG dbcounter [-] [92904] Writing DB stats nova_api:SELECT=2 {{(pid=92904) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:50:42 npc99a84aae6d34 devstack@n-api.service[92905]: DEBUG dbcounter [-] [92905] Writing DB stats nova_cell0:SELECT=1 {{(pid=92905) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:50:42 npc99a84aae6d34 devstack@n-api.service[92905]: DEBUG dbcounter [-] [92905] Writing DB stats nova_cell1:SELECT=1 {{(pid=92905) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:50:43 npc99a84aae6d34 devstack@n-api.service[92905]: DEBUG dbcounter [-] [92905] Writing DB stats nova_api:SELECT=2 {{(pid=92905) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} ● devstack@n-cond-cell1.service - Devstack devstack@n-cond-cell1.service Loaded: loaded (/etc/systemd/system/devstack@n-cond-cell1.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:34:37 UTC; 17min ago Invocation: 514ebcabbeca46eaa104edded12b9b15 Main PID: 103735 (nova-conductor:) Tasks: 137 (limit: 9256) Memory: 186.4M (peak: 187.1M) CPU: 15.668s CGroup: /system.slice/system-devstack.slice/devstack@n-cond-cell1.service ├─103735 "nova-conductor: master process [/opt/stack/data/venv/bin/nova-conductor --config-file /etc/nova/nova_cell1.conf]" ├─104902 "nova-conductor: ServiceWrapper worker(0)" └─104904 "nova-conductor: ServiceWrapper worker(1)" Apr 14 22:51:26 npc99a84aae6d34 nova-conductor[104904]: DEBUG oslo.service.backend._threading.loopingcall [-] Fixed interval looping call 'nova.servicegroup.drivers.db.DbDriver._report_state' sleeping for 9.99 seconds {{(pid=104904) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:51:36 npc99a84aae6d34 nova-conductor[104904]: DEBUG dbcounter [-] [104904] Writing DB stats nova_cell1:UPDATE=7,nova_cell1:SELECT=10 {{(pid=104904) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:51:36 npc99a84aae6d34 nova-conductor[104902]: DEBUG oslo.service.backend._threading.loopingcall [-] Fixed interval looping call 'nova.servicegroup.drivers.db.DbDriver._report_state' sleeping for 9.99 seconds {{(pid=104902) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:51:36 npc99a84aae6d34 nova-conductor[104904]: DEBUG oslo.service.backend._threading.loopingcall [-] Fixed interval looping call 'nova.servicegroup.drivers.db.DbDriver._report_state' sleeping for 9.97 seconds {{(pid=104904) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:51:46 npc99a84aae6d34 nova-conductor[104902]: DEBUG dbcounter [-] [104902] Writing DB stats nova_cell1:SELECT=8,nova_cell1:UPDATE=4 {{(pid=104902) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:51:46 npc99a84aae6d34 nova-conductor[104904]: DEBUG oslo.service.backend._threading.loopingcall [-] Fixed interval looping call 'nova.servicegroup.drivers.db.DbDriver._report_state' sleeping for 9.99 seconds {{(pid=104904) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:51:46 npc99a84aae6d34 nova-conductor[104902]: DEBUG oslo.service.backend._threading.loopingcall [-] Fixed interval looping call 'nova.servicegroup.drivers.db.DbDriver._report_state' sleeping for 9.99 seconds {{(pid=104902) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:51:54 npc99a84aae6d34 nova-conductor[104904]: DEBUG oslo.service.backend._threading.loopingcall [None req-14b21a59-40f9-4be6-aa67-f6fd98ded260 None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=104904) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:51:56 npc99a84aae6d34 nova-conductor[104904]: DEBUG oslo.service.backend._threading.loopingcall [-] Fixed interval looping call 'nova.servicegroup.drivers.db.DbDriver._report_state' sleeping for 9.99 seconds {{(pid=104904) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:51:56 npc99a84aae6d34 nova-conductor[104902]: DEBUG oslo.service.backend._threading.loopingcall [-] Fixed interval looping call 'nova.servicegroup.drivers.db.DbDriver._report_state' sleeping for 9.99 seconds {{(pid=104902) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} ● devstack@n-cpu.service - Devstack devstack@n-cpu.service Loaded: loaded (/etc/systemd/system/devstack@n-cpu.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:34:41 UTC; 17min ago Invocation: e5228bf7e62d473391a64ae34fb0ba05 Main PID: 104923 (nova-compute) Tasks: 10 (limit: 9256) Memory: 151.1M (peak: 184.4M) CPU: 31.484s CGroup: /system.slice/system-devstack.slice/devstack@n-cpu.service └─104923 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/nova-compute --config-file /etc/nova/nova-cpu.conf Apr 14 22:51:02 npc99a84aae6d34 nova-compute[104923]: DEBUG nova.scheduler.client.report [None req-df2c4159-497e-487a-8573-5150acef5506 None None] Inventory has not changed for provider 14e023c6-ad69-4588-bce8-864262a20ebe based on inventory data: {'VCPU': {'total': 8, 'reserved': 0, 'min_unit': 1, 'max_unit': 8, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 7756, 'reserved': 512, 'min_unit': 1, 'max_unit': 7756, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 29, 'reserved': 0, 'min_unit': 1, 'max_unit': 29, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=104923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} Apr 14 22:51:03 npc99a84aae6d34 nova-compute[104923]: DEBUG nova.compute.resource_tracker [None req-df2c4159-497e-487a-8573-5150acef5506 None None] Compute_service record updated for npc99a84aae6d34:npc99a84aae6d34 {{(pid=104923) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} Apr 14 22:51:03 npc99a84aae6d34 nova-compute[104923]: DEBUG oslo_concurrency.lockutils [None req-df2c4159-497e-487a-8573-5150acef5506 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.019s {{(pid=104923) inner /opt/stack/data/venv/lib/python3.13/site-packages/oslo_concurrency/lockutils.py:538}} Apr 14 22:51:03 npc99a84aae6d34 nova-compute[104923]: DEBUG oslo.service.backend._threading.loopingcall [None req-df2c4159-497e-487a-8573-5150acef5506 None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 1.00 seconds {{(pid=104923) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:51:04 npc99a84aae6d34 nova-compute[104923]: DEBUG oslo_service.periodic_task [None req-df2c4159-497e-487a-8573-5150acef5506 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=104923) run_periodic_tasks /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/periodic_task.py:210}} Apr 14 22:51:04 npc99a84aae6d34 nova-compute[104923]: DEBUG oslo_service.periodic_task [None req-df2c4159-497e-487a-8573-5150acef5506 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=104923) run_periodic_tasks /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/periodic_task.py:210}} Apr 14 22:51:04 npc99a84aae6d34 nova-compute[104923]: DEBUG oslo_service.periodic_task [None req-df2c4159-497e-487a-8573-5150acef5506 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=104923) run_periodic_tasks /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/periodic_task.py:210}} Apr 14 22:51:04 npc99a84aae6d34 nova-compute[104923]: DEBUG oslo_service.periodic_task [None req-df2c4159-497e-487a-8573-5150acef5506 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=104923) run_periodic_tasks /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/periodic_task.py:210}} Apr 14 22:51:04 npc99a84aae6d34 nova-compute[104923]: DEBUG oslo.service.backend._threading.loopingcall [None req-df2c4159-497e-487a-8573-5150acef5506 None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 52.74 seconds {{(pid=104923) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:51:18 npc99a84aae6d34 nova-compute[104923]: DEBUG oslo.service.backend._threading.loopingcall [-] Fixed interval looping call 'nova.servicegroup.drivers.db.DbDriver._report_state' sleeping for 119.49 seconds {{(pid=104923) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} ● devstack@n-novnc-cell1.service - Devstack devstack@n-novnc-cell1.service Loaded: loaded (/etc/systemd/system/devstack@n-novnc-cell1.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:34:32 UTC; 17min ago Invocation: 55993ee0bd2742ce86dfb280315b1b80 Main PID: 102318 (nova-novncproxy) Tasks: 8 (limit: 9256) Memory: 115M (peak: 120.5M, swap: 692K, swap peak: 692K, zswap: 178.7K) CPU: 4.290s CGroup: /system.slice/system-devstack.slice/devstack@n-novnc-cell1.service └─102318 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/nova-novncproxy --config-file /etc/nova/nova_cell1.conf --web /opt/stack/novnc Apr 14 22:34:33 npc99a84aae6d34 nova-novncproxy[102318]: we strongly recommend against using it for new projects. Apr 14 22:34:33 npc99a84aae6d34 nova-novncproxy[102318]: If you are already using Eventlet, we recommend migrating to a different Apr 14 22:34:33 npc99a84aae6d34 nova-novncproxy[102318]: framework. For more detail see Apr 14 22:34:33 npc99a84aae6d34 nova-novncproxy[102318]: https://eventlet.readthedocs.io/en/latest/asyncio/migration.html Apr 14 22:34:33 npc99a84aae6d34 nova-novncproxy[102318]: import eventlet Apr 14 22:34:36 npc99a84aae6d34 nova-novncproxy[102318]: INFO nova.console.websocketproxy [-] WebSocket server settings: Apr 14 22:34:36 npc99a84aae6d34 nova-novncproxy[102318]: INFO nova.console.websocketproxy [-]  - Listen on :::6080 Apr 14 22:34:36 npc99a84aae6d34 nova-novncproxy[102318]: INFO nova.console.websocketproxy [-]  - Web server (no directory listings). Web root: /opt/stack/novnc Apr 14 22:34:36 npc99a84aae6d34 nova-novncproxy[102318]: INFO nova.console.websocketproxy [-]  - No SSL/TLS support (no cert file) Apr 14 22:34:36 npc99a84aae6d34 nova-novncproxy[102318]: INFO nova.console.websocketproxy [-]  - proxying from :::6080 to None:None ● devstack@n-sch.service - Devstack devstack@n-sch.service Loaded: loaded (/etc/systemd/system/devstack@n-sch.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:34:27 UTC; 17min ago Invocation: 09db5691b6ba4cca98f0651aeb30290d Main PID: 100952 (nova-scheduler:) Tasks: 19 (limit: 9256) Memory: 164.1M (peak: 166.2M, swap: 3.4M, swap peak: 3.4M, zswap: 1.2M) CPU: 4.540s CGroup: /system.slice/system-devstack.slice/devstack@n-sch.service ├─100952 "nova-scheduler: master process [/opt/stack/data/venv/bin/nova-scheduler --config-file /etc/nova/nova.conf]" ├─102220 "nova-scheduler: ServiceWrapper worker(0)" └─102223 "nova-scheduler: ServiceWrapper worker(1)" Apr 14 22:49:28 npc99a84aae6d34 nova-scheduler[102220]: DEBUG oslo.service.backend._threading.loopingcall [None req-7cc405b2-3d31-43a3-b3f4-71783d15e5f2 None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=102220) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:49:33 npc99a84aae6d34 nova-scheduler[102223]: DEBUG oslo.service.backend._threading.loopingcall [None req-0fd2ef85-94bc-4111-932f-942a82ec16e8 None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=102223) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:50:28 npc99a84aae6d34 nova-scheduler[102220]: DEBUG oslo.service.backend._threading.loopingcall [None req-7cc405b2-3d31-43a3-b3f4-71783d15e5f2 None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=102220) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:50:33 npc99a84aae6d34 nova-scheduler[102223]: DEBUG oslo.service.backend._threading.loopingcall [None req-0fd2ef85-94bc-4111-932f-942a82ec16e8 None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=102223) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:50:37 npc99a84aae6d34 nova-scheduler[102220]: DEBUG oslo.service.backend._threading.loopingcall [-] Fixed interval looping call 'nova.servicegroup.drivers.db.DbDriver._report_state' sleeping for 119.99 seconds {{(pid=102220) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:50:37 npc99a84aae6d34 nova-scheduler[102223]: DEBUG oslo.service.backend._threading.loopingcall [-] Fixed interval looping call 'nova.servicegroup.drivers.db.DbDriver._report_state' sleeping for 119.99 seconds {{(pid=102223) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:50:47 npc99a84aae6d34 nova-scheduler[102223]: DEBUG dbcounter [-] [102223] Writing DB stats nova_cell0:SELECT=1,nova_cell0:UPDATE=1 {{(pid=102223) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:50:47 npc99a84aae6d34 nova-scheduler[102220]: DEBUG dbcounter [-] [102220] Writing DB stats nova_cell0:SELECT=1,nova_cell0:UPDATE=1 {{(pid=102220) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:51:28 npc99a84aae6d34 nova-scheduler[102220]: DEBUG oslo.service.backend._threading.loopingcall [None req-7cc405b2-3d31-43a3-b3f4-71783d15e5f2 None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=102220) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:51:33 npc99a84aae6d34 nova-scheduler[102223]: DEBUG oslo.service.backend._threading.loopingcall [None req-0fd2ef85-94bc-4111-932f-942a82ec16e8 None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=102223) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} ● devstack@n-super-cond.service - Devstack devstack@n-super-cond.service Loaded: loaded (/etc/systemd/system/devstack@n-super-cond.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:34:35 UTC; 17min ago Invocation: 89bec8ce55b545a08619c7e14cb01d87 Main PID: 103157 (nova-conductor:) Tasks: 19 (limit: 9256) Memory: 162M (peak: 162.6M, swap: 312K, swap peak: 312K, zswap: 94.8K) CPU: 4.144s CGroup: /system.slice/system-devstack.slice/devstack@n-super-cond.service ├─103157 "nova-conductor: master process [/opt/stack/data/venv/bin/nova-conductor --config-file /etc/nova/nova.conf]" ├─104721 "nova-conductor: ServiceWrapper worker(0)" └─104725 "nova-conductor: ServiceWrapper worker(1)" Apr 14 22:49:10 npc99a84aae6d34 nova-conductor[104721]: DEBUG oslo.service.backend._threading.loopingcall [None req-ef455d0a-f04f-4f86-b220-4a859cb22cee None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=104721) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:49:19 npc99a84aae6d34 nova-conductor[104725]: DEBUG oslo.service.backend._threading.loopingcall [None req-1f2a7cb7-55c6-4bf5-bf5b-aebd87f8656e None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=104725) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:50:10 npc99a84aae6d34 nova-conductor[104721]: DEBUG oslo.service.backend._threading.loopingcall [None req-ef455d0a-f04f-4f86-b220-4a859cb22cee None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=104721) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:50:19 npc99a84aae6d34 nova-conductor[104725]: DEBUG oslo.service.backend._threading.loopingcall [None req-1f2a7cb7-55c6-4bf5-bf5b-aebd87f8656e None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=104725) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:50:45 npc99a84aae6d34 nova-conductor[104725]: DEBUG oslo.service.backend._threading.loopingcall [-] Fixed interval looping call 'nova.servicegroup.drivers.db.DbDriver._report_state' sleeping for 119.99 seconds {{(pid=104725) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:50:45 npc99a84aae6d34 nova-conductor[104721]: DEBUG oslo.service.backend._threading.loopingcall [-] Fixed interval looping call 'nova.servicegroup.drivers.db.DbDriver._report_state' sleeping for 119.99 seconds {{(pid=104721) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:50:55 npc99a84aae6d34 nova-conductor[104725]: DEBUG dbcounter [-] [104725] Writing DB stats nova_cell0:SELECT=1,nova_cell0:UPDATE=1 {{(pid=104725) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:50:55 npc99a84aae6d34 nova-conductor[104721]: DEBUG dbcounter [-] [104721] Writing DB stats nova_cell0:SELECT=1 {{(pid=104721) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:51:10 npc99a84aae6d34 nova-conductor[104721]: DEBUG oslo.service.backend._threading.loopingcall [None req-ef455d0a-f04f-4f86-b220-4a859cb22cee None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=104721) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:51:19 npc99a84aae6d34 nova-conductor[104725]: DEBUG oslo.service.backend._threading.loopingcall [None req-1f2a7cb7-55c6-4bf5-bf5b-aebd87f8656e None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=104725) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} ● devstack@neutron-api.service - Devstack devstack@neutron-api.service Loaded: loaded (/etc/systemd/system/devstack@neutron-api.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:33:45 UTC; 18min ago Invocation: 4134b49eda9c43b687e207f2d8f4beca Main PID: 96269 (uwsgi) Status: "uWSGI is ready" Tasks: 101 (limit: 9256) Memory: 371.9M (peak: 389.7M, swap: 30.5M, swap peak: 30.6M, zswap: 9.7M) CPU: 41.869s CGroup: /system.slice/system-devstack.slice/devstack@neutron-api.service ├─96269 "neutron-apiuWSGI master" ├─96270 "neutron-apiuWSGI worker 1" └─96271 "neutron-apiuWSGI worker 2" Apr 14 22:51:37 npc99a84aae6d34 devstack@neutron-api.service[96270]: DEBUG futurist.periodics [-] Submitting periodic callback 'neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance.HashRingHealthCheckPeriodics.touch_hash_ring_node' {{(pid=96270) _process_scheduled /opt/stack/data/venv/lib/python3.13/site-packages/futurist/periodics.py:784}} Apr 14 22:51:37 npc99a84aae6d34 devstack@neutron-api.service[96270]: DEBUG neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance [-] Touching Hash Ring node "23e5c09fc6575826bd636a7824484a34" from periodic health check thread {{(pid=96270) touch_hash_ring_node /opt/stack/neutron/neutron/plugins/ml2/drivers/ovn/mech_driver/ovsdb/maintenance.py:1401}} Apr 14 22:51:38 npc99a84aae6d34 devstack@neutron-api.service[96271]: DEBUG futurist.periodics [-] Submitting periodic callback 'neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance.HashRingHealthCheckPeriodics.touch_hash_ring_node' {{(pid=96271) _process_scheduled /opt/stack/data/venv/lib/python3.13/site-packages/futurist/periodics.py:784}} Apr 14 22:51:38 npc99a84aae6d34 devstack@neutron-api.service[96271]: DEBUG neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance [-] Touching Hash Ring node "4e9ca10919e351f7b8e2b7a65a3ed572" from periodic health check thread {{(pid=96271) touch_hash_ring_node /opt/stack/neutron/neutron/plugins/ml2/drivers/ovn/mech_driver/ovsdb/maintenance.py:1401}} Apr 14 22:51:47 npc99a84aae6d34 devstack@neutron-api.service[96270]: DEBUG dbcounter [-] [96270] Writing DB stats neutron:UPDATE=1,neutron:SELECT=1 {{(pid=96270) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:51:48 npc99a84aae6d34 devstack@neutron-api.service[96271]: DEBUG dbcounter [-] [96271] Writing DB stats neutron:UPDATE=1,neutron:SELECT=1 {{(pid=96271) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:51:52 npc99a84aae6d34 devstack@neutron-api.service[96270]: DEBUG futurist.periodics [-] Submitting periodic callback 'neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance.HashRingHealthCheckPeriodics.touch_hash_ring_node' {{(pid=96270) _process_scheduled /opt/stack/data/venv/lib/python3.13/site-packages/futurist/periodics.py:784}} Apr 14 22:51:52 npc99a84aae6d34 devstack@neutron-api.service[96270]: DEBUG neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance [-] Touching Hash Ring node "23e5c09fc6575826bd636a7824484a34" from periodic health check thread {{(pid=96270) touch_hash_ring_node /opt/stack/neutron/neutron/plugins/ml2/drivers/ovn/mech_driver/ovsdb/maintenance.py:1401}} Apr 14 22:51:53 npc99a84aae6d34 devstack@neutron-api.service[96271]: DEBUG futurist.periodics [-] Submitting periodic callback 'neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance.HashRingHealthCheckPeriodics.touch_hash_ring_node' {{(pid=96271) _process_scheduled /opt/stack/data/venv/lib/python3.13/site-packages/futurist/periodics.py:784}} Apr 14 22:51:53 npc99a84aae6d34 devstack@neutron-api.service[96271]: DEBUG neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance [-] Touching Hash Ring node "4e9ca10919e351f7b8e2b7a65a3ed572" from periodic health check thread {{(pid=96271) touch_hash_ring_node /opt/stack/neutron/neutron/plugins/ml2/drivers/ovn/mech_driver/ovsdb/maintenance.py:1401}} ● devstack@neutron-ovn-maintenance-worker.service - Devstack devstack@neutron-ovn-maintenance-worker.service Loaded: loaded (/etc/systemd/system/devstack@neutron-ovn-maintenance-worker.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:33:52 UTC; 18min ago Invocation: 30f4c70613d14fcfb5b7b659a2e1aa4b Main PID: 97897 (neutron-ovn-mai) Tasks: 39 (limit: 9256) Memory: 216.3M (peak: 219M, swap: 3.4M, swap peak: 3.4M, zswap: 944.7K) CPU: 8.757s CGroup: /system.slice/system-devstack.slice/devstack@neutron-ovn-maintenance-worker.service ├─97897 "neutron-ovn-maintenance-worker: master process [/opt/stack/data/venv/bin/neutron-ovn-maintenance-worker --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" └─98765 "neutron-server: maintenance worker (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-ovn-maintenance-worker --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" Apr 14 22:44:16 npc99a84aae6d34 neutron-ovn-maintenance-worker[98765]: DEBUG neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance [-] OVN maintenance task: Checking Neutron and OVN revision consistency. {{(pid=98765) wrapper /opt/stack/neutron/neutron/plugins/ml2/drivers/ovn/mech_driver/ovsdb/maintenance.py:101}} Apr 14 22:44:16 npc99a84aae6d34 neutron-ovn-maintenance-worker[98765]: DEBUG neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance [None req-ba205fb8-e544-4386-872e-8c65d0f420e3 None None] Maintenance task: No inconsistencies found. Skipping {{(pid=98765) check_for_inconsistencies /opt/stack/neutron/neutron/plugins/ml2/drivers/ovn/mech_driver/ovsdb/maintenance.py:436}} Apr 14 22:44:16 npc99a84aae6d34 neutron-ovn-maintenance-worker[98765]: INFO neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance [None req-ba205fb8-e544-4386-872e-8c65d0f420e3 None None] OVN maintenance task check_for_inconsistencies finished in 0.015 seconds Apr 14 22:44:26 npc99a84aae6d34 neutron-ovn-maintenance-worker[98765]: DEBUG dbcounter [-] [98765] Writing DB stats neutron:SELECT=2 {{(pid=98765) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:49:16 npc99a84aae6d34 neutron-ovn-maintenance-worker[98765]: DEBUG futurist.periodics [-] Submitting periodic callback 'neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance.DBInconsistenciesPeriodics.check_for_inconsistencies' {{(pid=98765) _process_scheduled /opt/stack/data/venv/lib/python3.13/site-packages/futurist/periodics.py:784}} Apr 14 22:49:16 npc99a84aae6d34 neutron-ovn-maintenance-worker[98765]: DEBUG neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance [None req-ba205fb8-e544-4386-872e-8c65d0f420e3 None None] Starting OVN maintenance task: check_for_inconsistencies {{(pid=98765) wrapper /opt/stack/neutron/neutron/plugins/ml2/drivers/ovn/mech_driver/ovsdb/maintenance.py:99}} Apr 14 22:49:16 npc99a84aae6d34 neutron-ovn-maintenance-worker[98765]: DEBUG neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance [None req-ba205fb8-e544-4386-872e-8c65d0f420e3 None None] OVN maintenance task: Checking Neutron and OVN revision consistency. {{(pid=98765) wrapper /opt/stack/neutron/neutron/plugins/ml2/drivers/ovn/mech_driver/ovsdb/maintenance.py:101}} Apr 14 22:49:16 npc99a84aae6d34 neutron-ovn-maintenance-worker[98765]: DEBUG neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance [None req-ba205fb8-e544-4386-872e-8c65d0f420e3 None None] Maintenance task: No inconsistencies found. Skipping {{(pid=98765) check_for_inconsistencies /opt/stack/neutron/neutron/plugins/ml2/drivers/ovn/mech_driver/ovsdb/maintenance.py:436}} Apr 14 22:49:16 npc99a84aae6d34 neutron-ovn-maintenance-worker[98765]: INFO neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance [None req-ba205fb8-e544-4386-872e-8c65d0f420e3 None None] OVN maintenance task check_for_inconsistencies finished in 0.017 seconds Apr 14 22:49:26 npc99a84aae6d34 neutron-ovn-maintenance-worker[98765]: DEBUG dbcounter [-] [98765] Writing DB stats neutron:SELECT=2 {{(pid=98765) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} ● devstack@neutron-periodic-workers.service - Devstack devstack@neutron-periodic-workers.service Loaded: loaded (/etc/systemd/system/devstack@neutron-periodic-workers.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:33:50 UTC; 18min ago Invocation: 0b1d10da1ef444e191a21e8bf71af45f Main PID: 97349 (neutron-periodi) Tasks: 23 (limit: 9256) Memory: 207.5M (peak: 210.2M, swap: 17.2M, swap peak: 17.2M, zswap: 6.3M) CPU: 5.289s CGroup: /system.slice/system-devstack.slice/devstack@neutron-periodic-workers.service ├─97349 "neutron-periodic-workers: master process [/opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" ├─98512 "neutron-server: Periodic worker for \"AgentSchedulerDbMixin\" (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" ├─98520 "neutron-server: Periodic worker for \"AgentSchedulerDbMixin\" (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" ├─98524 "neutron-server: Periodic worker for \"DbQuotaNoLockDriver\" (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" └─98532 "neutron-server: Periodic worker for \"L3_NAT_dbonly_mixin\" (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" Apr 14 22:51:02 npc99a84aae6d34 neutron-periodic-workers[98520]: DEBUG neutron.db.agents_db [None req-b84310ad-4353-4aa2-bcaf-bdf5373cdc9e None None] Agent healthcheck: found 0 active agents {{(pid=98520) agent_health_check /opt/stack/neutron/neutron/db/agents_db.py:317}} Apr 14 22:51:02 npc99a84aae6d34 neutron-periodic-workers[98520]: DEBUG oslo.service.backend._threading.loopingcall [None req-b84310ad-4353-4aa2-bcaf-bdf5373cdc9e None None] Fixed interval looping call 'neutron.plugins.ml2.plugin.AgentDbMixin.agent_health_check' sleeping for 36.99 seconds {{(pid=98520) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:51:06 npc99a84aae6d34 neutron-periodic-workers[98512]: DEBUG dbcounter [-] [98512] Writing DB stats neutron:SELECT=3 {{(pid=98512) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:51:12 npc99a84aae6d34 neutron-periodic-workers[98520]: DEBUG dbcounter [-] [98520] Writing DB stats neutron:SELECT=1 {{(pid=98520) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:51:33 npc99a84aae6d34 neutron-periodic-workers[98512]: DEBUG oslo.service.backend._threading.loopingcall [None req-4d7fb020-fa14-426b-9f04-4d47a6379da4 None None] Fixed interval looping call 'neutron.plugins.ml2.plugin.DhcpAgentSchedulerDbMixin.remove_networks_from_down_agents' sleeping for 36.98 seconds {{(pid=98512) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:51:39 npc99a84aae6d34 neutron-periodic-workers[98520]: DEBUG neutron.db.agents_db [None req-b84310ad-4353-4aa2-bcaf-bdf5373cdc9e None None] Agent healthcheck: found 0 active agents {{(pid=98520) agent_health_check /opt/stack/neutron/neutron/db/agents_db.py:317}} Apr 14 22:51:39 npc99a84aae6d34 neutron-periodic-workers[98520]: DEBUG oslo.service.backend._threading.loopingcall [None req-b84310ad-4353-4aa2-bcaf-bdf5373cdc9e None None] Fixed interval looping call 'neutron.plugins.ml2.plugin.AgentDbMixin.agent_health_check' sleeping for 36.99 seconds {{(pid=98520) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 14 22:51:43 npc99a84aae6d34 neutron-periodic-workers[98512]: DEBUG dbcounter [-] [98512] Writing DB stats neutron:SELECT=3 {{(pid=98512) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:51:49 npc99a84aae6d34 neutron-periodic-workers[98520]: DEBUG dbcounter [-] [98520] Writing DB stats neutron:SELECT=1 {{(pid=98520) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:51:54 npc99a84aae6d34 neutron-periodic-workers[98524]: DEBUG oslo.service.backend._threading.loopingcall [None req-dfdfa238-f9a1-48e7-bf97-e4f73840af79 None None] Fixed interval looping call 'neutron.db.quota.driver_nolock.DbQuotaNoLockDriver._remove_expired_reservations' sleeping for 119.99 seconds {{(pid=98524) _run_loop /opt/stack/data/venv/lib/python3.13/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} ● devstack@neutron-rpc-server.service - Devstack devstack@neutron-rpc-server.service Loaded: loaded (/etc/systemd/system/devstack@neutron-rpc-server.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:33:47 UTC; 18min ago Invocation: 9b040c72c0e84f27931a5287effff9aa Main PID: 96802 (neutron-rpc-ser) Tasks: 16 (limit: 9256) Memory: 195.8M (peak: 196.9M, swap: 2.8M, swap peak: 2.8M, zswap: 964.8K) CPU: 5.225s CGroup: /system.slice/system-devstack.slice/devstack@neutron-rpc-server.service ├─96802 "neutron-rpc-server: master process [/opt/stack/data/venv/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" └─98148 "neutron-server: rpc worker (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" Apr 14 22:36:15 npc99a84aae6d34 neutron-rpc-server[98148]: DEBUG neutron.common.ovn.hash_ring_manager [-] Hash Ring loaded. 2 active nodes. 0 offline nodes {{(pid=98148) _load_hash_ring /opt/stack/neutron/neutron/common/ovn/hash_ring_manager.py:102}} Apr 14 22:36:18 npc99a84aae6d34 neutron-rpc-server[98148]: DEBUG neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.ovsdb_monitor [-] ChassisOVNAgentWriteEvent : Matched Chassis_Private, update, None None {{(pid=98148) matches /opt/stack/neutron/neutron/plugins/ml2/drivers/ovn/mech_driver/ovsdb/ovsdb_monitor.py:64}} Apr 14 22:36:25 npc99a84aae6d34 neutron-rpc-server[98148]: DEBUG dbcounter [-] [98148] Writing DB stats neutron:SELECT=4 {{(pid=98148) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:37:55 npc99a84aae6d34 neutron-rpc-server[98148]: DEBUG neutron.common.ovn.hash_ring_manager [-] Hash Ring loaded. 2 active nodes. 0 offline nodes {{(pid=98148) _load_hash_ring /opt/stack/neutron/neutron/common/ovn/hash_ring_manager.py:102}} Apr 14 22:37:56 npc99a84aae6d34 neutron-rpc-server[98148]: DEBUG neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.ovsdb_monitor [-] ChassisAgentWriteEvent : Matched Chassis_Private, update, None None {{(pid=98148) matches /opt/stack/neutron/neutron/plugins/ml2/drivers/ovn/mech_driver/ovsdb/ovsdb_monitor.py:64}} Apr 14 22:37:56 npc99a84aae6d34 neutron-rpc-server[98148]: DEBUG neutron.common.ovn.hash_ring_manager [-] Hash Ring loaded. 2 active nodes. 0 offline nodes {{(pid=98148) _load_hash_ring /opt/stack/neutron/neutron/common/ovn/hash_ring_manager.py:102}} Apr 14 22:38:02 npc99a84aae6d34 neutron-rpc-server[98148]: DEBUG neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.ovsdb_monitor [-] ChassisOVNAgentWriteEvent : Matched Chassis_Private, update, None None {{(pid=98148) matches /opt/stack/neutron/neutron/plugins/ml2/drivers/ovn/mech_driver/ovsdb/ovsdb_monitor.py:64}} Apr 14 22:38:06 npc99a84aae6d34 neutron-rpc-server[98148]: DEBUG dbcounter [-] [98148] Writing DB stats neutron:SELECT=4 {{(pid=98148) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:39:16 npc99a84aae6d34 neutron-rpc-server[98148]: DEBUG neutron.common.ovn.hash_ring_manager [-] Hash Ring loaded. 2 active nodes. 0 offline nodes {{(pid=98148) _load_hash_ring /opt/stack/neutron/neutron/common/ovn/hash_ring_manager.py:102}} Apr 14 22:39:26 npc99a84aae6d34 neutron-rpc-server[98148]: DEBUG dbcounter [-] [98148] Writing DB stats neutron:SELECT=2 {{(pid=98148) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} ● devstack@openstack-cli-server.service - Devstack devstack@openstack-cli-server.service Loaded: loaded (/etc/systemd/system/devstack@openstack-cli-server.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:27:27 UTC; 24min ago Invocation: 2a7cab51d84c42dda8369699f6cd1e03 Main PID: 61256 (python3) Tasks: 1 (limit: 9256) Memory: 161.5M (peak: 173.7M, swap: 372K, swap peak: 372K, zswap: 98.5K) CPU: 23.550s CGroup: /system.slice/system-devstack.slice/devstack@openstack-cli-server.service └─61256 /opt/stack/data/venv/bin/python3 /opt/stack/devstack/files/openstack-cli-server/openstack-cli-server Apr 14 22:36:01 npc99a84aae6d34 python3[61256]: openstack ['--os-cloud', 'devstack-admin', 'project', 'list'] Apr 14 22:36:01 npc99a84aae6d34 python3[61256]: openstack ['--os-cloud', 'devstack-admin', 'flavor', 'list'] Apr 14 22:36:02 npc99a84aae6d34 python3[61256]: openstack ['--os-cloud', 'devstack-admin', 'image', 'show', '38c5b16f-9596-4f4b-af78-5371b9296653', '-c', 'size', '-f', 'value'] Apr 14 22:36:02 npc99a84aae6d34 python3[61256]: openstack ['--os-cloud', 'devstack-admin', 'flavor', 'create', '--id', '42', '--ram', '192', '--disk', '1', '--vcpus', '1', '--property', 'hw_rng:allowed=True', 'm1.nano'] Apr 14 22:36:02 npc99a84aae6d34 python3[61256]: openstack ['--os-cloud', 'devstack-admin', 'image', 'show', 'd3170f3d-6e44-4717-a8b5-74b0b9638323', '-c', 'size', '-f', 'value'] Apr 14 22:36:03 npc99a84aae6d34 python3[61256]: openstack ['--os-cloud', 'devstack-admin', 'flavor', 'create', '--id', '84', '--ram', '256', '--disk', '1', '--vcpus', '1', '--property', 'hw_rng:allowed=True', 'm1.micro'] Apr 14 22:36:14 npc99a84aae6d34 python3[61256]: openstack ['--os-cloud', 'devstack-admin', '--os-region', 'RegionOne', 'extension', 'list', '--network', '-c', 'Alias', '-f', 'value'] Apr 14 22:36:14 npc99a84aae6d34 python3[61256]: openstack ['--os-cloud', 'devstack-admin', 'network', 'show', '-f', 'value', '-c', 'id', 'public'] Apr 14 22:36:15 npc99a84aae6d34 python3[61256]: openstack ['--os-cloud', 'devstack-admin', '--os-region', 'RegionOne', 'network', 'create', '--share', 'shared'] Apr 14 22:36:16 npc99a84aae6d34 python3[61256]: openstack ['--os-cloud', 'devstack-admin', '--os-region', 'RegionOne', 'subnet', 'create', '--description', 'shared-subnet', '--subnet-range', '192.168.233.0/24', '--network', 'shared', 'shared-subnet'] ● devstack@placement-api.service - Devstack devstack@placement-api.service Loaded: loaded (/etc/systemd/system/devstack@placement-api.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:33:55 UTC; 18min ago Invocation: 6868124620264d4da8eaab8d78b71d60 Main PID: 98699 (uwsgi) Status: "uWSGI is ready" Tasks: 5 (limit: 9256) Memory: 174.6M (peak: 176.8M, swap: 11.1M, swap peak: 11.1M, zswap: 887.5K) CPU: 5.707s CGroup: /system.slice/system-devstack.slice/devstack@placement-api.service ├─98699 "placementuWSGI master" ├─98700 "placementuWSGI worker 1" └─98701 "placementuWSGI worker 2" Apr 14 22:51:01 npc99a84aae6d34 devstack@placement-api.service[98701]: INFO placement.requestlog [req-55cc7c70-41be-48fb-b96d-5c1e4efca572 req-20e5e59d-ffcf-49e6-8fb3-93674da1a6cf service nova] 2001:41d0:302:1000::cbc "GET /placement/resource_providers/14e023c6-ad69-4588-bce8-864262a20ebe/aggregates" status: 200 len: 53 microversion: 1.19 Apr 14 22:51:01 npc99a84aae6d34 devstack@placement-api.service[98701]: [pid: 98701|app: 0|req: 27/54] 2001:41d0:302:1000::cbc () {68 vars in 1654 bytes} [Tue Apr 14 22:51:01 2026] GET /placement/resource_providers/14e023c6-ad69-4588-bce8-864262a20ebe/aggregates => generated 53 bytes in 9 msecs (HTTP/1.1 200) 8 headers in 295 bytes (1 switches on core 0) Apr 14 22:51:01 npc99a84aae6d34 devstack@placement-api.service[98700]: DEBUG placement.requestlog [req-55cc7c70-41be-48fb-b96d-5c1e4efca572 req-440fae88-b065-4102-9060-02e884e2a59c None None] Starting request: 2001:41d0:302:1000::cbc "GET /placement/resource_providers/14e023c6-ad69-4588-bce8-864262a20ebe/traits" {{(pid=98700) __call__ /opt/stack/placement/placement/requestlog.py:55}} Apr 14 22:51:01 npc99a84aae6d34 devstack@placement-api.service[98700]: INFO placement.requestlog [req-55cc7c70-41be-48fb-b96d-5c1e4efca572 req-440fae88-b065-4102-9060-02e884e2a59c service nova] 2001:41d0:302:1000::cbc "GET /placement/resource_providers/14e023c6-ad69-4588-bce8-864262a20ebe/traits" status: 200 len: 2026 microversion: 1.6 Apr 14 22:51:01 npc99a84aae6d34 devstack@placement-api.service[98700]: [pid: 98700|app: 0|req: 28/55] 2001:41d0:302:1000::cbc () {68 vars in 1641 bytes} [Tue Apr 14 22:51:01 2026] GET /placement/resource_providers/14e023c6-ad69-4588-bce8-864262a20ebe/traits => generated 2026 bytes in 16 msecs (HTTP/1.1 200) 6 headers in 225 bytes (1 switches on core 0) Apr 14 22:51:01 npc99a84aae6d34 devstack@placement-api.service[98701]: DEBUG placement.requestlog [req-55cc7c70-41be-48fb-b96d-5c1e4efca572 req-6b5484d8-6b79-49ce-a2d7-7e6d13d97ff4 None None] Starting request: 2001:41d0:302:1000::cbc "GET /placement/resource_providers/14e023c6-ad69-4588-bce8-864262a20ebe/allocations" {{(pid=98701) __call__ /opt/stack/placement/placement/requestlog.py:55}} Apr 14 22:51:01 npc99a84aae6d34 devstack@placement-api.service[98701]: INFO placement.requestlog [req-55cc7c70-41be-48fb-b96d-5c1e4efca572 req-6b5484d8-6b79-49ce-a2d7-7e6d13d97ff4 service nova] 2001:41d0:302:1000::cbc "GET /placement/resource_providers/14e023c6-ad69-4588-bce8-864262a20ebe/allocations" status: 200 len: 54 microversion: 1.0 Apr 14 22:51:01 npc99a84aae6d34 devstack@placement-api.service[98701]: [pid: 98701|app: 0|req: 28/56] 2001:41d0:302:1000::cbc () {66 vars in 1613 bytes} [Tue Apr 14 22:51:01 2026] GET /placement/resource_providers/14e023c6-ad69-4588-bce8-864262a20ebe/allocations => generated 54 bytes in 9 msecs (HTTP/1.1 200) 6 headers in 223 bytes (1 switches on core 0) Apr 14 22:51:11 npc99a84aae6d34 devstack@placement-api.service[98700]: DEBUG dbcounter [-] [98700] Writing DB stats placement:SELECT=8 {{(pid=98700) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} Apr 14 22:51:11 npc99a84aae6d34 devstack@placement-api.service[98701]: DEBUG dbcounter [-] [98701] Writing DB stats placement:SELECT=6 {{(pid=98701) stat_writer /opt/stack/data/venv/lib/python3.13/site-packages/dbcounter.py:115}} ● devstack@q-ovn-agent.service - Devstack devstack@q-ovn-agent.service Loaded: loaded (/etc/systemd/system/devstack@q-ovn-agent.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:33:41 UTC; 18min ago Invocation: 37e777bc145f4422a991804782b69cf9 Main PID: 95086 (neutron-ovn-age) Tasks: 20 (limit: 9256) Memory: 352.7M (peak: 355.7M, swap: 5.9M, swap peak: 5.9M, zswap: 1.8M) CPU: 8.831s CGroup: /system.slice/system-devstack.slice/devstack@q-ovn-agent.service ├─95086 "neutron-ovn-agent: master process [/opt/stack/data/venv/bin/neutron-ovn-agent --config-file /etc/neutron/plugins/ml2/ovn_agent.ini]" ├─95991 "neutron-ovn-agent: ServiceWrapper worker(0)" ├─96262 /opt/stack/data/venv/bin/python3.13 /usr/local/bin/privsep-helper --config-file /etc/neutron/plugins/ml2/ovn_agent.ini --privsep_context neutron.privileged.namespace_cmd --privsep_sock_path /tmp/tmp38ne4qkg/privsep.sock └─99558 /opt/stack/data/venv/bin/python3.13 /usr/local/bin/privsep-helper --config-file /etc/neutron/plugins/ml2/ovn_agent.ini --privsep_context neutron.privileged.default --privsep_sock_path /tmp/tmpqd4w1pg4/privsep.sock Apr 14 22:50:44 npc99a84aae6d34 neutron-ovn-agent[95991]: DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s {{(pid=95991) inner /opt/stack/data/venv/lib/python3.13/site-packages/oslo_concurrency/lockutils.py:538}} Apr 14 22:50:44 npc99a84aae6d34 neutron-ovn-agent[95991]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" {{(pid=95991) inner /opt/stack/data/venv/lib/python3.13/site-packages/oslo_concurrency/lockutils.py:506}} Apr 14 22:50:44 npc99a84aae6d34 neutron-ovn-agent[95991]: DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s {{(pid=95991) inner /opt/stack/data/venv/lib/python3.13/site-packages/oslo_concurrency/lockutils.py:519}} Apr 14 22:50:44 npc99a84aae6d34 neutron-ovn-agent[95991]: DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s {{(pid=95991) inner /opt/stack/data/venv/lib/python3.13/site-packages/oslo_concurrency/lockutils.py:538}} Apr 14 22:51:44 npc99a84aae6d34 neutron-ovn-agent[95991]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" {{(pid=95991) inner /opt/stack/data/venv/lib/python3.13/site-packages/oslo_concurrency/lockutils.py:506}} Apr 14 22:51:44 npc99a84aae6d34 neutron-ovn-agent[95991]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" {{(pid=95991) inner /opt/stack/data/venv/lib/python3.13/site-packages/oslo_concurrency/lockutils.py:506}} Apr 14 22:51:44 npc99a84aae6d34 neutron-ovn-agent[95991]: DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s {{(pid=95991) inner /opt/stack/data/venv/lib/python3.13/site-packages/oslo_concurrency/lockutils.py:519}} Apr 14 22:51:44 npc99a84aae6d34 neutron-ovn-agent[95991]: DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s {{(pid=95991) inner /opt/stack/data/venv/lib/python3.13/site-packages/oslo_concurrency/lockutils.py:538}} Apr 14 22:51:44 npc99a84aae6d34 neutron-ovn-agent[95991]: DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.002s {{(pid=95991) inner /opt/stack/data/venv/lib/python3.13/site-packages/oslo_concurrency/lockutils.py:519}} Apr 14 22:51:44 npc99a84aae6d34 neutron-ovn-agent[95991]: DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s {{(pid=95991) inner /opt/stack/data/venv/lib/python3.13/site-packages/oslo_concurrency/lockutils.py:538}} ● devstack@s-account.service - Devstack devstack@s-account.service Loaded: loaded (/etc/systemd/system/devstack@s-account.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:33:10 UTC; 18min ago Invocation: 42a5f79ec242485cb172371a7719080c Main PID: 90424 (swift-account-s) Tasks: 3 (limit: 9256) Memory: 20.3M (peak: 48.2M, swap: 45.1M, swap peak: 45.1M, zswap: 17.7M) CPU: 16.199s CGroup: /system.slice/system-devstack.slice/devstack@s-account.service ├─90424 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-account-server /etc/swift/account-server/1.conf -v ├─90719 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-account-server /etc/swift/account-server/1.conf -v └─90721 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-account-server /etc/swift/account-server/1.conf -v Apr 14 22:33:11 npc99a84aae6d34 swift-account-server[90424]: Eventlet is deprecated. It is currently being maintained in bugfix mode, and Apr 14 22:33:11 npc99a84aae6d34 swift-account-server[90424]: we strongly recommend against using it for new projects. Apr 14 22:33:11 npc99a84aae6d34 swift-account-server[90424]: If you are already using Eventlet, we recommend migrating to a different Apr 14 22:33:11 npc99a84aae6d34 swift-account-server[90424]: framework. For more detail see Apr 14 22:33:11 npc99a84aae6d34 swift-account-server[90424]: https://eventlet.readthedocs.io/en/latest/asyncio/migration.html Apr 14 22:33:11 npc99a84aae6d34 swift-account-server[90424]: from eventlet import Timeout Apr 14 22:33:11 npc99a84aae6d34 account-server[90424]: Started child 90719 from parent 90424 Apr 14 22:33:11 npc99a84aae6d34 swift-account-server[90424]: account-server: Started child 90719 from parent 90424 Apr 14 22:33:11 npc99a84aae6d34 account-server[90424]: Started child 90721 from parent 90424 Apr 14 22:33:11 npc99a84aae6d34 swift-account-server[90424]: account-server: Started child 90721 from parent 90424 ● devstack@s-container-sync.service - Devstack devstack@s-container-sync.service Loaded: loaded (/etc/systemd/system/devstack@s-container-sync.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:33:13 UTC; 18min ago Invocation: bafc20edde3f4237b9b4c8db4dac201e Main PID: 91065 (swift-container) Tasks: 1 (limit: 9256) Memory: 18M (peak: 46.2M, swap: 41.5M, swap peak: 41.5M, zswap: 16.1M) CPU: 826ms CGroup: /system.slice/system-devstack.slice/devstack@s-container-sync.service └─91065 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-container-sync /etc/swift/container-server/1.conf Apr 14 22:33:13 npc99a84aae6d34 systemd[1]: Started devstack@s-container-sync.service - Devstack devstack@s-container-sync.service. Apr 14 22:33:13 npc99a84aae6d34 swift-container-sync[91065]: /opt/stack/swift/swift/container/sync.py:24: EventletDeprecationWarning: Apr 14 22:33:13 npc99a84aae6d34 swift-container-sync[91065]: Eventlet is deprecated. It is currently being maintained in bugfix mode, and Apr 14 22:33:13 npc99a84aae6d34 swift-container-sync[91065]: we strongly recommend against using it for new projects. Apr 14 22:33:13 npc99a84aae6d34 swift-container-sync[91065]: If you are already using Eventlet, we recommend migrating to a different Apr 14 22:33:13 npc99a84aae6d34 swift-container-sync[91065]: framework. For more detail see Apr 14 22:33:13 npc99a84aae6d34 swift-container-sync[91065]: https://eventlet.readthedocs.io/en/latest/asyncio/migration.html Apr 14 22:33:13 npc99a84aae6d34 swift-container-sync[91065]: from eventlet import sleep, Timeout Apr 14 22:33:14 npc99a84aae6d34 container-sync[91065]: Starting 91065 Apr 14 22:33:14 npc99a84aae6d34 container-sync[91065]: Configuration option internal_client_conf_path not defined. Using default configuration, See internal-client.conf-sample for options ● devstack@s-container.service - Devstack devstack@s-container.service Loaded: loaded (/etc/systemd/system/devstack@s-container.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:33:08 UTC; 18min ago Invocation: 9bad394693424550b1b8471238f5630e Main PID: 89898 (swift-container) Tasks: 3 (limit: 9256) Memory: 22M (peak: 52.1M, swap: 48.5M, swap peak: 48.5M, zswap: 19.4M) CPU: 16.378s CGroup: /system.slice/system-devstack.slice/devstack@s-container.service ├─89898 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-container-server /etc/swift/container-server/1.conf -v ├─90188 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-container-server /etc/swift/container-server/1.conf -v └─90190 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-container-server /etc/swift/container-server/1.conf -v Apr 14 22:33:09 npc99a84aae6d34 swift-container-server[89898]: Eventlet is deprecated. It is currently being maintained in bugfix mode, and Apr 14 22:33:09 npc99a84aae6d34 swift-container-server[89898]: we strongly recommend against using it for new projects. Apr 14 22:33:09 npc99a84aae6d34 swift-container-server[89898]: If you are already using Eventlet, we recommend migrating to a different Apr 14 22:33:09 npc99a84aae6d34 swift-container-server[89898]: framework. For more detail see Apr 14 22:33:09 npc99a84aae6d34 swift-container-server[89898]: https://eventlet.readthedocs.io/en/latest/asyncio/migration.html Apr 14 22:33:09 npc99a84aae6d34 swift-container-server[89898]: from eventlet import Timeout Apr 14 22:33:09 npc99a84aae6d34 swift-container-server[89898]: container-server: Started child 90188 from parent 89898 Apr 14 22:33:09 npc99a84aae6d34 container-server[89898]: Started child 90188 from parent 89898 Apr 14 22:33:09 npc99a84aae6d34 swift-container-server[89898]: container-server: Started child 90190 from parent 89898 Apr 14 22:33:09 npc99a84aae6d34 container-server[89898]: Started child 90190 from parent 89898 ● devstack@s-object.service - Devstack devstack@s-object.service Loaded: loaded (/etc/systemd/system/devstack@s-object.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:33:06 UTC; 18min ago Invocation: 69e3edf1bff649b2a77cf3af53b70608 Main PID: 89369 (swift-object-se) Tasks: 3 (limit: 9256) Memory: 23.4M (peak: 56.4M, swap: 50.4M, swap peak: 50.4M, zswap: 20.5M) CPU: 16.747s CGroup: /system.slice/system-devstack.slice/devstack@s-object.service ├─89369 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-object-server /etc/swift/object-server/1.conf -v ├─89690 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-object-server /etc/swift/object-server/1.conf -v └─89691 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-object-server /etc/swift/object-server/1.conf -v Apr 14 22:33:07 npc99a84aae6d34 swift-object-server[89369]: Eventlet is deprecated. It is currently being maintained in bugfix mode, and Apr 14 22:33:07 npc99a84aae6d34 swift-object-server[89369]: we strongly recommend against using it for new projects. Apr 14 22:33:07 npc99a84aae6d34 swift-object-server[89369]: If you are already using Eventlet, we recommend migrating to a different Apr 14 22:33:07 npc99a84aae6d34 swift-object-server[89369]: framework. For more detail see Apr 14 22:33:07 npc99a84aae6d34 swift-object-server[89369]: https://eventlet.readthedocs.io/en/latest/asyncio/migration.html Apr 14 22:33:07 npc99a84aae6d34 swift-object-server[89369]: from eventlet import sleep, wsgi, Timeout, tpool Apr 14 22:33:07 npc99a84aae6d34 object-server[89369]: Started child 89690 from parent 89369 Apr 14 22:33:07 npc99a84aae6d34 swift-object-server[89369]: object-server: Started child 89690 from parent 89369 Apr 14 22:33:07 npc99a84aae6d34 object-server[89369]: Started child 89691 from parent 89369 Apr 14 22:33:07 npc99a84aae6d34 swift-object-server[89369]: object-server: Started child 89691 from parent 89369 ● devstack@s-proxy.service - Devstack devstack@s-proxy.service Loaded: loaded (/etc/systemd/system/devstack@s-proxy.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:33:16 UTC; 18min ago Invocation: d4c94aed10cc480ca6f8415134fbb13b Main PID: 91903 (swift-proxy-ser) Tasks: 2 (limit: 9256) Memory: 35.8M (peak: 74.8M, swap: 69.5M, swap peak: 69.6M, zswap: 23.6M) CPU: 17.128s CGroup: /system.slice/system-devstack.slice/devstack@s-proxy.service ├─91903 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-proxy-server /etc/swift/proxy-server.conf -v └─91982 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-proxy-server /etc/swift/proxy-server.conf -v Apr 14 22:33:19 npc99a84aae6d34 proxy-server[91982]: 2001:41d0:302:1000::cbc 2001:41d0:302:1000::cbc 14/Apr/2026/22/33/19 GET /info HTTP/1.0 200 - curl/8.14.1 - - 1883 - tx71174f50f31a4da980be1-0069dec0af - 0.0062 - - 1776205999.049824715 1776205999.056034803 - - Apr 14 22:33:19 npc99a84aae6d34 swift-proxy-server[91982]: proxy-server: 2001:41d0:302:1000::cbc 2001:41d0:302:1000::cbc 14/Apr/2026/22/33/19 GET /info HTTP/1.0 200 - curl/8.14.1 - - 1883 - tx71174f50f31a4da980be1-0069dec0af - 0.0062 - - 1776205999.049824715 1776205999.056034803 - - Apr 14 22:37:54 npc99a84aae6d34 proxy-server[91982]: Deferring reject downstream Apr 14 22:37:54 npc99a84aae6d34 swift-proxy-server[91982]: proxy-server: Deferring reject downstream Apr 14 22:37:54 npc99a84aae6d34 proxy-server[91982]: Received request from Apr 14 22:37:54 npc99a84aae6d34 swift-proxy-server[91982]: proxy-server: Received request from Apr 14 22:37:54 npc99a84aae6d34 proxy-server[91982]: Authorizing as anonymous (txn: txaf035b9446d94023af250-0069dec1c2) Apr 14 22:37:54 npc99a84aae6d34 swift-proxy-server[91982]: proxy-server: Authorizing as anonymous (txn: txaf035b9446d94023af250-0069dec1c2) Apr 14 22:37:54 npc99a84aae6d34 proxy-server[91982]: 2001:41d0:302:1000::cbc 2001:41d0:302:1000::cbc 14/Apr/2026/22/37/54 GET /info HTTP/1.0 200 - python-urllib3/2.6.3 - - 1883 - txaf035b9446d94023af250-0069dec1c2 - 0.0258 - - 1776206274.893903971 1776206274.919739246 - - Apr 14 22:37:54 npc99a84aae6d34 swift-proxy-server[91982]: proxy-server: 2001:41d0:302:1000::cbc 2001:41d0:302:1000::cbc 14/Apr/2026/22/37/54 GET /info HTTP/1.0 200 - python-urllib3/2.6.3 - - 1883 - txaf035b9446d94023af250-0069dec1c2 - 0.0258 - - 1776206274.893903971 1776206274.919739246 - - ○ dm-event.service - Device-mapper event daemon Loaded: loaded (/usr/lib/systemd/system/dm-event.service; static) Active: inactive (dead) TriggeredBy: ● dm-event.socket Docs: man:dmeventd(8) ○ dpkg-db-backup.service - Daily dpkg database backup service Loaded: loaded (/usr/lib/systemd/system/dpkg-db-backup.service; static) Active: inactive (dead) TriggeredBy: ● dpkg-db-backup.timer Docs: man:dpkg(1) ○ e2scrub_all.service - Online ext4 Metadata Check for All Filesystems Loaded: loaded (/usr/lib/systemd/system/e2scrub_all.service; static) Active: inactive (dead) TriggeredBy: ● e2scrub_all.timer Docs: man:e2scrub_all(8) ○ e2scrub_reap.service - Remove Stale Online ext4 Metadata Check Snapshots Loaded: loaded (/usr/lib/systemd/system/e2scrub_reap.service; enabled; preset: enabled) Active: inactive (dead) since Tue 2026-04-14 22:17:33 UTC; 34min ago Invocation: 2977f7ffc3be47038dead9a070de1328 Docs: man:e2scrub_all(8) Main PID: 14218 (code=exited, status=0/SUCCESS) Mem peak: 1.8M CPU: 49ms Apr 14 22:17:33 npc99a84aae6d34 systemd[1]: Starting e2scrub_reap.service - Remove Stale Online ext4 Metadata Check Snapshots... Apr 14 22:17:33 npc99a84aae6d34 systemd[1]: e2scrub_reap.service: Deactivated successfully. Apr 14 22:17:33 npc99a84aae6d34 systemd[1]: Finished e2scrub_reap.service - Remove Stale Online ext4 Metadata Check Snapshots. ○ emergency.service - Emergency Shell Loaded: loaded (/usr/lib/systemd/system/emergency.service; static) Active: inactive (dead) Docs: man:sulogin(8) ● epmd.service - Erlang Port Mapper Daemon Loaded: loaded (/usr/lib/systemd/system/epmd.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:19:26 UTC; 32min ago Invocation: 54f7be6c71ed4ad18581b83f106e1368 TriggeredBy: ● epmd.socket Main PID: 21889 (epmd) Tasks: 1 (limit: 9256) Memory: 336K (peak: 2M) CPU: 72ms CGroup: /system.slice/epmd.service └─21889 /usr/bin/epmd -systemd Apr 14 22:19:26 npc99a84aae6d34 systemd[1]: Started epmd.service - Erlang Port Mapper Daemon. ○ firewalld.service Loaded: masked (Reason: Unit firewalld.service is masked.) Active: inactive (dead) ○ fstrim.service - Discard unused blocks on filesystems from /etc/fstab Loaded: loaded (/usr/lib/systemd/system/fstrim.service; static) Active: inactive (dead) TriggeredBy: ● fstrim.timer Docs: man:fstrim(8) ○ getty-static.service - getty on tty2-tty6 if dbus and logind are not available Loaded: loaded (/usr/lib/systemd/system/getty-static.service; static) Active: inactive (dead) Condition: start condition unmet at Tue 2026-04-14 22:07:35 UTC; 44min ago Apr 14 22:07:35 debian systemd[1]: getty-static.service - getty on tty2-tty6 if dbus and logind are not available was skipped because of an unmet condition check (ConditionPathExists=!/usr/bin/dbus-daemon). ● getty@tty1.service - Getty on tty1 Loaded: loaded (/usr/lib/systemd/system/getty@.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:07:36 UTC; 44min ago Invocation: f089159a46474870be318ac592e326e1 Docs: man:agetty(8) man:systemd-getty-generator(8) https://0pointer.de/blog/projects/serial-console.html Main PID: 821 (agetty) Tasks: 1 (limit: 9256) Memory: 196K (peak: 2M, swap: 212K, swap peak: 212K, zswap: 41.5K) CPU: 32ms CGroup: /system.slice/system-getty.slice/getty@tty1.service └─821 /sbin/agetty -o "-- \\u" --noreset --noclear - linux Apr 14 22:07:36 npc99a84aae6d34 systemd[1]: Started getty@tty1.service - Getty on tty1. ● glean-early.service - Early glean execution Loaded: loaded (/usr/lib/systemd/system/glean-early.service; enabled; preset: enabled) Active: active (exited) since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: 8ec239806b0449ce8ea6113754d4d1c3 Main PID: 591 (code=exited, status=0/SUCCESS) Mem peak: 18.3M CPU: 201ms Apr 14 22:07:35 debian glean-early.sh[649]: DEBUG:glean:Detected distro : debian Apr 14 22:07:35 debian glean-early.sh[649]: DEBUG:glean:Configuring without NetworkManager Apr 14 22:07:35 npc99a84aae6d34 glean-early.sh[649]: DEBUG:glean:metadata loaded from: /mnt/config/openstack/latest/meta_data.json Apr 14 22:07:35 npc99a84aae6d34 glean-early.sh[649]: DEBUG:glean:Writing output files Apr 14 22:07:35 npc99a84aae6d34 glean-early.sh[649]: DEBUG:glean:Writing output file : /root/.ssh/authorized_keys Apr 14 22:07:35 npc99a84aae6d34 glean-early.sh[649]: DEBUG:glean: ... done Apr 14 22:07:35 npc99a84aae6d34 glean-early.sh[649]: DEBUG:glean:Got hostname from meta_data.json : npc99a84aae6d34 Apr 14 22:07:35 npc99a84aae6d34 glean-early.sh[649]: DEBUG:glean:Found network_info file /mnt/config/openstack/latest/network_data.json Apr 14 22:07:35 npc99a84aae6d34 glean-early.sh[649]: DEBUG:glean:Done! Apr 14 22:07:35 npc99a84aae6d34 systemd[1]: Finished glean-early.service - Early glean execution. ● glean@ens3.service - Glean for interface ens3 Loaded: loaded (/usr/lib/systemd/system/glean@.service; disabled; preset: enabled) Drop-In: /etc/systemd/system/glean@.service.d └─override.conf Active: active (exited) since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: abfc37fef32448f49e8324735f12fd27 Main PID: 726 (code=exited, status=0/SUCCESS) Tasks: 1 (limit: 9256) Memory: 1M (peak: 8.2M, swap: 28K, swap peak: 28K, zswap: 5.4K) CPU: 368ms CGroup: /system.slice/system-glean.slice/glean@ens3.service └─743 dhclient -4 -v -i -pf /run/dhclient.ens3.pid -lf /var/lib/dhcp/dhclient.ens3.leases -I -df /var/lib/dhcp/dhclient6.ens3.leases ens3 Apr 14 22:07:35 npc99a84aae6d34 dhclient[743]: DHCPREQUEST for 217.182.143.61 on ens3 to 255.255.255.255 port 67 Apr 14 22:07:35 npc99a84aae6d34 dhclient[743]: DHCPACK of 217.182.143.61 from 217.182.140.1 Apr 14 22:07:35 npc99a84aae6d34 ifup[743]: DHCPACK of 217.182.143.61 from 217.182.140.1 Apr 14 22:07:35 npc99a84aae6d34 ifup[743]: suspect value in domain_search option - discarded Apr 14 22:07:35 npc99a84aae6d34 dhclient[743]: suspect value in domain_search option - discarded Apr 14 22:07:35 npc99a84aae6d34 dhclient[743]: Error printing text. Apr 14 22:07:35 npc99a84aae6d34 ifup[743]: Error printing text. Apr 14 22:07:35 npc99a84aae6d34 dhclient[743]: bound to 217.182.143.61 -- renewal in 40097 seconds. Apr 14 22:07:35 npc99a84aae6d34 ifup[743]: bound to 217.182.143.61 -- renewal in 40097 seconds. Apr 14 22:07:35 npc99a84aae6d34 systemd[1]: Finished glean@ens3.service - Glean for interface ens3. ● glean@lo.service - Glean for interface lo Loaded: loaded (/usr/lib/systemd/system/glean@.service; disabled; preset: enabled) Drop-In: /etc/systemd/system/glean@.service.d └─override.conf Active: active (exited) since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: 7d3d4bb3124b4795a05370af83df34d7 Main PID: 722 (code=exited, status=0/SUCCESS) Mem peak: 8.4M CPU: 173ms Apr 14 22:07:35 npc99a84aae6d34 systemd[1]: Starting glean@lo.service - Glean for interface lo... Apr 14 22:07:35 npc99a84aae6d34 systemd[1]: Finished glean@lo.service - Glean for interface lo. ● growroot.service - Grow root partition Loaded: loaded (/usr/lib/systemd/system/growroot.service; enabled; preset: enabled) Active: active (exited) since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: aec5cf78f27a4e7a9450e1dd0a602b4d Main PID: 592 (code=exited, status=0/SUCCESS) Mem peak: 11.8M CPU: 243ms Apr 14 22:07:35 debian growroot[592]: + set +e Apr 14 22:07:35 debian growroot[592]: + growpart /dev/vda 1 Apr 14 22:07:35 npc99a84aae6d34 growroot[623]: CHANGED: partition=1 start=2048 old: size=42276864 end=42278911 new: size=167770079 end=167772126 Apr 14 22:07:35 npc99a84aae6d34 growroot[592]: + '[' 0 -le 1 ']' Apr 14 22:07:35 npc99a84aae6d34 growroot[592]: + resize2fs /dev/vda1 Apr 14 22:07:35 npc99a84aae6d34 growroot[725]: resize2fs 1.47.2 (1-Jan-2025) Apr 14 22:07:35 npc99a84aae6d34 growroot[725]: Filesystem at /dev/vda1 is mounted on /; on-line resizing required Apr 14 22:07:35 npc99a84aae6d34 growroot[725]: old_desc_blocks = 3, new_desc_blocks = 10 Apr 14 22:07:35 npc99a84aae6d34 growroot[725]: The filesystem on /dev/vda1 is now 20971259 (4k) blocks long. Apr 14 22:07:35 npc99a84aae6d34 systemd[1]: Finished growroot.service - Grow root partition. ○ grub-common.service - Record successful boot for GRUB Loaded: loaded (/usr/lib/systemd/system/grub-common.service; enabled; preset: enabled) Active: inactive (dead) since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: e0efd7a36c0542b8a1ff03963b56ba1c Main PID: 612 (code=exited, status=0/SUCCESS) Mem peak: 2.5M CPU: 73ms Apr 14 22:07:35 debian systemd[1]: Starting grub-common.service - Record successful boot for GRUB... Apr 14 22:07:35 debian systemd[1]: grub-common.service: Deactivated successfully. Apr 14 22:07:35 debian systemd[1]: Finished grub-common.service - Record successful boot for GRUB. ● haproxy.service - HAProxy Load Balancer Loaded: loaded (/usr/lib/systemd/system/haproxy.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:17:46 UTC; 34min ago Invocation: 5c55fd947c6a4e258d6988cded071932 Docs: man:haproxy(1) file:/usr/share/doc/haproxy/configuration.txt.gz Main PID: 15053 (haproxy) Status: "Ready." Tasks: 9 (limit: 9256) Memory: 4.1M (peak: 95M, swap: 42.8M, swap peak: 43.2M, zswap: 2.7M) CPU: 267ms CGroup: /system.slice/haproxy.service ├─15053 /usr/sbin/haproxy -Ws -f /etc/haproxy/haproxy.cfg -p /run/haproxy.pid -S /run/haproxy-master.sock └─15055 /usr/sbin/haproxy -Ws -f /etc/haproxy/haproxy.cfg -p /run/haproxy.pid -S /run/haproxy-master.sock Apr 14 22:17:46 npc99a84aae6d34 systemd[1]: Starting haproxy.service - HAProxy Load Balancer... Apr 14 22:17:46 npc99a84aae6d34 haproxy[15053]: [NOTICE] (15053) : New worker (15055) forked Apr 14 22:17:46 npc99a84aae6d34 haproxy[15053]: [NOTICE] (15053) : Loading success. Apr 14 22:17:46 npc99a84aae6d34 systemd[1]: Started haproxy.service - HAProxy Load Balancer. ● haveged.service - Entropy Daemon based on the HAVEGE algorithm Loaded: loaded (/usr/lib/systemd/system/haveged.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: f0b496346afc43df9d3ba3567a1b686e Docs: man:haveged(8) http://www.issihosts.com/haveged/ Main PID: 432 (haveged) Tasks: 1 (limit: 9256) Memory: 4.3M (peak: 7.3M, swap: 264K, swap peak: 524K, zswap: 161B) CPU: 418ms CGroup: /system.slice/haveged.service └─432 /usr/sbin/haveged --Foreground --verbose=1 Apr 14 22:07:34 debian systemd[1]: Started haveged.service - Entropy Daemon based on the HAVEGE algorithm. Apr 14 22:07:34 debian (haveged)[432]: haveged.service: Referenced but unset environment variable evaluates to an empty string: DAEMON_ARGS Apr 14 22:07:34 debian haveged[432]: haveged: command socket is listening at fd 3 Apr 14 22:07:34 debian haveged[432]: haveged starting up Apr 14 22:07:34 debian haveged[432]: haveged: ver: 1.9.19; arch: x86; vend: GenuineIntel; build: (gcc 14.2.0 ITV); collect: 128K Apr 14 22:07:34 debian haveged[432]: haveged: cpu: (L4 VC); data: 32K (L2 L4 V); inst: 32K (L2 L4 V); idx: 24/40; sz: 32172/54097 Apr 14 22:07:34 debian haveged[432]: haveged: tot tests(BA8): A:1/1 B:1/1 continuous tests(B): last entropy estimate 8.00115 Apr 14 22:07:34 debian haveged[432]: haveged: fills: 0, generated: 0 ● ifupdown-pre.service - Helper to synchronize boot up for ifupdown Loaded: loaded (/usr/lib/systemd/system/ifupdown-pre.service; static) Active: active (exited) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: a91625c70da24a3691117bd35a071a52 Main PID: 395 (code=exited, status=0/SUCCESS) Mem peak: 2M CPU: 20ms Apr 14 22:07:33 debian systemd[1]: Starting ifupdown-pre.service - Helper to synchronize boot up for ifupdown... Apr 14 22:07:33 debian systemd[1]: Finished ifupdown-pre.service - Helper to synchronize boot up for ifupdown. ○ initrd-cleanup.service - Cleaning Up and Shutting Down Daemons Loaded: loaded (/usr/lib/systemd/system/initrd-cleanup.service; static) Active: inactive (dead) ○ initrd-parse-etc.service - Mountpoints Configured in the Real Root Loaded: loaded (/usr/lib/systemd/system/initrd-parse-etc.service; static) Active: inactive (dead) ○ initrd-switch-root.service - Switch Root Loaded: loaded (/usr/lib/systemd/system/initrd-switch-root.service; static) Active: inactive (dead) ○ initrd-udevadm-cleanup-db.service - Cleanup udev Database Loaded: loaded (/usr/lib/systemd/system/initrd-udevadm-cleanup-db.service; static) Active: inactive (dead) ● iscsid.service - iSCSI initiator daemon (iscsid) Loaded: loaded (/usr/lib/systemd/system/iscsid.service; disabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:24:36 UTC; 27min ago Invocation: 119f94f6414643eda15b5731b1bf63de TriggeredBy: ● iscsid.socket Docs: man:iscsid(8) Main PID: 43269 (iscsid) Tasks: 2 (limit: 9256) Memory: 4M (peak: 4.6M, swap: 132K, swap peak: 132K, zswap: 26.1K) CPU: 151ms CGroup: /system.slice/iscsid.service ├─43268 /usr/sbin/iscsid └─43269 /usr/sbin/iscsid Apr 14 22:24:36 npc99a84aae6d34 systemd[1]: Starting iscsid.service - iSCSI initiator daemon (iscsid)... Apr 14 22:24:36 npc99a84aae6d34 iscsid[43266]: iSCSI logger with pid=43268 started! Apr 14 22:24:36 npc99a84aae6d34 systemd[1]: Started iscsid.service - iSCSI initiator daemon (iscsid). Apr 14 22:24:37 npc99a84aae6d34 iscsid[43268]: iSCSI daemon with pid=43269 started! ● kmod-static-nodes.service - Create List of Static Device Nodes Loaded: loaded (/usr/lib/systemd/system/kmod-static-nodes.service; static) Active: active (exited) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 517596902ba644c7862a9bc38eb67bca Main PID: 353 (code=exited, status=0/SUCCESS) Mem peak: 1.7M CPU: 47ms Notice: journal has been rotated since unit was started, output may be incomplete. ● ksm.service - Kernel Samepage Merging Loaded: loaded (/usr/lib/systemd/system/ksm.service; enabled; preset: enabled) Active: active (exited) since Tue 2026-04-14 22:15:57 UTC; 35min ago Invocation: af8f8f01487b4969a799112f5475ce41 Main PID: 6264 (code=exited, status=0/SUCCESS) Mem peak: 1.8M CPU: 10ms Apr 14 22:15:57 npc99a84aae6d34 systemd[1]: Starting ksm.service - Kernel Samepage Merging... Apr 14 22:15:57 npc99a84aae6d34 systemd[1]: Finished ksm.service - Kernel Samepage Merging. ● ksmtuned.service - Kernel Samepage Merging (KSM) Tuning Daemon Loaded: loaded (/usr/lib/systemd/system/ksmtuned.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:15:57 UTC; 35min ago Invocation: 902048771d3e4f73892eeabd1b1995b8 Main PID: 6269 (ksmtuned) Tasks: 2 (limit: 9256) Memory: 1.2M (peak: 3.9M, swap: 668K, swap peak: 668K, zswap: 17.6K) CPU: 2.326s CGroup: /system.slice/ksmtuned.service ├─ 6269 /bin/bash /usr/sbin/ksmtuned └─121050 sleep 60 Apr 14 22:15:57 npc99a84aae6d34 systemd[1]: Starting ksmtuned.service - Kernel Samepage Merging (KSM) Tuning Daemon... Apr 14 22:15:57 npc99a84aae6d34 systemd[1]: Started ksmtuned.service - Kernel Samepage Merging (KSM) Tuning Daemon. ● ldconfig.service - Rebuild Dynamic Linker Cache Loaded: loaded (/usr/lib/systemd/system/ldconfig.service; static) Active: active (exited) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: cce292b3d6ba42d286fb4dea66e0fd15 Docs: man:ldconfig(8) Main PID: 433 (code=exited, status=0/SUCCESS) Mem peak: 24.6M CPU: 87ms Apr 14 22:07:34 debian systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Apr 14 22:07:34 debian systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. ● libvirt-guests.service - libvirt guests suspend/resume service Loaded: loaded (/usr/lib/systemd/system/libvirt-guests.service; enabled; preset: enabled) Active: active (exited) since Tue 2026-04-14 22:24:03 UTC; 27min ago Invocation: dd3fcd32c0994879ae9eaba2bf6ee07b Docs: man:libvirt-guests(8) https://libvirt.org/ Main PID: 41834 (code=exited, status=0/SUCCESS) Mem peak: 1.9M CPU: 40ms Apr 14 22:24:03 npc99a84aae6d34 systemd[1]: Starting libvirt-guests.service - libvirt guests suspend/resume service... Apr 14 22:24:03 npc99a84aae6d34 systemd[1]: Finished libvirt-guests.service - libvirt guests suspend/resume service. ● libvirtd.service - libvirt legacy monolithic daemon Loaded: loaded (/usr/lib/systemd/system/libvirtd.service; enabled; preset: enabled) Drop-In: /etc/systemd/system/libvirtd.service.d └─coredump.conf Active: active (running) since Tue 2026-04-14 22:32:57 UTC; 18min ago Invocation: 3f5b7e4a36ba46c39f230456f6af6512 TriggeredBy: ● libvirtd.socket ● libvirtd-ro.socket ● libvirtd-admin.socket Docs: man:libvirtd(8) https://libvirt.org/ Main PID: 87645 (libvirtd) Tasks: 21 (limit: 32768) Memory: 346.8M (peak: 395.8M, swap: 2.8M, swap peak: 2.8M, zswap: 766.8K) CPU: 9.312s CGroup: /system.slice/libvirtd.service └─87645 /usr/sbin/libvirtd --timeout 120 Apr 14 22:32:57 npc99a84aae6d34 systemd[1]: Starting libvirtd.service - libvirt legacy monolithic daemon... Apr 14 22:32:57 npc99a84aae6d34 libvirtd[87645]: 2026-04-14 22:32:57.410+0000: 87645: info : libvirt version: 11.3.0, package: 11.3.0-3+deb13u2 (Debian) Apr 14 22:32:57 npc99a84aae6d34 libvirtd[87645]: 2026-04-14 22:32:57.410+0000: 87645: info : hostname: npc99a84aae6d34 Apr 14 22:32:57 npc99a84aae6d34 libvirtd[87645]: 2026-04-14 22:32:57.410+0000: 87645: debug : virLogParseOutputs:1638 : outputs=1:file:/var/log/libvirt/libvirtd.log Apr 14 22:32:57 npc99a84aae6d34 libvirtd[87645]: 2026-04-14 22:32:57.410+0000: 87645: debug : virLogParseOutput:1485 : output=1:file:/var/log/libvirt/libvirtd.log Apr 14 22:32:57 npc99a84aae6d34 systemd[1]: Started libvirtd.service - libvirt legacy monolithic daemon. ○ logrotate.service - Rotate log files Loaded: loaded (/usr/lib/systemd/system/logrotate.service; static) Active: inactive (dead) TriggeredBy: ● logrotate.timer Docs: man:logrotate(8) man:logrotate.conf(5) ○ lvm2-lvmpolld.service - LVM2 poll daemon Loaded: loaded (/usr/lib/systemd/system/lvm2-lvmpolld.service; static) Active: inactive (dead) TriggeredBy: ● lvm2-lvmpolld.socket Docs: man:lvmpolld(8) ● lvm2-monitor.service - Monitoring of LVM2 mirrors, snapshots etc. using dmeventd or progress polling Loaded: loaded (/usr/lib/systemd/system/lvm2-monitor.service; enabled; preset: enabled) Active: active (exited) since Tue 2026-04-14 22:18:00 UTC; 33min ago Invocation: 63eb85bbcce246fc96ee3db3e7763e73 Docs: man:dmeventd(8) man:lvcreate(8) man:lvchange(8) man:vgchange(8) Main PID: 16090 (code=exited, status=0/SUCCESS) Mem peak: 2.5M CPU: 33ms Apr 14 22:18:00 npc99a84aae6d34 systemd[1]: Starting lvm2-monitor.service - Monitoring of LVM2 mirrors, snapshots etc. using dmeventd or progress polling... Apr 14 22:18:00 npc99a84aae6d34 systemd[1]: Finished lvm2-monitor.service - Monitoring of LVM2 mirrors, snapshots etc. using dmeventd or progress polling. ○ man-db.service - Daily man-db regeneration Loaded: loaded (/usr/lib/systemd/system/man-db.service; static) Active: inactive (dead) TriggeredBy: ● man-db.timer Docs: man:mandb(8) ● mariadb.service - MariaDB 11.8.6 database server Loaded: loaded (/usr/lib/systemd/system/mariadb.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:27:43 UTC; 24min ago Invocation: 1bbc16433e5c41dfa28405b14b4e8129 Docs: man:mariadbd(8) https://mariadb.com/kb/en/library/systemd/ Main PID: 62213 (mariadbd) Status: "Taking your SQL requests now..." Tasks: 138 (limit: 61091) Memory: 193.1M (peak: 234.5M, swap: 21.9M, swap peak: 22M, zswap: 1M) CPU: 37.738s CGroup: /system.slice/mariadb.service └─62213 /usr/sbin/mariadbd Apr 14 22:32:13 npc99a84aae6d34 mariadbd[62213]: 2026-04-14 22:32:13 66 [Warning] Aborted connection 66 to db: 'nova_cell0' user: 'root' host: 'localhost' (Got an error reading communication packets) Apr 14 22:32:13 npc99a84aae6d34 mariadbd[62213]: 2026-04-14 22:32:13 65 [Warning] Aborted connection 65 to db: 'nova_api' user: 'root' host: 'localhost' (Got an error reading communication packets) Apr 14 22:32:17 npc99a84aae6d34 mariadbd[62213]: 2026-04-14 22:32:17 67 [Warning] Aborted connection 67 to db: 'nova_api' user: 'root' host: 'localhost' (Got an error reading communication packets) Apr 14 22:34:44 npc99a84aae6d34 mariadbd[62213]: 2026-04-14 22:34:44 144 [Warning] Aborted connection 144 to db: 'nova_api' user: 'root' host: 'localhost' (Got an error reading communication packets) Apr 14 22:35:35 npc99a84aae6d34 mariadbd[62213]: 2026-04-14 22:35:35 192 [Warning] Aborted connection 192 to db: 'nova_cell1' user: 'root' host: 'localhost' (Got an error reading communication packets) Apr 14 22:35:35 npc99a84aae6d34 mariadbd[62213]: 2026-04-14 22:35:35 191 [Warning] Aborted connection 191 to db: 'nova_api' user: 'root' host: 'localhost' (Got an error reading communication packets) Apr 14 22:35:40 npc99a84aae6d34 mariadbd[62213]: 2026-04-14 22:35:40 195 [Warning] Aborted connection 195 to db: 'nova_cell1' user: 'root' host: 'localhost' (Got an error reading communication packets) Apr 14 22:35:40 npc99a84aae6d34 mariadbd[62213]: 2026-04-14 22:35:40 193 [Warning] Aborted connection 193 to db: 'nova_api' user: 'root' host: 'localhost' (Got an error reading communication packets) Apr 14 22:35:40 npc99a84aae6d34 mariadbd[62213]: 2026-04-14 22:35:40 194 [Warning] Aborted connection 194 to db: 'nova_cell0' user: 'root' host: 'localhost' (Got an error reading communication packets) Apr 14 22:44:38 npc99a84aae6d34 mariadbd[62213]: 2026-04-14 22:44:38 165 [Warning] Aborted connection 165 to db: 'cinder' user: 'root' host: 'localhost' (Got an error reading communication packets) ● memcached.service - memcached daemon Loaded: loaded (/usr/lib/systemd/system/memcached.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:33:04 UTC; 18min ago Invocation: 5d311508ef1149fcac9e184ebbc4f64c Docs: man:memcached(1) Main PID: 88828 (memcached) Tasks: 10 (limit: 9256) Memory: 11M (peak: 11.7M, swap: 876K, swap peak: 876K, zswap: 185.1K) CPU: 870ms CGroup: /system.slice/memcached.service └─88828 /usr/bin/memcached -m 64 -p 11211 -u memcache -l 127.0.0.1 -l ::1 -P /var/run/memcached/memcached.pid Apr 14 22:33:04 npc99a84aae6d34 systemd[1]: Started memcached.service - memcached daemon. ○ modprobe@configfs.service - Load Kernel Module configfs Loaded: loaded (/usr/lib/systemd/system/modprobe@.service; static) Active: inactive (dead) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: e26fe19615a4426888ca97b4aad3098b Docs: man:modprobe(8) Main PID: 354 (code=exited, status=0/SUCCESS) Mem peak: 1.5M CPU: 42ms Notice: journal has been rotated since unit was started, output may be incomplete. ○ modprobe@drm.service - Load Kernel Module drm Loaded: loaded (/usr/lib/systemd/system/modprobe@.service; static) Active: inactive (dead) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 318202e85499407f809d9d35fa953589 Docs: man:modprobe(8) Main PID: 355 (code=exited, status=0/SUCCESS) Mem peak: 1.6M CPU: 160ms Apr 14 22:07:33 debian systemd[1]: modprobe@drm.service: Deactivated successfully. Apr 14 22:07:33 debian systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Notice: journal has been rotated since unit was started, output may be incomplete. ○ modprobe@efi_pstore.service - Load Kernel Module efi_pstore Loaded: loaded (/usr/lib/systemd/system/modprobe@.service; static) Active: inactive (dead) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 2bd626438c6149b5bf5a32cbfd8ade4c Docs: man:modprobe(8) Main PID: 495 (code=exited, status=0/SUCCESS) Mem peak: 1.7M CPU: 12ms Apr 14 22:07:34 debian systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 14 22:07:34 debian systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 14 22:07:34 debian systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. ○ modprobe@fuse.service - Load Kernel Module fuse Loaded: loaded (/usr/lib/systemd/system/modprobe@.service; static) Active: inactive (dead) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 39ccf42e47e64505aa8b5c2da7c6c9b0 Docs: man:modprobe(8) Main PID: 357 (code=exited, status=0/SUCCESS) Mem peak: 1.7M CPU: 23ms Notice: journal has been rotated since unit was started, output may be incomplete. ○ netavark-dhcp-proxy.service - Netavark DHCP proxy service Loaded: loaded (/usr/lib/systemd/system/netavark-dhcp-proxy.service; enabled; preset: enabled) Active: inactive (dead) since Tue 2026-04-14 22:17:55 UTC; 34min ago Duration: 30.008s Invocation: fa4dcd95fd4a4da19c2179b76b19d7fa TriggeredBy: ● netavark-dhcp-proxy.socket Main PID: 13082 (code=exited, status=0/SUCCESS) Mem peak: 2.5M CPU: 21ms Apr 14 22:17:25 npc99a84aae6d34 systemd[1]: Starting netavark-dhcp-proxy.service - Netavark DHCP proxy service... Apr 14 22:17:25 npc99a84aae6d34 systemd[1]: Started netavark-dhcp-proxy.service - Netavark DHCP proxy service. Apr 14 22:17:55 npc99a84aae6d34 netavark[13082]: timeout met: exiting after 30 secs of inactivity Apr 14 22:17:55 npc99a84aae6d34 systemd[1]: netavark-dhcp-proxy.service: Deactivated successfully. ● netavark-firewalld-reload.service - Listen for the firewalld reload event and reapply all netavark firewall rules. Loaded: loaded (/usr/lib/systemd/system/netavark-firewalld-reload.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:17:25 UTC; 34min ago Invocation: 01715d3020b84654abf532e128514e9a Main PID: 13083 (netavark) Tasks: 3 (limit: 9256) Memory: 528K (peak: 1.8M, swap: 940K, swap peak: 940K, zswap: 288K) CPU: 15ms CGroup: /system.slice/netavark-firewalld-reload.service └─13083 /usr/lib/podman/netavark firewalld-reload Apr 14 22:17:25 npc99a84aae6d34 systemd[1]: Started netavark-firewalld-reload.service - Listen for the firewalld reload event and reapply all netavark firewall rules.. ● netfilter-persistent.service - netfilter persistent configuration Loaded: loaded (/usr/lib/systemd/system/netfilter-persistent.service; enabled; preset: enabled) Drop-In: /usr/lib/systemd/system/netfilter-persistent.service.d └─iptables.conf Active: active (exited) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 1f1c4ac67fbe48fd80edfda9ba0e9949 Docs: man:netfilter-persistent(8) Main PID: 419 (code=exited, status=0/SUCCESS) Mem peak: 2.3M CPU: 63ms Apr 14 22:07:33 debian systemd[1]: Starting netfilter-persistent.service - netfilter persistent configuration... Apr 14 22:07:33 debian netfilter-persistent[425]: run-parts: executing /usr/share/netfilter-persistent/plugins.d/15-ip4tables start Apr 14 22:07:34 debian netfilter-persistent[425]: run-parts: executing /usr/share/netfilter-persistent/plugins.d/25-ip6tables start Apr 14 22:07:34 debian systemd[1]: Finished netfilter-persistent.service - netfilter persistent configuration. ● networking.service - Raise network interfaces Loaded: loaded (/usr/lib/systemd/system/networking.service; enabled; preset: enabled) Active: active (exited) since Tue 2026-04-14 22:07:36 UTC; 44min ago Invocation: 05160838ae90403b86b50384dbdb4dce Docs: man:interfaces(5) Main PID: 801 (code=exited, status=0/SUCCESS) Mem peak: 1.8M CPU: 35ms Apr 14 22:07:35 npc99a84aae6d34 systemd[1]: Starting networking.service - Raise network interfaces... Apr 14 22:07:36 npc99a84aae6d34 systemd[1]: Finished networking.service - Raise network interfaces. ○ open-iscsi.service - Login to default iSCSI targets Loaded: loaded (/usr/lib/systemd/system/open-iscsi.service; enabled; preset: enabled) Active: inactive (dead) Condition: start condition unmet at Tue 2026-04-14 22:17:36 UTC; 34min ago Docs: man:iscsiadm(8) man:iscsid(8) Apr 14 22:07:36 npc99a84aae6d34 systemd[1]: open-iscsi.service - Login to default iSCSI targets was skipped because no trigger condition checks were met. Apr 14 22:17:36 npc99a84aae6d34 systemd[1]: open-iscsi.service - Login to default iSCSI targets was skipped because no trigger condition checks were met. ● openvswitch-switch.service - Open vSwitch Loaded: loaded (/usr/lib/systemd/system/openvswitch-switch.service; enabled; preset: enabled) Active: active (exited) since Tue 2026-04-14 22:33:34 UTC; 18min ago Invocation: 803bf6da48c34ff6bed478057afe00f1 Main PID: 93705 (code=exited, status=0/SUCCESS) Mem peak: 1.7M CPU: 13ms Apr 14 22:33:34 npc99a84aae6d34 systemd[1]: Starting openvswitch-switch.service - Open vSwitch... Apr 14 22:33:34 npc99a84aae6d34 systemd[1]: Finished openvswitch-switch.service - Open vSwitch. ● ovn-central.service - Open Virtual Network central components Loaded: loaded (/usr/lib/systemd/system/ovn-central.service; enabled; preset: enabled) Active: active (exited) since Tue 2026-04-14 22:33:36 UTC; 18min ago Invocation: 1e4afe40777f4c52942164b7ac54b174 Main PID: 94060 (code=exited, status=0/SUCCESS) Mem peak: 1.7M CPU: 17ms Apr 14 22:33:36 npc99a84aae6d34 systemd[1]: Starting ovn-central.service - Open Virtual Network central components... Apr 14 22:33:36 npc99a84aae6d34 systemd[1]: Finished ovn-central.service - Open Virtual Network central components. ● ovn-controller-vtep.service - Open Virtual Network VTEP gateway controller daemon Loaded: loaded (/usr/lib/systemd/system/ovn-controller-vtep.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:33:34 UTC; 18min ago Invocation: a417c2f6ae8846db9b82cef9b66d7a49 Main PID: 93768 (ovn-controller-) Tasks: 1 (limit: 9256) Memory: 560K (peak: 3.3M, swap: 908K, swap peak: 908K, zswap: 238.7K) CPU: 82ms CGroup: /system.slice/ovn-controller-vtep.service └─93768 ovn-controller-vtep -vconsole:emer -vsyslog:err -vfile:info --vtep-db=/var/run/openvswitch/db.sock --ovnsb-db=/var/run/ovn/ovnsb_db.sock --no-chdir --log-file=/var/log/ovn/ovn-controller-vtep.log --pidfile=/var/run/ovn/ovn-controller-vtep.pid --detach Apr 14 22:33:34 npc99a84aae6d34 systemd[1]: Starting ovn-controller-vtep.service - Open Virtual Network VTEP gateway controller daemon... Apr 14 22:33:34 npc99a84aae6d34 (ovn-ctl)[93730]: ovn-controller-vtep.service: Referenced but unset environment variable evaluates to an empty string: OVN_CTL_OPTS Apr 14 22:33:34 npc99a84aae6d34 ovn-ctl[93730]: Starting ovn-controller-vtep. Apr 14 22:33:34 npc99a84aae6d34 systemd[1]: Started ovn-controller-vtep.service - Open Virtual Network VTEP gateway controller daemon. ● ovn-controller.service - Open Virtual Network host control daemon Loaded: loaded (/usr/lib/systemd/system/ovn-controller.service; static) Active: active (running) since Tue 2026-04-14 22:33:39 UTC; 18min ago Invocation: 511f55d113ee40eeb2c73a72f87dc8ef Main PID: 94497 (ovn-controller) Tasks: 5 (limit: 9256) Memory: 4.8M (peak: 7.1M, swap: 2.1M, swap peak: 2.1M, zswap: 660.7K) CPU: 829ms CGroup: /system.slice/ovn-controller.service └─94497 ovn-controller unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --no-chdir --log-file=/var/log/ovn/ovn-controller.log --pidfile=/var/run/ovn/ovn-controller.pid --detach Apr 14 22:33:39 npc99a84aae6d34 systemd[1]: Starting ovn-controller.service - Open Virtual Network host control daemon... Apr 14 22:33:39 npc99a84aae6d34 (ovn-ctl)[94468]: ovn-controller.service: Referenced but unset environment variable evaluates to an empty string: OVN_CTL_OPTS Apr 14 22:33:39 npc99a84aae6d34 ovn-ctl[94468]: Starting ovn-controller. Apr 14 22:33:39 npc99a84aae6d34 systemd[1]: Started ovn-controller.service - Open Virtual Network host control daemon. ● ovn-host.service - Open Virtual Network host components Loaded: loaded (/usr/lib/systemd/system/ovn-host.service; enabled; preset: enabled) Active: active (exited) since Tue 2026-04-14 22:23:24 UTC; 28min ago Invocation: b3656512da1d407bb69107e58854a10c Main PID: 39611 (code=exited, status=0/SUCCESS) Mem peak: 1.9M CPU: 16ms Apr 14 22:23:24 npc99a84aae6d34 systemd[1]: Starting ovn-host.service - Open Virtual Network host components... Apr 14 22:23:24 npc99a84aae6d34 systemd[1]: Finished ovn-host.service - Open Virtual Network host components. ● ovn-northd.service - Open Virtual Network central control daemon Loaded: loaded (/usr/lib/systemd/system/ovn-northd.service; static) Active: active (running) since Tue 2026-04-14 22:33:36 UTC; 18min ago Invocation: 3f7a1fa76fb44d039f95535f26b6eb45 Main PID: 94198 (ovn-northd) Tasks: 3 (limit: 9256) Memory: 2.6M (peak: 4.6M, swap: 1.3M, swap peak: 1.3M, zswap: 380.5K) CPU: 255ms CGroup: /system.slice/ovn-northd.service └─94198 ovn-northd -vconsole:emer -vsyslog:err -vfile:info --ovnnb-db=unix:/var/run/ovn/ovnnb_db.sock --ovnsb-db=unix:/var/run/ovn/ovnsb_db.sock --no-chdir --log-file=/var/log/ovn/ovn-northd.log --pidfile=/var/run/ovn/ovn-northd.pid --detach Apr 14 22:33:36 npc99a84aae6d34 systemd[1]: Starting ovn-northd.service - Open Virtual Network central control daemon... Apr 14 22:33:36 npc99a84aae6d34 (ovn-ctl)[94128]: ovn-northd.service: Referenced but unset environment variable evaluates to an empty string: OVN_CTL_OPTS Apr 14 22:33:36 npc99a84aae6d34 ovn-ctl[94128]: Starting ovn-northd. Apr 14 22:33:36 npc99a84aae6d34 systemd[1]: Started ovn-northd.service - Open Virtual Network central control daemon. ● ovn-ovsdb-server-nb.service - Open vSwitch database server for OVN Northbound database Loaded: loaded (/usr/lib/systemd/system/ovn-ovsdb-server-nb.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:33:36 UTC; 18min ago Invocation: ba7b8112225042d6bdb571a4452db2ba Main PID: 94125 (ovsdb-server) Tasks: 1 (limit: 9256) Memory: 2.7M (peak: 4.4M, swap: 1M, swap peak: 1M, zswap: 293.6K) CPU: 649ms CGroup: /system.slice/ovn-ovsdb-server-nb.service └─94125 ovsdb-server -vconsole:off -vfile:info --log-file=/var/log/ovn/ovsdb-server-nb.log --pidfile=/var/run/ovn/ovnnb_db.pid --remote=punix:/var/run/ovn/ovnnb_db.sock --unixctl=/var/run/ovn/ovnnb_db.ctl --remote=db:OVN_Northbound,NB_Global,connections --private-key=db:OVN_Northbound,SSL,private_key --certificate=db:OVN_Northbound,SSL,certificate --ca-cert=db:OVN_Northbound,SSL,ca_cert --ssl-protocols=db:OVN_Northbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Northbound,SSL,ssl_ciphers --ssl-ciphersuites=db:OVN_Northbound,SSL,ssl_ciphersuites /var/lib/ovn/ovnnb_db.db Apr 14 22:51:56 npc99a84aae6d34 ovsdb-server[94125]: ovs|03517|reconnect|DBG|ssl:[2001:41d0:302:1000::cbc]:56252: entering IDLE Apr 14 22:51:56 npc99a84aae6d34 ovsdb-server[94125]: ovs|03518|jsonrpc|DBG|ssl:[2001:41d0:302:1000::cbc]:56252: send request, method="echo", params=[], id="echo" Apr 14 22:51:56 npc99a84aae6d34 ovsdb-server[94125]: ovs|03519|stream_ssl|DBG|server7-->ssl:[2001:41d0:302:1000::cbc]:56252 type 256 (5 bytes) Apr 14 22:51:56 npc99a84aae6d34 ovsdb-server[94125]: ovs|03520|stream_ssl|DBG|server7-->ssl:[2001:41d0:302:1000::cbc]:56252 type 257 (1 bytes) Apr 14 22:51:56 npc99a84aae6d34 ovsdb-server[94125]: ovs|03521|poll_loop|DBG|wakeup due to [POLLIN] on fd 23 ([2001:41d0:302:1000::cbc]:6641<->[2001:41d0:302:1000::cbc]:56252) at ../lib/stream-ssl.c:835 (0% CPU usage) Apr 14 22:51:56 npc99a84aae6d34 ovsdb-server[94125]: ovs|03522|stream_ssl|DBG|server7<--ssl:[2001:41d0:302:1000::cbc]:56252 type 256 (5 bytes) Apr 14 22:51:56 npc99a84aae6d34 ovsdb-server[94125]: ovs|03523|stream_ssl|DBG|server7<--ssl:[2001:41d0:302:1000::cbc]:56252 type 257 (1 bytes) Apr 14 22:51:56 npc99a84aae6d34 ovsdb-server[94125]: ovs|03524|jsonrpc|DBG|ssl:[2001:41d0:302:1000::cbc]:56252: received reply, result=[], id="echo" Apr 14 22:51:56 npc99a84aae6d34 ovsdb-server[94125]: ovs|03525|reconnect|DBG|ssl:[2001:41d0:302:1000::cbc]:56252: entering ACTIVE Apr 14 22:51:56 npc99a84aae6d34 ovsdb-server[94125]: ovs|03526|poll_loop|DBG|wakeup due to 0-ms timeout at ../lib/stream-ssl.c:837 (0% CPU usage) ● ovn-ovsdb-server-sb.service - Open vSwitch database server for OVN Southbound database Loaded: loaded (/usr/lib/systemd/system/ovn-ovsdb-server-sb.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:33:36 UTC; 18min ago Invocation: 064c034707fd4c2dacfa3d5f94d97b93 Main PID: 94127 (ovsdb-server) Tasks: 1 (limit: 9256) Memory: 3.8M (peak: 5.7M, swap: 1.1M, swap peak: 1.1M, zswap: 321.8K) CPU: 919ms CGroup: /system.slice/ovn-ovsdb-server-sb.service └─94127 ovsdb-server -vconsole:off -vfile:info --log-file=/var/log/ovn/ovsdb-server-sb.log --pidfile=/var/run/ovn/ovnsb_db.pid --remote=punix:/var/run/ovn/ovnsb_db.sock --unixctl=/var/run/ovn/ovnsb_db.ctl --remote=db:OVN_Southbound,SB_Global,connections --private-key=db:OVN_Southbound,SSL,private_key --certificate=db:OVN_Southbound,SSL,certificate --ca-cert=db:OVN_Southbound,SSL,ca_cert --ssl-protocols=db:OVN_Southbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Southbound,SSL,ssl_ciphers --ssl-ciphersuites=db:OVN_Southbound,SSL,ssl_ciphersuites /var/lib/ovn/ovnsb_db.db Apr 14 22:51:56 npc99a84aae6d34 ovsdb-server[94127]: ovs|05052|reconnect|DBG|ssl:[2001:41d0:302:1000::cbc]:37994: entering IDLE Apr 14 22:51:56 npc99a84aae6d34 ovsdb-server[94127]: ovs|05053|jsonrpc|DBG|ssl:[2001:41d0:302:1000::cbc]:37994: send request, method="echo", params=[], id="echo" Apr 14 22:51:56 npc99a84aae6d34 ovsdb-server[94127]: ovs|05054|stream_ssl|DBG|server10-->ssl:[2001:41d0:302:1000::cbc]:37994 type 256 (5 bytes) Apr 14 22:51:56 npc99a84aae6d34 ovsdb-server[94127]: ovs|05055|stream_ssl|DBG|server10-->ssl:[2001:41d0:302:1000::cbc]:37994 type 257 (1 bytes) Apr 14 22:51:56 npc99a84aae6d34 ovsdb-server[94127]: ovs|05056|poll_loop|DBG|wakeup due to [POLLIN] on fd 26 ([2001:41d0:302:1000::cbc]:6642<->[2001:41d0:302:1000::cbc]:37994) at ../lib/stream-ssl.c:835 (0% CPU usage) Apr 14 22:51:56 npc99a84aae6d34 ovsdb-server[94127]: ovs|05057|stream_ssl|DBG|server10<--ssl:[2001:41d0:302:1000::cbc]:37994 type 256 (5 bytes) Apr 14 22:51:56 npc99a84aae6d34 ovsdb-server[94127]: ovs|05058|stream_ssl|DBG|server10<--ssl:[2001:41d0:302:1000::cbc]:37994 type 257 (1 bytes) Apr 14 22:51:56 npc99a84aae6d34 ovsdb-server[94127]: ovs|05059|jsonrpc|DBG|ssl:[2001:41d0:302:1000::cbc]:37994: received reply, result=[], id="echo" Apr 14 22:51:56 npc99a84aae6d34 ovsdb-server[94127]: ovs|05060|reconnect|DBG|ssl:[2001:41d0:302:1000::cbc]:37994: entering ACTIVE Apr 14 22:51:56 npc99a84aae6d34 ovsdb-server[94127]: ovs|05061|poll_loop|DBG|wakeup due to 0-ms timeout at ../lib/stream-ssl.c:837 (0% CPU usage) ● ovs-record-hostname.service - Open vSwitch Record Hostname Loaded: loaded (/usr/lib/systemd/system/ovs-record-hostname.service; enabled; preset: enabled) Active: active (exited) since Tue 2026-04-14 22:33:34 UTC; 18min ago Invocation: 4a28d463ff264da0a1fac48900889099 Main PID: 93706 (code=exited, status=0/SUCCESS) Mem peak: 2.1M CPU: 39ms Apr 14 22:33:34 npc99a84aae6d34 systemd[1]: Starting ovs-record-hostname.service - Open vSwitch Record Hostname... Apr 14 22:33:34 npc99a84aae6d34 ovs-vsctl[93745]: ovs|00001|vsctl|INFO|Called as ovs-vsctl --no-wait add Open_vSwitch . external-ids hostname=npc99a84aae6d34 Apr 14 22:33:34 npc99a84aae6d34 systemd[1]: Finished ovs-record-hostname.service - Open vSwitch Record Hostname. ● ovs-vswitchd.service - Open vSwitch Forwarding Unit Loaded: loaded (/usr/lib/systemd/system/ovs-vswitchd.service; static) Active: active (running) since Tue 2026-04-14 22:33:34 UTC; 18min ago Invocation: c694a890184b4b79ae756227e17efb26 Main PID: 93684 (ovs-vswitchd) Tasks: 13 (limit: 9256) Memory: 8.2M (peak: 8.8M) CPU: 3.511s CGroup: /system.slice/ovs-vswitchd.service └─93684 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach Apr 14 22:33:34 npc99a84aae6d34 systemd[1]: Starting ovs-vswitchd.service - Open vSwitch Forwarding Unit... Apr 14 22:33:34 npc99a84aae6d34 (ovs-ctl)[93646]: ovs-vswitchd.service: Referenced but unset environment variable evaluates to an empty string: OVS_CTL_OPTS Apr 14 22:33:34 npc99a84aae6d34 ovs-ctl[93646]: Starting ovs-vswitchd. Apr 14 22:33:34 npc99a84aae6d34 ovs-ctl[93646]: Enabling remote OVSDB managers. Apr 14 22:33:34 npc99a84aae6d34 systemd[1]: Started ovs-vswitchd.service - Open vSwitch Forwarding Unit. ● ovsdb-server.service - Open vSwitch Database Unit Loaded: loaded (/usr/lib/systemd/system/ovsdb-server.service; static) Active: active (running) since Tue 2026-04-14 22:33:33 UTC; 18min ago Invocation: 0957c00ca3be44c4bdace424f7367de3 Main PID: 93449 (ovsdb-server) Tasks: 1 (limit: 9256) Memory: 1.5M (peak: 6.5M, swap: 1.8M, swap peak: 1.8M, zswap: 505.4K) CPU: 514ms CGroup: /system.slice/ovsdb-server.service └─93449 ovsdb-server /etc/openvswitch/conf.db -vconsole:emer -vsyslog:err -vfile:info --remote=punix:/var/run/openvswitch/db.sock --private-key=db:Open_vSwitch,SSL,private_key --certificate=db:Open_vSwitch,SSL,certificate --bootstrap-ca-cert=db:Open_vSwitch,SSL,ca_cert --no-chdir --log-file=/var/log/openvswitch/ovsdb-server.log --pidfile=/var/run/openvswitch/ovsdb-server.pid --detach Apr 14 22:33:33 npc99a84aae6d34 systemd[1]: Starting ovsdb-server.service - Open vSwitch Database Unit... Apr 14 22:33:33 npc99a84aae6d34 (ovs-ctl)[93415]: ovsdb-server.service: Referenced but unset environment variable evaluates to an empty string: OVS_CTL_OPTS Apr 14 22:33:33 npc99a84aae6d34 ovs-ctl[93415]: Starting ovsdb-server. Apr 14 22:33:33 npc99a84aae6d34 ovs-vsctl[93450]: ovs|00001|vsctl|INFO|Called as ovs-vsctl --no-wait -- init -- set Open_vSwitch . db-version=8.8.0 Apr 14 22:33:33 npc99a84aae6d34 ovs-vsctl[93455]: ovs|00001|vsctl|INFO|Called as ovs-vsctl --no-wait set Open_vSwitch . ovs-version=3.5.0 "external-ids:system-id=\"2e18b751-66f6-4e90-a263-2d9ce97bd2ce\"" "external-ids:rundir=\"/var/run/openvswitch\"" "system-type=\"debian\"" "system-version=\"13\"" Apr 14 22:33:33 npc99a84aae6d34 ovs-ctl[93415]: Configuring Open vSwitch system IDs. Apr 14 22:33:33 npc99a84aae6d34 ovs-ctl[93415]: Enabling remote OVSDB managers. Apr 14 22:33:33 npc99a84aae6d34 systemd[1]: Started ovsdb-server.service - Open vSwitch Database Unit. ● polkit.service - Authorization Manager Loaded: loaded (/usr/lib/systemd/system/polkit.service; static) Active: active (running) since Tue 2026-04-14 22:34:47 UTC; 17min ago Invocation: 4ccada4f436d46fa88e77448218f4c64 Docs: man:polkit(8) Main PID: 105627 (polkitd) Status: "Processing requests..." Tasks: 4 (limit: 9256) Memory: 2.1M (peak: 3.1M, swap: 704K, swap peak: 716K, zswap: 154.4K) CPU: 161ms CGroup: /system.slice/polkit.service └─105627 /usr/lib/polkit-1/polkitd --no-debug --log-level=notice Apr 14 22:34:46 npc99a84aae6d34 polkitd[105627]: Started polkitd version 126 Apr 14 22:34:46 npc99a84aae6d34 polkitd[105627]: Loading rules from directory /etc/polkit-1/rules.d Apr 14 22:34:46 npc99a84aae6d34 polkitd[105627]: Loading rules from directory /run/polkit-1/rules.d Apr 14 22:34:46 npc99a84aae6d34 polkitd[105627]: Error opening rules directory: Error opening directory “/run/polkit-1/rules.d”: No such file or directory (g-file-error-quark, 4) Apr 14 22:34:46 npc99a84aae6d34 polkitd[105627]: Loading rules from directory /usr/local/share/polkit-1/rules.d Apr 14 22:34:46 npc99a84aae6d34 polkitd[105627]: Error opening rules directory: Error opening directory “/usr/local/share/polkit-1/rules.d”: No such file or directory (g-file-error-quark, 4) Apr 14 22:34:46 npc99a84aae6d34 polkitd[105627]: Loading rules from directory /usr/share/polkit-1/rules.d Apr 14 22:34:46 npc99a84aae6d34 polkitd[105627]: Finished loading, compiling and executing 3 rules Apr 14 22:34:46 npc99a84aae6d34 polkitd[105627]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Apr 14 22:34:47 npc99a84aae6d34 systemd[1]: Started polkit.service - Authorization Manager. ● postgresql.service - PostgreSQL RDBMS Loaded: loaded (/usr/lib/systemd/system/postgresql.service; enabled; preset: enabled) Active: active (exited) since Tue 2026-04-14 22:17:40 UTC; 34min ago Invocation: d6c600b249cd4c73b59227d6f00d88dd Main PID: 14788 (code=exited, status=0/SUCCESS) Mem peak: 1.7M CPU: 15ms Apr 14 22:17:40 npc99a84aae6d34 systemd[1]: Starting postgresql.service - PostgreSQL RDBMS... Apr 14 22:17:40 npc99a84aae6d34 systemd[1]: Finished postgresql.service - PostgreSQL RDBMS. ● rabbitmq-server.service - RabbitMQ Messaging Server Loaded: loaded (/usr/lib/systemd/system/rabbitmq-server.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:19:32 UTC; 32min ago Invocation: bd077e698be8487bb76228b23d90d8ed Main PID: 22008 (beam.smp) Tasks: 39 (limit: 9256) Memory: 101.5M (peak: 111.3M, swap: 9.6M, swap peak: 9.6M, zswap: 4.2M) CPU: 11.863s CGroup: /system.slice/rabbitmq-server.service ├─22008 /usr/lib/erlang/erts-15.2.7/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -- -root /usr/lib/erlang -bindir /usr/lib/erlang/erts-15.2.7/bin -progname erl -- -home /var/lib/rabbitmq -- -pa "" -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger "[]" -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -- ├─22018 erl_child_setup 65536 ├─22049 sh -s disksup ├─22051 /usr/lib/erlang/lib/os_mon-2.10.1/priv/bin/memsup ├─22052 /usr/lib/erlang/lib/os_mon-2.10.1/priv/bin/cpu_sup ├─22054 /usr/lib/erlang/erts-15.2.7/bin/inet_gethost 4 ├─22055 /usr/lib/erlang/erts-15.2.7/bin/inet_gethost 4 └─22058 /bin/sh -s rabbit_disk_monitor Apr 14 22:19:28 npc99a84aae6d34 systemd[1]: Starting rabbitmq-server.service - RabbitMQ Messaging Server... Apr 14 22:19:32 npc99a84aae6d34 systemd[1]: Started rabbitmq-server.service - RabbitMQ Messaging Server. ● rbdmap.service - Map RBD devices Loaded: loaded (/usr/lib/systemd/system/rbdmap.service; enabled; preset: enabled) Active: active (exited) since Tue 2026-04-14 22:17:56 UTC; 34min ago Invocation: 50f77fca807d44da8be6267fd7511bc6 Main PID: 15808 (code=exited, status=0/SUCCESS) Mem peak: 1.9M CPU: 16ms Apr 14 22:17:56 npc99a84aae6d34 systemd[1]: Starting rbdmap.service - Map RBD devices... Apr 14 22:17:56 npc99a84aae6d34 systemd[1]: Finished rbdmap.service - Map RBD devices. ● rc-local.service - /etc/rc.local Compatibility Loaded: loaded (/usr/lib/systemd/system/rc-local.service; enabled-runtime; preset: enabled) Drop-In: /usr/lib/systemd/system/rc-local.service.d └─debian.conf Active: active (exited) since Tue 2026-04-14 22:07:36 UTC; 44min ago Invocation: daa5d20e5d8c4037a741899410c11486 Docs: man:systemd-rc-local-generator(8) Mem peak: 1.7M CPU: 31ms Apr 14 22:07:36 npc99a84aae6d34 systemd[1]: Starting rc-local.service - /etc/rc.local Compatibility... Apr 14 22:07:36 npc99a84aae6d34 rc.local[806]: + set +e Apr 14 22:07:36 npc99a84aae6d34 rc.local[806]: + sed -i -e 's/^\(DNS[0-9]*=[.0-9]\+\)/#\1/g' '/etc/sysconfig/network-scripts/ifcfg-*' Apr 14 22:07:36 npc99a84aae6d34 rc.local[810]: sed: can't read /etc/sysconfig/network-scripts/ifcfg-*: No such file or directory Apr 14 22:07:36 npc99a84aae6d34 rc.local[806]: + sed -i -e 's/^NETCONFIG_DNS_POLICY=.*/NETCONFIG_DNS_POLICY=""/g' /etc/sysconfig/network/config Apr 14 22:07:36 npc99a84aae6d34 rc.local[812]: sed: can't read /etc/sysconfig/network/config: No such file or directory Apr 14 22:07:36 npc99a84aae6d34 rc.local[806]: + set -e Apr 14 22:07:36 npc99a84aae6d34 rc.local[806]: + echo 'nameserver 127.0.0.1' Apr 14 22:07:36 npc99a84aae6d34 rc.local[806]: + exit 0 Apr 14 22:07:36 npc99a84aae6d34 systemd[1]: Started rc-local.service - /etc/rc.local Compatibility. ○ rescue.service - Rescue Shell Loaded: loaded (/usr/lib/systemd/system/rescue.service; static) Active: inactive (dead) Docs: man:sulogin(8) ● rsync.service - fast remote file copy program daemon Loaded: loaded (/usr/lib/systemd/system/rsync.service; disabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:33:04 UTC; 18min ago Invocation: f80acd0a843247e9b281ecdb2c2ba55e Docs: man:rsync(1) man:rsyncd.conf(5) Main PID: 88855 (rsync) Tasks: 1 (limit: 9256) Memory: 380K (peak: 1.9M, swap: 760K, swap peak: 760K, zswap: 188.8K) CPU: 23ms CGroup: /system.slice/rsync.service └─88855 /usr/bin/rsync --daemon --no-detach Apr 14 22:33:04 npc99a84aae6d34 systemd[1]: Started rsync.service - fast remote file copy program daemon. ● rsyslog.service - System Logging Service Loaded: loaded (/usr/lib/systemd/system/rsyslog.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:18:16 UTC; 33min ago Invocation: c1426a00b2c54503bfcc17c861e511ff TriggeredBy: ● syslog.socket Docs: man:rsyslogd(8) man:rsyslog.conf(5) https://www.rsyslog.com/doc/ Main PID: 19101 (rsyslogd) Tasks: 4 (limit: 9256) Memory: 14.2M (peak: 19.3M, swap: 200K, swap peak: 200K, zswap: 48.1K) CPU: 7.606s CGroup: /system.slice/rsyslog.service └─19101 /usr/sbin/rsyslogd -n -iNONE Apr 14 22:18:16 npc99a84aae6d34 systemd[1]: Starting rsyslog.service - System Logging Service... Apr 14 22:18:16 npc99a84aae6d34 rsyslogd[19101]: imuxsock: Acquired UNIX socket '/run/systemd/journal/syslog' (fd 3) from systemd. [v8.2504.0] Apr 14 22:18:16 npc99a84aae6d34 rsyslogd[19101]: [origin software="rsyslogd" swVersion="8.2504.0" x-pid="19101" x-info="https://www.rsyslog.com"] start Apr 14 22:18:16 npc99a84aae6d34 systemd[1]: Started rsyslog.service - System Logging Service. ● rtslib-fb-targetctl.service - Restore LIO kernel target configuration Loaded: loaded (/usr/lib/systemd/system/rtslib-fb-targetctl.service; enabled; preset: enabled) Active: active (exited) since Tue 2026-04-14 22:22:40 UTC; 29min ago Invocation: 571c8666e2d04265915510a4d24ecc01 Main PID: 35320 (code=exited, status=0/SUCCESS) Mem peak: 8.8M CPU: 194ms Apr 14 22:22:39 npc99a84aae6d34 systemd[1]: Starting rtslib-fb-targetctl.service - Restore LIO kernel target configuration... Apr 14 22:22:40 npc99a84aae6d34 target[35320]: No saved config file at /etc/rtslib-fb-target/saveconfig.json, ok, exiting Apr 14 22:22:40 npc99a84aae6d34 systemd[1]: Finished rtslib-fb-targetctl.service - Restore LIO kernel target configuration. ● serial-getty@ttyS0.service - Serial Getty on ttyS0 Loaded: loaded (/usr/lib/systemd/system/serial-getty@.service; enabled-runtime; preset: enabled) Active: active (running) since Tue 2026-04-14 22:07:36 UTC; 44min ago Invocation: 7a5957561c17423d9ddc15901e9f3666 Docs: man:agetty(8) man:systemd-getty-generator(8) https://0pointer.de/blog/projects/serial-console.html Main PID: 823 (agetty) Tasks: 1 (limit: 9256) Memory: 272K (peak: 2M, swap: 192K, swap peak: 192K, zswap: 41K) CPU: 26ms CGroup: /system.slice/system-serial\x2dgetty.slice/serial-getty@ttyS0.service └─823 /sbin/agetty -o "-- \\u" --noreset --noclear --keep-baud 115200,57600,38400,9600 - vt220 Apr 14 22:07:36 npc99a84aae6d34 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. × ssh-keygen.service - OpenSSH Server Key Generation Loaded: loaded (/usr/lib/systemd/system/ssh-keygen.service; enabled; preset: enabled) Active: failed (Result: exit-code) since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: b979c20c25ff4bd58f9952b6a7252c52 Main PID: 595 (code=exited, status=255/EXCEPTION) Mem peak: 2.3M CPU: 48ms Apr 14 22:07:35 debian runtime-ssh-host-keys.sh[595]: + for key in dsa ecdsa ed25519 rsa Apr 14 22:07:35 debian runtime-ssh-host-keys.sh[595]: + FILE=/etc/ssh/ssh_host_dsa_key Apr 14 22:07:35 debian runtime-ssh-host-keys.sh[595]: + '[' -f /etc/ssh/ssh_host_dsa_key ']' Apr 14 22:07:35 debian runtime-ssh-host-keys.sh[610]: + /usr/bin/yes n Apr 14 22:07:35 debian runtime-ssh-host-keys.sh[611]: + /usr/bin/ssh-keygen -f /etc/ssh/ssh_host_dsa_key -N '' -t dsa Apr 14 22:07:35 debian runtime-ssh-host-keys.sh[611]: unknown key type dsa Apr 14 22:07:35 debian runtime-ssh-host-keys.sh[610]: yes: standard output: Broken pipe Apr 14 22:07:35 debian systemd[1]: ssh-keygen.service: Main process exited, code=exited, status=255/EXCEPTION Apr 14 22:07:35 debian systemd[1]: ssh-keygen.service: Failed with result 'exit-code'. Apr 14 22:07:35 debian systemd[1]: Failed to start ssh-keygen.service - OpenSSH Server Key Generation. ● ssh.service - OpenBSD Secure Shell server Loaded: loaded (/usr/lib/systemd/system/ssh.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:17:28 UTC; 34min ago Invocation: ecf39ce545544affadc8f0257b498830 Docs: man:sshd(8) man:sshd_config(5) Main PID: 13390 (sshd) Tasks: 1 (limit: 9256) Memory: 5.9M (peak: 24.6M, swap: 812K, swap peak: 812K, zswap: 188.3K) CPU: 3.211s CGroup: /system.slice/ssh.service └─13390 "sshd: /usr/sbin/sshd -D [listener] 0 of 30-100 startups" Apr 14 22:49:11 npc99a84aae6d34 sshd-session[119457]: Connection closed by invalid user sol 80.94.92.183 port 58050 [preauth] Apr 14 22:49:40 npc99a84aae6d34 sshd-session[119476]: Invalid user solv from 92.118.39.62 port 45882 Apr 14 22:49:40 npc99a84aae6d34 sshd-session[119476]: Connection closed by invalid user solv 92.118.39.62 port 45882 [preauth] Apr 14 22:50:12 npc99a84aae6d34 sshd-session[119590]: Accepted publickey for zuul from 104.130.127.165 port 41408 ssh2: RSA SHA256:c94tZpPtmWI4tk86GWYkvl4qtYyEmDbB3eg5vn0XjIc Apr 14 22:50:12 npc99a84aae6d34 sshd-session[119590]: pam_unix(sshd:session): session opened for user zuul(uid=1000) by zuul(uid=0) Apr 14 22:50:15 npc99a84aae6d34 sshd-session[119610]: Connection reset by authenticating user root 2.57.122.188 port 15710 [preauth] Apr 14 22:50:36 npc99a84aae6d34 sshd-session[120732]: Accepted publickey for zuul from 104.130.127.165 port 53642 ssh2: RSA SHA256:c94tZpPtmWI4tk86GWYkvl4qtYyEmDbB3eg5vn0XjIc Apr 14 22:50:36 npc99a84aae6d34 sshd-session[120732]: pam_unix(sshd:session): session opened for user zuul(uid=1000) by zuul(uid=0) Apr 14 22:50:46 npc99a84aae6d34 sshd-session[120787]: Accepted publickey for zuul from 104.130.127.165 port 34652 ssh2: RSA SHA256:c94tZpPtmWI4tk86GWYkvl4qtYyEmDbB3eg5vn0XjIc Apr 14 22:50:46 npc99a84aae6d34 sshd-session[120787]: pam_unix(sshd:session): session opened for user zuul(uid=1000) by zuul(uid=0) ○ sshd-keygen.service - Generate sshd host keys on first boot Loaded: loaded (/usr/lib/systemd/system/sshd-keygen.service; enabled; preset: enabled) Active: inactive (dead) since Tue 2026-04-14 22:17:30 UTC; 34min ago Duration: 9min 54.368s Invocation: 179c93dbe69c4dc6928046869b1d3f09 Condition: start condition unmet at Tue 2026-04-14 22:17:30 UTC; 34min ago Main PID: 602 (code=exited, status=0/SUCCESS) Apr 14 22:07:35 debian systemd[1]: Starting sshd-keygen.service - Generate sshd host keys on first boot... Apr 14 22:07:35 npc99a84aae6d34 ssh-keygen[602]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Apr 14 22:07:35 npc99a84aae6d34 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys on first boot. Apr 14 22:17:30 npc99a84aae6d34 systemd[1]: sshd-keygen.service: Deactivated successfully. Apr 14 22:17:30 npc99a84aae6d34 systemd[1]: Stopped sshd-keygen.service - Generate sshd host keys on first boot. Apr 14 22:17:30 npc99a84aae6d34 systemd[1]: Stopping sshd-keygen.service - Generate sshd host keys on first boot... Apr 14 22:17:30 npc99a84aae6d34 systemd[1]: sshd-keygen.service - Generate sshd host keys on first boot was skipped because of an unmet condition check (ConditionFirstBoot=yes). ○ sshd@sshd-keygen.service - OpenBSD Secure Shell server per-connection daemon Loaded: loaded (/usr/lib/systemd/system/sshd@.service; static) Active: inactive (dead) Docs: man:sshd(8) man:sshd_config(5) ● sysfsutils.service - Apply sysfs variables Loaded: loaded (/usr/lib/systemd/system/sysfsutils.service; enabled; preset: enabled) Active: active (exited) since Tue 2026-04-14 22:17:17 UTC; 34min ago Invocation: 0d2a06f060374fb5967b2a3b1cb50471 Docs: man:sysfs.conf(5) man:systool(1) Main PID: 12645 (code=exited, status=0/SUCCESS) Mem peak: 1.8M CPU: 31ms Apr 14 22:17:17 npc99a84aae6d34 systemd[1]: Starting sysfsutils.service - Apply sysfs variables... Apr 14 22:17:17 npc99a84aae6d34 systemd[1]: Finished sysfsutils.service - Apply sysfs variables. ○ systemd-ask-password-console.service - Dispatch Password Requests to Console Loaded: loaded (/usr/lib/systemd/system/systemd-ask-password-console.service; static) Active: inactive (dead) TriggeredBy: ● systemd-ask-password-console.path Docs: man:systemd-ask-password-console.service(8) ○ systemd-ask-password-wall.service - Forward Password Requests to Wall Loaded: loaded (/usr/lib/systemd/system/systemd-ask-password-wall.service; static) Active: inactive (dead) TriggeredBy: ● systemd-ask-password-wall.path Docs: man:systemd-ask-password-wall.service(8) ○ systemd-battery-check.service - Check battery level during early boot Loaded: loaded (/usr/lib/systemd/system/systemd-battery-check.service; static) Active: inactive (dead) Docs: man:systemd-battery-check.service(8) ● systemd-binfmt.service - Set Up Additional Binary Formats Loaded: loaded (/usr/lib/systemd/system/systemd-binfmt.service; static) Active: active (exited) since Tue 2026-04-14 22:18:16 UTC; 33min ago Invocation: 750ebf8a6ab4489096b04233106e738d Docs: man:systemd-binfmt.service(8) man:binfmt.d(5) https://docs.kernel.org/admin-guide/binfmt-misc.html https://systemd.io/API_FILE_SYSTEMS Main PID: 19114 (code=exited, status=0/SUCCESS) Mem peak: 1.7M CPU: 31ms Apr 14 22:18:16 npc99a84aae6d34 systemd[1]: Starting systemd-binfmt.service - Set Up Additional Binary Formats... Apr 14 22:18:16 npc99a84aae6d34 systemd[1]: Finished systemd-binfmt.service - Set Up Additional Binary Formats. ○ systemd-bsod.service - Display Boot-Time Emergency Messages In Full Screen Loaded: loaded (/usr/lib/systemd/system/systemd-bsod.service; static) Active: inactive (dead) Docs: man:systemd-bsod.service(8) ○ systemd-confext.service - Merge System Configuration Images into /etc/ Loaded: loaded (/usr/lib/systemd/system/systemd-confext.service; disabled; preset: enabled) Active: inactive (dead) Docs: man:systemd-confext.service(8) ● systemd-firstboot.service - First Boot Wizard Loaded: loaded (/usr/lib/systemd/system/systemd-firstboot.service; static) Active: active (exited) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 73ca77126b8e4d2cac19a8eb14ad49b2 Docs: man:systemd-firstboot(1) Main PID: 434 (code=exited, status=0/SUCCESS) Mem peak: 2M CPU: 29ms Apr 14 22:07:34 debian systemd[1]: Starting systemd-firstboot.service - First Boot Wizard... Apr 14 22:07:34 debian systemd[1]: Finished systemd-firstboot.service - First Boot Wizard. ● systemd-fsck-root.service - File System Check on Root Device Loaded: loaded (/usr/lib/systemd/system/systemd-fsck-root.service; enabled-runtime; preset: enabled) Drop-In: /usr/lib/systemd/system/systemd-fsck-root.service.d └─10-skip-fsck-initramfs.conf Active: active (exited) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 4c01e25e7a874122bd9edbd01f0239a8 Docs: man:systemd-fsck-root.service(8) Main PID: 358 (code=exited, status=0/SUCCESS) Mem peak: 3.4M CPU: 49ms Apr 14 22:07:33 debian systemd-fsck[366]: cloudimg-rootfs: clean, 648583/5285088 files, 3554854/5284608 blocks Notice: journal has been rotated since unit was started, output may be incomplete. ○ systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info Loaded: loaded (/usr/lib/systemd/system/systemd-hibernate-clear.service; static) Active: inactive (dead) Condition: start condition unmet at Tue 2026-04-14 22:07:34 UTC; 44min ago Docs: man:systemd-hibernate-clear.service(8) Apr 14 22:07:34 debian systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Apr 14 22:07:34 debian systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). ○ systemd-hibernate-resume.service - Resume from hibernation Loaded: loaded (/usr/lib/systemd/system/systemd-hibernate-resume.service; static) Active: inactive (dead) Docs: man:systemd-hibernate-resume.service(8) ○ systemd-hibernate.service - System Hibernate Loaded: loaded (/usr/lib/systemd/system/systemd-hibernate.service; static) Active: inactive (dead) Docs: man:systemd-hibernate.service(8) ○ systemd-hostnamed.service - Hostname Service Loaded: loaded (/usr/lib/systemd/system/systemd-hostnamed.service; static) Active: inactive (dead) TriggeredBy: ● systemd-hostnamed.socket Docs: man:systemd-hostnamed.service(8) man:hostname(5) man:machine-info(5) man:org.freedesktop.hostname1(5) ○ systemd-hwdb-update.service - Rebuild Hardware Database Loaded: loaded (/usr/lib/systemd/system/systemd-hwdb-update.service; static) Active: inactive (dead) Condition: start condition unmet at Tue 2026-04-14 22:07:34 UTC; 44min ago Docs: man:hwdb(7) man:systemd-hwdb(8) Apr 14 22:07:33 debian systemd[1]: systemd-hwdb-update.service - Rebuild Hardware Database was skipped because no trigger condition checks were met. Apr 14 22:07:34 debian systemd[1]: systemd-hwdb-update.service - Rebuild Hardware Database was skipped because no trigger condition checks were met. Apr 14 22:07:34 debian systemd[1]: systemd-hwdb-update.service - Rebuild Hardware Database was skipped because no trigger condition checks were met. ○ systemd-hybrid-sleep.service - System Hybrid Suspend+Hibernate Loaded: loaded (/usr/lib/systemd/system/systemd-hybrid-sleep.service; static) Active: inactive (dead) Docs: man:systemd-hybrid-sleep.service(8) ○ systemd-importd.service - Disk Image Download Service Loaded: loaded (/usr/lib/systemd/system/systemd-importd.service; static) Active: inactive (dead) TriggeredBy: ○ systemd-importd.socket Docs: man:systemd-importd.service(8) man:org.freedesktop.import1(5) ○ systemd-initctl.service - initctl Compatibility Daemon Loaded: loaded (/usr/lib/systemd/system/systemd-initctl.service; static) Active: inactive (dead) TriggeredBy: ● systemd-initctl.socket Docs: man:systemd-initctl.service(8) ● systemd-journal-catalog-update.service - Rebuild Journal Catalog Loaded: loaded (/usr/lib/systemd/system/systemd-journal-catalog-update.service; static) Active: active (exited) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 4d188b282f0a469ab45b939e21610245 Docs: man:systemd-journald.service(8) man:journald.conf(5) Main PID: 435 (code=exited, status=0/SUCCESS) Mem peak: 2.4M CPU: 58ms Apr 14 22:07:34 debian systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Apr 14 22:07:34 debian systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. ● systemd-journal-flush.service - Flush Journal to Persistent Storage Loaded: loaded (/usr/lib/systemd/system/systemd-journal-flush.service; static) Active: active (exited) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 65d1d93e043f4214add7984f67870a1d Docs: man:systemd-journald.service(8) man:journald.conf(5) Main PID: 403 (code=exited, status=0/SUCCESS) Mem peak: 1.7M CPU: 24ms Apr 14 22:07:33 debian systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Apr 14 22:07:33 debian systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. ● systemd-journald.service - Journal Service Loaded: loaded (/usr/lib/systemd/system/systemd-journald.service; static) Active: active (running) since Tue 2026-04-14 22:18:40 UTC; 33min ago Invocation: 51b63a9fa82f4963b8adbaaafd48247c TriggeredBy: ○ systemd-journald-audit.socket ● systemd-journald.socket ● systemd-journald-dev-log.socket Docs: man:systemd-journald.service(8) man:journald.conf(5) Main PID: 20737 (systemd-journal) Status: "Processing requests..." Tasks: 1 (limit: 9256) FD Store: 52 (limit: 4224) Memory: 67.4M (peak: 67.6M, swap: 860K, swap peak: 860K, zswap: 197.1K) CPU: 18.342s CGroup: /system.slice/systemd-journald.service └─20737 /usr/lib/systemd/systemd-journald Apr 14 22:18:40 npc99a84aae6d34 systemd-journald[20737]: Collecting audit messages is disabled. Apr 14 22:18:40 npc99a84aae6d34 systemd-journald[20737]: Journal started Apr 14 22:18:40 npc99a84aae6d34 systemd-journald[20737]: System Journal (/var/log/journal/05f98293711944c0a3294124e4af3d05) is 24M, max 4G, 3.9G free. ● systemd-logind.service - User Login Management Loaded: loaded (/usr/lib/systemd/system/systemd-logind.service; static) Drop-In: /usr/lib/systemd/system/systemd-logind.service.d └─dbus.conf Active: active (running) since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: 4f4c2378dd894e3eacc7d6e3398bc6ab Docs: man:sd-login(3) man:systemd-logind.service(8) man:logind.conf(5) man:org.freedesktop.login1(5) Main PID: 605 (systemd-logind) Status: "Processing requests..." Tasks: 1 (limit: 9256) FD Store: 2 (limit: 768) Memory: 2M (peak: 2.4M, swap: 4K, swap peak: 12K, zswap: 114B) CPU: 3.577s CGroup: /system.slice/systemd-logind.service └─605 /usr/lib/systemd/systemd-logind Apr 14 22:39:14 npc99a84aae6d34 systemd-logind[605]: Session 31 logged out. Waiting for processes to exit. Apr 14 22:39:14 npc99a84aae6d34 systemd-logind[605]: Removed session 31. Apr 14 22:40:23 npc99a84aae6d34 systemd-logind[605]: Session 1 logged out. Waiting for processes to exit. Apr 14 22:50:12 npc99a84aae6d34 systemd-logind[605]: New session 32 of user zuul. Apr 14 22:50:34 npc99a84aae6d34 systemd-logind[605]: Session 32 logged out. Waiting for processes to exit. Apr 14 22:50:34 npc99a84aae6d34 systemd-logind[605]: Removed session 32. Apr 14 22:50:36 npc99a84aae6d34 systemd-logind[605]: New session 33 of user zuul. Apr 14 22:50:43 npc99a84aae6d34 systemd-logind[605]: Session 33 logged out. Waiting for processes to exit. Apr 14 22:50:43 npc99a84aae6d34 systemd-logind[605]: Removed session 33. Apr 14 22:50:46 npc99a84aae6d34 systemd-logind[605]: New session 34 of user zuul. ● systemd-machine-id-commit.service - Save Transient machine-id to Disk Loaded: loaded (/usr/lib/systemd/system/systemd-machine-id-commit.service; static) Active: active (exited) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 548187d28b7f40adbbf3d15a8f0f0825 Docs: man:systemd-machine-id-commit.service(8) Main PID: 472 (code=exited, status=0/SUCCESS) Mem peak: 1.7M CPU: 42ms Apr 14 22:07:34 debian systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Apr 14 22:07:34 debian systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. ● systemd-machined.service - Virtual Machine and Container Registration Service Loaded: loaded (/usr/lib/systemd/system/systemd-machined.service; static) Active: active (running) since Tue 2026-04-14 22:24:05 UTC; 27min ago Invocation: 4e6d297bcae44c13844f415e58d54c65 Docs: man:systemd-machined.service(8) man:org.freedesktop.machine1(5) Main PID: 41980 (systemd-machine) Status: "Processing requests..." Tasks: 1 (limit: 9256) Memory: 696K (peak: 2.2M, swap: 940K, swap peak: 940K, zswap: 240.9K) CPU: 2.376s CGroup: /system.slice/systemd-machined.service └─41980 /usr/lib/systemd/systemd-machined Apr 14 22:24:05 npc99a84aae6d34 systemd[1]: Starting systemd-machined.service - Virtual Machine and Container Registration Service... Apr 14 22:24:05 npc99a84aae6d34 systemd[1]: Started systemd-machined.service - Virtual Machine and Container Registration Service. ● systemd-modules-load.service - Load Kernel Modules Loaded: loaded (/usr/lib/systemd/system/systemd-modules-load.service; static) Active: active (exited) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 750a9b1392d5478098d3e9bed2c62b32 Docs: man:systemd-modules-load.service(8) man:modules-load.d(5) Main PID: 360 (code=exited, status=0/SUCCESS) Mem peak: 1.7M CPU: 105ms Apr 14 22:07:33 debian systemd-modules-load[360]: Inserted module '8021q' Apr 14 22:07:33 debian systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Notice: journal has been rotated since unit was started, output may be incomplete. ○ systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd Loaded: loaded (/usr/lib/systemd/system/systemd-networkd-persistent-storage.service; static) Active: inactive (dead) Docs: man:networkctl(1) ○ systemd-networkd.service - Network Configuration Loaded: loaded (/usr/lib/systemd/system/systemd-networkd.service; disabled; preset: enabled) Active: inactive (dead) TriggeredBy: ○ systemd-networkd.socket Docs: man:systemd-networkd.service(8) man:org.freedesktop.network1(5) FD Store: 0 (limit: 512) ○ systemd-pcrmachine.service - TPM PCR Machine ID Measurement Loaded: loaded (/usr/lib/systemd/system/systemd-pcrmachine.service; static) Active: inactive (dead) Condition: start condition unmet at Tue 2026-04-14 22:07:34 UTC; 44min ago Docs: man:systemd-pcrmachine.service(8) Apr 14 22:07:34 debian systemd[1]: systemd-pcrmachine.service - TPM PCR Machine ID Measurement was skipped because of an unmet condition check (ConditionSecurity=measured-uki). Apr 14 22:07:34 debian systemd[1]: systemd-pcrmachine.service - TPM PCR Machine ID Measurement was skipped because of an unmet condition check (ConditionSecurity=measured-uki). ○ systemd-pcrphase-initrd.service - TPM PCR Barrier (initrd) Loaded: loaded (/usr/lib/systemd/system/systemd-pcrphase-initrd.service; static) Active: inactive (dead) Docs: man:systemd-pcrphase-initrd.service(8) ○ systemd-pcrphase-sysinit.service - TPM PCR Barrier (Initialization) Loaded: loaded (/usr/lib/systemd/system/systemd-pcrphase-sysinit.service; static) Active: inactive (dead) Condition: start condition unmet at Tue 2026-04-14 22:07:35 UTC; 44min ago Docs: man:systemd-pcrphase-sysinit.service(8) Apr 14 22:07:35 debian systemd[1]: systemd-pcrphase-sysinit.service - TPM PCR Barrier (Initialization) was skipped because of an unmet condition check (ConditionSecurity=measured-uki). ○ systemd-pcrphase.service - TPM PCR Barrier (User) Loaded: loaded (/usr/lib/systemd/system/systemd-pcrphase.service; static) Active: inactive (dead) Condition: start condition unmet at Tue 2026-04-14 22:07:36 UTC; 44min ago Docs: man:systemd-pcrphase.service(8) Apr 14 22:07:36 npc99a84aae6d34 systemd[1]: systemd-pcrphase.service - TPM PCR Barrier (User) was skipped because of an unmet condition check (ConditionSecurity=measured-uki). ○ systemd-pstore.service - Platform Persistent Storage Archival Loaded: loaded (/usr/lib/systemd/system/systemd-pstore.service; enabled; preset: enabled) Active: inactive (dead) Condition: start condition unmet at Tue 2026-04-14 22:07:34 UTC; 44min ago Docs: man:systemd-pstore(8) Apr 14 22:07:33 debian systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 14 22:07:34 debian systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). ○ systemd-quotacheck-root.service - Root File System Quota Check Loaded: loaded (/usr/lib/systemd/system/systemd-quotacheck-root.service; static) Active: inactive (dead) Docs: man:systemd-quotacheck.service(8) ● systemd-random-seed.service - Load/Save OS Random Seed Loaded: loaded (/usr/lib/systemd/system/systemd-random-seed.service; static) Active: active (exited) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 386f2b5f874047228e91eaf3ae93eee0 Docs: man:systemd-random-seed.service(8) man:random(4) Main PID: 404 (code=exited, status=0/SUCCESS) Mem peak: 1.7M CPU: 23ms Apr 14 22:07:33 debian systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Apr 14 22:07:33 debian systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. ● systemd-remount-fs.service - Remount Root and Kernel File Systems Loaded: loaded (/usr/lib/systemd/system/systemd-remount-fs.service; enabled-runtime; preset: enabled) Active: active (exited) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: bd242d92e59a47d7a8a5c34e096f79d2 Docs: man:systemd-remount-fs.service(8) https://systemd.io/API_FILE_SYSTEMS Main PID: 370 (code=exited, status=0/SUCCESS) Mem peak: 1.8M CPU: 29ms Apr 14 22:07:33 debian systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. ○ systemd-soft-reboot.service - Reboot System Userspace Loaded: loaded (/usr/lib/systemd/system/systemd-soft-reboot.service; static) Active: inactive (dead) Docs: man:systemd-soft-reboot.service(8) ○ systemd-suspend-then-hibernate.service - System Suspend then Hibernate Loaded: loaded (/usr/lib/systemd/system/systemd-suspend-then-hibernate.service; static) Active: inactive (dead) Docs: man:systemd-suspend-then-hibernate.service(8) ○ systemd-suspend.service - System Suspend Loaded: loaded (/usr/lib/systemd/system/systemd-suspend.service; static) Active: inactive (dead) Docs: man:systemd-suspend.service(8) ● systemd-sysctl.service - Apply Kernel Variables Loaded: loaded (/usr/lib/systemd/system/systemd-sysctl.service; static) Active: active (exited) since Tue 2026-04-14 22:24:00 UTC; 27min ago Invocation: e405b67a37af45ac9f220a18cf956a0e Docs: man:systemd-sysctl.service(8) man:sysctl.d(5) Main PID: 41586 (code=exited, status=0/SUCCESS) Mem peak: 1.9M CPU: 18ms Apr 14 22:24:00 npc99a84aae6d34 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 14 22:24:00 npc99a84aae6d34 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. ○ systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/ Loaded: loaded (/usr/lib/systemd/system/systemd-sysext.service; disabled; preset: enabled) Active: inactive (dead) Docs: man:systemd-sysext.service(8) ● systemd-sysusers.service - Create System Users Loaded: loaded (/usr/lib/systemd/system/systemd-sysusers.service; static) Active: active (exited) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 9a4162767389475fa0078f8274ff9df4 Docs: man:sysusers.d(5) man:systemd-sysusers.service(8) Main PID: 405 (code=exited, status=0/SUCCESS) Mem peak: 2M CPU: 21ms Apr 14 22:07:33 debian systemd[1]: Starting systemd-sysusers.service - Create System Users... Apr 14 22:07:33 debian systemd[1]: Finished systemd-sysusers.service - Create System Users. ● systemd-timesyncd.service - Network Time Synchronization Loaded: loaded (/usr/lib/systemd/system/systemd-timesyncd.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 2eeb70f87cbf43fbb61f02f633c10571 Docs: man:systemd-timesyncd.service(8) Main PID: 411 (systemd-timesyn) Status: "Contacted time server [2001:678:938:300:9780:3b76:7a8a:b9fa]:123 (2.debian.pool.ntp.org)." Tasks: 2 (limit: 9256) Memory: 1.6M (peak: 2.3M) CPU: 122ms CGroup: /system.slice/systemd-timesyncd.service └─411 /usr/lib/systemd/systemd-timesyncd Apr 14 22:07:33 debian systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Apr 14 22:07:33 debian systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Apr 14 22:07:34 debian systemd-timesyncd[411]: Network configuration changed, trying to establish connection. Apr 14 22:07:34 debian systemd-timesyncd[411]: Network configuration changed, trying to establish connection. Apr 14 22:07:34 debian systemd-timesyncd[411]: Network configuration changed, trying to establish connection. Apr 14 22:07:34 debian systemd-timesyncd[411]: Network configuration changed, trying to establish connection. Apr 14 22:08:05 npc99a84aae6d34 systemd-timesyncd[411]: Contacted time server [2001:678:938:300:9780:3b76:7a8a:b9fa]:123 (2.debian.pool.ntp.org). Apr 14 22:08:05 npc99a84aae6d34 systemd-timesyncd[411]: Initial clock synchronization to Tue 2026-04-14 22:08:05.330018 UTC. ○ systemd-tmpfiles-clean.service - Cleanup of Temporary Directories Loaded: loaded (/usr/lib/systemd/system/systemd-tmpfiles-clean.service; static) Active: inactive (dead) since Tue 2026-04-14 22:22:37 UTC; 29min ago Invocation: 11440435bfe94fb191e178f640c8aaeb TriggeredBy: ● systemd-tmpfiles-clean.timer Docs: man:tmpfiles.d(5) man:systemd-tmpfiles(8) Main PID: 35127 (code=exited, status=0/SUCCESS) Mem peak: 2M CPU: 42ms Apr 14 22:22:37 npc99a84aae6d34 systemd[1]: Starting systemd-tmpfiles-clean.service - Cleanup of Temporary Directories... Apr 14 22:22:37 npc99a84aae6d34 systemd-tmpfiles[35127]: /usr/lib/tmpfiles.d/legacy.conf:14: Duplicate line for path "/run/lock", ignoring. Apr 14 22:22:37 npc99a84aae6d34 systemd-tmpfiles[35127]: /etc/tmpfiles.d/uwsgi.conf:1: Line references path below legacy directory /var/run/, updating /var/run/uwsgi → /run/uwsgi; please update the tmpfiles.d/ drop-in file accordingly. Apr 14 22:22:37 npc99a84aae6d34 systemd[1]: systemd-tmpfiles-clean.service: Deactivated successfully. Apr 14 22:22:37 npc99a84aae6d34 systemd[1]: Finished systemd-tmpfiles-clean.service - Cleanup of Temporary Directories. ● systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully Loaded: loaded (/usr/lib/systemd/system/systemd-tmpfiles-setup-dev-early.service; static) Active: active (exited) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 12dd65fb6b634e86a02049d79dfbbe4f Docs: man:tmpfiles.d(5) man:systemd-tmpfiles(8) Main PID: 374 (code=exited, status=0/SUCCESS) Mem peak: 2.1M CPU: 36ms Apr 14 22:07:33 debian systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. ● systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev Loaded: loaded (/usr/lib/systemd/system/systemd-tmpfiles-setup-dev.service; static) Active: active (exited) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 43817d6e8a7946719e9fef087f73b24d Docs: man:tmpfiles.d(5) man:systemd-tmpfiles(8) Main PID: 412 (code=exited, status=0/SUCCESS) Mem peak: 2M CPU: 24ms Apr 14 22:07:33 debian systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Apr 14 22:07:33 debian systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. ● systemd-tmpfiles-setup.service - Create System Files and Directories Loaded: loaded (/usr/lib/systemd/system/systemd-tmpfiles-setup.service; static) Active: active (exited) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 65fa450570b5479bb0480b4351263ec1 Docs: man:tmpfiles.d(5) man:systemd-tmpfiles(8) Main PID: 421 (code=exited, status=0/SUCCESS) Mem peak: 1.9M CPU: 59ms Apr 14 22:07:33 debian systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Apr 14 22:07:33 debian systemd-tmpfiles[421]: /usr/lib/tmpfiles.d/legacy.conf:14: Duplicate line for path "/run/lock", ignoring. Apr 14 22:07:34 debian systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. ○ systemd-tpm2-setup-early.service - Early TPM SRK Setup Loaded: loaded (/usr/lib/systemd/system/systemd-tpm2-setup-early.service; static) Active: inactive (dead) Condition: start condition unmet at Tue 2026-04-14 22:07:34 UTC; 44min ago Docs: man:systemd-tpm2-setup.service(8) Apr 14 22:07:34 debian systemd[1]: systemd-tpm2-setup-early.service - Early TPM SRK Setup was skipped because of an unmet condition check (ConditionSecurity=measured-uki). Apr 14 22:07:34 debian systemd[1]: systemd-tpm2-setup-early.service - Early TPM SRK Setup was skipped because of an unmet condition check (ConditionSecurity=measured-uki). ○ systemd-tpm2-setup.service - TPM SRK Setup Loaded: loaded (/usr/lib/systemd/system/systemd-tpm2-setup.service; static) Active: inactive (dead) Condition: start condition unmet at Tue 2026-04-14 22:07:34 UTC; 44min ago Docs: man:systemd-tpm2-setup.service(8) Apr 14 22:07:33 debian systemd[1]: systemd-tpm2-setup.service - TPM SRK Setup was skipped because of an unmet condition check (ConditionSecurity=measured-uki). Apr 14 22:07:34 debian systemd[1]: systemd-tpm2-setup.service - TPM SRK Setup was skipped because of an unmet condition check (ConditionSecurity=measured-uki). Apr 14 22:07:34 debian systemd[1]: systemd-tpm2-setup.service - TPM SRK Setup was skipped because of an unmet condition check (ConditionSecurity=measured-uki). ● systemd-udev-load-credentials.service - Load udev Rules from Credentials Loaded: loaded (/usr/lib/systemd/system/systemd-udev-load-credentials.service; disabled; preset: enabled) Active: active (exited) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 358cca064f604e8f8545a132471eeb6d Docs: man:udevadm(8) man:udev(7) man:systemd.system-credentials(7) Main PID: 363 (code=exited, status=0/SUCCESS) Mem peak: 1.9M CPU: 34ms Notice: journal has been rotated since unit was started, output may be incomplete. ● systemd-udev-settle.service - Wait for udev To Complete Device Initialization Loaded: loaded (/usr/lib/systemd/system/systemd-udev-settle.service; static) Active: active (exited) since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: 27aefbded6d8499a9e048ab609c91ed6 Docs: man:systemd-udev-settle.service(8) Main PID: 397 (code=exited, status=0/SUCCESS) Mem peak: 1.9M CPU: 26ms Apr 14 22:07:33 debian systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Apr 14 22:07:33 debian udevadm[397]: systemd-udev-settle.service is deprecated. Please fix growroot.service not to pull it in. Apr 14 22:07:35 debian systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. ● systemd-udev-trigger.service - Coldplug All udev Devices Loaded: loaded (/usr/lib/systemd/system/systemd-udev-trigger.service; static) Active: active (exited) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 5eac7c72bafa4eec9e0e2b9d50bf3845 Docs: man:udev(7) man:systemd-udevd.service(8) Main PID: 365 (code=exited, status=0/SUCCESS) Mem peak: 3.3M CPU: 133ms Apr 14 22:07:33 debian systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Notice: journal has been rotated since unit was started, output may be incomplete. ● systemd-udevd.service - Rule-based Manager for Device Events and Files Loaded: loaded (/usr/lib/systemd/system/systemd-udevd.service; static) Drop-In: /usr/lib/systemd/system/systemd-udevd.service.d └─syscall-architecture.conf Active: active (running) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: f7f8c50f69914d08be1614d12ae95ddf TriggeredBy: ● systemd-udevd-kernel.socket ● systemd-udevd-control.socket Docs: man:systemd-udevd.service(8) man:udev(7) Main PID: 422 (systemd-udevd) Status: "Processing with 32 children at max" Tasks: 1 Memory: 5M (peak: 25.2M, swap: 292K, swap peak: 292K, zswap: 11.1K) CPU: 3.018s CGroup: /system.slice/systemd-udevd.service └─udev └─422 /usr/lib/systemd/systemd-udevd Apr 14 22:27:19 npc99a84aae6d34 lvm[60637]: PV /dev/loop0 online, VG ceph_vg is complete. Apr 14 22:27:19 npc99a84aae6d34 lvm[60637]: VG ceph_vg finished Apr 14 22:28:42 npc99a84aae6d34 lvm[68029]: PV /dev/loop0 online, VG ceph_vg is complete. Apr 14 22:28:42 npc99a84aae6d34 lvm[68029]: VG ceph_vg finished Apr 14 22:29:52 npc99a84aae6d34 lvm[73769]: PV /dev/loop0 online, VG ceph_vg is complete. Apr 14 22:29:52 npc99a84aae6d34 lvm[73769]: VG ceph_vg finished Apr 14 22:31:00 npc99a84aae6d34 lvm[76742]: PV /dev/loop0 online, VG ceph_vg is complete. Apr 14 22:31:00 npc99a84aae6d34 lvm[76742]: VG ceph_vg finished Apr 14 22:32:08 npc99a84aae6d34 lvm[83497]: PV /dev/loop0 online, VG ceph_vg is complete. Apr 14 22:32:08 npc99a84aae6d34 lvm[83497]: VG ceph_vg finished ● systemd-update-done.service - Update is Completed Loaded: loaded (/usr/lib/systemd/system/systemd-update-done.service; static) Active: active (exited) since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 3f8abee7601a4cd8ac9b4c67442a0929 Docs: man:systemd-update-done.service(8) Main PID: 500 (code=exited, status=0/SUCCESS) Mem peak: 1.7M CPU: 14ms Apr 14 22:07:34 debian systemd[1]: Starting systemd-update-done.service - Update is Completed... Apr 14 22:07:34 debian systemd[1]: Finished systemd-update-done.service - Update is Completed. ● systemd-user-sessions.service - Permit User Sessions Loaded: loaded (/usr/lib/systemd/system/systemd-user-sessions.service; static) Active: active (exited) since Tue 2026-04-14 22:07:36 UTC; 44min ago Invocation: 2b80e96f2c9743d89ee12efbd8dd9c0b Docs: man:systemd-user-sessions.service(8) Main PID: 808 (code=exited, status=0/SUCCESS) Mem peak: 1.7M CPU: 28ms Apr 14 22:07:36 npc99a84aae6d34 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Apr 14 22:07:36 npc99a84aae6d34 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. ○ unbound-resolvconf.service - Unbound asyncronous resolvconf update helper Loaded: loaded (/usr/lib/systemd/system/unbound-resolvconf.service; enabled; preset: enabled) Active: inactive (dead) Condition: start condition unmet at Tue 2026-04-14 22:10:32 UTC; 41min ago Apr 14 22:07:36 npc99a84aae6d34 systemd[1]: unbound-resolvconf.service - Unbound asyncronous resolvconf update helper was skipped because of an unmet condition check (ConditionFileIsExecutable=/sbin/resolvconf). Apr 14 22:10:32 npc99a84aae6d34 systemd[1]: unbound-resolvconf.service - Unbound asyncronous resolvconf update helper was skipped because of an unmet condition check (ConditionFileIsExecutable=/sbin/resolvconf). ● unbound.service - Unbound DNS server Loaded: loaded (/usr/lib/systemd/system/unbound.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:10:32 UTC; 41min ago Invocation: 8437f26198b54289823ca28fc5854e35 Docs: man:unbound(8) Main PID: 1203 (unbound) Tasks: 1 (limit: 9256) Memory: 10.9M (peak: 13.6M, swap: 1.8M, swap peak: 1.8M, zswap: 501.1K) CPU: 247ms CGroup: /system.slice/unbound.service └─1203 /usr/sbin/unbound -d -p Apr 14 22:10:32 npc99a84aae6d34 unbound[1203]: [1776204632] unbound[1203:0] debug: creating udp6 socket ::1 53 Apr 14 22:10:32 npc99a84aae6d34 unbound[1203]: [1776204632] unbound[1203:0] debug: creating tcp6 socket ::1 53 Apr 14 22:10:32 npc99a84aae6d34 unbound[1203]: [1776204632] unbound[1203:0] debug: creating udp4 socket 127.0.0.1 53 Apr 14 22:10:32 npc99a84aae6d34 unbound[1203]: [1776204632] unbound[1203:0] debug: creating tcp4 socket 127.0.0.1 53 Apr 14 22:10:32 npc99a84aae6d34 unbound[1203]: [1776204632] unbound[1203:0] debug: creating unix socket /run/unbound.ctl Apr 14 22:10:32 npc99a84aae6d34 unbound[1203]: [1776204632] unbound[1203:0] debug: module config: "subnetcache validator iterator" Apr 14 22:10:32 npc99a84aae6d34 unbound[1203]: [1776204632] unbound[1203:0] debug: chdir to /etc/unbound Apr 14 22:10:32 npc99a84aae6d34 unbound[1203]: [1776204632] unbound[1203:0] debug: drop user privileges, run as unbound Apr 14 22:10:32 npc99a84aae6d34 unbound[1203]: [1776204632] unbound[1203:0] debug: switching log to /var/log/unbound.log Apr 14 22:10:32 npc99a84aae6d34 systemd[1]: Started unbound.service - Unbound DNS server. ● user-runtime-dir@1000.service - User Runtime Directory /run/user/1000 Loaded: loaded (/usr/lib/systemd/system/user-runtime-dir@.service; static) Active: active (exited) since Tue 2026-04-14 22:09:41 UTC; 42min ago Invocation: dda4765f807f4154882a86f1f15eaa15 Docs: man:user@.service(5) Main PID: 857 (code=exited, status=0/SUCCESS) Mem peak: 1.7M CPU: 31ms Apr 14 22:09:41 npc99a84aae6d34 systemd[1]: Starting user-runtime-dir@1000.service - User Runtime Directory /run/user/1000... Apr 14 22:09:41 npc99a84aae6d34 systemd[1]: Finished user-runtime-dir@1000.service - User Runtime Directory /run/user/1000. ● user@1000.service - User Manager for UID 1000 Loaded: loaded (/usr/lib/systemd/system/user@.service; static) Drop-In: /usr/lib/systemd/system/user@.service.d └─10-login-barrier.conf Active: active (running) since Tue 2026-04-14 22:09:41 UTC; 42min ago Invocation: 12596e08a0b9411890c28d3730d0ecdb Docs: man:user@.service(5) Main PID: 859 (systemd) Status: "Ready." Tasks: 3 Memory: 4.9M (peak: 11.5M, swap: 1M, swap peak: 1M, zswap: 252K) CPU: 1.560s CGroup: /user.slice/user-1000.slice/user@1000.service ├─init.scope │ ├─859 /usr/lib/systemd/systemd --user │ └─862 "(sd-pam)" └─session.slice └─dbus.service └─12675 /usr/bin/dbus-daemon --session --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only Apr 14 22:09:41 npc99a84aae6d34 systemd[859]: Reached target default.target - Main User Target. Apr 14 22:09:41 npc99a84aae6d34 systemd[859]: Startup finished in 242ms. Apr 14 22:09:41 npc99a84aae6d34 systemd[1]: Started user@1000.service - User Manager for UID 1000. Apr 14 22:17:17 npc99a84aae6d34 systemd[859]: Created slice session.slice - User Core Session Slice. Apr 14 22:17:17 npc99a84aae6d34 systemd[859]: Starting dbus.service - D-Bus User Message Bus... Apr 14 22:17:17 npc99a84aae6d34 systemd[859]: Started dbus.service - D-Bus User Message Bus. Apr 14 22:17:17 npc99a84aae6d34 dbus-daemon[12675]: [session uid=1000 pid=12675 pidfd=5] Successfully activated service 'org.freedesktop.systemd1' Apr 14 22:17:17 npc99a84aae6d34 systemd[859]: Reload requested from client PID 12672 ('systemd-stdio-b')... Apr 14 22:17:17 npc99a84aae6d34 systemd[859]: Reloading... Apr 14 22:17:17 npc99a84aae6d34 systemd[859]: Reloading finished in 87 ms. ● uuidd.service - Daemon for generating UUIDs Loaded: loaded (/usr/lib/systemd/system/uuidd.service; indirect; preset: enabled) Active: active (running) since Tue 2026-04-14 22:33:44 UTC; 18min ago Invocation: 759a70486ade4684a6f0d98b756357f5 TriggeredBy: ● uuidd.socket Docs: man:uuidd(8) Main PID: 96046 (uuidd) Tasks: 1 (limit: 9256) Memory: 376K (peak: 2.4M, swap: 152K, swap peak: 152K, zswap: 34.2K) CPU: 95ms CGroup: /system.slice/uuidd.service └─96046 /usr/sbin/uuidd --socket-activation --cont-clock Apr 14 22:33:44 npc99a84aae6d34 systemd[1]: Started uuidd.service - Daemon for generating UUIDs. ● uwsgi.service - LSB: Start/stop uWSGI server instance(s) Loaded: loaded (/etc/init.d/uwsgi; generated) Active: active (exited) since Tue 2026-04-14 22:20:11 UTC; 31min ago Invocation: 62b24bca6dc04e62ac145d647a8fe7c7 Docs: man:systemd-sysv-generator(8) Mem peak: 3.1M CPU: 87ms Apr 14 22:20:11 npc99a84aae6d34 systemd[1]: Starting uwsgi.service - LSB: Start/stop uWSGI server instance(s)... Apr 14 22:20:11 npc99a84aae6d34 uwsgi[25797]: Starting app server(s): uwsgi (omitted; missing conffile(s) in /etc/uwsgi/apps-enabled). Apr 14 22:20:11 npc99a84aae6d34 systemd[1]: Started uwsgi.service - LSB: Start/stop uWSGI server instance(s). ● virtlogd.service - libvirt logging daemon Loaded: loaded (/usr/lib/systemd/system/virtlogd.service; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:24:49 UTC; 27min ago Invocation: c2e915f95a65445faebe778f3d41a3ca TriggeredBy: ● virtlogd-admin.socket ● virtlogd.socket Docs: man:virtlogd(8) https://libvirt.org/ Main PID: 47685 (virtlogd) Tasks: 1 (limit: 9256) Memory: 1.2M (peak: 2.8M, swap: 1.9M, swap peak: 1.9M, zswap: 583.6K) CPU: 57ms CGroup: /system.slice/virtlogd.service └─47685 /usr/sbin/virtlogd Apr 14 22:24:49 npc99a84aae6d34 systemd[1]: Starting virtlogd.service - libvirt logging daemon... Apr 14 22:24:49 npc99a84aae6d34 systemd[1]: Started virtlogd.service - libvirt logging daemon. ○ xfs_scrub_all.service - Online XFS Metadata Check for All Filesystems Loaded: loaded (/usr/lib/systemd/system/xfs_scrub_all.service; static) Active: inactive (dead) TriggeredBy: ● xfs_scrub_all.timer Docs: man:xfs_scrub_all(8) ○ xfs_scrub_all_fail.service - Online XFS Metadata Check for All Filesystems Failure Reporting Loaded: loaded (/usr/lib/systemd/system/xfs_scrub_all_fail.service; static) Active: inactive (dead) Docs: man:xfs_scrub_all(8) ● -.slice - Root Slice Loaded: loaded Active: active since Tue 2026-04-14 22:07:33 UTC; 44min ago Docs: man:systemd.special(7) Tasks: 1307 Memory: 5.9G CPU: 32min 48.990s CGroup: / ├─init.scope │ └─1 /sbin/init nofb ├─system.slice │ ├─apache-htcacheclean.service │ │ └─15613 /usr/bin/htcacheclean -d 120 -p /var/cache/apache2/mod_cache_disk -l 300M -n │ ├─apache2.service │ │ ├─91335 /usr/sbin/apache2 -k start │ │ ├─91340 /usr/sbin/apache2 -k start │ │ └─91341 /usr/sbin/apache2 -k start │ ├─atd.service │ │ └─814 /usr/sbin/atd -f │ ├─cron.service │ │ └─805 /usr/sbin/cron -f │ ├─dbus.service │ │ └─589 /usr/bin/dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─epmd.service │ │ └─21889 /usr/bin/epmd -systemd │ ├─haproxy.service │ │ ├─15053 /usr/sbin/haproxy -Ws -f /etc/haproxy/haproxy.cfg -p /run/haproxy.pid -S /run/haproxy-master.sock │ │ └─15055 /usr/sbin/haproxy -Ws -f /etc/haproxy/haproxy.cfg -p /run/haproxy.pid -S /run/haproxy-master.sock │ ├─haveged.service │ │ └─432 /usr/sbin/haveged --Foreground --verbose=1 │ ├─iscsid.service │ │ ├─43268 /usr/sbin/iscsid │ │ └─43269 /usr/sbin/iscsid │ ├─ksmtuned.service │ │ ├─ 6269 /bin/bash /usr/sbin/ksmtuned │ │ └─121050 sleep 60 │ ├─libvirtd.service │ │ └─87645 /usr/sbin/libvirtd --timeout 120 │ ├─mariadb.service │ │ └─62213 /usr/sbin/mariadbd │ ├─memcached.service │ │ └─88828 /usr/bin/memcached -m 64 -p 11211 -u memcache -l 127.0.0.1 -l ::1 -P /var/run/memcached/memcached.pid │ ├─netavark-firewalld-reload.service │ │ └─13083 /usr/lib/podman/netavark firewalld-reload │ ├─ovn-controller-vtep.service │ │ └─93768 ovn-controller-vtep -vconsole:emer -vsyslog:err -vfile:info --vtep-db=/var/run/openvswitch/db.sock --ovnsb-db=/var/run/ovn/ovnsb_db.sock --no-chdir --log-file=/var/log/ovn/ovn-controller-vtep.log --pidfile=/var/run/ovn/ovn-controller-vtep.pid --detach │ ├─ovn-controller.service │ │ └─94497 ovn-controller unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --no-chdir --log-file=/var/log/ovn/ovn-controller.log --pidfile=/var/run/ovn/ovn-controller.pid --detach │ ├─ovn-northd.service │ │ └─94198 ovn-northd -vconsole:emer -vsyslog:err -vfile:info --ovnnb-db=unix:/var/run/ovn/ovnnb_db.sock --ovnsb-db=unix:/var/run/ovn/ovnsb_db.sock --no-chdir --log-file=/var/log/ovn/ovn-northd.log --pidfile=/var/run/ovn/ovn-northd.pid --detach │ ├─ovn-ovsdb-server-nb.service │ │ └─94125 ovsdb-server -vconsole:off -vfile:info --log-file=/var/log/ovn/ovsdb-server-nb.log --pidfile=/var/run/ovn/ovnnb_db.pid --remote=punix:/var/run/ovn/ovnnb_db.sock --unixctl=/var/run/ovn/ovnnb_db.ctl --remote=db:OVN_Northbound,NB_Global,connections --private-key=db:OVN_Northbound,SSL,private_key --certificate=db:OVN_Northbound,SSL,certificate --ca-cert=db:OVN_Northbound,SSL,ca_cert --ssl-protocols=db:OVN_Northbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Northbound,SSL,ssl_ciphers --ssl-ciphersuites=db:OVN_Northbound,SSL,ssl_ciphersuites /var/lib/ovn/ovnnb_db.db │ ├─ovn-ovsdb-server-sb.service │ │ └─94127 ovsdb-server -vconsole:off -vfile:info --log-file=/var/log/ovn/ovsdb-server-sb.log --pidfile=/var/run/ovn/ovnsb_db.pid --remote=punix:/var/run/ovn/ovnsb_db.sock --unixctl=/var/run/ovn/ovnsb_db.ctl --remote=db:OVN_Southbound,SB_Global,connections --private-key=db:OVN_Southbound,SSL,private_key --certificate=db:OVN_Southbound,SSL,certificate --ca-cert=db:OVN_Southbound,SSL,ca_cert --ssl-protocols=db:OVN_Southbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Southbound,SSL,ssl_ciphers --ssl-ciphersuites=db:OVN_Southbound,SSL,ssl_ciphersuites /var/lib/ovn/ovnsb_db.db │ ├─ovs-vswitchd.service │ │ └─93684 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach │ ├─ovsdb-server.service │ │ └─93449 ovsdb-server /etc/openvswitch/conf.db -vconsole:emer -vsyslog:err -vfile:info --remote=punix:/var/run/openvswitch/db.sock --private-key=db:Open_vSwitch,SSL,private_key --certificate=db:Open_vSwitch,SSL,certificate --bootstrap-ca-cert=db:Open_vSwitch,SSL,ca_cert --no-chdir --log-file=/var/log/openvswitch/ovsdb-server.log --pidfile=/var/run/openvswitch/ovsdb-server.pid --detach │ ├─polkit.service │ │ └─105627 /usr/lib/polkit-1/polkitd --no-debug --log-level=notice │ ├─rabbitmq-server.service │ │ ├─22008 /usr/lib/erlang/erts-15.2.7/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -- -root /usr/lib/erlang -bindir /usr/lib/erlang/erts-15.2.7/bin -progname erl -- -home /var/lib/rabbitmq -- -pa "" -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger "[]" -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -- │ │ ├─22018 erl_child_setup 65536 │ │ ├─22049 sh -s disksup │ │ ├─22051 /usr/lib/erlang/lib/os_mon-2.10.1/priv/bin/memsup │ │ ├─22052 /usr/lib/erlang/lib/os_mon-2.10.1/priv/bin/cpu_sup │ │ ├─22054 /usr/lib/erlang/erts-15.2.7/bin/inet_gethost 4 │ │ ├─22055 /usr/lib/erlang/erts-15.2.7/bin/inet_gethost 4 │ │ └─22058 /bin/sh -s rabbit_disk_monitor │ ├─rsync.service │ │ └─88855 /usr/bin/rsync --daemon --no-detach │ ├─rsyslog.service │ │ └─19101 /usr/sbin/rsyslogd -n -iNONE │ ├─ssh.service │ │ └─13390 "sshd: /usr/sbin/sshd -D [listener] 0 of 30-100 startups" │ ├─system-ceph\x2d929e05b1\x2d4e8c\x2d4c8f\x2db13f\x2d4521b44f10e1.slice │ │ ├─ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@crash.npc99a84aae6d34.service │ │ │ ├─libpod-payload-91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc │ │ │ │ ├─54984 /run/podman-init -- /usr/bin/ceph-crash -n client.crash.npc99a84aae6d34 │ │ │ │ └─54986 /usr/bin/python3 -s /usr/bin/ceph-crash -n client.crash.npc99a84aae6d34 │ │ │ └─runtime │ │ │ └─54981 /usr/bin/conmon --api-version 1 -c 91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc -u 91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc/userdata -p /run/containers/storage/overlay-containers/91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc/userdata/pidfile -n ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1-crash-npc99a84aae6d34 --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc --full-attach -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc/userdata/oci-log --conmon-pidfile /run/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@crash.npc99a84aae6d34.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg "" --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg --rm --exit-command-arg 91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc │ │ ├─ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mgr.npc99a84aae6d34.krobmi.service │ │ │ ├─libpod-payload-3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274 │ │ │ │ ├─55422 /run/podman-init -- /usr/bin/ceph-mgr -n mgr.npc99a84aae6d34.krobmi -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false │ │ │ │ └─55424 /usr/bin/ceph-mgr -n mgr.npc99a84aae6d34.krobmi -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false │ │ │ └─runtime │ │ │ └─55417 /usr/bin/conmon --api-version 1 -c 3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274 -u 3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274 -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274/userdata -p /run/containers/storage/overlay-containers/3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274/userdata/pidfile -n ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1-mgr-npc99a84aae6d34-krobmi --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274 --full-attach -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274/userdata/oci-log --conmon-pidfile /run/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mgr.npc99a84aae6d34.krobmi.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg "" --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg --rm --exit-command-arg 3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274 │ │ ├─ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mgr.npc99a84aae6d34.mxlymt.service │ │ │ ├─libpod-payload-c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73 │ │ │ │ ├─51960 /run/podman-init -- /usr/bin/ceph-mgr -n mgr.npc99a84aae6d34.mxlymt -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false │ │ │ │ └─51962 /usr/bin/ceph-mgr -n mgr.npc99a84aae6d34.mxlymt -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false │ │ │ └─runtime │ │ │ └─51958 /usr/bin/conmon --api-version 1 -c c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73 -u c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73 -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73/userdata -p /run/containers/storage/overlay-containers/c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73/userdata/pidfile -n ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1-mgr-npc99a84aae6d34-mxlymt --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73 --full-attach -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73/userdata/oci-log --conmon-pidfile /run/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mgr.npc99a84aae6d34.mxlymt.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg "" --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg --rm --exit-command-arg c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73 │ │ ├─ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mon.npc99a84aae6d34.service │ │ │ ├─libpod-payload-b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942 │ │ │ │ ├─51752 /run/podman-init -- /usr/bin/ceph-mon -n mon.npc99a84aae6d34 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false │ │ │ │ └─51754 /usr/bin/ceph-mon -n mon.npc99a84aae6d34 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false │ │ │ └─runtime │ │ │ └─51750 /usr/bin/conmon --api-version 1 -c b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942 -u b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942 -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942/userdata -p /run/containers/storage/overlay-containers/b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942/userdata/pidfile -n ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1-mon-npc99a84aae6d34 --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942 --full-attach -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942/userdata/oci-log --conmon-pidfile /run/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mon.npc99a84aae6d34.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg "" --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg --rm --exit-command-arg b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942 │ │ └─ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@osd.0.service │ │ ├─libpod-payload-45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee │ │ │ ├─57930 /run/podman-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --osd-objectstore=bluestore │ │ │ └─57932 /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --osd-objectstore=bluestore │ │ └─runtime │ │ └─57928 /usr/bin/conmon --api-version 1 -c 45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee -u 45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee/userdata -p /run/containers/storage/overlay-containers/45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee/userdata/pidfile -n ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1-osd-0 --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee --full-attach -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee/userdata/oci-log --conmon-pidfile /run/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@osd.0.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg "" --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg --rm --exit-command-arg 45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee │ ├─system-devstack.slice │ │ ├─devstack@br-ex-tcpdump.service │ │ │ └─110919 /bin/tcpdump -i br-ex arp or rarp or icmp or icmp6 -enlX │ │ ├─devstack@br-int-flows.service │ │ │ ├─111525 /bin/sh -c "set +e; while true; do echo ovs-ofctl dump-flows br-int; ovs-ofctl dump-flows br-int ; sleep 30; done; " │ │ │ └─122648 sleep 30 │ │ ├─devstack@c-api.service │ │ │ ├─105804 "cinder-apiuWSGI master" │ │ │ ├─105808 "cinder-apiuWSGI worker 1" │ │ │ └─105809 "cinder-apiuWSGI worker 2" │ │ ├─devstack@c-bak.service │ │ │ └─107200 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/cinder-backup --config-file /etc/cinder/cinder.conf │ │ ├─devstack@c-sch.service │ │ │ └─106585 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/cinder-scheduler --config-file /etc/cinder/cinder.conf │ │ ├─devstack@c-vol.service │ │ │ ├─107810 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/cinder-volume --config-file /etc/cinder/cinder.conf │ │ │ └─108126 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/cinder-volume --config-file /etc/cinder/cinder.conf │ │ ├─devstack@etcd.service │ │ │ └─64223 /opt/stack/bin/etcd --name npc99a84aae6d34 --data-dir /opt/stack/data/etcd --initial-cluster-state new --initial-cluster-token etcd-cluster-01 --initial-cluster "npc99a84aae6d34=http://[2001:41d0:302:1000::cbc]:2380" --initial-advertise-peer-urls "http://[2001:41d0:302:1000::cbc]:2380" --advertise-client-urls "http://[2001:41d0:302:1000::cbc]:2379" --listen-peer-urls http://0.0.0.0:2380 --listen-client-urls "http://[2001:41d0:302:1000::cbc]:2379" --log-level=debug │ │ ├─devstack@file_tracker.service │ │ │ ├─ 63498 /bin/bash /opt/stack/devstack/tools/file_tracker.sh │ │ │ └─122592 sleep 20 │ │ ├─devstack@g-api.service │ │ │ ├─108887 "glance-apiuWSGI master" │ │ │ ├─108888 "glance-apiuWSGI worker 1" │ │ │ └─108889 "glance-apiuWSGI worker 2" │ │ ├─devstack@keystone.service │ │ │ ├─65424 "keystoneuWSGI master" │ │ │ ├─65425 "keystoneuWSGI worker 1" │ │ │ └─65426 "keystoneuWSGI worker 2" │ │ ├─devstack@memory_tracker.service │ │ │ ├─ 62947 /bin/bash /opt/stack/devstack/tools/memory_tracker.sh │ │ │ └─122596 sleep 20 │ │ ├─devstack@n-api-meta.service │ │ │ ├─101619 "nova-api-metauWSGI master" │ │ │ ├─101621 "nova-api-metauWSGI worker 1" │ │ │ ├─101622 "nova-api-metauWSGI worker 2" │ │ │ └─101623 "nova-api-metauWSGI http 1" │ │ ├─devstack@n-api.service │ │ │ ├─92903 "nova-apiuWSGI master" │ │ │ ├─92904 "nova-apiuWSGI worker 1" │ │ │ └─92905 "nova-apiuWSGI worker 2" │ │ ├─devstack@n-cond-cell1.service │ │ │ ├─103735 "nova-conductor: master process [/opt/stack/data/venv/bin/nova-conductor --config-file /etc/nova/nova_cell1.conf]" │ │ │ ├─104902 "nova-conductor: ServiceWrapper worker(0)" │ │ │ └─104904 "nova-conductor: ServiceWrapper worker(1)" │ │ ├─devstack@n-cpu.service │ │ │ └─104923 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/nova-compute --config-file /etc/nova/nova-cpu.conf │ │ ├─devstack@n-novnc-cell1.service │ │ │ └─102318 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/nova-novncproxy --config-file /etc/nova/nova_cell1.conf --web /opt/stack/novnc │ │ ├─devstack@n-sch.service │ │ │ ├─100952 "nova-scheduler: master process [/opt/stack/data/venv/bin/nova-scheduler --config-file /etc/nova/nova.conf]" │ │ │ ├─102220 "nova-scheduler: ServiceWrapper worker(0)" │ │ │ └─102223 "nova-scheduler: ServiceWrapper worker(1)" │ │ ├─devstack@n-super-cond.service │ │ │ ├─103157 "nova-conductor: master process [/opt/stack/data/venv/bin/nova-conductor --config-file /etc/nova/nova.conf]" │ │ │ ├─104721 "nova-conductor: ServiceWrapper worker(0)" │ │ │ └─104725 "nova-conductor: ServiceWrapper worker(1)" │ │ ├─devstack@neutron-api.service │ │ │ ├─96269 "neutron-apiuWSGI master" │ │ │ ├─96270 "neutron-apiuWSGI worker 1" │ │ │ └─96271 "neutron-apiuWSGI worker 2" │ │ ├─devstack@neutron-ovn-maintenance-worker.service │ │ │ ├─97897 "neutron-ovn-maintenance-worker: master process [/opt/stack/data/venv/bin/neutron-ovn-maintenance-worker --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" │ │ │ └─98765 "neutron-server: maintenance worker (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-ovn-maintenance-worker --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ ├─devstack@neutron-periodic-workers.service │ │ │ ├─97349 "neutron-periodic-workers: master process [/opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" │ │ │ ├─98512 "neutron-server: Periodic worker for \"AgentSchedulerDbMixin\" (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ │ ├─98520 "neutron-server: Periodic worker for \"AgentSchedulerDbMixin\" (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ │ ├─98524 "neutron-server: Periodic worker for \"DbQuotaNoLockDriver\" (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ │ └─98532 "neutron-server: Periodic worker for \"L3_NAT_dbonly_mixin\" (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ ├─devstack@neutron-rpc-server.service │ │ │ ├─96802 "neutron-rpc-server: master process [/opt/stack/data/venv/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" │ │ │ └─98148 "neutron-server: rpc worker (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ ├─devstack@openstack-cli-server.service │ │ │ └─61256 /opt/stack/data/venv/bin/python3 /opt/stack/devstack/files/openstack-cli-server/openstack-cli-server │ │ ├─devstack@placement-api.service │ │ │ ├─98699 "placementuWSGI master" │ │ │ ├─98700 "placementuWSGI worker 1" │ │ │ └─98701 "placementuWSGI worker 2" │ │ ├─devstack@q-ovn-agent.service │ │ │ ├─95086 "neutron-ovn-agent: master process [/opt/stack/data/venv/bin/neutron-ovn-agent --config-file /etc/neutron/plugins/ml2/ovn_agent.ini]" │ │ │ ├─95991 "neutron-ovn-agent: ServiceWrapper worker(0)" │ │ │ ├─96262 /opt/stack/data/venv/bin/python3.13 /usr/local/bin/privsep-helper --config-file /etc/neutron/plugins/ml2/ovn_agent.ini --privsep_context neutron.privileged.namespace_cmd --privsep_sock_path /tmp/tmp38ne4qkg/privsep.sock │ │ │ └─99558 /opt/stack/data/venv/bin/python3.13 /usr/local/bin/privsep-helper --config-file /etc/neutron/plugins/ml2/ovn_agent.ini --privsep_context neutron.privileged.default --privsep_sock_path /tmp/tmpqd4w1pg4/privsep.sock │ │ ├─devstack@s-account.service │ │ │ ├─90424 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-account-server /etc/swift/account-server/1.conf -v │ │ │ ├─90719 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-account-server /etc/swift/account-server/1.conf -v │ │ │ └─90721 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-account-server /etc/swift/account-server/1.conf -v │ │ ├─devstack@s-container-sync.service │ │ │ └─91065 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-container-sync /etc/swift/container-server/1.conf │ │ ├─devstack@s-container.service │ │ │ ├─89898 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-container-server /etc/swift/container-server/1.conf -v │ │ │ ├─90188 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-container-server /etc/swift/container-server/1.conf -v │ │ │ └─90190 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-container-server /etc/swift/container-server/1.conf -v │ │ ├─devstack@s-object.service │ │ │ ├─89369 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-object-server /etc/swift/object-server/1.conf -v │ │ │ ├─89690 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-object-server /etc/swift/object-server/1.conf -v │ │ │ └─89691 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-object-server /etc/swift/object-server/1.conf -v │ │ └─devstack@s-proxy.service │ │ ├─91903 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-proxy-server /etc/swift/proxy-server.conf -v │ │ └─91982 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-proxy-server /etc/swift/proxy-server.conf -v │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─821 /sbin/agetty -o "-- \\u" --noreset --noclear - linux │ ├─system-glean.slice │ │ └─glean@ens3.service │ │ └─743 dhclient -4 -v -i -pf /run/dhclient.ens3.pid -lf /var/lib/dhcp/dhclient.ens3.leases -I -df /var/lib/dhcp/dhclient6.ens3.leases ens3 │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─823 /sbin/agetty -o "-- \\u" --noreset --noclear --keep-baud 115200,57600,38400,9600 - vt220 │ ├─systemd-journald.service │ │ └─20737 /usr/lib/systemd/systemd-journald │ ├─systemd-logind.service │ │ └─605 /usr/lib/systemd/systemd-logind │ ├─systemd-machined.service │ │ └─41980 /usr/lib/systemd/systemd-machined │ ├─systemd-timesyncd.service │ │ └─411 /usr/lib/systemd/systemd-timesyncd │ ├─systemd-udevd.service │ │ └─udev │ │ └─422 /usr/lib/systemd/systemd-udevd │ ├─unbound.service │ │ └─1203 /usr/sbin/unbound -d -p │ ├─uuidd.service │ │ └─96046 /usr/sbin/uuidd --socket-activation --cont-clock │ └─virtlogd.service │ └─47685 /usr/sbin/virtlogd └─user.slice └─user-1000.slice ├─session-1.scope │ └─1087 /usr/bin/python3.13 ├─session-34.scope │ ├─120787 "sshd-session: zuul [priv]" │ ├─120794 "sshd-session: zuul@notty" │ ├─122625 /bin/sh -c "sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-mjejnubreskapnsupqhyuesybxywnovx ; /usr/bin/python3.13' && sleep 0" │ ├─122626 sudo -H -S -n -u root /bin/sh -c "echo BECOME-SUCCESS-mjejnubreskapnsupqhyuesybxywnovx ; /usr/bin/python3.13" │ ├─122628 /bin/sh -c "echo BECOME-SUCCESS-mjejnubreskapnsupqhyuesybxywnovx ; /usr/bin/python3.13" │ ├─122629 /usr/bin/python3.13 │ ├─122630 /bin/bash -c "sudo iptables-save > /home/zuul/iptables.txt\n\n# NOTE(sfernand): Run 'df' with a 60s timeout to prevent hangs from\n# stale NFS mounts.\ntimeout -s 9 60s df -h > /home/zuul/df.txt || true\n# If 'df' times out, the mount output helps debug which NFS share\n# is unresponsive.\nmount > /home/zuul/mount.txt\n\nfor py_ver in 2 3; do\n if [[ \`which python\${py_ver}\` ]]; then\n python\${py_ver} -m pip freeze > /home/zuul/pip\${py_ver}-freeze.txt\n fi\ndone\n\nif [ \`command -v dpkg\` ]; then\n dpkg -l> /home/zuul/dpkg-l.txt\nfi\nif [ \`command -v rpm\` ]; then\n rpm -qa | sort > /home/zuul/rpm-qa.txt\nfi\n\n# Services status\nsudo systemctl status --all > services.txt 2>/dev/null\n\n# NOTE(kchamart) The 'audit.log' can be useful in cases when QEMU\n# failed to start due to denials from SELinux — useful for CentOS\n# and Fedora machines. For Ubuntu (which runs AppArmor), DevStack\n# already captures the contents of /var/log/kern.log (via\n# \`journalctl -t kernel\` redirected into syslog.txt.gz), which\n# contains AppArmor-related messages.\nif [ -f /var/log/audit/audit.log ] ; then\n sudo cp /var/log/audit/audit.log /home/zuul/audit.log &&\n chmod +r /home/zuul/audit.log;\nfi\n\n# gzip and save any coredumps in /var/core\nif [ -d /var/core ]; then\n sudo gzip -r /var/core\n sudo cp -r /var/core /home/zuul/\nfi\n\nsudo ss -lntup | grep ':53' > /home/zuul/listen53.txt\n\n# NOTE(andreaf) Service logs are already in logs/ thanks for the\n# export-devstack-journal log. Apache logs are under apache/ thans to the\n# apache-logs-conf role.\ngrep -i deprecat /home/zuul/logs/*.txt /home/zuul/apache/*.log | \\\n sed -r 's/[0-9]{1,2}\\:[0-9]{1,2}\\:[0-9]{1,2}\\.[0-9]{1,3}/ /g' | \\\n sed -r 's/[0-9]{1,2}\\:[0-9]{1,2}\\:[0-9]{1,2}/ /g' | \\\n sed -r 's/[0-9]{1,4}-[0-9]{1,2}-[0-9]{1,4}/ /g' |\n sed -r 's/\\[.*\\]/ /g' | \\\n sed -r 's/\\s[0-9]+\\s/ /g' | \\\n awk '{if (\$0 in seen) {seen[\$0]++} else {out[++n]=\$0;seen[\$0]=1}} END { for (i=1; i<=n; i++) print seen[out[i]]\" :: \" out[i] }' > /home/zuul/deprecations.log\n" │ ├─122644 sudo systemctl status --all │ └─122646 systemctl status --all └─user@1000.service ├─init.scope │ ├─859 /usr/lib/systemd/systemd --user │ └─862 "(sd-pam)" └─session.slice └─dbus.service └─12675 /usr/bin/dbus-daemon --session --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only Apr 14 22:39:10 npc99a84aae6d34 systemd[1]: Started session-30.scope - Session 30 of User zuul. Apr 14 22:39:11 npc99a84aae6d34 systemd[1]: session-30.scope: Deactivated successfully. Apr 14 22:39:13 npc99a84aae6d34 systemd[1]: Started session-31.scope - Session 31 of User zuul. Apr 14 22:39:14 npc99a84aae6d34 systemd[1]: session-31.scope: Deactivated successfully. Apr 14 22:50:12 npc99a84aae6d34 systemd[1]: Started session-32.scope - Session 32 of User zuul. Apr 14 22:50:34 npc99a84aae6d34 systemd[1]: session-32.scope: Deactivated successfully. Apr 14 22:50:34 npc99a84aae6d34 systemd[1]: session-32.scope: Consumed 11.856s CPU time, 123.4M memory peak. Apr 14 22:50:36 npc99a84aae6d34 systemd[1]: Started session-33.scope - Session 33 of User zuul. Apr 14 22:50:43 npc99a84aae6d34 systemd[1]: session-33.scope: Deactivated successfully. Apr 14 22:50:46 npc99a84aae6d34 systemd[1]: Started session-34.scope - Session 34 of User zuul. ● machine.slice - Virtual Machine and Container Slice Loaded: loaded (/usr/lib/systemd/system/machine.slice; static) Active: active since Tue 2026-04-14 22:24:05 UTC; 27min ago Invocation: decfb6e24bb34654af913b40ce2e9f5b Docs: man:systemd.special(7) Tasks: 0 Memory: 400K (peak: 192.7M, swap: 16K, swap peak: 24K, zswap: 7.3K) CPU: 59.569s CGroup: /machine.slice Apr 14 22:33:04 npc99a84aae6d34 distracted_bhabha[88488]: "ceph.vdo": "0", Apr 14 22:33:04 npc99a84aae6d34 distracted_bhabha[88488]: "ceph.with_tpm": "0" Apr 14 22:33:04 npc99a84aae6d34 distracted_bhabha[88488]: }, Apr 14 22:33:04 npc99a84aae6d34 distracted_bhabha[88488]: "type": "block", Apr 14 22:33:04 npc99a84aae6d34 distracted_bhabha[88488]: "vg_name": "ceph_vg" Apr 14 22:33:04 npc99a84aae6d34 distracted_bhabha[88488]: } Apr 14 22:33:04 npc99a84aae6d34 distracted_bhabha[88488]: ] Apr 14 22:33:04 npc99a84aae6d34 distracted_bhabha[88488]: } Apr 14 22:33:04 npc99a84aae6d34 podman[88740]: 2026-04-14 22:33:04.156606616 +0000 UTC m=+0.047876913 container died fa3f1b8b9cc83c8dba1abe6cc3c0415a9ff7fb0d0bfcf40f8c87931a2c218f9f (image=quay.io/ceph/ceph@sha256:0bae386bc859cd9a05b804d1ca16cca8853a64f90809044e2bf43095419dc337, name=distracted_bhabha, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.43.0, org.label-schema.schema-version=1.0, ceph=True, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260331, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph.git, CEPH_REF=tentacle, CEPH_SHA1=6a49aff47758778a5f5951e731d437c317f72fb2, org.label-schema.vendor=CentOS, OSD_FLAVOR=default, org.label-schema.license=GPLv2) Apr 14 22:33:04 npc99a84aae6d34 podman[88740]: 2026-04-14 22:33:04.198408581 +0000 UTC m=+0.089678880 container remove fa3f1b8b9cc83c8dba1abe6cc3c0415a9ff7fb0d0bfcf40f8c87931a2c218f9f (image=quay.io/ceph/ceph@sha256:0bae386bc859cd9a05b804d1ca16cca8853a64f90809044e2bf43095419dc337, name=distracted_bhabha, OSD_FLAVOR=default, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, FROM_IMAGE=quay.io/centos/centos:stream9, io.buildah.version=1.43.0, org.label-schema.build-date=20260331, org.opencontainers.image.documentation=https://docs.ceph.com/, ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph.git, org.label-schema.schema-version=1.0, CEPH_REF=tentacle, CEPH_SHA1=6a49aff47758778a5f5951e731d437c317f72fb2, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) ● system-ceph\x2d929e05b1\x2d4e8c\x2d4c8f\x2db13f\x2d4521b44f10e1.slice - Slice /system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1 Loaded: loaded Active: active since Tue 2026-04-14 22:25:51 UTC; 26min ago Invocation: a1079fea0de142fda262ce17560ff9f9 Tasks: 200 Memory: 1G (peak: 1.4G, swap: 148.1M, swap peak: 148.3M, zswap: 49.2M) CPU: 1min 43.405s CGroup: /system.slice/system-ceph\x2d929e05b1\x2d4e8c\x2d4c8f\x2db13f\x2d4521b44f10e1.slice ├─ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@crash.npc99a84aae6d34.service │ ├─libpod-payload-91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc │ │ ├─54984 /run/podman-init -- /usr/bin/ceph-crash -n client.crash.npc99a84aae6d34 │ │ └─54986 /usr/bin/python3 -s /usr/bin/ceph-crash -n client.crash.npc99a84aae6d34 │ └─runtime │ └─54981 /usr/bin/conmon --api-version 1 -c 91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc -u 91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc/userdata -p /run/containers/storage/overlay-containers/91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc/userdata/pidfile -n ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1-crash-npc99a84aae6d34 --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc --full-attach -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc/userdata/oci-log --conmon-pidfile /run/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@crash.npc99a84aae6d34.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg "" --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg --rm --exit-command-arg 91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc ├─ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mgr.npc99a84aae6d34.krobmi.service │ ├─libpod-payload-3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274 │ │ ├─55422 /run/podman-init -- /usr/bin/ceph-mgr -n mgr.npc99a84aae6d34.krobmi -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false │ │ └─55424 /usr/bin/ceph-mgr -n mgr.npc99a84aae6d34.krobmi -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false │ └─runtime │ └─55417 /usr/bin/conmon --api-version 1 -c 3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274 -u 3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274 -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274/userdata -p /run/containers/storage/overlay-containers/3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274/userdata/pidfile -n ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1-mgr-npc99a84aae6d34-krobmi --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274 --full-attach -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274/userdata/oci-log --conmon-pidfile /run/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mgr.npc99a84aae6d34.krobmi.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg "" --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg --rm --exit-command-arg 3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274 ├─ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mgr.npc99a84aae6d34.mxlymt.service │ ├─libpod-payload-c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73 │ │ ├─51960 /run/podman-init -- /usr/bin/ceph-mgr -n mgr.npc99a84aae6d34.mxlymt -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false │ │ └─51962 /usr/bin/ceph-mgr -n mgr.npc99a84aae6d34.mxlymt -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false │ └─runtime │ └─51958 /usr/bin/conmon --api-version 1 -c c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73 -u c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73 -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73/userdata -p /run/containers/storage/overlay-containers/c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73/userdata/pidfile -n ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1-mgr-npc99a84aae6d34-mxlymt --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73 --full-attach -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73/userdata/oci-log --conmon-pidfile /run/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mgr.npc99a84aae6d34.mxlymt.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg "" --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg --rm --exit-command-arg c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73 ├─ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mon.npc99a84aae6d34.service │ ├─libpod-payload-b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942 │ │ ├─51752 /run/podman-init -- /usr/bin/ceph-mon -n mon.npc99a84aae6d34 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false │ │ └─51754 /usr/bin/ceph-mon -n mon.npc99a84aae6d34 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false │ └─runtime │ └─51750 /usr/bin/conmon --api-version 1 -c b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942 -u b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942 -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942/userdata -p /run/containers/storage/overlay-containers/b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942/userdata/pidfile -n ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1-mon-npc99a84aae6d34 --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942 --full-attach -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942/userdata/oci-log --conmon-pidfile /run/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mon.npc99a84aae6d34.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg "" --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg --rm --exit-command-arg b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942 └─ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@osd.0.service ├─libpod-payload-45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee │ ├─57930 /run/podman-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --osd-objectstore=bluestore │ └─57932 /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --osd-objectstore=bluestore └─runtime └─57928 /usr/bin/conmon --api-version 1 -c 45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee -u 45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee/userdata -p /run/containers/storage/overlay-containers/45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee/userdata/pidfile -n ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1-osd-0 --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee --full-attach -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee/userdata/oci-log --conmon-pidfile /run/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@osd.0.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg "" --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg --rm --exit-command-arg 45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee Apr 14 22:51:48 npc99a84aae6d34 ceph-mgr[51962]: log_channel(cluster) log [DBG] : pgmap v574: 129 pgs: 129 active+clean; 62 MiB data, 88 MiB used, 30 GiB / 30 GiB avail Apr 14 22:51:50 npc99a84aae6d34 ceph-mon[51754]: pgmap v574: 129 pgs: 129 active+clean; 62 MiB data, 88 MiB used, 30 GiB / 30 GiB avail Apr 14 22:51:50 npc99a84aae6d34 ceph-mgr[51962]: log_channel(cluster) log [DBG] : pgmap v575: 129 pgs: 129 active+clean; 62 MiB data, 88 MiB used, 30 GiB / 30 GiB avail Apr 14 22:51:52 npc99a84aae6d34 ceph-mon[51754]: pgmap v575: 129 pgs: 129 active+clean; 62 MiB data, 88 MiB used, 30 GiB / 30 GiB avail Apr 14 22:51:52 npc99a84aae6d34 ceph-mon[51754]: mon.npc99a84aae6d34@0(leader).osd e40 _set_new_cache_sizes cache_size:1020054731 inc_alloc: 348127232 full_alloc: 348127232 kv_alloc: 322961408 Apr 14 22:51:52 npc99a84aae6d34 ceph-mgr[51962]: log_channel(cluster) log [DBG] : pgmap v576: 129 pgs: 129 active+clean; 62 MiB data, 88 MiB used, 30 GiB / 30 GiB avail Apr 14 22:51:54 npc99a84aae6d34 ceph-mon[51754]: pgmap v576: 129 pgs: 129 active+clean; 62 MiB data, 88 MiB used, 30 GiB / 30 GiB avail Apr 14 22:51:54 npc99a84aae6d34 ceph-mgr[51962]: log_channel(cluster) log [DBG] : pgmap v577: 129 pgs: 129 active+clean; 62 MiB data, 88 MiB used, 30 GiB / 30 GiB avail Apr 14 22:51:56 npc99a84aae6d34 ceph-mon[51754]: pgmap v577: 129 pgs: 129 active+clean; 62 MiB data, 88 MiB used, 30 GiB / 30 GiB avail Apr 14 22:51:56 npc99a84aae6d34 ceph-mgr[51962]: log_channel(cluster) log [DBG] : pgmap v578: 129 pgs: 129 active+clean; 62 MiB data, 88 MiB used, 30 GiB / 30 GiB avail ● system-devstack.slice - Slice /system/devstack Loaded: loaded Active: active since Tue 2026-04-14 22:27:27 UTC; 24min ago Invocation: bca2e92d53a44c7d8a1a962091d99546 Tasks: 571 Memory: 4G (peak: 4.2G, swap: 623.6M, swap peak: 630.1M, zswap: 177.9M) CPU: 7min 5.896s CGroup: /system.slice/system-devstack.slice ├─devstack@br-ex-tcpdump.service │ └─110919 /bin/tcpdump -i br-ex arp or rarp or icmp or icmp6 -enlX ├─devstack@br-int-flows.service │ ├─111525 /bin/sh -c "set +e; while true; do echo ovs-ofctl dump-flows br-int; ovs-ofctl dump-flows br-int ; sleep 30; done; " │ └─122648 sleep 30 ├─devstack@c-api.service │ ├─105804 "cinder-apiuWSGI master" │ ├─105808 "cinder-apiuWSGI worker 1" │ └─105809 "cinder-apiuWSGI worker 2" ├─devstack@c-bak.service │ └─107200 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/cinder-backup --config-file /etc/cinder/cinder.conf ├─devstack@c-sch.service │ └─106585 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/cinder-scheduler --config-file /etc/cinder/cinder.conf ├─devstack@c-vol.service │ ├─107810 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/cinder-volume --config-file /etc/cinder/cinder.conf │ └─108126 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/cinder-volume --config-file /etc/cinder/cinder.conf ├─devstack@etcd.service │ └─64223 /opt/stack/bin/etcd --name npc99a84aae6d34 --data-dir /opt/stack/data/etcd --initial-cluster-state new --initial-cluster-token etcd-cluster-01 --initial-cluster "npc99a84aae6d34=http://[2001:41d0:302:1000::cbc]:2380" --initial-advertise-peer-urls "http://[2001:41d0:302:1000::cbc]:2380" --advertise-client-urls "http://[2001:41d0:302:1000::cbc]:2379" --listen-peer-urls http://0.0.0.0:2380 --listen-client-urls "http://[2001:41d0:302:1000::cbc]:2379" --log-level=debug ├─devstack@file_tracker.service │ ├─ 63498 /bin/bash /opt/stack/devstack/tools/file_tracker.sh │ └─122592 sleep 20 ├─devstack@g-api.service │ ├─108887 "glance-apiuWSGI master" │ ├─108888 "glance-apiuWSGI worker 1" │ └─108889 "glance-apiuWSGI worker 2" ├─devstack@keystone.service │ ├─65424 "keystoneuWSGI master" │ ├─65425 "keystoneuWSGI worker 1" │ └─65426 "keystoneuWSGI worker 2" ├─devstack@memory_tracker.service │ ├─ 62947 /bin/bash /opt/stack/devstack/tools/memory_tracker.sh │ └─122596 sleep 20 ├─devstack@n-api-meta.service │ ├─101619 "nova-api-metauWSGI master" │ ├─101621 "nova-api-metauWSGI worker 1" │ ├─101622 "nova-api-metauWSGI worker 2" │ └─101623 "nova-api-metauWSGI http 1" ├─devstack@n-api.service │ ├─92903 "nova-apiuWSGI master" │ ├─92904 "nova-apiuWSGI worker 1" │ └─92905 "nova-apiuWSGI worker 2" ├─devstack@n-cond-cell1.service │ ├─103735 "nova-conductor: master process [/opt/stack/data/venv/bin/nova-conductor --config-file /etc/nova/nova_cell1.conf]" │ ├─104902 "nova-conductor: ServiceWrapper worker(0)" │ └─104904 "nova-conductor: ServiceWrapper worker(1)" ├─devstack@n-cpu.service │ └─104923 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/nova-compute --config-file /etc/nova/nova-cpu.conf ├─devstack@n-novnc-cell1.service │ └─102318 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/nova-novncproxy --config-file /etc/nova/nova_cell1.conf --web /opt/stack/novnc ├─devstack@n-sch.service │ ├─100952 "nova-scheduler: master process [/opt/stack/data/venv/bin/nova-scheduler --config-file /etc/nova/nova.conf]" │ ├─102220 "nova-scheduler: ServiceWrapper worker(0)" │ └─102223 "nova-scheduler: ServiceWrapper worker(1)" ├─devstack@n-super-cond.service │ ├─103157 "nova-conductor: master process [/opt/stack/data/venv/bin/nova-conductor --config-file /etc/nova/nova.conf]" │ ├─104721 "nova-conductor: ServiceWrapper worker(0)" │ └─104725 "nova-conductor: ServiceWrapper worker(1)" ├─devstack@neutron-api.service │ ├─96269 "neutron-apiuWSGI master" │ ├─96270 "neutron-apiuWSGI worker 1" │ └─96271 "neutron-apiuWSGI worker 2" ├─devstack@neutron-ovn-maintenance-worker.service │ ├─97897 "neutron-ovn-maintenance-worker: master process [/opt/stack/data/venv/bin/neutron-ovn-maintenance-worker --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" │ └─98765 "neutron-server: maintenance worker (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-ovn-maintenance-worker --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" ├─devstack@neutron-periodic-workers.service │ ├─97349 "neutron-periodic-workers: master process [/opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" │ ├─98512 "neutron-server: Periodic worker for \"AgentSchedulerDbMixin\" (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ ├─98520 "neutron-server: Periodic worker for \"AgentSchedulerDbMixin\" (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ ├─98524 "neutron-server: Periodic worker for \"DbQuotaNoLockDriver\" (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ └─98532 "neutron-server: Periodic worker for \"L3_NAT_dbonly_mixin\" (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" ├─devstack@neutron-rpc-server.service │ ├─96802 "neutron-rpc-server: master process [/opt/stack/data/venv/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" │ └─98148 "neutron-server: rpc worker (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" ├─devstack@openstack-cli-server.service │ └─61256 /opt/stack/data/venv/bin/python3 /opt/stack/devstack/files/openstack-cli-server/openstack-cli-server ├─devstack@placement-api.service │ ├─98699 "placementuWSGI master" │ ├─98700 "placementuWSGI worker 1" │ └─98701 "placementuWSGI worker 2" ├─devstack@q-ovn-agent.service │ ├─95086 "neutron-ovn-agent: master process [/opt/stack/data/venv/bin/neutron-ovn-agent --config-file /etc/neutron/plugins/ml2/ovn_agent.ini]" │ ├─95991 "neutron-ovn-agent: ServiceWrapper worker(0)" │ ├─96262 /opt/stack/data/venv/bin/python3.13 /usr/local/bin/privsep-helper --config-file /etc/neutron/plugins/ml2/ovn_agent.ini --privsep_context neutron.privileged.namespace_cmd --privsep_sock_path /tmp/tmp38ne4qkg/privsep.sock │ └─99558 /opt/stack/data/venv/bin/python3.13 /usr/local/bin/privsep-helper --config-file /etc/neutron/plugins/ml2/ovn_agent.ini --privsep_context neutron.privileged.default --privsep_sock_path /tmp/tmpqd4w1pg4/privsep.sock ├─devstack@s-account.service │ ├─90424 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-account-server /etc/swift/account-server/1.conf -v │ ├─90719 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-account-server /etc/swift/account-server/1.conf -v │ └─90721 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-account-server /etc/swift/account-server/1.conf -v ├─devstack@s-container-sync.service │ └─91065 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-container-sync /etc/swift/container-server/1.conf ├─devstack@s-container.service │ ├─89898 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-container-server /etc/swift/container-server/1.conf -v │ ├─90188 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-container-server /etc/swift/container-server/1.conf -v │ └─90190 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-container-server /etc/swift/container-server/1.conf -v ├─devstack@s-object.service │ ├─89369 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-object-server /etc/swift/object-server/1.conf -v │ ├─89690 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-object-server /etc/swift/object-server/1.conf -v │ └─89691 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-object-server /etc/swift/object-server/1.conf -v └─devstack@s-proxy.service ├─91903 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-proxy-server /etc/swift/proxy-server.conf -v └─91982 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-proxy-server /etc/swift/proxy-server.conf -v Apr 14 22:51:57 npc99a84aae6d34 sh[122647]: cookie=0xa59e6423, duration=1060.282s, table=65, n_packets=0, n_bytes=0, idle_age=1060, priority=100,reg15=0x1,metadata=0x2 actions=clone(ct_clear,load:0->NXM_NX_REG11[],load:0->NXM_NX_REG12[],load:0->NXM_NX_REG13[0..15],load:0x5->NXM_NX_REG11[],load:0x6->NXM_NX_REG12[],load:0x1->OXM_OF_METADATA[],load:0x2->NXM_NX_REG14[],load:0->NXM_NX_REG10[],load:0->NXM_NX_REG15[],load:0->NXM_NX_REG0[],load:0->NXM_NX_REG1[],load:0->NXM_NX_REG2[],load:0->NXM_NX_REG3[],load:0->NXM_NX_REG4[],load:0->NXM_NX_REG5[],load:0->NXM_NX_REG6[],load:0->NXM_NX_REG7[],load:0->NXM_NX_REG8[],load:0->NXM_NX_REG9[],resubmit(,8)) Apr 14 22:51:57 npc99a84aae6d34 sh[122647]: cookie=0xfba4b1cc, duration=1060.282s, table=65, n_packets=0, n_bytes=0, idle_age=1060, priority=100,reg15=0x2,metadata=0x2 actions=clone(ct_clear,load:0->NXM_NX_REG11[],load:0->NXM_NX_REG12[],load:0->NXM_NX_REG13[0..15],load:0x1->NXM_NX_REG11[],load:0x2->NXM_NX_REG12[],load:0x3->OXM_OF_METADATA[],load:0x3->NXM_NX_REG14[],load:0->NXM_NX_REG10[],load:0->NXM_NX_REG15[],load:0->NXM_NX_REG0[],load:0->NXM_NX_REG1[],load:0->NXM_NX_REG2[],load:0->NXM_NX_REG3[],load:0->NXM_NX_REG4[],load:0->NXM_NX_REG5[],load:0->NXM_NX_REG6[],load:0->NXM_NX_REG7[],load:0->NXM_NX_REG8[],load:0->NXM_NX_REG9[],resubmit(,8)) Apr 14 22:51:57 npc99a84aae6d34 sh[122647]: cookie=0x94247a9, duration=1060.254s, table=65, n_packets=5, n_bytes=210, idle_age=1044, priority=100,reg15=0x1,metadata=0x3 actions=output:1 Apr 14 22:51:57 npc99a84aae6d34 sh[122647]: cookie=0x595271ba, duration=1056.403s, table=65, n_packets=0, n_bytes=0, idle_age=1056, priority=100,reg15=0x4,metadata=0x2 actions=clone(ct_clear,load:0->NXM_NX_REG11[],load:0->NXM_NX_REG12[],load:0->NXM_NX_REG13[0..15],load:0x5->NXM_NX_REG11[],load:0x6->NXM_NX_REG12[],load:0x1->OXM_OF_METADATA[],load:0x3->NXM_NX_REG14[],load:0->NXM_NX_REG10[],load:0->NXM_NX_REG15[],load:0->NXM_NX_REG0[],load:0->NXM_NX_REG1[],load:0->NXM_NX_REG2[],load:0->NXM_NX_REG3[],load:0->NXM_NX_REG4[],load:0->NXM_NX_REG5[],load:0->NXM_NX_REG6[],load:0->NXM_NX_REG7[],load:0->NXM_NX_REG8[],load:0->NXM_NX_REG9[],resubmit(,8)) Apr 14 22:51:57 npc99a84aae6d34 sh[122647]: cookie=0xcc9b5b27, duration=1056.403s, table=65, n_packets=0, n_bytes=0, idle_age=1056, priority=100,reg15=0x3,metadata=0x1 actions=clone(ct_clear,load:0->NXM_NX_REG11[],load:0->NXM_NX_REG12[],load:0->NXM_NX_REG13[0..15],load:0x4->NXM_NX_REG11[],load:0x3->NXM_NX_REG12[],load:0x2->OXM_OF_METADATA[],load:0x4->NXM_NX_REG14[],load:0->NXM_NX_REG10[],load:0->NXM_NX_REG15[],load:0->NXM_NX_REG0[],load:0->NXM_NX_REG1[],load:0->NXM_NX_REG2[],load:0->NXM_NX_REG3[],load:0->NXM_NX_REG4[],load:0->NXM_NX_REG5[],load:0->NXM_NX_REG6[],load:0->NXM_NX_REG7[],load:0->NXM_NX_REG8[],load:0->NXM_NX_REG9[],resubmit(,8)) Apr 14 22:51:57 npc99a84aae6d34 sh[122647]: cookie=0x0, duration=1097.994s, table=65, n_packets=17, n_bytes=1482, idle_age=1044, priority=0 actions=drop Apr 14 22:51:57 npc99a84aae6d34 sh[122647]: cookie=0x0, duration=1097.994s, table=80, n_packets=0, n_bytes=0, idle_age=1097, priority=0 actions=resubmit(,8) Apr 14 22:51:57 npc99a84aae6d34 sh[122647]: cookie=0x0, duration=1097.994s, table=81, n_packets=0, n_bytes=0, idle_age=1097, priority=100,ct_state=+est+trk,ip actions=move:NXM_NX_CT_NW_DST[]->NXM_NX_REG1[] Apr 14 22:51:57 npc99a84aae6d34 sh[122647]: cookie=0x0, duration=1097.994s, table=82, n_packets=0, n_bytes=0, idle_age=1097, priority=100,ct_state=+est+trk,ipv6 actions=move:NXM_NX_CT_IPV6_DST[]->NXM_NX_XXREG1[] Apr 14 22:51:57 npc99a84aae6d34 sh[122647]: cookie=0x0, duration=1097.994s, table=83, n_packets=0, n_bytes=0, idle_age=1097, priority=100,ct_state=+est+trk actions=move:NXM_NX_CT_TP_DST[]->NXM_NX_REG2[0..15] ● system-getty.slice - Slice /system/getty Loaded: loaded Active: active since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 8695cee70a2a40b6b68b88e5e009a1f2 Tasks: 1 Memory: 216K (peak: 2M, swap: 212K, swap peak: 212K, zswap: 41.5K) CPU: 32ms CGroup: /system.slice/system-getty.slice └─getty@tty1.service └─821 /sbin/agetty -o "-- \\u" --noreset --noclear - linux Notice: journal has been rotated since unit was started, output may be incomplete. ● system-glean.slice - Slice /system/glean Loaded: loaded Active: active since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: f1b48cee78cc4b4d9988fca4e68daebf Tasks: 1 Memory: 1.1M (peak: 16.6M, swap: 32K, swap peak: 32K, zswap: 5.4K) CPU: 542ms CGroup: /system.slice/system-glean.slice └─glean@ens3.service └─743 dhclient -4 -v -i -pf /run/dhclient.ens3.pid -lf /var/lib/dhcp/dhclient.ens3.leases -I -df /var/lib/dhcp/dhclient6.ens3.leases ens3 Apr 14 22:07:35 npc99a84aae6d34 ifup[743]: DHCPREQUEST for 217.182.143.61 on ens3 to 255.255.255.255 port 67 Apr 14 22:07:35 npc99a84aae6d34 dhclient[743]: DHCPREQUEST for 217.182.143.61 on ens3 to 255.255.255.255 port 67 Apr 14 22:07:35 npc99a84aae6d34 dhclient[743]: DHCPACK of 217.182.143.61 from 217.182.140.1 Apr 14 22:07:35 npc99a84aae6d34 ifup[743]: DHCPACK of 217.182.143.61 from 217.182.140.1 Apr 14 22:07:35 npc99a84aae6d34 ifup[743]: suspect value in domain_search option - discarded Apr 14 22:07:35 npc99a84aae6d34 dhclient[743]: suspect value in domain_search option - discarded Apr 14 22:07:35 npc99a84aae6d34 dhclient[743]: Error printing text. Apr 14 22:07:35 npc99a84aae6d34 ifup[743]: Error printing text. Apr 14 22:07:35 npc99a84aae6d34 dhclient[743]: bound to 217.182.143.61 -- renewal in 40097 seconds. Apr 14 22:07:35 npc99a84aae6d34 ifup[743]: bound to 217.182.143.61 -- renewal in 40097 seconds. ● system-modprobe.slice - Slice /system/modprobe Loaded: loaded Active: active since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: a15c0411345345c1983490f3c1bd165b Tasks: 0 Memory: 24K (peak: 5.8M) CPU: 293ms CGroup: /system.slice/system-modprobe.slice Notice: journal has been rotated since unit was started, output may be incomplete. ● system-serial\x2dgetty.slice - Slice /system/serial-getty Loaded: loaded Active: active since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: e1d6cc1fd9bc404e9dcaa2ca336905da Tasks: 1 Memory: 292K (peak: 2M, swap: 192K, swap peak: 192K, zswap: 41K) CPU: 26ms CGroup: /system.slice/system-serial\x2dgetty.slice └─serial-getty@ttyS0.service └─823 /sbin/agetty -o "-- \\u" --noreset --noclear --keep-baud 115200,57600,38400,9600 - vt220 Notice: journal has been rotated since unit was started, output may be incomplete. ○ system-sshd.slice - Slice /system/sshd Loaded: loaded Active: inactive (dead) ● system-xfs_scrub.slice - xfs_scrub background service slice Loaded: loaded (/usr/lib/systemd/system/system-xfs_scrub.slice; enabled; preset: enabled) Active: active since Tue 2026-04-14 22:17:41 UTC; 34min ago Invocation: 5091ba4649754c09855c47cdbe0da674 Tasks: 0 Memory: 0B (peak: 0B) CPU: 0 CGroup: /system.slice/system-xfs_scrub.slice Apr 14 22:17:41 npc99a84aae6d34 systemd[1]: Created slice system-xfs_scrub.slice - xfs_scrub background service slice. ● system.slice - System Slice Loaded: loaded Active: active since Tue 2026-04-14 22:07:33 UTC; 44min ago Docs: man:systemd.special(7) Tasks: 1117 Memory: 5.9G (peak: 6.2G, swap: 872.6M, swap peak: 879.6M, zswap: 241.8M) CPU: 11min 24.904s CGroup: /system.slice ├─apache-htcacheclean.service │ └─15613 /usr/bin/htcacheclean -d 120 -p /var/cache/apache2/mod_cache_disk -l 300M -n ├─apache2.service │ ├─91335 /usr/sbin/apache2 -k start │ ├─91340 /usr/sbin/apache2 -k start │ └─91341 /usr/sbin/apache2 -k start ├─atd.service │ └─814 /usr/sbin/atd -f ├─cron.service │ └─805 /usr/sbin/cron -f ├─dbus.service │ └─589 /usr/bin/dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only ├─epmd.service │ └─21889 /usr/bin/epmd -systemd ├─haproxy.service │ ├─15053 /usr/sbin/haproxy -Ws -f /etc/haproxy/haproxy.cfg -p /run/haproxy.pid -S /run/haproxy-master.sock │ └─15055 /usr/sbin/haproxy -Ws -f /etc/haproxy/haproxy.cfg -p /run/haproxy.pid -S /run/haproxy-master.sock ├─haveged.service │ └─432 /usr/sbin/haveged --Foreground --verbose=1 ├─iscsid.service │ ├─43268 /usr/sbin/iscsid │ └─43269 /usr/sbin/iscsid ├─ksmtuned.service │ ├─ 6269 /bin/bash /usr/sbin/ksmtuned │ └─121050 sleep 60 ├─libvirtd.service │ └─87645 /usr/sbin/libvirtd --timeout 120 ├─mariadb.service │ └─62213 /usr/sbin/mariadbd ├─memcached.service │ └─88828 /usr/bin/memcached -m 64 -p 11211 -u memcache -l 127.0.0.1 -l ::1 -P /var/run/memcached/memcached.pid ├─netavark-firewalld-reload.service │ └─13083 /usr/lib/podman/netavark firewalld-reload ├─ovn-controller-vtep.service │ └─93768 ovn-controller-vtep -vconsole:emer -vsyslog:err -vfile:info --vtep-db=/var/run/openvswitch/db.sock --ovnsb-db=/var/run/ovn/ovnsb_db.sock --no-chdir --log-file=/var/log/ovn/ovn-controller-vtep.log --pidfile=/var/run/ovn/ovn-controller-vtep.pid --detach ├─ovn-controller.service │ └─94497 ovn-controller unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --no-chdir --log-file=/var/log/ovn/ovn-controller.log --pidfile=/var/run/ovn/ovn-controller.pid --detach ├─ovn-northd.service │ └─94198 ovn-northd -vconsole:emer -vsyslog:err -vfile:info --ovnnb-db=unix:/var/run/ovn/ovnnb_db.sock --ovnsb-db=unix:/var/run/ovn/ovnsb_db.sock --no-chdir --log-file=/var/log/ovn/ovn-northd.log --pidfile=/var/run/ovn/ovn-northd.pid --detach ├─ovn-ovsdb-server-nb.service │ └─94125 ovsdb-server -vconsole:off -vfile:info --log-file=/var/log/ovn/ovsdb-server-nb.log --pidfile=/var/run/ovn/ovnnb_db.pid --remote=punix:/var/run/ovn/ovnnb_db.sock --unixctl=/var/run/ovn/ovnnb_db.ctl --remote=db:OVN_Northbound,NB_Global,connections --private-key=db:OVN_Northbound,SSL,private_key --certificate=db:OVN_Northbound,SSL,certificate --ca-cert=db:OVN_Northbound,SSL,ca_cert --ssl-protocols=db:OVN_Northbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Northbound,SSL,ssl_ciphers --ssl-ciphersuites=db:OVN_Northbound,SSL,ssl_ciphersuites /var/lib/ovn/ovnnb_db.db ├─ovn-ovsdb-server-sb.service │ └─94127 ovsdb-server -vconsole:off -vfile:info --log-file=/var/log/ovn/ovsdb-server-sb.log --pidfile=/var/run/ovn/ovnsb_db.pid --remote=punix:/var/run/ovn/ovnsb_db.sock --unixctl=/var/run/ovn/ovnsb_db.ctl --remote=db:OVN_Southbound,SB_Global,connections --private-key=db:OVN_Southbound,SSL,private_key --certificate=db:OVN_Southbound,SSL,certificate --ca-cert=db:OVN_Southbound,SSL,ca_cert --ssl-protocols=db:OVN_Southbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Southbound,SSL,ssl_ciphers --ssl-ciphersuites=db:OVN_Southbound,SSL,ssl_ciphersuites /var/lib/ovn/ovnsb_db.db ├─ovs-vswitchd.service │ └─93684 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach ├─ovsdb-server.service │ └─93449 ovsdb-server /etc/openvswitch/conf.db -vconsole:emer -vsyslog:err -vfile:info --remote=punix:/var/run/openvswitch/db.sock --private-key=db:Open_vSwitch,SSL,private_key --certificate=db:Open_vSwitch,SSL,certificate --bootstrap-ca-cert=db:Open_vSwitch,SSL,ca_cert --no-chdir --log-file=/var/log/openvswitch/ovsdb-server.log --pidfile=/var/run/openvswitch/ovsdb-server.pid --detach ├─polkit.service │ └─105627 /usr/lib/polkit-1/polkitd --no-debug --log-level=notice ├─rabbitmq-server.service │ ├─22008 /usr/lib/erlang/erts-15.2.7/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -- -root /usr/lib/erlang -bindir /usr/lib/erlang/erts-15.2.7/bin -progname erl -- -home /var/lib/rabbitmq -- -pa "" -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger "[]" -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -- │ ├─22018 erl_child_setup 65536 │ ├─22049 sh -s disksup │ ├─22051 /usr/lib/erlang/lib/os_mon-2.10.1/priv/bin/memsup │ ├─22052 /usr/lib/erlang/lib/os_mon-2.10.1/priv/bin/cpu_sup │ ├─22054 /usr/lib/erlang/erts-15.2.7/bin/inet_gethost 4 │ ├─22055 /usr/lib/erlang/erts-15.2.7/bin/inet_gethost 4 │ └─22058 /bin/sh -s rabbit_disk_monitor ├─rsync.service │ └─88855 /usr/bin/rsync --daemon --no-detach ├─rsyslog.service │ └─19101 /usr/sbin/rsyslogd -n -iNONE ├─ssh.service │ └─13390 "sshd: /usr/sbin/sshd -D [listener] 0 of 30-100 startups" ├─system-ceph\x2d929e05b1\x2d4e8c\x2d4c8f\x2db13f\x2d4521b44f10e1.slice │ ├─ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@crash.npc99a84aae6d34.service │ │ ├─libpod-payload-91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc │ │ │ ├─54984 /run/podman-init -- /usr/bin/ceph-crash -n client.crash.npc99a84aae6d34 │ │ │ └─54986 /usr/bin/python3 -s /usr/bin/ceph-crash -n client.crash.npc99a84aae6d34 │ │ └─runtime │ │ └─54981 /usr/bin/conmon --api-version 1 -c 91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc -u 91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc/userdata -p /run/containers/storage/overlay-containers/91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc/userdata/pidfile -n ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1-crash-npc99a84aae6d34 --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc --full-attach -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc/userdata/oci-log --conmon-pidfile /run/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@crash.npc99a84aae6d34.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg "" --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg --rm --exit-command-arg 91d9b7fb809a783a31c9c29f49a745d174f2dccac8048221e8703bc91d39fabc │ ├─ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mgr.npc99a84aae6d34.krobmi.service │ │ ├─libpod-payload-3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274 │ │ │ ├─55422 /run/podman-init -- /usr/bin/ceph-mgr -n mgr.npc99a84aae6d34.krobmi -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false │ │ │ └─55424 /usr/bin/ceph-mgr -n mgr.npc99a84aae6d34.krobmi -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false │ │ └─runtime │ │ └─55417 /usr/bin/conmon --api-version 1 -c 3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274 -u 3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274 -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274/userdata -p /run/containers/storage/overlay-containers/3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274/userdata/pidfile -n ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1-mgr-npc99a84aae6d34-krobmi --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274 --full-attach -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274/userdata/oci-log --conmon-pidfile /run/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mgr.npc99a84aae6d34.krobmi.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg "" --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg --rm --exit-command-arg 3000dc89d3c7ad165b6563438b6b02737d6dd351a555d07021bf9bc74cfbe274 │ ├─ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mgr.npc99a84aae6d34.mxlymt.service │ │ ├─libpod-payload-c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73 │ │ │ ├─51960 /run/podman-init -- /usr/bin/ceph-mgr -n mgr.npc99a84aae6d34.mxlymt -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false │ │ │ └─51962 /usr/bin/ceph-mgr -n mgr.npc99a84aae6d34.mxlymt -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false │ │ └─runtime │ │ └─51958 /usr/bin/conmon --api-version 1 -c c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73 -u c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73 -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73/userdata -p /run/containers/storage/overlay-containers/c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73/userdata/pidfile -n ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1-mgr-npc99a84aae6d34-mxlymt --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73 --full-attach -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73/userdata/oci-log --conmon-pidfile /run/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mgr.npc99a84aae6d34.mxlymt.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg "" --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg --rm --exit-command-arg c38a9e87667ad07ddacba982e408b3f002065e99ece0fa902bb4c99930684b73 │ ├─ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mon.npc99a84aae6d34.service │ │ ├─libpod-payload-b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942 │ │ │ ├─51752 /run/podman-init -- /usr/bin/ceph-mon -n mon.npc99a84aae6d34 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false │ │ │ └─51754 /usr/bin/ceph-mon -n mon.npc99a84aae6d34 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false │ │ └─runtime │ │ └─51750 /usr/bin/conmon --api-version 1 -c b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942 -u b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942 -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942/userdata -p /run/containers/storage/overlay-containers/b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942/userdata/pidfile -n ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1-mon-npc99a84aae6d34 --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942 --full-attach -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942/userdata/oci-log --conmon-pidfile /run/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@mon.npc99a84aae6d34.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg "" --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg --rm --exit-command-arg b511208d952f6972779d0bb1d369fd9eaafc16c43f63e22697405aa27f7dd942 │ └─ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@osd.0.service │ ├─libpod-payload-45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee │ │ ├─57930 /run/podman-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --osd-objectstore=bluestore │ │ └─57932 /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --osd-objectstore=bluestore │ └─runtime │ └─57928 /usr/bin/conmon --api-version 1 -c 45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee -u 45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee/userdata -p /run/containers/storage/overlay-containers/45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee/userdata/pidfile -n ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1-osd-0 --exit-dir /run/libpod/exits --persist-dir /run/libpod/persist/45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee --full-attach -l journald --log-level warning --syslog --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee/userdata/oci-log --conmon-pidfile /run/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1@osd.0.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg "" --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg --rm --exit-command-arg 45dceece14449a710937e634757ffce23d35cc9119b73983f30ab85cb4e3b5ee ├─system-devstack.slice │ ├─devstack@br-ex-tcpdump.service │ │ └─110919 /bin/tcpdump -i br-ex arp or rarp or icmp or icmp6 -enlX │ ├─devstack@br-int-flows.service │ │ ├─111525 /bin/sh -c "set +e; while true; do echo ovs-ofctl dump-flows br-int; ovs-ofctl dump-flows br-int ; sleep 30; done; " │ │ └─122648 sleep 30 │ ├─devstack@c-api.service │ │ ├─105804 "cinder-apiuWSGI master" │ │ ├─105808 "cinder-apiuWSGI worker 1" │ │ └─105809 "cinder-apiuWSGI worker 2" │ ├─devstack@c-bak.service │ │ └─107200 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/cinder-backup --config-file /etc/cinder/cinder.conf │ ├─devstack@c-sch.service │ │ └─106585 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/cinder-scheduler --config-file /etc/cinder/cinder.conf │ ├─devstack@c-vol.service │ │ ├─107810 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/cinder-volume --config-file /etc/cinder/cinder.conf │ │ └─108126 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/cinder-volume --config-file /etc/cinder/cinder.conf │ ├─devstack@etcd.service │ │ └─64223 /opt/stack/bin/etcd --name npc99a84aae6d34 --data-dir /opt/stack/data/etcd --initial-cluster-state new --initial-cluster-token etcd-cluster-01 --initial-cluster "npc99a84aae6d34=http://[2001:41d0:302:1000::cbc]:2380" --initial-advertise-peer-urls "http://[2001:41d0:302:1000::cbc]:2380" --advertise-client-urls "http://[2001:41d0:302:1000::cbc]:2379" --listen-peer-urls http://0.0.0.0:2380 --listen-client-urls "http://[2001:41d0:302:1000::cbc]:2379" --log-level=debug │ ├─devstack@file_tracker.service │ │ ├─ 63498 /bin/bash /opt/stack/devstack/tools/file_tracker.sh │ │ └─122592 sleep 20 │ ├─devstack@g-api.service │ │ ├─108887 "glance-apiuWSGI master" │ │ ├─108888 "glance-apiuWSGI worker 1" │ │ └─108889 "glance-apiuWSGI worker 2" │ ├─devstack@keystone.service │ │ ├─65424 "keystoneuWSGI master" │ │ ├─65425 "keystoneuWSGI worker 1" │ │ └─65426 "keystoneuWSGI worker 2" │ ├─devstack@memory_tracker.service │ │ ├─ 62947 /bin/bash /opt/stack/devstack/tools/memory_tracker.sh │ │ └─122596 sleep 20 │ ├─devstack@n-api-meta.service │ │ ├─101619 "nova-api-metauWSGI master" │ │ ├─101621 "nova-api-metauWSGI worker 1" │ │ ├─101622 "nova-api-metauWSGI worker 2" │ │ └─101623 "nova-api-metauWSGI http 1" │ ├─devstack@n-api.service │ │ ├─92903 "nova-apiuWSGI master" │ │ ├─92904 "nova-apiuWSGI worker 1" │ │ └─92905 "nova-apiuWSGI worker 2" │ ├─devstack@n-cond-cell1.service │ │ ├─103735 "nova-conductor: master process [/opt/stack/data/venv/bin/nova-conductor --config-file /etc/nova/nova_cell1.conf]" │ │ ├─104902 "nova-conductor: ServiceWrapper worker(0)" │ │ └─104904 "nova-conductor: ServiceWrapper worker(1)" │ ├─devstack@n-cpu.service │ │ └─104923 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/nova-compute --config-file /etc/nova/nova-cpu.conf │ ├─devstack@n-novnc-cell1.service │ │ └─102318 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/nova-novncproxy --config-file /etc/nova/nova_cell1.conf --web /opt/stack/novnc │ ├─devstack@n-sch.service │ │ ├─100952 "nova-scheduler: master process [/opt/stack/data/venv/bin/nova-scheduler --config-file /etc/nova/nova.conf]" │ │ ├─102220 "nova-scheduler: ServiceWrapper worker(0)" │ │ └─102223 "nova-scheduler: ServiceWrapper worker(1)" │ ├─devstack@n-super-cond.service │ │ ├─103157 "nova-conductor: master process [/opt/stack/data/venv/bin/nova-conductor --config-file /etc/nova/nova.conf]" │ │ ├─104721 "nova-conductor: ServiceWrapper worker(0)" │ │ └─104725 "nova-conductor: ServiceWrapper worker(1)" │ ├─devstack@neutron-api.service │ │ ├─96269 "neutron-apiuWSGI master" │ │ ├─96270 "neutron-apiuWSGI worker 1" │ │ └─96271 "neutron-apiuWSGI worker 2" │ ├─devstack@neutron-ovn-maintenance-worker.service │ │ ├─97897 "neutron-ovn-maintenance-worker: master process [/opt/stack/data/venv/bin/neutron-ovn-maintenance-worker --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" │ │ └─98765 "neutron-server: maintenance worker (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-ovn-maintenance-worker --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ ├─devstack@neutron-periodic-workers.service │ │ ├─97349 "neutron-periodic-workers: master process [/opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" │ │ ├─98512 "neutron-server: Periodic worker for \"AgentSchedulerDbMixin\" (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ ├─98520 "neutron-server: Periodic worker for \"AgentSchedulerDbMixin\" (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ ├─98524 "neutron-server: Periodic worker for \"DbQuotaNoLockDriver\" (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ └─98532 "neutron-server: Periodic worker for \"L3_NAT_dbonly_mixin\" (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ ├─devstack@neutron-rpc-server.service │ │ ├─96802 "neutron-rpc-server: master process [/opt/stack/data/venv/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" │ │ └─98148 "neutron-server: rpc worker (/opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ ├─devstack@openstack-cli-server.service │ │ └─61256 /opt/stack/data/venv/bin/python3 /opt/stack/devstack/files/openstack-cli-server/openstack-cli-server │ ├─devstack@placement-api.service │ │ ├─98699 "placementuWSGI master" │ │ ├─98700 "placementuWSGI worker 1" │ │ └─98701 "placementuWSGI worker 2" │ ├─devstack@q-ovn-agent.service │ │ ├─95086 "neutron-ovn-agent: master process [/opt/stack/data/venv/bin/neutron-ovn-agent --config-file /etc/neutron/plugins/ml2/ovn_agent.ini]" │ │ ├─95991 "neutron-ovn-agent: ServiceWrapper worker(0)" │ │ ├─96262 /opt/stack/data/venv/bin/python3.13 /usr/local/bin/privsep-helper --config-file /etc/neutron/plugins/ml2/ovn_agent.ini --privsep_context neutron.privileged.namespace_cmd --privsep_sock_path /tmp/tmp38ne4qkg/privsep.sock │ │ └─99558 /opt/stack/data/venv/bin/python3.13 /usr/local/bin/privsep-helper --config-file /etc/neutron/plugins/ml2/ovn_agent.ini --privsep_context neutron.privileged.default --privsep_sock_path /tmp/tmpqd4w1pg4/privsep.sock │ ├─devstack@s-account.service │ │ ├─90424 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-account-server /etc/swift/account-server/1.conf -v │ │ ├─90719 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-account-server /etc/swift/account-server/1.conf -v │ │ └─90721 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-account-server /etc/swift/account-server/1.conf -v │ ├─devstack@s-container-sync.service │ │ └─91065 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-container-sync /etc/swift/container-server/1.conf │ ├─devstack@s-container.service │ │ ├─89898 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-container-server /etc/swift/container-server/1.conf -v │ │ ├─90188 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-container-server /etc/swift/container-server/1.conf -v │ │ └─90190 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-container-server /etc/swift/container-server/1.conf -v │ ├─devstack@s-object.service │ │ ├─89369 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-object-server /etc/swift/object-server/1.conf -v │ │ ├─89690 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-object-server /etc/swift/object-server/1.conf -v │ │ └─89691 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-object-server /etc/swift/object-server/1.conf -v │ └─devstack@s-proxy.service │ ├─91903 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-proxy-server /etc/swift/proxy-server.conf -v │ └─91982 /opt/stack/data/venv/bin/python3.13 /opt/stack/data/venv/bin/swift-proxy-server /etc/swift/proxy-server.conf -v ├─system-getty.slice │ └─getty@tty1.service │ └─821 /sbin/agetty -o "-- \\u" --noreset --noclear - linux ├─system-glean.slice │ └─glean@ens3.service │ └─743 dhclient -4 -v -i -pf /run/dhclient.ens3.pid -lf /var/lib/dhcp/dhclient.ens3.leases -I -df /var/lib/dhcp/dhclient6.ens3.leases ens3 ├─system-serial\x2dgetty.slice │ └─serial-getty@ttyS0.service │ └─823 /sbin/agetty -o "-- \\u" --noreset --noclear --keep-baud 115200,57600,38400,9600 - vt220 ├─systemd-journald.service │ └─20737 /usr/lib/systemd/systemd-journald ├─systemd-logind.service │ └─605 /usr/lib/systemd/systemd-logind ├─systemd-machined.service │ └─41980 /usr/lib/systemd/systemd-machined ├─systemd-timesyncd.service │ └─411 /usr/lib/systemd/systemd-timesyncd ├─systemd-udevd.service │ └─udev │ └─422 /usr/lib/systemd/systemd-udevd ├─unbound.service │ └─1203 /usr/sbin/unbound -d -p ├─uuidd.service │ └─96046 /usr/sbin/uuidd --socket-activation --cont-clock └─virtlogd.service └─47685 /usr/sbin/virtlogd Apr 14 22:51:56 npc99a84aae6d34 ovsdb-server[94127]: ovs|05060|reconnect|DBG|ssl:[2001:41d0:302:1000::cbc]:37994: entering ACTIVE Apr 14 22:51:56 npc99a84aae6d34 ovsdb-server[94127]: ovs|05061|poll_loop|DBG|wakeup due to 0-ms timeout at ../lib/stream-ssl.c:837 (0% CPU usage) Apr 14 22:51:57 npc99a84aae6d34 ovsdb-server[94127]: ovs|05062|poll_loop|DBG|wakeup due to [POLLIN] on fd 21 ([2001:41d0:302:1000::cbc]:6642<->[2001:41d0:302:1000::cbc]:50690) at ../lib/stream-ssl.c:835 (0% CPU usage) Apr 14 22:51:57 npc99a84aae6d34 ovsdb-server[94127]: ovs|05063|stream_ssl|DBG|server0<--ssl:[2001:41d0:302:1000::cbc]:50690 type 256 (5 bytes) Apr 14 22:51:57 npc99a84aae6d34 ovsdb-server[94127]: ovs|05064|stream_ssl|DBG|server0<--ssl:[2001:41d0:302:1000::cbc]:50690 type 257 (1 bytes) Apr 14 22:51:57 npc99a84aae6d34 ovsdb-server[94127]: ovs|05065|jsonrpc|DBG|ssl:[2001:41d0:302:1000::cbc]:50690: received request, method="echo", params=[], id="echo" Apr 14 22:51:57 npc99a84aae6d34 ovsdb-server[94127]: ovs|05066|jsonrpc|DBG|ssl:[2001:41d0:302:1000::cbc]:50690: send reply, result=[], id="echo" Apr 14 22:51:57 npc99a84aae6d34 ovsdb-server[94127]: ovs|05067|stream_ssl|DBG|server0-->ssl:[2001:41d0:302:1000::cbc]:50690 type 256 (5 bytes) Apr 14 22:51:57 npc99a84aae6d34 ovsdb-server[94127]: ovs|05068|stream_ssl|DBG|server0-->ssl:[2001:41d0:302:1000::cbc]:50690 type 257 (1 bytes) Apr 14 22:51:57 npc99a84aae6d34 ovsdb-server[94127]: ovs|05069|poll_loop|DBG|wakeup due to 0-ms timeout at ../lib/stream-ssl.c:837 (0% CPU usage) ● user-1000.slice - User Slice of UID 1000 Loaded: loaded Drop-In: /usr/lib/systemd/system/user-.slice.d └─10-defaults.conf Active: active since Tue 2026-04-14 22:09:41 UTC; 42min ago Invocation: f7782b286003405ab0ef0ac37303fef0 Docs: man:user@.service(5) Tasks: 15 (limit: 20363) Memory: 1.1G (peak: 7.1G, swap: 1M, swap peak: 1.1M, zswap: 253.3K) CPU: 17min 42.914s CGroup: /user.slice/user-1000.slice ├─session-1.scope │ └─1087 /usr/bin/python3.13 ├─session-34.scope │ ├─120787 "sshd-session: zuul [priv]" │ ├─120794 "sshd-session: zuul@notty" │ ├─122625 /bin/sh -c "sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-mjejnubreskapnsupqhyuesybxywnovx ; /usr/bin/python3.13' && sleep 0" │ ├─122626 sudo -H -S -n -u root /bin/sh -c "echo BECOME-SUCCESS-mjejnubreskapnsupqhyuesybxywnovx ; /usr/bin/python3.13" │ ├─122628 /bin/sh -c "echo BECOME-SUCCESS-mjejnubreskapnsupqhyuesybxywnovx ; /usr/bin/python3.13" │ ├─122629 /usr/bin/python3.13 │ ├─122630 /bin/bash -c "sudo iptables-save > /home/zuul/iptables.txt\n\n# NOTE(sfernand): Run 'df' with a 60s timeout to prevent hangs from\n# stale NFS mounts.\ntimeout -s 9 60s df -h > /home/zuul/df.txt || true\n# If 'df' times out, the mount output helps debug which NFS share\n# is unresponsive.\nmount > /home/zuul/mount.txt\n\nfor py_ver in 2 3; do\n if [[ \`which python\${py_ver}\` ]]; then\n python\${py_ver} -m pip freeze > /home/zuul/pip\${py_ver}-freeze.txt\n fi\ndone\n\nif [ \`command -v dpkg\` ]; then\n dpkg -l> /home/zuul/dpkg-l.txt\nfi\nif [ \`command -v rpm\` ]; then\n rpm -qa | sort > /home/zuul/rpm-qa.txt\nfi\n\n# Services status\nsudo systemctl status --all > services.txt 2>/dev/null\n\n# NOTE(kchamart) The 'audit.log' can be useful in cases when QEMU\n# failed to start due to denials from SELinux — useful for CentOS\n# and Fedora machines. For Ubuntu (which runs AppArmor), DevStack\n# already captures the contents of /var/log/kern.log (via\n# \`journalctl -t kernel\` redirected into syslog.txt.gz), which\n# contains AppArmor-related messages.\nif [ -f /var/log/audit/audit.log ] ; then\n sudo cp /var/log/audit/audit.log /home/zuul/audit.log &&\n chmod +r /home/zuul/audit.log;\nfi\n\n# gzip and save any coredumps in /var/core\nif [ -d /var/core ]; then\n sudo gzip -r /var/core\n sudo cp -r /var/core /home/zuul/\nfi\n\nsudo ss -lntup | grep ':53' > /home/zuul/listen53.txt\n\n# NOTE(andreaf) Service logs are already in logs/ thanks for the\n# export-devstack-journal log. Apache logs are under apache/ thans to the\n# apache-logs-conf role.\ngrep -i deprecat /home/zuul/logs/*.txt /home/zuul/apache/*.log | \\\n sed -r 's/[0-9]{1,2}\\:[0-9]{1,2}\\:[0-9]{1,2}\\.[0-9]{1,3}/ /g' | \\\n sed -r 's/[0-9]{1,2}\\:[0-9]{1,2}\\:[0-9]{1,2}/ /g' | \\\n sed -r 's/[0-9]{1,4}-[0-9]{1,2}-[0-9]{1,4}/ /g' |\n sed -r 's/\\[.*\\]/ /g' | \\\n sed -r 's/\\s[0-9]+\\s/ /g' | \\\n awk '{if (\$0 in seen) {seen[\$0]++} else {out[++n]=\$0;seen[\$0]=1}} END { for (i=1; i<=n; i++) print seen[out[i]]\" :: \" out[i] }' > /home/zuul/deprecations.log\n" │ ├─122644 sudo systemctl status --all │ └─122646 systemctl status --all └─user@1000.service ├─init.scope │ ├─859 /usr/lib/systemd/systemd --user │ └─862 "(sd-pam)" └─session.slice └─dbus.service └─12675 /usr/bin/dbus-daemon --session --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only Apr 14 22:51:53 npc99a84aae6d34 python3.13[122619]: ansible-ansible.legacy.command Invoked with _raw_params=cp -pRL /etc/openstack /home/zuul/etc/ zuul_no_log=False zuul_log_id=bc764e01-af26-4e37-8285-000000000030-1-controller zuul_output_max_bytes=1073741824 zuul_ansible_split_streams=False _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Apr 14 22:51:53 npc99a84aae6d34 sudo[122616]: pam_unix(sudo:session): session closed for user root Apr 14 22:51:54 npc99a84aae6d34 sudo[122626]: zuul : PWD=/home/zuul ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-mjejnubreskapnsupqhyuesybxywnovx ; /usr/bin/python3.13' Apr 14 22:51:54 npc99a84aae6d34 sudo[122626]: pam_unix(sudo:session): session opened for user root(uid=0) by zuul(uid=1000) Apr 14 22:51:55 npc99a84aae6d34 python3.13[122629]: ansible-ansible.legacy.command Invoked with executable=/bin/bash _raw_params=sudo iptables-save > /home/zuul/iptables.txt # NOTE(sfernand): Run 'df' with a 60s timeout to prevent hangs from # stale NFS mounts. timeout -s 9 60s df -h > /home/zuul/df.txt || true # If 'df' times out, the mount output helps debug which NFS share # is unresponsive. mount > /home/zuul/mount.txt for py_ver in 2 3; do if [[ `which python${py_ver}` ]]; then python${py_ver} -m pip freeze > /home/zuul/pip${py_ver}-freeze.txt fi done if [ `command -v dpkg` ]; then dpkg -l> /home/zuul/dpkg-l.txt fi if [ `command -v rpm` ]; then rpm -qa | sort > /home/zuul/rpm-qa.txt fi # Services status sudo systemctl status --all > services.txt 2>/dev/null # NOTE(kchamart) The 'audit.log' can be useful in cases when QEMU # failed to start due to denials from SELinux — useful for CentOS # and Fedora machines. For Ubuntu (which runs AppArmor), DevStack # already captures the contents of /var/log/kern.log (via # `journalctl -t kernel` redirected into syslog.txt.gz), which # contains AppArmor-related messages. if [ -f /var/log/audit/audit.log ] ; then sudo cp /var/log/audit/audit.log /home/zuul/audit.log && chmod +r /home/zuul/audit.log; fi # gzip and save any coredumps in /var/core if [ -d /var/core ]; then sudo gzip -r /var/core sudo cp -r /var/core /home/zuul/ fi sudo ss -lntup | grep ':53' > /home/zuul/listen53.txt # NOTE(andreaf) Service logs are already in logs/ thanks for the # export-devstack-journal log. Apache logs are under apache/ thans to the # apache-logs-conf role. grep -i deprecat /home/zuul/logs/*.txt /home/zuul/apache/*.log | \ sed -r 's/[0-9]{1,2}\:[0-9]{1,2}\:[0-9]{1,2}\.[0-9]{1,3}/ /g' | \ sed -r 's/[0-9]{1,2}\:[0-9]{1,2}\:[0-9]{1,2}/ /g' | \ sed -r 's/[0-9]{1,4}-[0-9]{1,2}-[0-9]{1,4}/ /g' | sed -r 's/\[.*\]/ /g' | \ sed -r 's/\s[0-9]+\s/ /g' | \ awk '{if ($0 in seen) {seen[$0]++} else {out[++n]=$0;seen[$0]=1}} END { for (i=1; i<=n; i++) print seen[out[i]]" :: " out[i] }' > /home/zuul/deprecations.log _uses_shell=True zuul_no_log=False zuul_log_id=bc764e01-af26-4e37-8285-000000000034-1-controller zuul_output_max_bytes=1073741824 zuul_ansible_split_streams=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None creates=None removes=None stdin=None Apr 14 22:51:55 npc99a84aae6d34 sudo[122632]: root : PWD=/home/zuul ; USER=root ; COMMAND=/usr/sbin/iptables-save Apr 14 22:51:55 npc99a84aae6d34 sudo[122632]: pam_unix(sudo:session): session opened for user root(uid=0) by zuul(uid=0) Apr 14 22:51:55 npc99a84aae6d34 sudo[122632]: pam_unix(sudo:session): session closed for user root Apr 14 22:51:55 npc99a84aae6d34 sudo[122644]: root : PWD=/home/zuul ; USER=root ; COMMAND=/usr/bin/systemctl status --all Apr 14 22:51:55 npc99a84aae6d34 sudo[122644]: pam_unix(sudo:session): session opened for user root(uid=0) by zuul(uid=0) ● user.slice - User and Session Slice Loaded: loaded (/usr/lib/systemd/system/user.slice; static) Active: active since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 13ebf72bfd764fa2bf717ea215a89558 Docs: man:systemd.special(7) Tasks: 15 Memory: 1.1G (peak: 7.1G, swap: 1.1M, swap peak: 1.1M, zswap: 253.3K) CPU: 18min 43.994s CGroup: /user.slice └─user-1000.slice ├─session-1.scope │ └─1087 /usr/bin/python3.13 ├─session-34.scope │ ├─120787 "sshd-session: zuul [priv]" │ ├─120794 "sshd-session: zuul@notty" │ ├─122625 /bin/sh -c "sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-mjejnubreskapnsupqhyuesybxywnovx ; /usr/bin/python3.13' && sleep 0" │ ├─122626 sudo -H -S -n -u root /bin/sh -c "echo BECOME-SUCCESS-mjejnubreskapnsupqhyuesybxywnovx ; /usr/bin/python3.13" │ ├─122628 /bin/sh -c "echo BECOME-SUCCESS-mjejnubreskapnsupqhyuesybxywnovx ; /usr/bin/python3.13" │ ├─122629 /usr/bin/python3.13 │ ├─122630 /bin/bash -c "sudo iptables-save > /home/zuul/iptables.txt\n\n# NOTE(sfernand): Run 'df' with a 60s timeout to prevent hangs from\n# stale NFS mounts.\ntimeout -s 9 60s df -h > /home/zuul/df.txt || true\n# If 'df' times out, the mount output helps debug which NFS share\n# is unresponsive.\nmount > /home/zuul/mount.txt\n\nfor py_ver in 2 3; do\n if [[ \`which python\${py_ver}\` ]]; then\n python\${py_ver} -m pip freeze > /home/zuul/pip\${py_ver}-freeze.txt\n fi\ndone\n\nif [ \`command -v dpkg\` ]; then\n dpkg -l> /home/zuul/dpkg-l.txt\nfi\nif [ \`command -v rpm\` ]; then\n rpm -qa | sort > /home/zuul/rpm-qa.txt\nfi\n\n# Services status\nsudo systemctl status --all > services.txt 2>/dev/null\n\n# NOTE(kchamart) The 'audit.log' can be useful in cases when QEMU\n# failed to start due to denials from SELinux — useful for CentOS\n# and Fedora machines. For Ubuntu (which runs AppArmor), DevStack\n# already captures the contents of /var/log/kern.log (via\n# \`journalctl -t kernel\` redirected into syslog.txt.gz), which\n# contains AppArmor-related messages.\nif [ -f /var/log/audit/audit.log ] ; then\n sudo cp /var/log/audit/audit.log /home/zuul/audit.log &&\n chmod +r /home/zuul/audit.log;\nfi\n\n# gzip and save any coredumps in /var/core\nif [ -d /var/core ]; then\n sudo gzip -r /var/core\n sudo cp -r /var/core /home/zuul/\nfi\n\nsudo ss -lntup | grep ':53' > /home/zuul/listen53.txt\n\n# NOTE(andreaf) Service logs are already in logs/ thanks for the\n# export-devstack-journal log. Apache logs are under apache/ thans to the\n# apache-logs-conf role.\ngrep -i deprecat /home/zuul/logs/*.txt /home/zuul/apache/*.log | \\\n sed -r 's/[0-9]{1,2}\\:[0-9]{1,2}\\:[0-9]{1,2}\\.[0-9]{1,3}/ /g' | \\\n sed -r 's/[0-9]{1,2}\\:[0-9]{1,2}\\:[0-9]{1,2}/ /g' | \\\n sed -r 's/[0-9]{1,4}-[0-9]{1,2}-[0-9]{1,4}/ /g' |\n sed -r 's/\\[.*\\]/ /g' | \\\n sed -r 's/\\s[0-9]+\\s/ /g' | \\\n awk '{if (\$0 in seen) {seen[\$0]++} else {out[++n]=\$0;seen[\$0]=1}} END { for (i=1; i<=n; i++) print seen[out[i]]\" :: \" out[i] }' > /home/zuul/deprecations.log\n" │ ├─122644 sudo systemctl status --all │ └─122646 systemctl status --all └─user@1000.service ├─init.scope │ ├─859 /usr/lib/systemd/systemd --user │ └─862 "(sd-pam)" └─session.slice └─dbus.service └─12675 /usr/bin/dbus-daemon --session --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only Notice: journal has been rotated since unit was started, output may be incomplete. ● acpid.socket - ACPID Listen Socket Loaded: loaded (/usr/lib/systemd/system/acpid.socket; enabled; preset: enabled) Active: active (listening) since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: 3b73820805914bceb004241462937618 Triggers: ● acpid.service Listen: /run/acpid.socket (Stream) CGroup: /system.slice/acpid.socket Apr 14 22:07:35 debian systemd[1]: Listening on acpid.socket - ACPID Listen Socket. ● dbus.socket - D-Bus System Message Bus Socket Loaded: loaded (/usr/lib/systemd/system/dbus.socket; static) Active: active (running) since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: 1c088f2ae6fa4edb97bc30362feceb13 Triggers: ● dbus.service Listen: /run/dbus/system_bus_socket (Stream) CGroup: /system.slice/dbus.socket Apr 14 22:07:35 debian systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. ● dm-event.socket - Device-mapper event daemon FIFOs Loaded: loaded (/usr/lib/systemd/system/dm-event.socket; enabled; preset: enabled) Active: active (listening) since Tue 2026-04-14 22:17:57 UTC; 34min ago Invocation: 428c45622399434db4c63ff20f8309c2 Triggers: ● dm-event.service Docs: man:dmeventd(8) Listen: /run/dmeventd-server (FIFO) /run/dmeventd-client (FIFO) CGroup: /system.slice/dm-event.socket Apr 14 22:17:57 npc99a84aae6d34 systemd[1]: Listening on dm-event.socket - Device-mapper event daemon FIFOs. ● epmd.socket - Erlang Port Mapper Daemon Activation Socket Loaded: loaded (/usr/lib/systemd/system/epmd.socket; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:19:26 UTC; 32min ago Invocation: 64f0d89cf67348ae9143953ac58ef30f Triggers: ● epmd.service Listen: [::]:4369 (Stream) Tasks: 0 (limit: 9256) Memory: 8K (peak: 256K) CPU: 703us CGroup: /system.slice/epmd.socket Apr 14 22:19:26 npc99a84aae6d34 systemd[1]: Listening on epmd.socket - Erlang Port Mapper Daemon Activation Socket. ● iscsid.socket - Open-iSCSI iscsid Socket Loaded: loaded (/usr/lib/systemd/system/iscsid.socket; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: 2478634303bb495d9fe1e37d71a0f40d Triggers: ● iscsid.service Docs: man:iscsid(8) man:iscsiadm(8) Listen: @ISCSIADM_ABSTRACT_NAMESPACE (Stream) CGroup: /system.slice/iscsid.socket Apr 14 22:07:35 debian systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. ● libvirtd-admin.socket - libvirt legacy monolithic daemon admin socket Loaded: loaded (/usr/lib/systemd/system/libvirtd-admin.socket; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:24:05 UTC; 27min ago Invocation: bef8444c297447e49779d368a978d380 Triggers: ● libvirtd.service Listen: /run/libvirt/libvirt-admin-sock (Stream) CGroup: /system.slice/libvirtd-admin.socket Apr 14 22:24:05 npc99a84aae6d34 systemd[1]: Listening on libvirtd-admin.socket - libvirt legacy monolithic daemon admin socket. ● libvirtd-ro.socket - libvirt legacy monolithic daemon read-only socket Loaded: loaded (/usr/lib/systemd/system/libvirtd-ro.socket; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:24:05 UTC; 27min ago Invocation: e717e2dea2bb4a77b0c9812a5ba1586a Triggers: ● libvirtd.service Listen: /run/libvirt/libvirt-sock-ro (Stream) CGroup: /system.slice/libvirtd-ro.socket Apr 14 22:24:05 npc99a84aae6d34 systemd[1]: Listening on libvirtd-ro.socket - libvirt legacy monolithic daemon read-only socket. ● libvirtd.socket - libvirt legacy monolithic daemon socket Loaded: loaded (/usr/lib/systemd/system/libvirtd.socket; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:24:05 UTC; 27min ago Invocation: b30b648b12934269b01f6ac4d86eb2c8 Triggers: ● libvirtd.service Listen: /run/libvirt/libvirt-sock (Stream) CGroup: /system.slice/libvirtd.socket Apr 14 22:24:05 npc99a84aae6d34 systemd[1]: Listening on libvirtd.socket - libvirt legacy monolithic daemon socket. ● lvm2-lvmpolld.socket - LVM2 poll daemon socket Loaded: loaded (/usr/lib/systemd/system/lvm2-lvmpolld.socket; enabled; preset: enabled) Active: active (listening) since Tue 2026-04-14 22:18:00 UTC; 33min ago Invocation: 855f83a6b46548a688f6b57d18106c96 Triggers: ● lvm2-lvmpolld.service Docs: man:lvmpolld(8) Listen: /run/lvm/lvmpolld.socket (Stream) CGroup: /system.slice/lvm2-lvmpolld.socket Apr 14 22:18:00 npc99a84aae6d34 systemd[1]: Listening on lvm2-lvmpolld.socket - LVM2 poll daemon socket. ● netavark-dhcp-proxy.socket - Netavark DHCP proxy socket Loaded: loaded (/usr/lib/systemd/system/netavark-dhcp-proxy.socket; enabled; preset: enabled) Active: active (listening) since Tue 2026-04-14 22:17:25 UTC; 34min ago Invocation: 3ed62878463c4f13b318bf5de7b8c352 Triggers: ● netavark-dhcp-proxy.service Listen: /run/podman/nv-proxy.sock (Stream) CGroup: /system.slice/netavark-dhcp-proxy.socket Apr 14 22:17:25 npc99a84aae6d34 systemd[1]: Listening on netavark-dhcp-proxy.socket - Netavark DHCP proxy socket. ● sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local) Loaded: loaded (/run/systemd/generator/sshd-unix-local.socket; generated) Active: active (listening) since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: ead1be8e519745bdac273066b2b0fbbd Docs: man:systemd-ssh-generator(8) Listen: /run/ssh-unix-local/socket (Stream) Accepted: 0; Connected: 0; CGroup: /system.slice/sshd-unix-local.socket Apr 14 22:07:35 debian systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). ● sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK) Loaded: loaded (/run/systemd/generator/sshd-vsock.socket; generated) Active: active (listening) since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: f53cd1f9df9f454789d96868035e9c3a Docs: man:systemd-ssh-generator(8) Listen: vsock::22 (Stream) Accepted: 0; Connected: 0; CGroup: /system.slice/sshd-vsock.socket Apr 14 22:07:35 debian systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). ● syslog.socket - Syslog Socket Loaded: loaded (/usr/lib/systemd/system/syslog.socket; static) Active: active (running) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 6b8de91952194b2b9f918819135eac38 Triggers: ● rsyslog.service Docs: man:systemd.special(7) https://systemd.io/SYSLOG Listen: /run/systemd/journal/syslog (Datagram) CGroup: /system.slice/syslog.socket Notice: journal has been rotated since unit was started, output may be incomplete. ● systemd-coredump.socket - Process Core Dump Socket Loaded: loaded (/usr/lib/systemd/system/systemd-coredump.socket; static) Active: active (listening) since Tue 2026-04-14 22:23:55 UTC; 28min ago Invocation: 2eccda478e3140af88981d20ba84e4a6 Docs: man:systemd-coredump(8) Listen: /run/systemd/coredump (SequentialPacket) Accepted: 0; Connected: 0; CGroup: /system.slice/systemd-coredump.socket Apr 14 22:23:55 npc99a84aae6d34 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. ● systemd-creds.socket - Credential Encryption/Decryption Loaded: loaded (/usr/lib/systemd/system/systemd-creds.socket; static) Active: active (listening) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 0ba9decb3b3e448f8e12543c493d177c Docs: man:systemd-creds(1) Listen: /run/systemd/io.systemd.Credentials (Stream) Accepted: 0; Connected: 0; CGroup: /system.slice/systemd-creds.socket Notice: journal has been rotated since unit was started, output may be incomplete. ● systemd-hostnamed.socket - Hostname Service Socket Loaded: loaded (/usr/lib/systemd/system/systemd-hostnamed.socket; static) Active: active (listening) since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: 61d6bbb417df4368a93070b9de52eec8 Triggers: ● systemd-hostnamed.service Docs: man:systemd-hostnamed.service(8) man:hostname(5) man:machine-info(5) Listen: /run/systemd/io.systemd.Hostname (Stream) CGroup: /system.slice/systemd-hostnamed.socket Apr 14 22:07:35 debian systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. ○ systemd-importd.socket - Disk Image Download Service Socket Loaded: loaded (/usr/lib/systemd/system/systemd-importd.socket; static) Active: inactive (dead) Triggers: ● systemd-importd.service Docs: man:systemd-importd.service(8) man:org.freedesktop.import1(5) Listen: /run/systemd/io.systemd.Import (Stream) ● systemd-initctl.socket - initctl Compatibility Named Pipe Loaded: loaded (/usr/lib/systemd/system/systemd-initctl.socket; static) Active: active (listening) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: f83bebd123744c109989f963a2248abf Triggers: ● systemd-initctl.service Docs: man:systemd-initctl.socket(8) Listen: /run/initctl (FIFO) CGroup: /system.slice/systemd-initctl.socket Notice: journal has been rotated since unit was started, output may be incomplete. ○ systemd-journald-audit.socket - Journal Audit Socket Loaded: loaded (/usr/lib/systemd/system/systemd-journald-audit.socket; disabled; preset: enabled) Active: inactive (dead) Triggers: ● systemd-journald.service Docs: man:systemd-journald.service(8) man:journald.conf(5) Listen: audit 1 (Netlink) ● systemd-journald-dev-log.socket - Journal Socket (/dev/log) Loaded: loaded (/usr/lib/systemd/system/systemd-journald-dev-log.socket; static) Active: active (running) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 8b8b286dad744ca197be77b3895d9fb3 Triggers: ● systemd-journald.service Docs: man:systemd-journald.service(8) man:journald.conf(5) Listen: /run/systemd/journal/dev-log (Datagram) CGroup: /system.slice/systemd-journald-dev-log.socket Notice: journal has been rotated since unit was started, output may be incomplete. ● systemd-journald.socket - Journal Sockets Loaded: loaded (/usr/lib/systemd/system/systemd-journald.socket; static) Active: active (running) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: a5e6c61fdce343d1868b2e0e3fd133ae Triggers: ● systemd-journald.service Docs: man:systemd-journald.service(8) man:journald.conf(5) Listen: /run/systemd/journal/socket (Datagram) /run/systemd/journal/stdout (Stream) CGroup: /system.slice/systemd-journald.socket Notice: journal has been rotated since unit was started, output may be incomplete. ○ systemd-networkd.socket - Network Service Netlink Socket Loaded: loaded (/usr/lib/systemd/system/systemd-networkd.socket; disabled; preset: enabled) Active: inactive (dead) Triggers: ● systemd-networkd.service Docs: man:systemd-networkd.service(8) man:rtnetlink(7) Listen: route 1361 (Netlink) ○ systemd-pcrextend.socket - TPM PCR Measurements Loaded: loaded (/usr/lib/systemd/system/systemd-pcrextend.socket; disabled; preset: enabled) Active: inactive (dead) Condition: start condition unmet at Tue 2026-04-14 22:07:33 UTC; 44min ago Docs: man:systemd-pcrextend(8) Listen: /run/systemd/io.systemd.PCRExtend (Stream) Accepted: 0; Connected: 0; ○ systemd-pcrlock.socket - Make TPM PCR Policy Loaded: loaded (/usr/lib/systemd/system/systemd-pcrlock.socket; disabled; preset: enabled) Active: inactive (dead) Condition: start condition unmet at Tue 2026-04-14 22:07:33 UTC; 44min ago Docs: man:systemd-pcrlock(8) Listen: /run/systemd/io.systemd.PCRLock (Stream) Accepted: 0; Connected: 0; ● systemd-sysext.socket - System Extension Image Management Loaded: loaded (/usr/lib/systemd/system/systemd-sysext.socket; disabled; preset: enabled) Active: active (listening) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 2fabe6e5dec8450383d8a2271d0e2b17 Docs: man:systemd-sysext(8) Listen: /run/systemd/io.systemd.sysext (Stream) Accepted: 0; Connected: 0; CGroup: /system.slice/systemd-sysext.socket Apr 14 22:07:33 debian systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. ● systemd-udevd-control.socket - udev Control Socket Loaded: loaded (/usr/lib/systemd/system/systemd-udevd-control.socket; static) Active: active (running) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: e50e5382289b446d8ba132773dba2498 Triggers: ● systemd-udevd.service Docs: man:systemd-udevd-control.socket(8) man:udev(7) Listen: /run/udev/control (SequentialPacket) CGroup: /system.slice/systemd-udevd-control.socket Notice: journal has been rotated since unit was started, output may be incomplete. ● systemd-udevd-kernel.socket - udev Kernel Socket Loaded: loaded (/usr/lib/systemd/system/systemd-udevd-kernel.socket; static) Active: active (running) since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 8533f6bb2aff468c8ca833016bda55f4 Triggers: ● systemd-udevd.service Docs: man:systemd-udevd-kernel.socket(8) man:udev(7) Listen: kobject-uevent 1 (Netlink) CGroup: /system.slice/systemd-udevd-kernel.socket Notice: journal has been rotated since unit was started, output may be incomplete. ● uuidd.socket - UUID daemon activation socket Loaded: loaded (/usr/lib/systemd/system/uuidd.socket; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: 36d200dbacbb4342a611479d2d1d1b18 Triggers: ● uuidd.service Listen: /run/uuidd/request (Stream) CGroup: /system.slice/uuidd.socket Apr 14 22:07:35 debian systemd[1]: Listening on uuidd.socket - UUID daemon activation socket. ● virtlogd-admin.socket - libvirt logging daemon admin socket Loaded: loaded (/usr/lib/systemd/system/virtlogd-admin.socket; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:23:58 UTC; 27min ago Invocation: 0abc8cdda23a40cb9c1105682de46e35 Triggers: ● virtlogd.service Listen: /run/libvirt/virtlogd-admin-sock (Stream) CGroup: /system.slice/virtlogd-admin.socket Apr 14 22:23:58 npc99a84aae6d34 systemd[1]: Listening on virtlogd-admin.socket - libvirt logging daemon admin socket. ● virtlogd.socket - libvirt logging daemon socket Loaded: loaded (/usr/lib/systemd/system/virtlogd.socket; enabled; preset: enabled) Active: active (running) since Tue 2026-04-14 22:23:58 UTC; 27min ago Invocation: 6e7491464008473aa93d2183813dc184 Triggers: ● virtlogd.service Listen: /run/libvirt/virtlogd-sock (Stream) CGroup: /system.slice/virtlogd.socket Apr 14 22:23:58 npc99a84aae6d34 systemd[1]: Listening on virtlogd.socket - libvirt logging daemon socket. ● root-swapfile.swap - /root/swapfile Loaded: loaded (/etc/fstab; generated) Active: active since Tue 2026-04-14 22:14:39 UTC; 37min ago Invocation: 94309083e0b84688ad8e337c00ac2cc5 What: /root/swapfile Docs: man:fstab(5) man:systemd-fstab-generator(8) ● basic.target - Basic System Loaded: loaded (/usr/lib/systemd/system/basic.target; static) Active: active since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: e1669ccad2da47f1ba7d5f6f92ad68db Docs: man:systemd.special(7) Apr 14 22:07:35 debian systemd[1]: Reached target basic.target - Basic System. ○ blockdev@dev-disk-by\x2dlabel-cloudimg\x2drootfs.target - Block Device Preparation for /dev/disk/by-label/cloudimg-rootfs Loaded: loaded (/usr/lib/systemd/system/blockdev@.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ blockdev@dev-loop1.target - Block Device Preparation for /dev/loop1 Loaded: loaded (/usr/lib/systemd/system/blockdev@.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ blockdev@dev-sr0.target - Block Device Preparation for /dev/sr0 Loaded: loaded (/usr/lib/systemd/system/blockdev@.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ blockdev@dev-vda1.target - Block Device Preparation for /dev/vda1 Loaded: loaded (/usr/lib/systemd/system/blockdev@.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ● ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1.target - Ceph cluster 929e05b1-4e8c-4c8f-b13f-4521b44f10e1 Loaded: loaded (/etc/systemd/system/ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1.target; enabled; preset: enabled) Active: active since Tue 2026-04-14 22:25:51 UTC; 26min ago Invocation: 945a74cc5eee4cc4bdbeff058ccfd793 Apr 14 22:25:51 npc99a84aae6d34 systemd[1]: Reached target ceph-929e05b1-4e8c-4c8f-b13f-4521b44f10e1.target - Ceph cluster 929e05b1-4e8c-4c8f-b13f-4521b44f10e1. ● ceph.target - All Ceph clusters and services Loaded: loaded (/etc/systemd/system/ceph.target; enabled; preset: enabled) Active: active since Tue 2026-04-14 22:17:56 UTC; 34min ago Invocation: b0a9dc2726084255994bbdc6fa6fe5b0 Apr 14 22:17:56 npc99a84aae6d34 systemd[1]: Reached target ceph.target - ceph target allowing to start/stop all ceph*@.service instances at once. ○ cryptsetup-pre.target - Local Encrypted Volumes (Pre) Loaded: loaded (/usr/lib/systemd/system/cryptsetup-pre.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ● cryptsetup.target - Local Encrypted Volumes Loaded: loaded (/usr/lib/systemd/system/cryptsetup.target; static) Active: active since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 16c4a1fc24eb475292dbafdc467a91b9 Docs: man:systemd.special(7) Notice: journal has been rotated since unit was started, output may be incomplete. ○ emergency.target - Emergency Mode Loaded: loaded (/usr/lib/systemd/system/emergency.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ final.target - Late Shutdown Services Loaded: loaded (/usr/lib/systemd/system/final.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ● first-boot-complete.target - First Boot Complete Loaded: loaded (/usr/lib/systemd/system/first-boot-complete.target; static) Active: active since Tue 2026-04-14 22:07:34 UTC; 44min ago Invocation: 1d8d2d17edb84e62ae2a64bd3cc19e00 Docs: man:systemd.special(7) Apr 14 22:07:34 debian systemd[1]: Reached target first-boot-complete.target - First Boot Complete. ○ getty-pre.target - Preparation for Logins Loaded: loaded (/usr/lib/systemd/system/getty-pre.target; static) Active: inactive (dead) Docs: man:systemd.special(7) man:systemd-getty-generator(8) https://0pointer.de/blog/projects/serial-console.html ● getty.target - Login Prompts Loaded: loaded (/usr/lib/systemd/system/getty.target; static) Active: active since Tue 2026-04-14 22:07:36 UTC; 44min ago Invocation: 2612fd4738074b8a8cddbeb285843e65 Docs: man:systemd.special(7) man:systemd-getty-generator(8) https://0pointer.de/blog/projects/serial-console.html Apr 14 22:07:36 npc99a84aae6d34 systemd[1]: Reached target getty.target - Login Prompts. ● graphical.target - Graphical Interface Loaded: loaded (/usr/lib/systemd/system/graphical.target; static) Active: active since Tue 2026-04-14 22:07:36 UTC; 44min ago Invocation: c90e11a8a17d42e5bee93603b89a176b Docs: man:systemd.special(7) Apr 14 22:07:36 npc99a84aae6d34 systemd[1]: Reached target graphical.target - Graphical Interface. ○ hibernate.target - System Hibernation Loaded: loaded (/usr/lib/systemd/system/hibernate.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ hybrid-sleep.target - Hybrid Suspend+Hibernate Loaded: loaded (/usr/lib/systemd/system/hybrid-sleep.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ initrd-fs.target - Initrd File Systems Loaded: loaded (/usr/lib/systemd/system/initrd-fs.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ initrd-root-device.target - Initrd Root Device Loaded: loaded (/usr/lib/systemd/system/initrd-root-device.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ initrd-root-fs.target - Initrd Root File System Loaded: loaded (/usr/lib/systemd/system/initrd-root-fs.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ initrd-switch-root.target - Switch Root Loaded: loaded (/usr/lib/systemd/system/initrd-switch-root.target; static) Active: inactive (dead) ○ initrd-usr-fs.target - Initrd /usr File System Loaded: loaded (/usr/lib/systemd/system/initrd-usr-fs.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ initrd.target - Initrd Default Target Loaded: loaded (/usr/lib/systemd/system/initrd.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ● integritysetup.target - Local Integrity Protected Volumes Loaded: loaded (/usr/lib/systemd/system/integritysetup.target; static) Active: active since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: b488cd9bacbe4540b4a99864c1760fdc Docs: man:systemd.special(7) Notice: journal has been rotated since unit was started, output may be incomplete. ● local-fs-pre.target - Preparation for Local File Systems Loaded: loaded (/usr/lib/systemd/system/local-fs-pre.target; static) Active: active since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: f39b187175114a4c80a4ff0200c12e80 Docs: man:systemd.special(7) Apr 14 22:07:33 debian systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. ● local-fs.target - Local File Systems Loaded: loaded (/usr/lib/systemd/system/local-fs.target; static) Active: active since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: c3029dfacc5745d98ffd3f4466faf9ba Docs: man:systemd.special(7) Apr 14 22:07:33 debian systemd[1]: Reached target local-fs.target - Local File Systems. ● machines.target - Containers Loaded: loaded (/usr/lib/systemd/system/machines.target; enabled; preset: enabled) Active: active since Tue 2026-04-14 22:23:56 UTC; 28min ago Invocation: a56cfb7d77d8486484ac4837155fe358 Docs: man:systemd.special(7) Apr 14 22:23:56 npc99a84aae6d34 systemd[1]: Reached target machines.target - Containers. ● multi-user.target - Multi-User System Loaded: loaded (/usr/lib/systemd/system/multi-user.target; static) Active: active since Tue 2026-04-14 22:07:36 UTC; 44min ago Invocation: 2caa43e6ee4a4e96afaad832cd6c0e0b Docs: man:systemd.special(7) Apr 14 22:07:36 npc99a84aae6d34 systemd[1]: Reached target multi-user.target - Multi-User System. ● network-online.target - Network is Online Loaded: loaded (/usr/lib/systemd/system/network-online.target; static) Active: active since Tue 2026-04-14 22:07:36 UTC; 44min ago Invocation: 2bd647d284cb42d18e3cd641d5d42d0d Docs: man:systemd.special(7) https://systemd.io/NETWORK_ONLINE Apr 14 22:07:36 npc99a84aae6d34 systemd[1]: Reached target network-online.target - Network is Online. ● network-pre.target - Preparation for Network Loaded: loaded (/usr/lib/systemd/system/network-pre.target; static) Active: active since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: 397c75dbe1824a8197e47ff32139e36f Docs: man:systemd.special(7) https://systemd.io/NETWORK_ONLINE Apr 14 22:07:35 npc99a84aae6d34 systemd[1]: Reached target network-pre.target - Preparation for Network. ● network.target - Network Loaded: loaded (/usr/lib/systemd/system/network.target; static) Active: active since Tue 2026-04-14 22:07:36 UTC; 44min ago Invocation: a82fe8db5eea4b339da327a9ed862d7f Docs: man:systemd.special(7) https://systemd.io/NETWORK_ONLINE Apr 14 22:07:36 npc99a84aae6d34 systemd[1]: Reached target network.target - Network. ● nss-lookup.target - Host and Network Name Lookups Loaded: loaded (/usr/lib/systemd/system/nss-lookup.target; static) Active: active since Tue 2026-04-14 22:07:36 UTC; 44min ago Invocation: 6f031301b0874a3197214f41e1690d94 Docs: man:systemd.special(7) Apr 14 22:07:36 npc99a84aae6d34 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. ○ nss-user-lookup.target - User and Group Name Lookups Loaded: loaded (/usr/lib/systemd/system/nss-user-lookup.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ● paths.target - Path Units Loaded: loaded (/usr/lib/systemd/system/paths.target; static) Active: active since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: c64408177549490fa89d3628bfc2e569 Docs: man:systemd.special(7) Apr 14 22:07:35 debian systemd[1]: Reached target paths.target - Path Units. ○ remote-cryptsetup.target - Remote Encrypted Volumes Loaded: loaded (/usr/lib/systemd/system/remote-cryptsetup.target; disabled; preset: enabled) Active: inactive (dead) Docs: man:systemd.special(7) ● remote-fs-pre.target - Preparation for Remote File Systems Loaded: loaded (/usr/lib/systemd/system/remote-fs-pre.target; static) Active: active since Tue 2026-04-14 22:07:36 UTC; 44min ago Invocation: 92ea5be0394c4d41abf53c56d5419b61 Docs: man:systemd.special(7) Apr 14 22:07:36 npc99a84aae6d34 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. ● remote-fs.target - Remote File Systems Loaded: loaded (/usr/lib/systemd/system/remote-fs.target; enabled; preset: enabled) Active: active since Tue 2026-04-14 22:07:36 UTC; 44min ago Invocation: f2a9bfb96cf045608a6c4cb81eb84087 Docs: man:systemd.special(7) Apr 14 22:07:36 npc99a84aae6d34 systemd[1]: Reached target remote-fs.target - Remote File Systems. ○ remote-veritysetup.target - Remote Verity Protected Volumes Loaded: loaded (/usr/lib/systemd/system/remote-veritysetup.target; disabled; preset: enabled) Active: inactive (dead) Docs: man:systemd.special(7) ○ rescue.target - Rescue Mode Loaded: loaded (/usr/lib/systemd/system/rescue.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ shutdown.target - System Shutdown Loaded: loaded (/usr/lib/systemd/system/shutdown.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ sleep.target - Sleep Loaded: loaded (/usr/lib/systemd/system/sleep.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ● slices.target - Slice Units Loaded: loaded (/usr/lib/systemd/system/slices.target; static) Active: active since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 0a73840f2c584684b56957575d13c768 Docs: man:systemd.special(7) Notice: journal has been rotated since unit was started, output may be incomplete. ● sockets.target - Socket Units Loaded: loaded (/usr/lib/systemd/system/sockets.target; static) Active: active since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: 184dea2b36ad461aba92b6cf5dc8cb67 Docs: man:systemd.special(7) Apr 14 22:07:35 debian systemd[1]: Reached target sockets.target - Socket Units. ○ soft-reboot.target - Reboot System Userspace Loaded: loaded (/usr/lib/systemd/system/soft-reboot.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ● ssh-access.target - SSH Access Available Loaded: loaded (/usr/lib/systemd/system/ssh-access.target; static) Active: active since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: 8ceae2d4f61f4702891f3ea5ab2ff5ff Docs: man:systemd.special(7) Apr 14 22:07:35 debian systemd[1]: Reached target ssh-access.target - SSH Access Available. ○ suspend-then-hibernate.target - Suspend; Hibernate if not used for a period of time Loaded: loaded (/usr/lib/systemd/system/suspend-then-hibernate.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ suspend.target - Suspend Loaded: loaded (/usr/lib/systemd/system/suspend.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ● swap.target - Swaps Loaded: loaded (/usr/lib/systemd/system/swap.target; static) Active: active since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: dd0f545a6ff34b45aea3a14aee257bac Docs: man:systemd.special(7) Notice: journal has been rotated since unit was started, output may be incomplete. ● sysinit.target - System Initialization Loaded: loaded (/usr/lib/systemd/system/sysinit.target; static) Active: active since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: b7027667818547e6a5860977bea6f7be Docs: man:systemd.special(7) Apr 14 22:07:35 debian systemd[1]: Reached target sysinit.target - System Initialization. ● time-set.target - System Time Set Loaded: loaded (/usr/lib/systemd/system/time-set.target; static) Active: active since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 4235156acf284f6a8c60df64955c2303 Docs: man:systemd.special(7) Apr 14 22:07:33 debian systemd[1]: Reached target time-set.target - System Time Set. ● time-sync.target - System Time Synchronized Loaded: loaded (/usr/lib/systemd/system/time-sync.target; static) Active: active since Tue 2026-04-14 22:25:51 UTC; 26min ago Invocation: e03fb2d7b713437d8d13e601b9858ce1 Docs: man:systemd.special(7) Apr 14 22:25:51 npc99a84aae6d34 systemd[1]: Reached target time-sync.target - System Time Synchronized. ● timers.target - Timer Units Loaded: loaded (/usr/lib/systemd/system/timers.target; static) Active: active since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: 0eeadb59e9e3414f80ee08804e2c2810 Docs: man:systemd.special(7) Apr 14 22:07:35 debian systemd[1]: Reached target timers.target - Timer Units. ○ tpm2.target - Trusted Platform Module Loaded: loaded (/usr/lib/systemd/system/tpm2.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ umount.target - Unmount All Filesystems Loaded: loaded (/usr/lib/systemd/system/umount.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ veritysetup-pre.target - Local Verity Protected Volumes (Pre) Loaded: loaded (/usr/lib/systemd/system/veritysetup-pre.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ● veritysetup.target - Local Verity Protected Volumes Loaded: loaded (/usr/lib/systemd/system/veritysetup.target; static) Active: active since Tue 2026-04-14 22:07:33 UTC; 44min ago Invocation: 3ab814be9e464d128325186eb59a5ec3 Docs: man:systemd.special(7) Notice: journal has been rotated since unit was started, output may be incomplete. ● virt-guest-shutdown.target - libvirt guests shutdown target Loaded: loaded (/usr/lib/systemd/system/virt-guest-shutdown.target; static) Active: active since Tue 2026-04-14 22:24:03 UTC; 27min ago Invocation: f83dead10ece4b6bb538bbe8b5b742f6 Docs: https://libvirt.org/ Apr 14 22:24:03 npc99a84aae6d34 systemd[1]: Reached target virt-guest-shutdown.target - libvirt guests shutdown target. ● apt-daily-upgrade.timer - Daily apt upgrade and clean activities Loaded: loaded (/usr/lib/systemd/system/apt-daily-upgrade.timer; enabled; preset: enabled) Active: active (waiting) since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: c0e2808789e941e2910bc0ae2b2bb7b0 Trigger: Wed 2026-04-15 06:30:00 UTC; 7h left Triggers: ● apt-daily-upgrade.service Apr 14 22:07:35 debian systemd[1]: Started apt-daily-upgrade.timer - Daily apt upgrade and clean activities. ● apt-daily.timer - Daily apt download activities Loaded: loaded (/usr/lib/systemd/system/apt-daily.timer; enabled; preset: enabled) Active: active (waiting) since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: dcb0c137563c4f0eba04f7dd2cabf52d Trigger: Wed 2026-04-15 13:05:41 UTC; 14h left Triggers: ● apt-daily.service Apr 14 22:07:35 debian systemd[1]: Started apt-daily.timer - Daily apt download activities. ● dpkg-db-backup.timer - Daily dpkg database backup timer Loaded: loaded (/usr/lib/systemd/system/dpkg-db-backup.timer; enabled; preset: enabled) Active: active (waiting) since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: f44216a9288a43809aabfb324190708c Trigger: Wed 2026-04-15 00:00:00 UTC; 1h 8min left Triggers: ● dpkg-db-backup.service Docs: man:dpkg(1) Apr 14 22:07:35 debian systemd[1]: Started dpkg-db-backup.timer - Daily dpkg database backup timer. ● e2scrub_all.timer - Periodic ext4 Online Metadata Check for All Filesystems Loaded: loaded (/usr/lib/systemd/system/e2scrub_all.timer; enabled; preset: enabled) Active: active (waiting) since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: 94f16961271c4efe8e8a9fc19853c237 Trigger: Sun 2026-04-19 03:10:02 UTC; 4 days left Triggers: ● e2scrub_all.service Apr 14 22:07:35 debian systemd[1]: Started e2scrub_all.timer - Periodic ext4 Online Metadata Check for All Filesystems. ● fstrim.timer - Discard unused filesystem blocks once a week Loaded: loaded (/usr/lib/systemd/system/fstrim.timer; enabled; preset: enabled) Active: active (waiting) since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: df96ca4f310e4ac0964f9d6d18c81ad3 Trigger: Mon 2026-04-20 00:04:26 UTC; 5 days left Triggers: ● fstrim.service Docs: man:fstrim Apr 14 22:07:35 debian systemd[1]: Started fstrim.timer - Discard unused filesystem blocks once a week. ● logrotate.timer - Daily rotation of log files Loaded: loaded (/usr/lib/systemd/system/logrotate.timer; enabled; preset: enabled) Active: active (waiting) since Tue 2026-04-14 22:23:55 UTC; 28min ago Invocation: b968d4da95a94dd4823e718a55daca86 Trigger: Wed 2026-04-15 00:09:23 UTC; 1h 17min left Triggers: ● logrotate.service Docs: man:logrotate(8) man:logrotate.conf(5) Apr 14 22:23:55 npc99a84aae6d34 systemd[1]: Started logrotate.timer - Daily rotation of log files. ● man-db.timer - Daily man-db regeneration Loaded: loaded (/usr/lib/systemd/system/man-db.timer; enabled; preset: enabled) Active: active (waiting) since Tue 2026-04-14 22:17:50 UTC; 34min ago Invocation: c972f589fea6425a8f30b0d3f3eaa791 Trigger: Wed 2026-04-15 05:04:31 UTC; 6h left Triggers: ● man-db.service Docs: man:mandb(8) Apr 14 22:17:50 npc99a84aae6d34 systemd[1]: Started man-db.timer - Daily man-db regeneration. ● systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories Loaded: loaded (/usr/lib/systemd/system/systemd-tmpfiles-clean.timer; static) Active: active (waiting) since Tue 2026-04-14 22:07:35 UTC; 44min ago Invocation: 54870956f5bc49a7b754896e85c94708 Trigger: Wed 2026-04-15 22:22:37 UTC; 23h left Triggers: ● systemd-tmpfiles-clean.service Docs: man:tmpfiles.d(5) man:systemd-tmpfiles(8) Apr 14 22:07:35 debian systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. ● xfs_scrub_all.timer - Periodic XFS Online Metadata Check for All Filesystems Loaded: loaded (/usr/lib/systemd/system/xfs_scrub_all.timer; enabled; preset: enabled) Active: active (waiting) since Tue 2026-04-14 22:17:41 UTC; 34min ago Invocation: 617ccdcde49643749eb66ba92d921ac4 Trigger: Sun 2026-04-19 03:10:09 UTC; 4 days left Triggers: ● xfs_scrub_all.service Apr 14 22:17:41 npc99a84aae6d34 systemd[1]: Started xfs_scrub_all.timer - Periodic XFS Online Metadata Check for All Filesystems.