F   UID     PID    PPID PRI  NI    VSZ   RSS WCHAN  STAT TTY        TIME COMMAND
4     0       1       0  20   0 176380 20824 ep_pol Ss   ?          1:48 /usr/lib/systemd/systemd --system --deserialize 33
1     0       2       0  20   0      0     0 kthrea S    ?          0:00 [kthreadd]
1     0       3       2   0 -20      0     0 rescue I<   ?          0:00 [rcu_gp]
1     0       4       2   0 -20      0     0 rescue I<   ?          0:00 [rcu_par_gp]
1     0       5       2   0 -20      0     0 rescue I<   ?          0:00 [slub_flushwq]
1     0       6       2   0 -20      0     0 rescue I<   ?          0:00 [netns]
1     0       8       2   0 -20      0     0 worker I<   ?          0:00 [kworker/0:0H-events_highpri]
1     0      10       2   0 -20      0     0 rescue I<   ?          0:00 [mm_percpu_wq]
1     0      12       2  20   0      0     0 rcu_ta I    ?          0:00 [rcu_tasks_kthre]
1     0      13       2  20   0      0     0 rcu_ta I    ?          0:00 [rcu_tasks_rude_]
1     0      14       2  20   0      0     0 rcu_ta I    ?          0:00 [rcu_tasks_trace]
1     0      15       2  20   0      0     0 smpboo S    ?          0:00 [ksoftirqd/0]
1     0      16       2  20   0      0     0 rcu_gp I    ?          0:18 [rcu_preempt]
1     0      17       2 -100  -      0     0 smpboo S    ?          0:00 [migration/0]
1     0      19       2  20   0      0     0 smpboo S    ?          0:00 [cpuhp/0]
1     0      20       2  20   0      0     0 smpboo S    ?          0:00 [cpuhp/1]
1     0      21       2 -100  -      0     0 smpboo S    ?          0:00 [migration/1]
1     0      22       2  20   0      0     0 smpboo S    ?          0:00 [ksoftirqd/1]
1     0      24       2   0 -20      0     0 worker I<   ?          0:00 [kworker/1:0H-events_highpri]
1     0      25       2  20   0      0     0 smpboo S    ?          0:00 [cpuhp/2]
1     0      26       2 -100  -      0     0 smpboo S    ?          0:00 [migration/2]
1     0      27       2  20   0      0     0 smpboo S    ?          0:00 [ksoftirqd/2]
1     0      29       2   0 -20      0     0 worker I<   ?          0:00 [kworker/2:0H-events_highpri]
1     0      30       2  20   0      0     0 smpboo S    ?          0:00 [cpuhp/3]
1     0      31       2 -100  -      0     0 smpboo S    ?          0:00 [migration/3]
1     0      32       2  20   0      0     0 smpboo S    ?          0:00 [ksoftirqd/3]
1     0      34       2   0 -20      0     0 worker I<   ?          0:00 [kworker/3:0H-events_highpri]
1     0      35       2  20   0      0     0 smpboo S    ?          0:00 [cpuhp/4]
1     0      36       2 -100  -      0     0 smpboo S    ?          0:00 [migration/4]
1     0      37       2  20   0      0     0 smpboo S    ?          0:00 [ksoftirqd/4]
1     0      39       2   0 -20      0     0 worker I<   ?          0:00 [kworker/4:0H-events_highpri]
1     0      40       2  20   0      0     0 smpboo S    ?          0:00 [cpuhp/5]
1     0      41       2 -100  -      0     0 smpboo S    ?          0:00 [migration/5]
1     0      42       2  20   0      0     0 smpboo S    ?          0:00 [ksoftirqd/5]
1     0      44       2   0 -20      0     0 worker I<   ?          0:00 [kworker/5:0H]
1     0      45       2  20   0      0     0 smpboo S    ?          0:00 [cpuhp/6]
1     0      46       2 -100  -      0     0 smpboo S    ?          0:00 [migration/6]
1     0      47       2  20   0      0     0 smpboo S    ?          0:02 [ksoftirqd/6]
1     0      49       2   0 -20      0     0 worker I<   ?          0:00 [kworker/6:0H-events_highpri]
1     0      50       2  20   0      0     0 smpboo S    ?          0:00 [cpuhp/7]
1     0      51       2 -100  -      0     0 smpboo S    ?          0:00 [migration/7]
1     0      52       2  20   0      0     0 smpboo S    ?          0:00 [ksoftirqd/7]
1     0      54       2   0 -20      0     0 worker I<   ?          0:00 [kworker/7:0H-events_highpri]
5     0      62       2  20   0      0     0 devtmp S    ?          0:00 [kdevtmpfs]
1     0      63       2   0 -20      0     0 rescue I<   ?          0:00 [inet_frag_wq]
1     0      64       2  20   0      0     0 kaudit S    ?          0:01 [kauditd]
1     0      65       2  20   0      0     0 watchd S    ?          0:00 [khungtaskd]
1     0      66       2  20   0      0     0 oom_re S    ?          0:00 [oom_reaper]
1     0      67       2   0 -20      0     0 rescue I<   ?          0:00 [writeback]
1     0      68       2  20   0      0     0 kcompa S    ?          0:00 [kcompactd0]
1     0      69       2  25   5      0     0 ksm_sc SN   ?          0:00 [ksmd]
1     0      70       2  39  19      0     0 khugep SN   ?          0:00 [khugepaged]
1     0      71       2   0 -20      0     0 rescue I<   ?          0:00 [cryptd]
1     0      72       2   0 -20      0     0 rescue I<   ?          0:00 [kintegrityd]
1     0      73       2   0 -20      0     0 rescue I<   ?          0:00 [kblockd]
1     0      74       2   0 -20      0     0 rescue I<   ?          0:00 [blkcg_punt_bio]
1     0      75       2   0 -20      0     0 rescue I<   ?          0:00 [tpm_dev_wq]
1     0      76       2   0 -20      0     0 rescue I<   ?          0:00 [md]
1     0      77       2   0 -20      0     0 rescue I<   ?          0:00 [edac-poller]
1     0      78       2 -51   -      0     0 kthrea S    ?          0:00 [watchdogd]
1     0      80       2   0 -20      0     0 worker I<   ?          0:01 [kworker/0:1H-kblockd]
1     0      82       2  20   0      0     0 kswapd S    ?          0:00 [kswapd0]
1     0      89       2   0 -20      0     0 rescue I<   ?          0:00 [kthrotld]
1     0      94       2   0 -20      0     0 rescue I<   ?          0:00 [acpi_thermal_pm]
1     0      96       2  20   0      0     0 add_hw S    ?          0:00 [hwrng]
1     0      97       2   0 -20      0     0 worker I<   ?          0:01 [kworker/1:1H-events_highpri]
1     0      98       2   0 -20      0     0 rescue I<   ?          0:00 [kmpath_rdacd]
1     0      99       2   0 -20      0     0 rescue I<   ?          0:00 [kaluad]
1     0     100       2   0 -20      0     0 worker I<   ?          0:01 [kworker/3:1H-xfs-log/vda4]
1     0     101       2   0 -20      0     0 worker I<   ?          0:01 [kworker/4:1H-events_highpri]
1     0     102       2   0 -20      0     0 rescue I<   ?          0:00 [mld]
1     0     103       2   0 -20      0     0 rescue I<   ?          0:00 [ipv6_addrconf]
1     0     104       2   0 -20      0     0 rescue I<   ?          0:00 [kstrp]
1     0     118       2   0 -20      0     0 rescue I<   ?          0:00 [zswap-shrink]
1     0     120       2   0 -20      0     0 worker I<   ?          0:01 [kworker/2:1H-kblockd]
1     0     240       2   0 -20      0     0 worker I<   ?          0:00 [kworker/u17:0]
1     0     241       2   0 -20      0     0 worker I<   ?          0:00 [kworker/6:1H-events_highpri]
1     0     393       2   0 -20      0     0 rescue I<   ?          0:00 [rpciod]
1     0     394       2   0 -20      0     0 rescue I<   ?          0:00 [xprtiod]
1     0     487       2   0 -20      0     0 rescue I<   ?          0:00 [ata_sff]
1     0     488       2  20   0      0     0 scsi_e S    ?          0:00 [scsi_eh_0]
1     0     489       2   0 -20      0     0 rescue I<   ?          0:00 [scsi_tmf_0]
1     0     490       2  20   0      0     0 scsi_e S    ?          0:00 [scsi_eh_1]
1     0     491       2   0 -20      0     0 rescue I<   ?          0:00 [scsi_tmf_1]
1     0     493       2   0 -20      0     0 worker I<   ?          0:01 [kworker/7:1H-kblockd]
1     0     496       2   0 -20      0     0 worker I<   ?          0:01 [kworker/5:1H-events_highpri]
1     0     515       2   0 -20      0     0 rescue I<   ?          0:00 [xfsalloc]
1     0     516       2   0 -20      0     0 rescue I<   ?          0:00 [xfs_mru_cache]
1     0     517       2   0 -20      0     0 rescue I<   ?          0:00 [xfs-buf/vda4]
1     0     518       2   0 -20      0     0 rescue I<   ?          0:00 [xfs-conv/vda4]
1     0     519       2   0 -20      0     0 rescue I<   ?          0:00 [xfs-reclaim/vda]
1     0     520       2   0 -20      0     0 rescue I<   ?          0:00 [xfs-blockgc/vda]
1     0     521       2   0 -20      0     0 rescue I<   ?          0:00 [xfs-inodegc/vda]
1     0     522       2   0 -20      0     0 rescue I<   ?          0:00 [xfs-log/vda4]
1     0     523       2   0 -20      0     0 rescue I<   ?          0:00 [xfs-cil/vda4]
1     0     524       2  20   0      0     0 xfsail S    ?          0:07 [xfsaild/vda4]
1     0     697       2   0 -20      0     0 rescue I<   ?          0:00 [xfs-buf/vda3]
1     0     698       2   0 -20      0     0 rescue I<   ?          0:00 [xfs-conv/vda3]
1     0     699       2   0 -20      0     0 rescue I<   ?          0:00 [xfs-reclaim/vda]
1     0     700       2   0 -20      0     0 rescue I<   ?          0:00 [xfs-blockgc/vda]
1     0     701       2   0 -20      0     0 rescue I<   ?          0:00 [xfs-inodegc/vda]
1     0     702       2   0 -20      0     0 rescue I<   ?          0:00 [xfs-log/vda3]
1     0     703       2   0 -20      0     0 rescue I<   ?          0:00 [xfs-cil/vda3]
1     0     704       2  20   0      0     0 xfsail S    ?          0:00 [xfsaild/vda3]
4    32     723       1  20   0  13240  5548 do_pol Ss   ?          0:00 /usr/bin/rpcbind -w -f
5     0     725       1  16  -4  94868  6808 do_sel S<sl ?          0:07 /sbin/auditd
4     0     727     725  16  -4   8276  3912 do_sel S<   ?          0:02 /usr/sbin/sedispatch
4    81     750       1  20   0  11284  4888 ep_pol Ss   ?          0:00 /usr/bin/dbus-broker-launch --scope system --audit
4    81     754     750  20   0   5432  3232 ep_pol S    ?          0:11 dbus-broker --log 4 --controller 9 --machine-id c0212a8b024a111cfc61293864f36c87 --max-bytes 536870912 --max-fds 4096 --max-matches 131072 --audit
4     0     757       1  20   0  79200  4824 do_pol Ssl  ?          0:00 /usr/sbin/irqbalance --foreground
4     0     758       1  20   0 269840 15608 do_sel Ssl  ?          0:19 /usr/sbin/rsyslogd -n
4     0     759       1  20   0  32408 15104 ep_pol Ss   ?          0:07 /usr/lib/systemd/systemd-logind
1     0     797       1  20   0  52452  5336 ep_pol Ssl  ?          0:00 /usr/sbin/gssproxy -D
1     0     801       1  20   0  78528  4160 do_pol Ssl  ?          0:00 /usr/bin/rhsmcertd
4   998    1033       1  20   0 2917176 22744 do_pol Ssl ?          0:02 /usr/lib/polkit-1/polkitd --no-debug
4     0    1134       1  20   0   6108  3592 hrtime Ss   ?          0:00 /usr/sbin/crond -n
4     0    1136       1  20   0   3044  1088 do_sel Ss+  tty1       0:00 /sbin/agetty -o -p -- \u --noclear - linux
4     0    1140       1  20   0   3088  1084 do_sel Ss+  ttyS0      0:00 /sbin/agetty -o -p -- \u --keep-baud 115200,57600,38400,9600 - vt220
4  1000    4176       1  20   0  23756 15004 ep_pol Ss   ?          0:17 /usr/lib/systemd/systemd --user
5  1000    4178    4176  20   0  25776  5444 do_sig S    ?          0:00 (sd-pam)
1  1000    4334       1  20   0 324300 18692 inet_c S    ?          1:59 /usr/bin/python3
4     0    5974       1  20   0 255276 15828 do_pol Ssl  ?          1:00 /usr/sbin/NetworkManager --no-daemon
4     0    6591       1  20   0 271520 98076 do_pol Ssl  ?          0:04 /usr/bin/python3 /usr/libexec/rhsm-service
1     0    6659       2   0 -20      0     0 rescue I<   ?          0:00 [tls-strp]
0  1000   18420       1  20   0  50088  9736 do_sys S    ?          0:00 podman
0  1000   18433    4176  20   0  10584  4196 ep_pol Ss   ?          0:00 /usr/bin/dbus-broker-launch --scope user
0  1000   18434   18433  20   0   4864  1392 ep_pol S    ?          0:00 dbus-broker --log 4 --controller 9 --machine-id c0212a8b024a111cfc61293864f36c87 --max-bytes 100000000000000 --max-fds 25000000000000 --max-matches 5000000000
1   986   22278       1  10 -10  19440  9972 do_pol S<s  ?          3:00 ovsdb-server /etc/openvswitch/conf.db -vconsole:emer -vsyslog:err -vfile:info --remote=punix:/var/run/openvswitch/db.sock --private-key=db:Open_vSwitch,SSL,private_key --certificate=db:Open_vSwitch,SSL,certificate --bootstrap-ca-cert=db:Open_vSwitch,SSL,ca_cert --user openvswitch:hugetlbfs --no-chdir --log-file=/var/log/openvswitch/ovsdb-server.log --pidfile=/var/run/openvswitch/ovsdb-server.pid --detach
1     0   22284       2   0 -20      0     0 rescue I<   ?          0:00 [ib-comp-wq]
1     0   22285       2   0 -20      0     0 rescue I<   ?          0:00 [ib-comp-unb-wq]
1     0   22286       2   0 -20      0     0 rescue I<   ?          0:00 [ib_mcast]
1     0   22287       2   0 -20      0     0 rescue I<   ?          0:00 [ib_nl_sa_wq]
5   986   22359       1  10 -10 1130312 279608 do_pol S<Lsl ?       0:51 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --user openvswitch:hugetlbfs --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach
1     0   24868       2   0 -20      0     0 rescue I<   ?          0:00 [kdmflush/253:0]
1     0   25118       2   0 -20      0     0 rescue I<   ?          0:00 [kdmflush/253:1]
4  1002   26179       1  20   0  22644 13816 ep_pol Ss   ?          0:16 /usr/lib/systemd/systemd --user
5  1002   26181   26179  20   0 176580  9016 do_sig S    ?          0:00 (sd-pam)
5     0   28907       1  20   0   8304  1920 do_pol Ss   ?          0:00 /usr/bin/conmon --api-version 1 -c fdf07215f0388d0ebc44f1f3744080ba594441e647c300d0dade62ff5beba234 -u fdf07215f0388d0ebc44f1f3744080ba594441e647c300d0dade62ff5beba234 -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/fdf07215f0388d0ebc44f1f3744080ba594441e647c300d0dade62ff5beba234/userdata -p /run/containers/storage/overlay-containers/fdf07215f0388d0ebc44f1f3744080ba594441e647c300d0dade62ff5beba234/userdata/pidfile -n ceph-f1fea371-cb69-578d-a3d0-b5c472a84b46-crash-np0005626463 --exit-dir /run/libpod/exits --full-attach -l journald --log-level warning --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/fdf07215f0388d0ebc44f1f3744080ba594441e647c300d0dade62ff5beba234/userdata/oci-log --conmon-pidfile /run/ceph-f1fea371-cb69-578d-a3d0-b5c472a84b46@crash.np0005626463.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg  --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --storage-driver --exit-command-arg overlay --exit-command-arg --storage-opt --exit-command-arg overlay.mountopt=nodev,metacopy=on --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --rm --exit-command-arg fdf07215f0388d0ebc44f1f3744080ba594441e647c300d0dade62ff5beba234
4     0   28909   28907  20   0   1084     4 signal Ss   ?          0:00 /run/podman-init -- /usr/bin/ceph-crash -n client.crash.np0005626463
4   167   28911   28909  20   0  14652 12600 do_sel S    ?          0:00 /usr/bin/python3 -s /usr/bin/ceph-crash -n client.crash.np0005626463
5     0   31629       1  20   0   8304  1952 do_pol Ss   ?          0:00 /usr/bin/conmon --api-version 1 -c 862eadaff641589ceb245e67477cf75d6f44dd8a2e370794aa63510852a63e9d -u 862eadaff641589ceb245e67477cf75d6f44dd8a2e370794aa63510852a63e9d -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/862eadaff641589ceb245e67477cf75d6f44dd8a2e370794aa63510852a63e9d/userdata -p /run/containers/storage/overlay-containers/862eadaff641589ceb245e67477cf75d6f44dd8a2e370794aa63510852a63e9d/userdata/pidfile -n ceph-f1fea371-cb69-578d-a3d0-b5c472a84b46-osd-2 --exit-dir /run/libpod/exits --full-attach -l journald --log-level warning --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/862eadaff641589ceb245e67477cf75d6f44dd8a2e370794aa63510852a63e9d/userdata/oci-log --conmon-pidfile /run/ceph-f1fea371-cb69-578d-a3d0-b5c472a84b46@osd.2.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg  --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --storage-driver --exit-command-arg overlay --exit-command-arg --storage-opt --exit-command-arg overlay.mountopt=nodev,metacopy=on --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --rm --exit-command-arg 862eadaff641589ceb245e67477cf75d6f44dd8a2e370794aa63510852a63e9d
4     0   31631   31629  20   0   1084     4 signal Ss   ?          0:00 /run/podman-init -- /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false
4   167   31633   31631  20   0 743180 210952 futex_ Sl  ?          1:22 /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false
5     0   32571       1  20   0   8304  1916 do_pol Ss   ?          0:00 /usr/bin/conmon --api-version 1 -c 6682f631389f5fea34334a28d29db8aea85f2971ab07e24b48970944e80cac0e -u 6682f631389f5fea34334a28d29db8aea85f2971ab07e24b48970944e80cac0e -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/6682f631389f5fea34334a28d29db8aea85f2971ab07e24b48970944e80cac0e/userdata -p /run/containers/storage/overlay-containers/6682f631389f5fea34334a28d29db8aea85f2971ab07e24b48970944e80cac0e/userdata/pidfile -n ceph-f1fea371-cb69-578d-a3d0-b5c472a84b46-osd-5 --exit-dir /run/libpod/exits --full-attach -l journald --log-level warning --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/6682f631389f5fea34334a28d29db8aea85f2971ab07e24b48970944e80cac0e/userdata/oci-log --conmon-pidfile /run/ceph-f1fea371-cb69-578d-a3d0-b5c472a84b46@osd.5.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg  --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --storage-driver --exit-command-arg overlay --exit-command-arg --storage-opt --exit-command-arg overlay.mountopt=nodev,metacopy=on --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --rm --exit-command-arg 6682f631389f5fea34334a28d29db8aea85f2971ab07e24b48970944e80cac0e
4     0   32573   32571  20   0   1084     4 signal Ss   ?          0:00 /run/podman-init -- /usr/bin/ceph-osd -n osd.5 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false
4   167   32575   32573  20   0 711728 195412 futex_ Sl  ?          1:22 /usr/bin/ceph-osd -n osd.5 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false
4     0   47710       1  20   0 134124 44368 ep_pol Ss   ?          0:17 /usr/lib/systemd/systemd-journald
4     0   47718       1  20   0  34872 12972 ep_pol Ss   ?          0:00 /usr/lib/systemd/systemd-udevd
1     0   61760       2   0 -20      0     0 rescue I<   ?          0:00 [iscsi_conn_clea]
4     0   67690       1  20   0  24432 14484 do_sel Ss   ?          0:08 /usr/sbin/snmpd -LS0-5d -f
4     0   84014       1  20   0  18592  9044 ep_pol Ss   ?          0:04 /usr/lib/systemd/systemd-machined
1     0   93454       2  20   0      0     0 worker I    ?          0:02 [kworker/u16:1-flush-252:0]
4     0  127485       1  20   0 256696 26276 futex_ Ssl  ?          0:01 /usr/bin/python3 -Es /usr/sbin/tuned -l -P
5   993  140850       1  20   0  84168  2536 do_sel S    ?          0:00 /usr/sbin/chronyd -F 2
5     0  157695       1  20   0   8304  1980 do_pol Ss   ?          0:00 /usr/bin/conmon --api-version 1 -c 83e0dfad8e11fc0edee47d0ecab9337f343a62cd8dd545e1b2fa3b528410a3fc -u 83e0dfad8e11fc0edee47d0ecab9337f343a62cd8dd545e1b2fa3b528410a3fc -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/83e0dfad8e11fc0edee47d0ecab9337f343a62cd8dd545e1b2fa3b528410a3fc/userdata -p /run/containers/storage/overlay-containers/83e0dfad8e11fc0edee47d0ecab9337f343a62cd8dd545e1b2fa3b528410a3fc/userdata/pidfile -n ovn_controller --exit-dir /run/libpod/exits --full-attach -s -l journald --log-level warning --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/83e0dfad8e11fc0edee47d0ecab9337f343a62cd8dd545e1b2fa3b528410a3fc/userdata/oci-log --conmon-pidfile /run/ovn_controller.pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg  --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --storage-driver --exit-command-arg overlay --exit-command-arg --storage-opt --exit-command-arg overlay.mountopt=nodev,metacopy=on --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg 83e0dfad8e11fc0edee47d0ecab9337f343a62cd8dd545e1b2fa3b528410a3fc
4     0  157697  157695  20   0   2512   940 do_sig Ss   ?          0:00 dumb-init --single-child -- kolla_start
4     0  157700  157697  20   0 328276 21000 do_pol Sl   ?          0:08 /usr/bin/ovn-controller --pidfile unix:/run/openvswitch/db.sock
5     0  163567       1  20   0   8304  1952 do_pol Ss   ?          0:00 /usr/bin/conmon --api-version 1 -c 11c0efd0e0b3e96f28e1667bfcc14ff7c3c7ffd1c98bac3f3df14178958e3739 -u 11c0efd0e0b3e96f28e1667bfcc14ff7c3c7ffd1c98bac3f3df14178958e3739 -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/11c0efd0e0b3e96f28e1667bfcc14ff7c3c7ffd1c98bac3f3df14178958e3739/userdata -p /run/containers/storage/overlay-containers/11c0efd0e0b3e96f28e1667bfcc14ff7c3c7ffd1c98bac3f3df14178958e3739/userdata/pidfile -n ovn_metadata_agent --exit-dir /run/libpod/exits --full-attach -s -l journald --log-level warning --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/11c0efd0e0b3e96f28e1667bfcc14ff7c3c7ffd1c98bac3f3df14178958e3739/userdata/oci-log --conmon-pidfile /run/ovn_metadata_agent.pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg  --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --storage-driver --exit-command-arg overlay --exit-command-arg --storage-opt --exit-command-arg overlay.mountopt=nodev,metacopy=on --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg 11c0efd0e0b3e96f28e1667bfcc14ff7c3c7ffd1c98bac3f3df14178958e3739
4     0  163569  163567  20   0   2512   960 do_sig Ss   ?          0:00 dumb-init --single-child -- kolla_start
4     0  163572  163569  20   0 145092 133312 ep_pol S   ?          0:11 neutron-ovn-metadata-agent (/usr/bin/python3 /usr/bin/neutron-ovn-metadata-agent)
1     0  163670  163572  20   0 141604 119744 ep_pol S   ?          0:05 neutron-ovn-metadata-agent (/usr/bin/python3 /usr/bin/neutron-ovn-metadata-agent)
5     0  163675  163567  20   0 175056 105052 unix_s Sl  ?          0:01 /usr/bin/python3 /bin/privsep-helper --config-file /etc/neutron/neutron.conf --config-dir /etc/neutron.conf.d --privsep_context neutron.privileged.default --privsep_sock_path /tmp/tmpckuxnm6t/privsep.sock
1     0  163808  163567  20   0 174544 104796 unix_s Sl  ?          0:00 /usr/bin/python3 /bin/privsep-helper --config-file /etc/neutron/neutron.conf --config-dir /etc/neutron.conf.d --privsep_context neutron.privileged.link_cmd --privsep_sock_path /tmp/tmprtzwo8ob/privsep.sock
1     0  163964  163567  20   0 174284 103156 unix_s Sl  ?          0:00 /usr/bin/python3 /bin/privsep-helper --config-file /etc/neutron/neutron.conf --config-dir /etc/neutron.conf.d --privsep_context neutron.privileged.namespace_cmd --privsep_sock_path /tmp/tmp8829ngjy/privsep.sock
1     0  184036       2  20   0      0     0 worker I    ?          0:00 [kworker/3:0-ata_sff]
4     0  187670       1  20   0  15816  9064 do_sel Ss   ?          0:00 sshd: /usr/sbin/sshd -D [listener] 0 of 10-100 startups
4     0  207018       1  20   0  32888 14496 do_pol Ss   ?          0:01 /usr/sbin/virtlogd
4     0  207530       1  20   0 1512544 46000 do_pol Ssl ?          0:04 /usr/sbin/virtqemud --timeout 120
4     0  214560       1  10 -10  13440 12324 do_pol S<Ls ?          0:00 /usr/sbin/iscsid -f
1     0  215445       2   0 -20      0     0 rescue I<   ?          0:00 [kmpathd]
1     0  215446       2   0 -20      0     0 rescue I<   ?          0:00 [kmpath_handlerd]
1     0  217882       2   0 -20      0     0 rescue I<   ?          0:00 [nvme-wq]
1     0  217883       2   0 -20      0     0 rescue I<   ?          0:00 [nvme-reset-wq]
1     0  217884       2   0 -20      0     0 rescue I<   ?          0:00 [nvme-delete-wq]
4     0  219874       1 -100  - 289236 26964 futex_ SLsl ?          0:00 /sbin/multipathd -d -s
1     0  230498       2  20   0      0     0 worker I    ?          0:00 [kworker/2:0-events]
4     0  231253       1  20   0 1508916 21164 do_pol Ssl ?          0:03 /usr/sbin/virtnodedevd --timeout 120
5     0  238244       1  20   0   8304  1896 do_pol Ss   ?          0:00 /usr/bin/conmon --api-version 1 -c be69a98bc8c30737a588f2504e3463b9254e63c980be4e376d2f104225355f5c -u be69a98bc8c30737a588f2504e3463b9254e63c980be4e376d2f104225355f5c -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/be69a98bc8c30737a588f2504e3463b9254e63c980be4e376d2f104225355f5c/userdata -p /run/containers/storage/overlay-containers/be69a98bc8c30737a588f2504e3463b9254e63c980be4e376d2f104225355f5c/userdata/pidfile -n ceilometer_agent_compute --exit-dir /run/libpod/exits --full-attach -s -l journald --log-level warning --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/be69a98bc8c30737a588f2504e3463b9254e63c980be4e376d2f104225355f5c/userdata/oci-log --conmon-pidfile /run/ceilometer_agent_compute.pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg  --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --storage-driver --exit-command-arg overlay --exit-command-arg --storage-opt --exit-command-arg overlay.mountopt=nodev,metacopy=on --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg be69a98bc8c30737a588f2504e3463b9254e63c980be4e376d2f104225355f5c
4 42405  238246  238244  20   0   2512   992 do_sig Ss   ?          0:00 dumb-init --single-child -- kolla_start
4 42405  238249  238246  20   0 179668 85328 do_sel Ssl  ?          0:00 ceilometer-polling: master process [/usr/bin/ceilometer-polling --polling-namespaces compute --logfile /dev/stdout]
5 42405  238404  238249  20   0 340752 79896 do_sel Sl   ?          0:01 ceilometer-polling: AgentManager worker(0)
5     0  240526       1  20   0   8304  1884 do_pol Ss   ?          0:00 /usr/bin/conmon --api-version 1 -c bee8ba29aeb0c3b6a916d1366f5aac446e01450955724b0f85ae8ab1d4d64b3d -u bee8ba29aeb0c3b6a916d1366f5aac446e01450955724b0f85ae8ab1d4d64b3d -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/bee8ba29aeb0c3b6a916d1366f5aac446e01450955724b0f85ae8ab1d4d64b3d/userdata -p /run/containers/storage/overlay-containers/bee8ba29aeb0c3b6a916d1366f5aac446e01450955724b0f85ae8ab1d4d64b3d/userdata/pidfile -n node_exporter --exit-dir /run/libpod/exits --full-attach -s -l journald --log-level warning --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/bee8ba29aeb0c3b6a916d1366f5aac446e01450955724b0f85ae8ab1d4d64b3d/userdata/oci-log --conmon-pidfile /run/node_exporter.pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg  --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --storage-driver --exit-command-arg overlay --exit-command-arg --storage-opt --exit-command-arg overlay.mountopt=nodev,metacopy=on --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg bee8ba29aeb0c3b6a916d1366f5aac446e01450955724b0f85ae8ab1d4d64b3d
4     0  240528  240526  20   0 727104 28388 ep_pol Ssl  ?          0:04 /bin/node_exporter --web.disable-exporter-metrics --collector.systemd --collector.systemd.unit-include=(edpm_.*|ovs.*|openvswitch|virt.*|rsyslog)\.service --no-collector.dmi --no-collector.entropy --no-collector.thermal_zone --no-collector.time --no-collector.timex --no-collector.uname --no-collector.stat --no-collector.hwmon --no-collector.os --no-collector.selinux --no-collector.textfile --no-collector.powersupplyclass --no-collector.pressure --no-collector.rapl --path.rootfs=/rootfs
5     0  242941       1  20   0   8304  2012 do_pol Ss   ?          0:00 /usr/bin/conmon --api-version 1 -c da3017db5cf229fdef3f2bfb4085c381427336b5c58ec1f7e94d11cfbeb2c8eb -u da3017db5cf229fdef3f2bfb4085c381427336b5c58ec1f7e94d11cfbeb2c8eb -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/da3017db5cf229fdef3f2bfb4085c381427336b5c58ec1f7e94d11cfbeb2c8eb/userdata -p /run/containers/storage/overlay-containers/da3017db5cf229fdef3f2bfb4085c381427336b5c58ec1f7e94d11cfbeb2c8eb/userdata/pidfile -n podman_exporter --exit-dir /run/libpod/exits --full-attach -s -l journald --log-level warning --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/da3017db5cf229fdef3f2bfb4085c381427336b5c58ec1f7e94d11cfbeb2c8eb/userdata/oci-log --conmon-pidfile /run/podman_exporter.pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg  --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --storage-driver --exit-command-arg overlay --exit-command-arg --storage-opt --exit-command-arg overlay.mountopt=nodev,metacopy=on --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg da3017db5cf229fdef3f2bfb4085c381427336b5c58ec1f7e94d11cfbeb2c8eb
4     0  242943  242941  20   0 1247012 26772 futex_ Ssl ?          0:03 /bin/podman_exporter
4     0  242954       1  20   0 2088532 90788 futex_ Ssl ?          5:58 /usr/bin/podman --log-level=info system service
5     0  245358       1  20   0   8304  1892 do_pol Ss   ?          0:00 /usr/bin/conmon --api-version 1 -c 6cb5576e612c42c1aa5a9a12d920a8ac121c3bbe97b17ae6c4c7aa9081ec5b43 -u 6cb5576e612c42c1aa5a9a12d920a8ac121c3bbe97b17ae6c4c7aa9081ec5b43 -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/6cb5576e612c42c1aa5a9a12d920a8ac121c3bbe97b17ae6c4c7aa9081ec5b43/userdata -p /run/containers/storage/overlay-containers/6cb5576e612c42c1aa5a9a12d920a8ac121c3bbe97b17ae6c4c7aa9081ec5b43/userdata/pidfile -n openstack_network_exporter --exit-dir /run/libpod/exits --full-attach -s -l journald --log-level warning --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/6cb5576e612c42c1aa5a9a12d920a8ac121c3bbe97b17ae6c4c7aa9081ec5b43/userdata/oci-log --conmon-pidfile /run/openstack_network_exporter.pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg  --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --storage-driver --exit-command-arg overlay --exit-command-arg --storage-opt --exit-command-arg overlay.mountopt=nodev,metacopy=on --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg 6cb5576e612c42c1aa5a9a12d920a8ac121c3bbe97b17ae6c4c7aa9081ec5b43
4     0  245360  245358  20   0 2202004 29032 futex_ Ssl ?          0:04 /app/openstack-network-exporter
1     0  253883       2  20   0      0     0 worker I    ?          0:01 [kworker/u16:3-writeback]
5     0  258207       1  20   0   8304  1976 do_pol Ss   ?          0:00 /usr/bin/conmon --api-version 1 -c 8f2ea6310cd353f33c1478e7503e6f9c52ea7620eec6612361cd1c39bc0392ae -u 8f2ea6310cd353f33c1478e7503e6f9c52ea7620eec6612361cd1c39bc0392ae -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/8f2ea6310cd353f33c1478e7503e6f9c52ea7620eec6612361cd1c39bc0392ae/userdata -p /run/containers/storage/overlay-containers/8f2ea6310cd353f33c1478e7503e6f9c52ea7620eec6612361cd1c39bc0392ae/userdata/pidfile -n neutron_sriov_agent --exit-dir /run/libpod/exits --full-attach -s -l journald --log-level warning --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/8f2ea6310cd353f33c1478e7503e6f9c52ea7620eec6612361cd1c39bc0392ae/userdata/oci-log --conmon-pidfile /run/neutron_sriov_agent.pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg  --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --storage-driver --exit-command-arg overlay --exit-command-arg --storage-opt --exit-command-arg overlay.mountopt=nodev,metacopy=on --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg 8f2ea6310cd353f33c1478e7503e6f9c52ea7620eec6612361cd1c39bc0392ae
4 42435  258209  258207  20   0   2512   932 do_sig Ss   ?          0:00 dumb-init --single-child -- kolla_start
4 42435  258211  258209  20   0 213912 132208 ep_pol S   ?          0:06 neutron-sriov-nic-agent (/usr/bin/python3 /usr/bin/neutron-sriov-nic-agent)
5     0  265537       1  20   0   8304  2012 do_pol Ss   ?          0:00 /usr/bin/conmon --api-version 1 -c 1b2c23eafd230afc8d091ee7b7cedbb3afbedca0796f62e8cc47ca513d5981f8 -u 1b2c23eafd230afc8d091ee7b7cedbb3afbedca0796f62e8cc47ca513d5981f8 -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/1b2c23eafd230afc8d091ee7b7cedbb3afbedca0796f62e8cc47ca513d5981f8/userdata -p /run/containers/storage/overlay-containers/1b2c23eafd230afc8d091ee7b7cedbb3afbedca0796f62e8cc47ca513d5981f8/userdata/pidfile -n neutron_dhcp_agent --exit-dir /run/libpod/exits --full-attach -s -l journald --log-level warning --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/1b2c23eafd230afc8d091ee7b7cedbb3afbedca0796f62e8cc47ca513d5981f8/userdata/oci-log --conmon-pidfile /run/neutron_dhcp_agent.pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg  --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --storage-driver --exit-command-arg overlay --exit-command-arg --storage-opt --exit-command-arg overlay.mountopt=nodev,metacopy=on --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg 1b2c23eafd230afc8d091ee7b7cedbb3afbedca0796f62e8cc47ca513d5981f8
4     0  265539  265537  20   0   2512   956 do_sig Ss   ?          0:00 dumb-init --single-child -- kolla_start
4     0  265541  265539  20   0 226296 138184 ep_pol S   ?          0:16 neutron-dhcp-agent (/usr/bin/python3 /usr/bin/neutron-dhcp-agent)
1     0  267636       2  20   0      0     0 worker I    ?          0:00 [kworker/1:2-events]
5     0  282206       1  20   0   8304  1952 do_pol Ss   ?          0:00 /usr/bin/conmon --api-version 1 -c 2129e353a5d4171f586ce5c762891941b44f7611471f84d84a42a19d9df53b3b -u 2129e353a5d4171f586ce5c762891941b44f7611471f84d84a42a19d9df53b3b -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/2129e353a5d4171f586ce5c762891941b44f7611471f84d84a42a19d9df53b3b/userdata -p /run/containers/storage/overlay-containers/2129e353a5d4171f586ce5c762891941b44f7611471f84d84a42a19d9df53b3b/userdata/pidfile -n nova_compute --exit-dir /run/libpod/exits --full-attach -s -l journald --log-level warning --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/2129e353a5d4171f586ce5c762891941b44f7611471f84d84a42a19d9df53b3b/userdata/oci-log --conmon-pidfile /run/nova_compute.pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg  --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --storage-driver --exit-command-arg overlay --exit-command-arg --storage-opt --exit-command-arg overlay.mountopt=nodev,metacopy=on --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg 2129e353a5d4171f586ce5c762891941b44f7611471f84d84a42a19d9df53b3b
4 42436  282208  282206  20   0   2512   960 do_sig Ss   ?          0:00 dumb-init --single-child -- kolla_start
4 42436  282211  282208  20   0 1833268 184056 ep_pol Sl ?          0:18 /usr/bin/python3 /usr/bin/nova-compute
5     0  282522  282206  20   0 136784 46732 unix_s Sl   ?          0:00 /usr/bin/python3 /bin/privsep-helper --config-file /etc/nova/nova.conf --config-file /etc/nova/nova-compute.conf --config-dir /etc/nova/nova.conf.d --privsep_context vif_plug_ovs.privsep.vif_plug --privsep_sock_path /tmp/tmp5o6kqvot/privsep.sock
6   107  285069       1  20   0 3554848 363628 do_pol Sl ?          1:23 /usr/libexec/qemu-kvm -name guest=instance-00000003,debug-threads=on -S -object {"qom-type":"secret","id":"masterKey0","format":"raw","file":"/var/lib/libvirt/qemu/domain-2-instance-00000003/master-key.aes"} -machine pc-q35-rhel9.0.0,usb=off,dump-guest-core=off,memory-backend=pc.ram,hpet=off,acpi=on -accel kvm -cpu EPYC-Rome,x2apic=on,tsc-deadline=on,hypervisor=on,tsc-adjust=on,spec-ctrl=on,stibp=on,ssbd=on,cmp-legacy=on,overflow-recov=on,succor=on,ibrs=on,amd-ssbd=on,virt-ssbd=on,lbrv=on,tsc-scale=on,vmcb-clean=on,pause-filter=on,pfthreshold=on,svme-addr-chk=on,lfence-always-serializing=on,xsaves=off -m size=524288k -object {"qom-type":"memory-backend-ram","id":"pc.ram","size":536870912} -overcommit mem-lock=off -smp 1,sockets=1,dies=1,clusters=1,cores=1,threads=1 -uuid c2a7d92b-952f-46a7-8a6a-3322a48fcf4b -smbios type=1,manufacturer=RDO,product=OpenStack Compute,version=27.5.2-0.20260220085704.5cfeecb.el9,serial=c2a7d92b-952f-46a7-8a6a-3322a48fcf4b,uuid=c2a7d92b-952f-46a7-8a6a-3322a48fcf4b,family=Virtual Machine -no-user-config -nodefaults -chardev socket,id=charmonitor,fd=26,server=on,wait=off -mon chardev=charmonitor,id=monitor,mode=control -rtc base=utc,driftfix=slew -global kvm-pit.lost_tick_policy=delay -no-shutdown -boot strict=on -device {"driver":"pcie-root-port","port":16,"chassis":1,"id":"pci.1","bus":"pcie.0","multifunction":true,"addr":"0x2"} -device {"driver":"pcie-root-port","port":17,"chassis":2,"id":"pci.2","bus":"pcie.0","addr":"0x2.0x1"} -device {"driver":"pcie-root-port","port":18,"chassis":3,"id":"pci.3","bus":"pcie.0","addr":"0x2.0x2"} -device {"driver":"pcie-root-port","port":19,"chassis":4,"id":"pci.4","bus":"pcie.0","addr":"0x2.0x3"} -device {"driver":"pcie-root-port","port":20,"chassis":5,"id":"pci.5","bus":"pcie.0","addr":"0x2.0x4"} -device {"driver":"pcie-root-port","port":21,"chassis":6,"id":"pci.6","bus":"pcie.0","addr":"0x2.0x5"} -device {"driver":"pcie-root-port","port":22,"chassis":7,"id":"pci.7","bus":"pcie.0","addr":"0x2.0x6"} -device {"driver":"pcie-root-port","port":23,"chassis":8,"id":"pci.8","bus":"pcie.0","addr":"0x2.0x7"} -device {"driver":"pcie-root-port","port":24,"chassis":9,"id":"pci.9","bus":"pcie.0","multifunction":true,"addr":"0x3"} -device {"driver":"pcie-root-port","port":25,"chassis":10,"id":"pci.10","bus":"pcie.0","addr":"0x3.0x1"} -device {"driver":"pcie-root-port","port":26,"chassis":11,"id":"pci.11","bus":"pcie.0","addr":"0x3.0x2"} -device {"driver":"pcie-root-port","port":27,"chassis":12,"id":"pci.12","bus":"pcie.0","addr":"0x3.0x3"} -device {"driver":"pcie-root-port","port":28,"chassis":13,"id":"pci.13","bus":"pcie.0","addr":"0x3.0x4"} -device {"driver":"pcie-root-port","port":29,"chassis":14,"id":"pci.14","bus":"pcie.0","addr":"0x3.0x5"} -device {"driver":"pcie-root-port","port":30,"chassis":15,"id":"pci.15","bus":"pcie.0","addr":"0x3.0x6"} -device {"driver":"pcie-root-port","port":31,"chassis":16,"id":"pci.16","bus":"pcie.0","addr":"0x3.0x7"} -device {"driver":"pcie-root-port","port":32,"chassis":17,"id":"pci.17","bus":"pcie.0","multifunction":true,"addr":"0x4"} -device {"driver":"pcie-root-port","port":33,"chassis":18,"id":"pci.18","bus":"pcie.0","addr":"0x4.0x1"} -device {"driver":"pcie-root-port","port":34,"chassis":19,"id":"pci.19","bus":"pcie.0","addr":"0x4.0x2"} -device {"driver":"pcie-root-port","port":35,"chassis":20,"id":"pci.20","bus":"pcie.0","addr":"0x4.0x3"} -device {"driver":"pcie-root-port","port":36,"chassis":21,"id":"pci.21","bus":"pcie.0","addr":"0x4.0x4"} -device {"driver":"pcie-root-port","port":37,"chassis":22,"id":"pci.22","bus":"pcie.0","addr":"0x4.0x5"} -device {"driver":"pcie-root-port","port":38,"chassis":23,"id":"pci.23","bus":"pcie.0","addr":"0x4.0x6"} -device {"driver":"pcie-root-port","port":39,"chassis":24,"id":"pci.24","bus":"pcie.0","addr":"0x4.0x7"} -device {"driver":"pcie-root-port","port":40,"chassis":25,"id":"pci.25","bus":"pcie.0","addr":"0x5"} -device {"driver":"pcie-pci-bridge","id":"pci.26","bus":"pci.1","addr":"0x0"} -device {"driver":"piix3-usb-uhci","id":"usb","bus":"pci.26","addr":"0x1"} -object {"qom-type":"secret","id":"libvirt-2-storage-auth-secret0","data":"jSqig74hSLcfqsgToHB3IYupI998eLlixhm64NOz9qk=","keyid":"masterKey0","iv":"tcZ+ld0uHPUPboZVj3OyQg==","format":"base64"} -blockdev {"driver":"rbd","pool":"vms","image":"c2a7d92b-952f-46a7-8a6a-3322a48fcf4b_disk","server":[{"host":"172.18.0.103","port":"6789"},{"host":"172.18.0.105","port":"6789"},{"host":"172.18.0.104","port":"6789"}],"user":"openstack","auth-client-required":["cephx","none"],"key-secret":"libvirt-2-storage-auth-secret0","node-name":"libvirt-2-storage","read-only":false,"cache":{"direct":true,"no-flush":false}} -device {"driver":"virtio-blk-pci","bus":"pci.3","addr":"0x0","drive":"libvirt-2-storage","id":"virtio-disk0","bootindex":1,"write-cache":"on"} -object {"qom-type":"secret","id":"libvirt-1-storage-auth-secret0","data":"QdIXPQqzZpSRgv8JK15jOwQ7pmRz5BXwm9kX54KZJwU=","keyid":"masterKey0","iv":"SaaNh9oQPew3Qd7O54RdpQ==","format":"base64"} -blockdev {"driver":"rbd","pool":"vms","image":"c2a7d92b-952f-46a7-8a6a-3322a48fcf4b_disk.eph0","server":[{"host":"172.18.0.103","port":"6789"},{"host":"172.18.0.105","port":"6789"},{"host":"172.18.0.104","port":"6789"}],"user":"openstack","auth-client-required":["cephx","none"],"key-secret":"libvirt-1-storage-auth-secret0","node-name":"libvirt-1-storage","read-only":false,"cache":{"direct":true,"no-flush":false}} -device {"driver":"virtio-blk-pci","bus":"pci.4","addr":"0x0","drive":"libvirt-1-storage","id":"virtio-disk1","write-cache":"on"} -netdev {"type":"tap","fd":"28","vhost":true,"vhostfd":"30","id":"hostnet0"} -device {"driver":"virtio-net-pci","rx_queue_size":512,"host_mtu":1292,"netdev":"hostnet0","id":"net0","mac":"fa:16:3e:a0:9d:00","bus":"pci.2","addr":"0x0"} -add-fd set=0,fd=27,opaque=serial0-log -chardev pty,id=charserial0,logfile=/dev/fdset/0,logappend=on -device {"driver":"isa-serial","chardev":"charserial0","id":"serial0","index":0} -device {"driver":"usb-tablet","id":"input0","bus":"usb.0","port":"1"} -device {"driver":"usb-kbd","id":"input1","bus":"usb.0","port":"2"} -audiodev {"id":"audio1","driver":"none"} -vnc [::0]:0,audiodev=audio1 -device {"driver":"virtio-vga","id":"video0","max_outputs":1,"bus":"pcie.0","addr":"0x1"} -global ICH9-LPC.noreboot=off -watchdog-action reset -device {"driver":"virtio-balloon-pci","id":"balloon0","bus":"pci.5","addr":"0x0"} -object {"qom-type":"rng-random","id":"objrng0","filename":"/dev/urandom"} -device {"driver":"virtio-rng-pci","rng":"objrng0","id":"rng0","bus":"pci.6","addr":"0x0"} -device {"driver":"vmcoreinfo"} -sandbox on,obsolete=deny,elevateprivileges=deny,spawn=deny,resourcecontrol=deny -msg timestamp=on
1     0  285145       2   0 -20      0     0 rescue I<   ?          0:00 [kvm]
1     0  285146       2  20   0      0     0 kvm_nx S    ?          0:00 [kvm-nx-lpage-re]
1     0  285147       2  20   0      0     0 vhost_ S    ?          0:00 [vhost-285069]
1     0  285150       2  20   0      0     0 kthrea S    ?          0:00 [kvm-pit/285069]
5     0  285191  163567  20   0   8304  1976 do_pol Ss   ?          0:00 /usr/bin/conmon --api-version 1 -c 983cee9e3a46a32c62717c639a883cc183806af2504f520a888928b166a0a907 -u 983cee9e3a46a32c62717c639a883cc183806af2504f520a888928b166a0a907 -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/983cee9e3a46a32c62717c639a883cc183806af2504f520a888928b166a0a907/userdata -p /run/containers/storage/overlay-containers/983cee9e3a46a32c62717c639a883cc183806af2504f520a888928b166a0a907/userdata/pidfile -n neutron-haproxy-ovnmeta-9da5b53d-3184-450f-9a5b-bdba1a6c9f6d --exit-dir /run/libpod/exits --full-attach -s -l journald --log-level warning --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/983cee9e3a46a32c62717c639a883cc183806af2504f520a888928b166a0a907/userdata/oci-log --conmon-pidfile /run/containers/storage/overlay-containers/983cee9e3a46a32c62717c639a883cc183806af2504f520a888928b166a0a907/userdata/conmon.pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg  --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --storage-driver --exit-command-arg overlay --exit-command-arg --storage-opt --exit-command-arg overlay.mountopt=nodev,metacopy=on --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg 983cee9e3a46a32c62717c639a883cc183806af2504f520a888928b166a0a907
4     0  285193  285191  20   0   2512   912 do_sig Ss   ?          0:00 dumb-init --single-child -- /bin/bash -c exec /usr/sbin/haproxy -Ws -f /var/lib/neutron/ovn-metadata-proxy/9da5b53d-3184-450f-9a5b-bdba1a6c9f6d.conf
4     0  285195  285193  20   0  97596 12544 ep_pol S    ?          0:00 /usr/sbin/haproxy -Ws -f /var/lib/neutron/ovn-metadata-proxy/9da5b53d-3184-450f-9a5b-bdba1a6c9f6d.conf
5     0  285197  285195  20   0 541504  8144 ep_pol Sl   ?          0:00 /usr/sbin/haproxy -Ws -f /var/lib/neutron/ovn-metadata-proxy/9da5b53d-3184-450f-9a5b-bdba1a6c9f6d.conf
1     0  286094       2  20   0      0     0 worker I    ?          0:00 [kworker/6:0-xfs-conv/vda4]
1     0  286374       2  20   0      0     0 worker I    ?          0:00 [kworker/7:2-events]
5     0  286873       1  20   0   8304  1848 do_pol Ss   ?          0:00 /usr/bin/conmon --api-version 1 -c 35c397f376b989389f5487b314924f02dd848f945c70656bc276f291652231c2 -u 35c397f376b989389f5487b314924f02dd848f945c70656bc276f291652231c2 -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/35c397f376b989389f5487b314924f02dd848f945c70656bc276f291652231c2/userdata -p /run/containers/storage/overlay-containers/35c397f376b989389f5487b314924f02dd848f945c70656bc276f291652231c2/userdata/pidfile -n ceph-f1fea371-cb69-578d-a3d0-b5c472a84b46-mds-mds-np0005626463-qcthuc --exit-dir /run/libpod/exits --full-attach -l journald --log-level warning --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/35c397f376b989389f5487b314924f02dd848f945c70656bc276f291652231c2/userdata/oci-log --conmon-pidfile /run/ceph-f1fea371-cb69-578d-a3d0-b5c472a84b46@mds.mds.np0005626463.qcthuc.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg  --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --storage-driver --exit-command-arg overlay --exit-command-arg --storage-opt --exit-command-arg overlay.mountopt=nodev,metacopy=on --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --rm --exit-command-arg 35c397f376b989389f5487b314924f02dd848f945c70656bc276f291652231c2
4     0  286875  286873  20   0   1084     4 signal Ss   ?          0:00 /run/podman-init -- /usr/bin/ceph-mds -n mds.mds.np0005626463.qcthuc -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false
4   167  286877  286875  20   0 320892 58640 futex_ Sl   ?          0:08 /usr/bin/ceph-mds -n mds.mds.np0005626463.qcthuc -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false
1     0  287534       2  20   0      0     0 worker I    ?          0:00 [kworker/5:2-cgroup_destroy]
5     0  288032       1  20   0   8304  1952 do_pol Ss   ?          0:00 /usr/bin/conmon --api-version 1 -c bb57b00fb4ffcc023092512e750b21cd585f41b4202d5c160e4cae01fa164fb4 -u bb57b00fb4ffcc023092512e750b21cd585f41b4202d5c160e4cae01fa164fb4 -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/bb57b00fb4ffcc023092512e750b21cd585f41b4202d5c160e4cae01fa164fb4/userdata -p /run/containers/storage/overlay-containers/bb57b00fb4ffcc023092512e750b21cd585f41b4202d5c160e4cae01fa164fb4/userdata/pidfile -n ceph-f1fea371-cb69-578d-a3d0-b5c472a84b46-mgr-np0005626463-wtksup --exit-dir /run/libpod/exits --full-attach -l journald --log-level warning --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/bb57b00fb4ffcc023092512e750b21cd585f41b4202d5c160e4cae01fa164fb4/userdata/oci-log --conmon-pidfile /run/ceph-f1fea371-cb69-578d-a3d0-b5c472a84b46@mgr.np0005626463.wtksup.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg  --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --storage-driver --exit-command-arg overlay --exit-command-arg --storage-opt --exit-command-arg overlay.mountopt=nodev,metacopy=on --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --rm --exit-command-arg bb57b00fb4ffcc023092512e750b21cd585f41b4202d5c160e4cae01fa164fb4
4     0  288034  288032  20   0   1084     0 signal Ss   ?          0:00 /run/podman-init -- /usr/bin/ceph-mgr -n mgr.np0005626463.wtksup -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false
4   167  288036  288034  20   0 1793692 484200 futex_ Sl ?          0:20 /usr/bin/ceph-mgr -n mgr.np0005626463.wtksup -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false
1     0  291481       2  20   0      0     0 worker I    ?          0:00 [kworker/7:3-events]
1     0  294155       1  20   0   8304  1980 do_pol Ss   ?          0:00 /usr/bin/conmon --api-version 1 -c a517a74ed21c459483d3bfd4abd622efb0a723f6a4c7629a8c105935a56ca753 -u a517a74ed21c459483d3bfd4abd622efb0a723f6a4c7629a8c105935a56ca753 -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/a517a74ed21c459483d3bfd4abd622efb0a723f6a4c7629a8c105935a56ca753/userdata -p /run/containers/storage/overlay-containers/a517a74ed21c459483d3bfd4abd622efb0a723f6a4c7629a8c105935a56ca753/userdata/pidfile -n ceph-f1fea371-cb69-578d-a3d0-b5c472a84b46-mon-np0005626463 --exit-dir /run/libpod/exits --full-attach -l journald --log-level warning --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/a517a74ed21c459483d3bfd4abd622efb0a723f6a4c7629a8c105935a56ca753/userdata/oci-log --conmon-pidfile /run/ceph-f1fea371-cb69-578d-a3d0-b5c472a84b46@mon.np0005626463.service-pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg  --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --storage-driver --exit-command-arg overlay --exit-command-arg --storage-opt --exit-command-arg overlay.mountopt=nodev,metacopy=on --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --rm --exit-command-arg a517a74ed21c459483d3bfd4abd622efb0a723f6a4c7629a8c105935a56ca753
4     0  294158  294155  20   0   1084     4 signal Ss   ?          0:00 /run/podman-init -- /usr/bin/ceph-mon -n mon.np0005626463 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false
4   167  294160  294158  20   0 358740 141928 futex_ Sl  ?          0:36 /usr/bin/ceph-mon -n mon.np0005626463 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false
4     0  306377  187670  20   0  18948 10384 do_pol Ss   ?          0:00 sshd: ceph-admin [priv]
5  1002  306380  306377  20   0  19060  7404 do_sel S    ?          0:00 sshd: ceph-admin@notty
5     0  307935  265537  20   0 324312 113064 unix_s Sl  ?          0:15 /usr/bin/python3 /bin/privsep-helper --config-file /etc/neutron/neutron.conf --config-dir /etc/neutron.conf.d --privsep_context neutron.privileged.default --privsep_sock_path /tmp/tmpu3w1xssi/privsep.sock
1     0  307963  265537  20   0 174284 103072 unix_s Sl  ?          0:01 /usr/bin/python3 /bin/privsep-helper --config-file /etc/neutron/neutron.conf --config-dir /etc/neutron.conf.d --privsep_context neutron.privileged.namespace_cmd --privsep_sock_path /tmp/tmpmznh5rez/privsep.sock
1     0  307975  265537  20   0 249556 108016 unix_s Sl  ?          0:03 /usr/bin/python3 /bin/privsep-helper --config-file /etc/neutron/neutron.conf --config-dir /etc/neutron.conf.d --privsep_context neutron.privileged.link_cmd --privsep_sock_path /tmp/tmpu8soh03c/privsep.sock
1     0  308097  265537  20   0   8304  1976 do_pol Ss   ?          0:00 /usr/bin/conmon --api-version 1 -c f4fb933ef414244081c84e33388df31570fc83c922a1140ac94ef0bca735ba11 -u f4fb933ef414244081c84e33388df31570fc83c922a1140ac94ef0bca735ba11 -r /usr/bin/crun -b /var/lib/containers/storage/overlay-containers/f4fb933ef414244081c84e33388df31570fc83c922a1140ac94ef0bca735ba11/userdata -p /run/containers/storage/overlay-containers/f4fb933ef414244081c84e33388df31570fc83c922a1140ac94ef0bca735ba11/userdata/pidfile -n neutron-dnsmasq-qdhcp-f52ac7ca-e197-490d-a7bf-412806b20437 --exit-dir /run/libpod/exits --full-attach -s -l journald --log-level warning --runtime-arg --log-format=json --runtime-arg --log --runtime-arg=/run/containers/storage/overlay-containers/f4fb933ef414244081c84e33388df31570fc83c922a1140ac94ef0bca735ba11/userdata/oci-log --conmon-pidfile /run/containers/storage/overlay-containers/f4fb933ef414244081c84e33388df31570fc83c922a1140ac94ef0bca735ba11/userdata/conmon.pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /var/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /run/containers/storage --exit-command-arg --log-level --exit-command-arg warning --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg  --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /var/lib/containers/storage/volumes --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --storage-driver --exit-command-arg overlay --exit-command-arg --storage-opt --exit-command-arg overlay.mountopt=nodev,metacopy=on --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg container --exit-command-arg cleanup --exit-command-arg f4fb933ef414244081c84e33388df31570fc83c922a1140ac94ef0bca735ba11
4     0  308099  308097  20   0   2512   964 do_sig Ss   ?          0:00 dumb-init --single-child -- /usr/sbin/dnsmasq -k --no-hosts --no-resolv --pid-file=/var/lib/neutron/dhcp/f52ac7ca-e197-490d-a7bf-412806b20437/pid --dhcp-hostsfile=/var/lib/neutron/dhcp/f52ac7ca-e197-490d-a7bf-412806b20437/host --addn-hosts=/var/lib/neutron/dhcp/f52ac7ca-e197-490d-a7bf-412806b20437/addn_hosts --dhcp-optsfile=/var/lib/neutron/dhcp/f52ac7ca-e197-490d-a7bf-412806b20437/opts --dhcp-leasefile=/var/lib/neutron/dhcp/f52ac7ca-e197-490d-a7bf-412806b20437/leases --dhcp-match=set:ipxe,175 --dhcp-userclass=set:ipxe6,iPXE --local-service --bind-dynamic --dhcp-range=set:subnet-3c8bc423-a56a-416a-8ca9-609d2525abe5,192.168.199.0,static,255.255.255.0,86400s --dhcp-option-force=option:mtu,1442 --dhcp-lease-max=256 --conf-file=/dev/null --domain=openstacklocal
4   997  308101  308099  20   0  10228  4472 do_pol S    ?          0:00 /usr/sbin/dnsmasq -k --no-hosts --no-resolv --pid-file=/var/lib/neutron/dhcp/f52ac7ca-e197-490d-a7bf-412806b20437/pid --dhcp-hostsfile=/var/lib/neutron/dhcp/f52ac7ca-e197-490d-a7bf-412806b20437/host --addn-hosts=/var/lib/neutron/dhcp/f52ac7ca-e197-490d-a7bf-412806b20437/addn_hosts --dhcp-optsfile=/var/lib/neutron/dhcp/f52ac7ca-e197-490d-a7bf-412806b20437/opts --dhcp-leasefile=/var/lib/neutron/dhcp/f52ac7ca-e197-490d-a7bf-412806b20437/leases --dhcp-match=set:ipxe,175 --dhcp-userclass=set:ipxe6,iPXE --local-service --bind-dynamic --dhcp-range=set:subnet-3c8bc423-a56a-416a-8ca9-609d2525abe5,192.168.199.0,static,255.255.255.0,86400s --dhcp-option-force=option:mtu,1442 --dhcp-lease-max=256 --conf-file=/dev/null --domain=openstacklocal
1     0  308469       2  20   0      0     0 worker I    ?          0:00 [kworker/u16:2-writeback]
1     0  308470       2  20   0      0     0 worker I    ?          0:01 [kworker/u16:5-flush-252:0]
1     0  310305       2  20   0      0     0 worker I    ?          0:00 [kworker/0:1-rcu_gp]
1     0  311411       2  20   0      0     0 worker I    ?          0:01 [kworker/u16:6-events_unbound]
1     0  315762       2  20   0      0     0 worker I    ?          0:00 [kworker/6:2-xfs-conv/vda4]
1     0  317893       2  20   0      0     0 worker I    ?          0:00 [kworker/5:1-events]
1     0  318717       2  20   0      0     0 worker I    ?          0:00 [kworker/2:2-events]
1     0  318881       2  20   0      0     0 worker I    ?          0:00 [kworker/1:1-ata_sff]
1     0  318921       2  20   0      0     0 worker I    ?          0:01 [kworker/4:2-ata_sff]
1     0  319063       2  20   0      0     0 worker I    ?          0:00 [kworker/2:3-cgroup_destroy]
1     0  319295       2  20   0      0     0 worker I    ?          0:00 [kworker/2:4-cgroup_destroy]
1     0  319781       2  20   0      0     0 worker I    ?          0:00 [kworker/0:0-events]
1     0  320737       2  20   0      0     0 worker I    ?          0:00 [kworker/3:3-events_freezable_power_]
1     0  322412       2  20   0      0     0 worker I    ?          0:02 [kworker/4:3-events]
1     0  323784       2  20   0      0     0 worker I    ?          0:00 [kworker/5:0-events]
1     0  323866       2  20   0      0     0 worker I    ?          0:00 [kworker/0:2-ata_sff]
1     0  323915       2  20   0      0     0 worker I    ?          0:00 [kworker/1:0-events]
1     0  323916       2  20   0      0     0 worker I    ?          0:00 [kworker/2:1-events]
1     0  324528       2  20   0      0     0 worker I    ?          0:00 [kworker/5:3-events]
1     0  324739       2  20   0      0     0 worker I    ?          0:00 [kworker/6:1-xfs-conv/vda4]
1     0  324823       2  20   0      0     0 worker I    ?          0:00 [kworker/4:0-ata_sff]
4     0  324865  187670  20   0  18796 11392 do_pol Ss   ?          0:00 sshd: zuul [priv]
5  1000  324868  324865  20   0  19000  6604 do_sel S    ?          0:00 sshd: zuul@notty
4     0  324869  324868  20   0  16772  8392 do_pol Ss   ?          0:00 sudo bash -c rm -rf /var/tmp/sos-osp && mkdir /var/tmp/sos-osp && sos report --batch --all-logs --tmp-dir=/var/tmp/sos-osp  -p container,openstack_edpm,system,storage,virt
4     0  324886  324869  20   0 663668 66376 futex_ Sl   ?          0:11 /usr/bin/python3 -s /sbin/sos report --batch --all-logs --tmp-dir=/var/tmp/sos-osp -p container,openstack_edpm,system,storage,virt
1     0  325315       2  20   0      0     0 worker I    ?          0:00 [kworker/3:1-xfs-inodegc/vda4]
1     0  325355       2  20   0      0     0 worker I    ?          0:00 [kworker/7:0-cgroup_destroy]
1     0  325360       2   0 -20      0     0 worker I<   ?          0:00 [kworker/3:2H]
1     0  325703       2  20   0      0     0 worker I    ?          0:00 [kworker/7:1-events]
1     0  327163       2  20   0      0     0 worker I    ?          0:00 [kworker/6:3-xfs-conv/vda4]
1     0  327164       2  20   0      0     0 worker I    ?          0:00 [kworker/6:4-xfs-conv/vda4]
1     0  327165       2  20   0      0     0 worker I    ?          0:00 [kworker/6:5-xfs-conv/vda4]
1     0  327166       2  20   0      0     0 worker I    ?          0:00 [kworker/6:6-xfs-conv/vda4]
1     0  327167       2  20   0      0     0 worker I    ?          0:00 [kworker/6:7-xfs-conv/vda4]
1     0  327168       2  20   0      0     0 worker I    ?          0:00 [kworker/6:8-xfs-conv/vda4]
1     0  327169       2  20   0      0     0 worker I    ?          0:00 [kworker/6:9-xfs-conv/vda4]
1     0  327170       2  20   0      0     0 worker I    ?          0:00 [kworker/6:10-xfs-conv/vda4]
1     0  327171       2  20   0      0     0 worker I    ?          0:00 [kworker/6:11-xfs-conv/vda4]
1     0  327172       2  20   0      0     0 worker I    ?          0:00 [kworker/6:12-xfs-conv/vda4]
1     0  327173       2  20   0      0     0 worker I    ?          0:00 [kworker/6:13-xfs-conv/vda4]
1     0  327174       2  20   0      0     0 worker I    ?          0:00 [kworker/6:14-xfs-conv/vda4]
1     0  327175       2  20   0      0     0 worker I    ?          0:00 [kworker/6:15-xfs-conv/vda4]
1     0  327176       2  20   0      0     0 worker I    ?          0:00 [kworker/6:16-xfs-conv/vda4]
1     0  327177       2  20   0      0     0 worker I    ?          0:00 [kworker/6:17-xfs-conv/vda4]
1     0  327178       2  20   0      0     0 worker I    ?          0:00 [kworker/6:18-xfs-conv/vda4]
1     0  327179       2  20   0      0     0 worker I    ?          0:00 [kworker/6:19-xfs-conv/vda4]
1     0  327180       2  20   0      0     0 worker I    ?          0:00 [kworker/6:20-xfs-conv/vda4]
1     0  327181       2  20   0      0     0 worker I    ?          0:00 [kworker/6:21-xfs-conv/vda4]
1     0  327182       2  20   0      0     0 worker I    ?          0:00 [kworker/6:22-xfs-conv/vda4]
1     0  327183       2  20   0      0     0 worker I    ?          0:00 [kworker/6:23-xfs-conv/vda4]
1     0  327184       2  20   0      0     0 worker I    ?          0:00 [kworker/6:24-xfs-conv/vda4]
1     0  327185       2  20   0      0     0 worker I    ?          0:00 [kworker/6:25-xfs-conv/vda4]
1     0  327186       2  20   0      0     0 worker I    ?          0:00 [kworker/6:26-xfs-conv/vda4]
1     0  327187       2  20   0      0     0 worker I    ?          0:00 [kworker/6:27-xfs-conv/vda4]
1     0  327188       2  20   0      0     0 worker I    ?          0:00 [kworker/6:28-xfs-conv/vda4]
1     0  327189       2  20   0      0     0 worker I    ?          0:00 [kworker/6:29-xfs-conv/vda4]
1     0  327190       2  20   0      0     0 worker I    ?          0:00 [kworker/6:30-xfs-conv/vda4]
1     0  327191       2  20   0      0     0 worker I    ?          0:00 [kworker/6:31-xfs-conv/vda4]
1     0  327192       2  20   0      0     0 worker I    ?          0:00 [kworker/6:32-xfs-conv/vda4]
1     0  327193       2  20   0      0     0 worker I    ?          0:00 [kworker/6:33-xfs-conv/vda4]
1     0  327194       2  20   0      0     0 worker I    ?          0:00 [kworker/6:34-xfs-conv/vda4]
1     0  327195       2  20   0      0     0 worker I    ?          0:00 [kworker/6:35-xfs-conv/vda4]
1     0  327196       2  20   0      0     0 worker I    ?          0:00 [kworker/6:36-xfs-conv/vda4]
1     0  327197       2  20   0      0     0 worker I    ?          0:00 [kworker/6:37-xfs-conv/vda4]
1     0  327198       2  20   0      0     0 worker I    ?          0:00 [kworker/6:38-xfs-conv/vda4]
1     0  327199       2  20   0      0     0 worker I    ?          0:00 [kworker/6:39-xfs-conv/vda4]
1     0  327200       2  20   0      0     0 worker I    ?          0:00 [kworker/6:40-xfs-conv/vda4]
1     0  327201       2  20   0      0     0 worker I    ?          0:00 [kworker/6:41-xfs-conv/vda4]
1     0  327202       2  20   0      0     0 worker I    ?          0:00 [kworker/6:42-xfs-conv/vda4]
1     0  327203       2  20   0      0     0 worker I    ?          0:00 [kworker/6:43-xfs-conv/vda4]
1     0  327204       2  20   0      0     0 worker I    ?          0:00 [kworker/6:44-xfs-conv/vda4]
1     0  327205       2  20   0      0     0 worker I    ?          0:00 [kworker/6:45-xfs-conv/vda4]
1     0  327206       2  20   0      0     0 worker I    ?          0:00 [kworker/6:46-xfs-conv/vda4]
1     0  327207       2  20   0      0     0 worker I    ?          0:00 [kworker/6:47-xfs-conv/vda4]
1     0  327208       2  20   0      0     0 worker I    ?          0:00 [kworker/6:48-xfs-conv/vda4]
1     0  327209       2  20   0      0     0 worker I    ?          0:00 [kworker/6:49-xfs-conv/vda4]
1     0  327210       2  20   0      0     0 worker I    ?          0:00 [kworker/6:50-xfs-conv/vda4]
1     0  327211       2  20   0      0     0 worker I    ?          0:00 [kworker/6:51-xfs-conv/vda4]
1     0  327212       2  20   0      0     0 worker I    ?          0:00 [kworker/6:52-xfs-conv/vda4]
1     0  327213       2  20   0      0     0 worker I    ?          0:00 [kworker/6:53-xfs-conv/vda4]
1     0  327215       2  20   0      0     0 worker I    ?          0:00 [kworker/6:54-xfs-conv/vda4]
1     0  327216       2  20   0      0     0 worker I    ?          0:00 [kworker/6:55-xfs-conv/vda4]
1     0  327217       2  20   0      0     0 worker I    ?          0:00 [kworker/6:56-xfs-conv/vda4]
1     0  327218       2  20   0      0     0 worker I    ?          0:00 [kworker/6:57-xfs-conv/vda4]
1     0  327219       2  20   0      0     0 worker I    ?          0:00 [kworker/6:58-xfs-conv/vda4]
1     0  327220       2  20   0      0     0 worker I    ?          0:00 [kworker/6:59-xfs-conv/vda4]
1     0  327221       2  20   0      0     0 worker I    ?          0:00 [kworker/6:60-xfs-conv/vda4]
1     0  327222       2  20   0      0     0 worker I    ?          0:00 [kworker/6:61-xfs-conv/vda4]
1     0  327223       2  20   0      0     0 worker I    ?          0:00 [kworker/6:62-xfs-conv/vda4]
1     0  327224       2  20   0      0     0 worker I    ?          0:00 [kworker/6:63-xfs-conv/vda4]
1     0  327225       2  20   0      0     0 worker I    ?          0:00 [kworker/6:64-xfs-conv/vda4]
1     0  327226       2  20   0      0     0 worker I    ?          0:00 [kworker/6:65-xfs-conv/vda4]
1     0  327227       2  20   0      0     0 worker I    ?          0:00 [kworker/6:66-xfs-conv/vda4]
1     0  327228       2  20   0      0     0 worker I    ?          0:00 [kworker/6:67-xfs-conv/vda4]
1     0  327229       2  20   0      0     0 worker I    ?          0:00 [kworker/6:68-xfs-conv/vda4]
1     0  327230       2  20   0      0     0 worker I    ?          0:00 [kworker/6:69-xfs-conv/vda4]
1     0  327231       2  20   0      0     0 worker I    ?          0:00 [kworker/6:70-xfs-conv/vda4]
1     0  327232       2  20   0      0     0 worker I    ?          0:00 [kworker/6:71-xfs-conv/vda4]
1     0  327233       2  20   0      0     0 worker I    ?          0:00 [kworker/6:72-xfs-conv/vda4]
1     0  327234       2  20   0      0     0 worker I    ?          0:00 [kworker/6:73-xfs-conv/vda4]
1     0  327235       2  20   0      0     0 worker I    ?          0:00 [kworker/6:74-xfs-conv/vda4]
1     0  327236       2  20   0      0     0 worker I    ?          0:00 [kworker/6:75-xfs-conv/vda4]
1     0  327237       2  20   0      0     0 worker I    ?          0:00 [kworker/6:76-xfs-conv/vda4]
1     0  327238       2  20   0      0     0 worker I    ?          0:00 [kworker/6:77-xfs-conv/vda4]
1     0  327239       2  20   0      0     0 worker I    ?          0:00 [kworker/6:78-xfs-conv/vda4]
1     0  327240       2  20   0      0     0 worker I    ?          0:00 [kworker/6:79-xfs-conv/vda4]
1     0  327241       2  20   0      0     0 worker I    ?          0:00 [kworker/6:80-xfs-conv/vda4]
1     0  327242       2  20   0      0     0 worker I    ?          0:00 [kworker/6:81-xfs-conv/vda4]
1     0  327243       2  20   0      0     0 worker I    ?          0:00 [kworker/6:82-xfs-conv/vda4]
1     0  327244       2  20   0      0     0 worker I    ?          0:00 [kworker/6:83-xfs-conv/vda4]
1     0  327245       2  20   0      0     0 worker I    ?          0:00 [kworker/6:84-xfs-conv/vda4]
1     0  327246       2  20   0      0     0 worker I    ?          0:00 [kworker/6:85-xfs-conv/vda4]
1     0  327247       2  20   0      0     0 worker I    ?          0:00 [kworker/6:86-xfs-conv/vda4]
1     0  327248       2  20   0      0     0 worker I    ?          0:00 [kworker/6:87-xfs-conv/vda4]
1     0  327249       2  20   0      0     0 worker I    ?          0:00 [kworker/6:88-xfs-conv/vda4]
1     0  327250       2  20   0      0     0 worker I    ?          0:00 [kworker/6:89-xfs-conv/vda4]
1     0  327251       2  20   0      0     0 worker I    ?          0:00 [kworker/6:90-xfs-conv/vda4]
1     0  327252       2  20   0      0     0 worker I    ?          0:00 [kworker/6:91-xfs-conv/vda4]
1     0  327253       2  20   0      0     0 worker I    ?          0:00 [kworker/6:92-xfs-conv/vda4]
1     0  327254       2  20   0      0     0 worker I    ?          0:00 [kworker/6:93-xfs-conv/vda4]
1     0  327255       2  20   0      0     0 worker I    ?          0:00 [kworker/6:94-xfs-conv/vda4]
1     0  327256       2  20   0      0     0 worker I    ?          0:00 [kworker/6:95-xfs-conv/vda4]
1     0  327257       2  20   0      0     0 worker I    ?          0:00 [kworker/6:96-xfs-conv/vda4]
1     0  327258       2  20   0      0     0 worker I    ?          0:00 [kworker/6:97-xfs-conv/vda4]
1     0  327259       2  20   0      0     0 worker I    ?          0:00 [kworker/6:98-xfs-conv/vda4]
1     0  327260       2  20   0      0     0 worker I    ?          0:00 [kworker/6:99-xfs-conv/vda4]
1     0  327261       2  20   0      0     0 worker I    ?          0:00 [kworker/6:100-xfs-conv/vda4]
1     0  327262       2  20   0      0     0 worker I    ?          0:00 [kworker/6:101-xfs-conv/vda4]
1     0  327263       2  20   0      0     0 worker I    ?          0:00 [kworker/6:102-xfs-conv/vda4]
1     0  327264       2  20   0      0     0 worker I    ?          0:00 [kworker/6:103-xfs-conv/vda4]
1     0  327265       2  20   0      0     0 worker I    ?          0:00 [kworker/6:104-xfs-conv/vda4]
1     0  327266       2  20   0      0     0 worker I    ?          0:00 [kworker/6:105-xfs-conv/vda4]
1     0  327267       2  20   0      0     0 worker I    ?          0:00 [kworker/6:106-xfs-conv/vda4]
1     0  327268       2  20   0      0     0 worker I    ?          0:00 [kworker/6:107-xfs-conv/vda4]
1     0  327269       2  20   0      0     0 worker I    ?          0:00 [kworker/6:108-xfs-conv/vda4]
1     0  327270       2  20   0      0     0 worker I    ?          0:00 [kworker/6:109-xfs-conv/vda4]
1     0  327271       2  20   0      0     0 worker I    ?          0:00 [kworker/6:110-xfs-conv/vda4]
1     0  327272       2  20   0      0     0 worker I    ?          0:00 [kworker/6:111-xfs-conv/vda4]
1     0  327273       2  20   0      0     0 worker I    ?          0:00 [kworker/6:112-xfs-conv/vda4]
1     0  327274       2  20   0      0     0 worker I    ?          0:00 [kworker/6:113-xfs-conv/vda4]
1     0  327275       2  20   0      0     0 worker I    ?          0:00 [kworker/6:114-xfs-conv/vda4]
1     0  327276       2  20   0      0     0 worker I    ?          0:00 [kworker/6:115-xfs-conv/vda4]
1     0  327277       2  20   0      0     0 worker I    ?          0:00 [kworker/6:116-xfs-conv/vda4]
1     0  327278       2  20   0      0     0 worker I    ?          0:00 [kworker/6:117-xfs-conv/vda4]
1     0  327279       2  20   0      0     0 worker I    ?          0:00 [kworker/6:118-xfs-conv/vda4]
1     0  327280       2  20   0      0     0 worker I    ?          0:00 [kworker/6:119-xfs-conv/vda4]
1     0  327281       2  20   0      0     0 worker I    ?          0:00 [kworker/6:120-xfs-conv/vda4]
1     0  327291       2  20   0      0     0 worker I    ?          0:00 [kworker/6:121-xfs-conv/vda4]
1     0  327292       2  20   0      0     0 worker I    ?          0:00 [kworker/6:122-xfs-conv/vda4]
1     0  327293       2  20   0      0     0 worker I    ?          0:00 [kworker/6:123-xfs-conv/vda4]
1     0  327294       2  20   0      0     0 worker I    ?          0:00 [kworker/6:124-xfs-conv/vda4]
1     0  327295       2  20   0      0     0 worker I    ?          0:00 [kworker/6:125-xfs-conv/vda4]
1     0  327297       2  20   0      0     0 worker I    ?          0:00 [kworker/6:126-xfs-conv/vda4]
1     0  327299       2  20   0      0     0 worker I    ?          0:00 [kworker/6:127-xfs-conv/vda4]
1     0  327300       2  20   0      0     0 worker I    ?          0:00 [kworker/6:128-xfs-conv/vda4]
1     0  327302       2  20   0      0     0 worker I    ?          0:00 [kworker/6:129-xfs-conv/vda4]
1     0  327303       2  20   0      0     0 worker I    ?          0:00 [kworker/6:130-xfs-conv/vda4]
1     0  327304       2  20   0      0     0 worker I    ?          0:00 [kworker/6:131-xfs-conv/vda4]
1     0  327305       2  20   0      0     0 worker I    ?          0:00 [kworker/6:132-xfs-conv/vda4]
1     0  327306       2  20   0      0     0 worker I    ?          0:00 [kworker/6:133-xfs-conv/vda4]
1     0  327307       2  20   0      0     0 worker I    ?          0:00 [kworker/6:134-xfs-conv/vda4]
1     0  327308       2  20   0      0     0 worker I    ?          0:00 [kworker/6:135-xfs-conv/vda4]
1     0  327309       2  20   0      0     0 worker I    ?          0:00 [kworker/6:136-xfs-conv/vda4]
1     0  327310       2  20   0      0     0 worker I    ?          0:00 [kworker/6:137-xfs-conv/vda4]
1     0  327311       2  20   0      0     0 worker I    ?          0:00 [kworker/6:138-xfs-conv/vda4]
1     0  327312       2  20   0      0     0 worker I    ?          0:00 [kworker/6:139-xfs-conv/vda4]
1     0  327313       2  20   0      0     0 worker I    ?          0:00 [kworker/6:140-xfs-conv/vda4]
1     0  327314       2  20   0      0     0 worker I    ?          0:00 [kworker/6:141-xfs-conv/vda4]
1     0  327315       2  20   0      0     0 worker I    ?          0:00 [kworker/6:142-xfs-conv/vda4]
1     0  327316       2  20   0      0     0 worker I    ?          0:00 [kworker/6:143-xfs-conv/vda4]
1     0  327317       2  20   0      0     0 worker I    ?          0:00 [kworker/6:144-xfs-conv/vda4]
1     0  327318       2  20   0      0     0 worker I    ?          0:00 [kworker/6:145-xfs-conv/vda4]
1     0  327319       2  20   0      0     0 worker I    ?          0:00 [kworker/6:146-xfs-conv/vda4]
1     0  327320       2  20   0      0     0 worker I    ?          0:00 [kworker/6:147-xfs-conv/vda4]
1     0  327321       2  20   0      0     0 worker I    ?          0:00 [kworker/6:148-xfs-conv/vda4]
1     0  327322       2  20   0      0     0 worker I    ?          0:00 [kworker/6:149-xfs-conv/vda4]
1     0  327323       2  20   0      0     0 worker I    ?          0:00 [kworker/6:150-xfs-conv/vda4]
1     0  327324       2  20   0      0     0 worker I    ?          0:00 [kworker/6:151-xfs-conv/vda4]
1     0  327325       2  20   0      0     0 worker I    ?          0:00 [kworker/6:152-xfs-conv/vda4]
1     0  327326       2  20   0      0     0 worker I    ?          0:00 [kworker/6:153-xfs-conv/vda4]
1     0  327327       2  20   0      0     0 worker I    ?          0:00 [kworker/6:154-xfs-conv/vda4]
1     0  327328       2  20   0      0     0 worker I    ?          0:00 [kworker/6:155-xfs-conv/vda4]
1     0  327329       2  20   0      0     0 worker I    ?          0:00 [kworker/6:156-xfs-conv/vda4]
1     0  327330       2  20   0      0     0 worker I    ?          0:00 [kworker/6:157-xfs-conv/vda4]
1     0  327331       2  20   0      0     0 worker I    ?          0:00 [kworker/6:158-xfs-conv/vda4]
1     0  327332       2  20   0      0     0 worker I    ?          0:00 [kworker/6:159-xfs-conv/vda4]
1     0  327333       2  20   0      0     0 worker I    ?          0:00 [kworker/6:160-xfs-conv/vda4]
1     0  327334       2  20   0      0     0 worker I    ?          0:00 [kworker/6:161-xfs-conv/vda4]
1     0  327335       2  20   0      0     0 worker I    ?          0:00 [kworker/6:162-xfs-conv/vda4]
1     0  327336       2  20   0      0     0 worker I    ?          0:00 [kworker/6:163-xfs-conv/vda4]
1     0  327337       2  20   0      0     0 worker I    ?          0:00 [kworker/6:164-xfs-conv/vda4]
1     0  327338       2  20   0      0     0 worker I    ?          0:00 [kworker/6:165-xfs-conv/vda4]
1     0  327339       2  20   0      0     0 worker I    ?          0:00 [kworker/6:166-xfs-conv/vda4]
1     0  327340       2  20   0      0     0 worker I    ?          0:00 [kworker/6:167-xfs-conv/vda4]
1     0  327341       2  20   0      0     0 worker I    ?          0:00 [kworker/6:168-xfs-conv/vda4]
1     0  327342       2  20   0      0     0 worker I    ?          0:00 [kworker/6:169-xfs-conv/vda4]
1     0  327343       2  20   0      0     0 worker I    ?          0:00 [kworker/6:170-xfs-conv/vda4]
1     0  327344       2  20   0      0     0 worker I    ?          0:00 [kworker/6:171-xfs-conv/vda4]
1     0  327345       2  20   0      0     0 worker I    ?          0:00 [kworker/6:172-xfs-conv/vda4]
1     0  327346       2  20   0      0     0 worker I    ?          0:00 [kworker/6:173-xfs-conv/vda4]
1     0  327347       2  20   0      0     0 worker I    ?          0:00 [kworker/6:174-xfs-conv/vda4]
1     0  327348       2  20   0      0     0 worker I    ?          0:00 [kworker/6:175-xfs-conv/vda4]
1     0  327349       2  20   0      0     0 worker I    ?          0:00 [kworker/6:176-xfs-conv/vda4]
1     0  327350       2  20   0      0     0 worker I    ?          0:00 [kworker/6:177-xfs-conv/vda4]
1     0  327351       2  20   0      0     0 worker I    ?          0:00 [kworker/6:178-xfs-conv/vda4]
1     0  327352       2  20   0      0     0 worker I    ?          0:00 [kworker/6:179-xfs-conv/vda4]
1     0  327353       2  20   0      0     0 worker I    ?          0:00 [kworker/6:180-xfs-conv/vda4]
1     0  327354       2  20   0      0     0 worker I    ?          0:00 [kworker/6:181-xfs-conv/vda4]
1     0  327355       2  20   0      0     0 worker I    ?          0:00 [kworker/6:182-xfs-conv/vda4]
1     0  327356       2  20   0      0     0 worker I    ?          0:00 [kworker/6:183-xfs-conv/vda4]
1     0  327357       2  20   0      0     0 worker I    ?          0:00 [kworker/6:184-xfs-conv/vda4]
1     0  327358       2  20   0      0     0 worker I    ?          0:00 [kworker/6:185-xfs-conv/vda4]
1     0  327360       2  20   0      0     0 worker I    ?          0:00 [kworker/6:186-xfs-conv/vda4]
1     0  327361       2  20   0      0     0 worker I    ?          0:00 [kworker/6:187-xfs-conv/vda4]
1     0  327362       2  20   0      0     0 worker I    ?          0:00 [kworker/6:188-xfs-conv/vda4]
1     0  327363       2  20   0      0     0 worker I    ?          0:00 [kworker/6:189-xfs-conv/vda4]
1     0  327364       2  20   0      0     0 worker I    ?          0:00 [kworker/6:190-xfs-conv/vda4]
1     0  327365       2  20   0      0     0 worker I    ?          0:00 [kworker/6:191-xfs-conv/vda4]
1     0  327368       2  20   0      0     0 worker I    ?          0:00 [kworker/6:192-xfs-conv/vda4]
1     0  327369       2  20   0      0     0 worker I    ?          0:00 [kworker/6:193-xfs-conv/vda4]
1     0  327370       2  20   0      0     0 worker I    ?          0:00 [kworker/6:194-xfs-conv/vda4]
1     0  327371       2  20   0      0     0 worker I    ?          0:00 [kworker/6:195-xfs-conv/vda4]
1     0  327372       2  20   0      0     0 worker I    ?          0:00 [kworker/6:196-xfs-conv/vda4]
1     0  327374       2  20   0      0     0 worker I    ?          0:00 [kworker/6:197-xfs-conv/vda4]
1     0  327375       2  20   0      0     0 worker I    ?          0:00 [kworker/6:198-xfs-conv/vda4]
1     0  327376       2  20   0      0     0 worker I    ?          0:00 [kworker/6:199-xfs-conv/vda4]
1     0  327378       2  20   0      0     0 worker I    ?          0:00 [kworker/6:200-xfs-conv/vda4]
1     0  327380       2  20   0      0     0 worker I    ?          0:00 [kworker/6:201-xfs-conv/vda4]
1     0  327381       2  20   0      0     0 worker I    ?          0:00 [kworker/6:202-xfs-conv/vda4]
1     0  327382       2  20   0      0     0 worker I    ?          0:00 [kworker/6:203-xfs-conv/vda4]
1     0  327383       2  20   0      0     0 worker I    ?          0:00 [kworker/6:204-xfs-conv/vda4]
1     0  327385       2  20   0      0     0 worker I    ?          0:00 [kworker/6:205-xfs-conv/vda4]
1     0  327386       2  20   0      0     0 worker I    ?          0:00 [kworker/6:206-xfs-conv/vda4]
1     0  327394       2  20   0      0     0 worker I    ?          0:00 [kworker/6:207-xfs-conv/vda4]
1     0  327395       2  20   0      0     0 worker I    ?          0:00 [kworker/6:208-xfs-conv/vda4]
1     0  327396       2  20   0      0     0 worker I    ?          0:00 [kworker/6:209-xfs-conv/vda4]
1     0  327397       2  20   0      0     0 worker I    ?          0:00 [kworker/6:210-xfs-conv/vda4]
1     0  327398       2  20   0      0     0 worker I    ?          0:00 [kworker/6:211-xfs-conv/vda4]
1     0  327399       2  20   0      0     0 worker I    ?          0:00 [kworker/6:212-xfs-conv/vda4]
1     0  327400       2  20   0      0     0 worker I    ?          0:00 [kworker/6:213-xfs-conv/vda4]
1     0  327401       2  20   0      0     0 worker I    ?          0:00 [kworker/6:214-xfs-conv/vda4]
1     0  327402       2  20   0      0     0 worker I    ?          0:00 [kworker/6:215-xfs-conv/vda4]
1     0  327403       2  20   0      0     0 worker I    ?          0:00 [kworker/6:216-xfs-conv/vda4]
1     0  327404       2  20   0      0     0 worker I    ?          0:00 [kworker/6:217-xfs-conv/vda4]
1     0  327405       2  20   0      0     0 worker I    ?          0:00 [kworker/6:218-xfs-conv/vda4]
1     0  327406       2  20   0      0     0 worker I    ?          0:00 [kworker/6:219-xfs-conv/vda4]
1     0  327407       2  20   0      0     0 worker I    ?          0:00 [kworker/6:220-xfs-conv/vda4]
1     0  327408       2  20   0      0     0 worker I    ?          0:00 [kworker/6:221-xfs-conv/vda4]
1     0  327409       2  20   0      0     0 worker I    ?          0:00 [kworker/6:222-xfs-conv/vda4]
1     0  327410       2  20   0      0     0 worker I    ?          0:00 [kworker/6:223-xfs-conv/vda4]
1     0  327411       2  20   0      0     0 worker I    ?          0:00 [kworker/6:224-xfs-conv/vda4]
1     0  327412       2  20   0      0     0 worker I    ?          0:00 [kworker/6:225-xfs-conv/vda4]
1     0  327413       2  20   0      0     0 worker I    ?          0:00 [kworker/6:226-xfs-conv/vda4]
1     0  327414       2  20   0      0     0 worker I    ?          0:00 [kworker/6:227-xfs-conv/vda4]
1     0  327415       2  20   0      0     0 worker I    ?          0:00 [kworker/6:228-events]
1     0  327416       2  20   0      0     0 worker I    ?          0:00 [kworker/6:229-xfs-conv/vda4]
1     0  327417       2  20   0      0     0 worker I    ?          0:00 [kworker/6:230-xfs-conv/vda4]
1     0  327418       2  20   0      0     0 worker I    ?          0:00 [kworker/6:231-xfs-conv/vda4]
1     0  327419       2  20   0      0     0 worker I    ?          0:00 [kworker/6:232-xfs-conv/vda4]
1     0  327420       2  20   0      0     0 worker I    ?          0:00 [kworker/6:233-xfs-conv/vda4]
1     0  327421       2  20   0      0     0 worker I    ?          0:00 [kworker/6:234-xfs-conv/vda4]
1     0  327424       2  20   0      0     0 worker I    ?          0:00 [kworker/6:235-xfs-conv/vda4]
1     0  327425       2  20   0      0     0 worker I    ?          0:00 [kworker/6:236-events]
1     0  327426       2  20   0      0     0 worker I    ?          0:00 [kworker/6:237-xfs-conv/vda4]
1     0  327427       2  20   0      0     0 worker I    ?          0:00 [kworker/6:238-xfs-conv/vda4]
1     0  327428       2  20   0      0     0 worker I    ?          0:00 [kworker/6:239]
1     0  327430       2  20   0      0     0 worker I    ?          0:00 [kworker/6:240]
4     0  329584       1  20   0  19448  9320 ep_pol Ss   ?          0:00 /usr/lib/systemd/systemd-hostnamed
1     0  330259       2   0 -20      0     0 rescue I<   ?          0:00 [cfg80211]
5     0  331701   47718  20   0  34872  7072 ep_pol S    ?          0:00 /usr/lib/systemd/systemd-udevd
0     0  331795  324886  20   0   3068  1028 sigsus S    ?          0:00 timeout 300s ceph config dump --format json-pretty
0     0  331796  331795  20   0 1018756 53384 -     Rl   ?          0:00 /usr/bin/python3 -s /usr/bin/ceph config dump --format json-pretty
0     0  331890  324886  20   0   3068  1028 sigsus S    ?          0:00 timeout 300s ps alxwww
4     0  331891  331890  20   0   7560  3216 -      R    ?          0:00 ps alxwww
